static inline bool sk_acceptq_is_full(const struct sock *sk)
 {
-       return READ_ONCE(sk->sk_ack_backlog) > sk->sk_max_ack_backlog;
+       return READ_ONCE(sk->sk_ack_backlog) > READ_ONCE(sk->sk_max_ack_backlog);
 }
 
 /*
 
        if (!((1 << old_state) & (DCCPF_CLOSED | DCCPF_LISTEN)))
                goto out;
 
-       sk->sk_max_ack_backlog = backlog;
+       WRITE_ONCE(sk->sk_max_ack_backlog, backlog);
        /* Really, if the socket is already in listen state
         * we can only allow the backlog to be adjusted.
         */
 
        if (!((1 << old_state) & (TCPF_CLOSE | TCPF_LISTEN)))
                goto out;
 
-       sk->sk_max_ack_backlog = backlog;
+       WRITE_ONCE(sk->sk_max_ack_backlog, backlog);
        /* Really, if the socket is already in listen state
         * we can only allow the backlog to be adjusted.
         */
 
         * ones are about to clog our table.
         */
        qlen = reqsk_queue_len(queue);
-       if ((qlen << 1) > max(8U, sk_listener->sk_max_ack_backlog)) {
+       if ((qlen << 1) > max(8U, READ_ONCE(sk_listener->sk_max_ack_backlog))) {
                int young = reqsk_queue_len_young(queue) << 1;
 
                while (thresh > 2) {
 
                 * tcpi_sacked  -> max backlog
                 */
                info->tcpi_unacked = READ_ONCE(sk->sk_ack_backlog);
-               info->tcpi_sacked = sk->sk_max_ack_backlog;
+               info->tcpi_sacked = READ_ONCE(sk->sk_max_ack_backlog);
                return;
        }
 
 
 
        if (inet_sk_state_load(sk) == TCP_LISTEN) {
                r->idiag_rqueue = READ_ONCE(sk->sk_ack_backlog);
-               r->idiag_wqueue = sk->sk_max_ack_backlog;
+               r->idiag_wqueue = READ_ONCE(sk->sk_max_ack_backlog);
        } else if (sk->sk_type == SOCK_STREAM) {
                const struct tcp_sock *tp = tcp_sk(sk);
 
 
                *err = -1;
                return;
        }
-       dst->value = sk->sk_max_ack_backlog;
+       dst->value = READ_ONCE(sk->sk_max_ack_backlog);
 }
 
 META_COLLECTOR(int_sk_prio)
 
                r->idiag_wqueue = infox->asoc->sndbuf_used;
        } else {
                r->idiag_rqueue = READ_ONCE(sk->sk_ack_backlog);
-               r->idiag_wqueue = sk->sk_max_ack_backlog;
+               r->idiag_wqueue = READ_ONCE(sk->sk_max_ack_backlog);
        }
        if (infox->sctpinfo)
                sctp_get_sctp_info(sk, infox->asoc, infox->sctpinfo);
 
                }
        }
 
-       sk->sk_max_ack_backlog = backlog;
+       WRITE_ONCE(sk->sk_max_ack_backlog, backlog);
        return sctp_hash_endpoint(ep);
 }
 
 
        /* If we are already listening, just update the backlog */
        if (sctp_sstate(sk, LISTENING))
-               sk->sk_max_ack_backlog = backlog;
+               WRITE_ONCE(sk->sk_max_ack_backlog, backlog);
        else {
                err = sctp_listen_start(sk, backlog);
                if (err)