u32 dot11RSNAStatsTKIPLocalMICFailures;
 
        int key_idx;
-       struct crypto_skcipher *rx_tfm_arc4;
+       struct crypto_sync_skcipher *rx_tfm_arc4;
        struct crypto_shash *rx_tfm_michael;
-       struct crypto_skcipher *tx_tfm_arc4;
+       struct crypto_sync_skcipher *tx_tfm_arc4;
        struct crypto_shash *tx_tfm_michael;
        /* scratch buffers for virt_to_page() (crypto API) */
        u8 rx_hdr[16];
        if (priv == NULL)
                goto fail;
        priv->key_idx = key_idx;
-       priv->tx_tfm_arc4 = crypto_alloc_skcipher("ecb(arc4)", 0,
-                                                 CRYPTO_ALG_ASYNC);
+       priv->tx_tfm_arc4 = crypto_alloc_sync_skcipher("ecb(arc4)", 0, 0);
        if (IS_ERR(priv->tx_tfm_arc4)) {
                pr_debug("Could not allocate crypto API arc4\n");
                priv->tx_tfm_arc4 = NULL;
                goto fail;
        }
 
-       priv->rx_tfm_arc4 = crypto_alloc_skcipher("ecb(arc4)", 0,
-                                                 CRYPTO_ALG_ASYNC);
+       priv->rx_tfm_arc4 = crypto_alloc_sync_skcipher("ecb(arc4)", 0, 0);
        if (IS_ERR(priv->rx_tfm_arc4)) {
                pr_debug("Could not allocate crypto API arc4\n");
                priv->rx_tfm_arc4 = NULL;
 fail:
        if (priv) {
                crypto_free_shash(priv->tx_tfm_michael);
-               crypto_free_skcipher(priv->tx_tfm_arc4);
+               crypto_free_sync_skcipher(priv->tx_tfm_arc4);
                crypto_free_shash(priv->rx_tfm_michael);
-               crypto_free_skcipher(priv->rx_tfm_arc4);
+               crypto_free_sync_skcipher(priv->rx_tfm_arc4);
                kfree(priv);
        }
 
 
        if (_priv) {
                crypto_free_shash(_priv->tx_tfm_michael);
-               crypto_free_skcipher(_priv->tx_tfm_arc4);
+               crypto_free_sync_skcipher(_priv->tx_tfm_arc4);
                crypto_free_shash(_priv->rx_tfm_michael);
-               crypto_free_skcipher(_priv->rx_tfm_arc4);
+               crypto_free_sync_skcipher(_priv->rx_tfm_arc4);
        }
        kfree(priv);
 }
        *pos++ = (tkey->tx_iv32 >> 24) & 0xff;
 
        if (!tcb_desc->bHwSec) {
-               SKCIPHER_REQUEST_ON_STACK(req, tkey->tx_tfm_arc4);
+               SYNC_SKCIPHER_REQUEST_ON_STACK(req, tkey->tx_tfm_arc4);
 
                icv = skb_put(skb, 4);
                crc = ~crc32_le(~0, pos, len);
                sg_init_one(&sg, pos, len+4);
 
 
-               crypto_skcipher_setkey(tkey->tx_tfm_arc4, rc4key, 16);
-               skcipher_request_set_tfm(req, tkey->tx_tfm_arc4);
+               crypto_sync_skcipher_setkey(tkey->tx_tfm_arc4, rc4key, 16);
+               skcipher_request_set_sync_tfm(req, tkey->tx_tfm_arc4);
                skcipher_request_set_callback(req, 0, NULL, NULL);
                skcipher_request_set_crypt(req, &sg, &sg, len + 4, NULL);
                ret = crypto_skcipher_encrypt(req);
        pos += 8;
 
        if (!tcb_desc->bHwSec || (skb->cb[0] == 1)) {
-               SKCIPHER_REQUEST_ON_STACK(req, tkey->rx_tfm_arc4);
+               SYNC_SKCIPHER_REQUEST_ON_STACK(req, tkey->rx_tfm_arc4);
 
                if ((iv32 < tkey->rx_iv32 ||
                    (iv32 == tkey->rx_iv32 && iv16 <= tkey->rx_iv16)) &&
 
                sg_init_one(&sg, pos, plen+4);
 
-               crypto_skcipher_setkey(tkey->rx_tfm_arc4, rc4key, 16);
-               skcipher_request_set_tfm(req, tkey->rx_tfm_arc4);
+               crypto_sync_skcipher_setkey(tkey->rx_tfm_arc4, rc4key, 16);
+               skcipher_request_set_sync_tfm(req, tkey->rx_tfm_arc4);
                skcipher_request_set_callback(req, 0, NULL, NULL);
                skcipher_request_set_crypt(req, &sg, &sg, plen + 4, NULL);
                err = crypto_skcipher_decrypt(req);
        struct rtllib_tkip_data *tkey = priv;
        int keyidx;
        struct crypto_shash *tfm = tkey->tx_tfm_michael;
-       struct crypto_skcipher *tfm2 = tkey->tx_tfm_arc4;
+       struct crypto_sync_skcipher *tfm2 = tkey->tx_tfm_arc4;
        struct crypto_shash *tfm3 = tkey->rx_tfm_michael;
-       struct crypto_skcipher *tfm4 = tkey->rx_tfm_arc4;
+       struct crypto_sync_skcipher *tfm4 = tkey->rx_tfm_arc4;
 
        keyidx = tkey->key_idx;
        memset(tkey, 0, sizeof(*tkey));
 
        u8 key[WEP_KEY_LEN + 1];
        u8 key_len;
        u8 key_idx;
-       struct crypto_skcipher *tx_tfm;
-       struct crypto_skcipher *rx_tfm;
+       struct crypto_sync_skcipher *tx_tfm;
+       struct crypto_sync_skcipher *rx_tfm;
 };
 
 
                goto fail;
        priv->key_idx = keyidx;
 
-       priv->tx_tfm = crypto_alloc_skcipher("ecb(arc4)", 0, CRYPTO_ALG_ASYNC);
+       priv->tx_tfm = crypto_alloc_sync_skcipher("ecb(arc4)", 0, 0);
        if (IS_ERR(priv->tx_tfm)) {
                pr_debug("rtllib_crypt_wep: could not allocate crypto API arc4\n");
                priv->tx_tfm = NULL;
                goto fail;
        }
-       priv->rx_tfm = crypto_alloc_skcipher("ecb(arc4)", 0, CRYPTO_ALG_ASYNC);
+       priv->rx_tfm = crypto_alloc_sync_skcipher("ecb(arc4)", 0, 0);
        if (IS_ERR(priv->rx_tfm)) {
                pr_debug("rtllib_crypt_wep: could not allocate crypto API arc4\n");
                priv->rx_tfm = NULL;
 
 fail:
        if (priv) {
-               crypto_free_skcipher(priv->tx_tfm);
-               crypto_free_skcipher(priv->rx_tfm);
+               crypto_free_sync_skcipher(priv->tx_tfm);
+               crypto_free_sync_skcipher(priv->rx_tfm);
                kfree(priv);
        }
        return NULL;
        struct prism2_wep_data *_priv = priv;
 
        if (_priv) {
-               crypto_free_skcipher(_priv->tx_tfm);
-               crypto_free_skcipher(_priv->rx_tfm);
+               crypto_free_sync_skcipher(_priv->tx_tfm);
+               crypto_free_sync_skcipher(_priv->rx_tfm);
        }
        kfree(priv);
 }
        memcpy(key + 3, wep->key, wep->key_len);
 
        if (!tcb_desc->bHwSec) {
-               SKCIPHER_REQUEST_ON_STACK(req, wep->tx_tfm);
+               SYNC_SKCIPHER_REQUEST_ON_STACK(req, wep->tx_tfm);
 
                /* Append little-endian CRC32 and encrypt it to produce ICV */
                crc = ~crc32_le(~0, pos, len);
                icv[3] = crc >> 24;
 
                sg_init_one(&sg, pos, len+4);
-               crypto_skcipher_setkey(wep->tx_tfm, key, klen);
-               skcipher_request_set_tfm(req, wep->tx_tfm);
+               crypto_sync_skcipher_setkey(wep->tx_tfm, key, klen);
+               skcipher_request_set_sync_tfm(req, wep->tx_tfm);
                skcipher_request_set_callback(req, 0, NULL, NULL);
                skcipher_request_set_crypt(req, &sg, &sg, len + 4, NULL);
                err = crypto_skcipher_encrypt(req);
        plen = skb->len - hdr_len - 8;
 
        if (!tcb_desc->bHwSec) {
-               SKCIPHER_REQUEST_ON_STACK(req, wep->rx_tfm);
+               SYNC_SKCIPHER_REQUEST_ON_STACK(req, wep->rx_tfm);
 
                sg_init_one(&sg, pos, plen+4);
-               crypto_skcipher_setkey(wep->rx_tfm, key, klen);
-               skcipher_request_set_tfm(req, wep->rx_tfm);
+               crypto_sync_skcipher_setkey(wep->rx_tfm, key, klen);
+               skcipher_request_set_sync_tfm(req, wep->rx_tfm);
                skcipher_request_set_callback(req, 0, NULL, NULL);
                skcipher_request_set_crypt(req, &sg, &sg, plen + 4, NULL);
                err = crypto_skcipher_decrypt(req);
 
 
        int key_idx;
 
-       struct crypto_skcipher *rx_tfm_arc4;
+       struct crypto_sync_skcipher *rx_tfm_arc4;
        struct crypto_shash *rx_tfm_michael;
-       struct crypto_skcipher *tx_tfm_arc4;
+       struct crypto_sync_skcipher *tx_tfm_arc4;
        struct crypto_shash *tx_tfm_michael;
 
        /* scratch buffers for virt_to_page() (crypto API) */
                goto fail;
        priv->key_idx = key_idx;
 
-       priv->tx_tfm_arc4 = crypto_alloc_skcipher("ecb(arc4)", 0,
-                       CRYPTO_ALG_ASYNC);
+       priv->tx_tfm_arc4 = crypto_alloc_sync_skcipher("ecb(arc4)", 0, 0);
        if (IS_ERR(priv->tx_tfm_arc4)) {
                printk(KERN_DEBUG "ieee80211_crypt_tkip: could not allocate "
                                "crypto API arc4\n");
                goto fail;
        }
 
-       priv->rx_tfm_arc4 = crypto_alloc_skcipher("ecb(arc4)", 0,
-                       CRYPTO_ALG_ASYNC);
+       priv->rx_tfm_arc4 = crypto_alloc_sync_skcipher("ecb(arc4)", 0, 0);
        if (IS_ERR(priv->rx_tfm_arc4)) {
                printk(KERN_DEBUG "ieee80211_crypt_tkip: could not allocate "
                                "crypto API arc4\n");
 fail:
        if (priv) {
                crypto_free_shash(priv->tx_tfm_michael);
-               crypto_free_skcipher(priv->tx_tfm_arc4);
+               crypto_free_sync_skcipher(priv->tx_tfm_arc4);
                crypto_free_shash(priv->rx_tfm_michael);
-               crypto_free_skcipher(priv->rx_tfm_arc4);
+               crypto_free_sync_skcipher(priv->rx_tfm_arc4);
                kfree(priv);
        }
 
 
        if (_priv) {
                crypto_free_shash(_priv->tx_tfm_michael);
-               crypto_free_skcipher(_priv->tx_tfm_arc4);
+               crypto_free_sync_skcipher(_priv->tx_tfm_arc4);
                crypto_free_shash(_priv->rx_tfm_michael);
-               crypto_free_skcipher(_priv->rx_tfm_arc4);
+               crypto_free_sync_skcipher(_priv->rx_tfm_arc4);
        }
        kfree(priv);
 }
        *pos++ = (tkey->tx_iv32 >> 24) & 0xff;
 
        if (!tcb_desc->bHwSec) {
-               SKCIPHER_REQUEST_ON_STACK(req, tkey->tx_tfm_arc4);
+               SYNC_SKCIPHER_REQUEST_ON_STACK(req, tkey->tx_tfm_arc4);
 
                icv = skb_put(skb, 4);
                crc = ~crc32_le(~0, pos, len);
                icv[1] = crc >> 8;
                icv[2] = crc >> 16;
                icv[3] = crc >> 24;
-               crypto_skcipher_setkey(tkey->tx_tfm_arc4, rc4key, 16);
+               crypto_sync_skcipher_setkey(tkey->tx_tfm_arc4, rc4key, 16);
                sg_init_one(&sg, pos, len+4);
-               skcipher_request_set_tfm(req, tkey->tx_tfm_arc4);
+               skcipher_request_set_sync_tfm(req, tkey->tx_tfm_arc4);
                skcipher_request_set_callback(req, 0, NULL, NULL);
                skcipher_request_set_crypt(req, &sg, &sg, len + 4, NULL);
                ret = crypto_skcipher_encrypt(req);
        pos += 8;
 
        if (!tcb_desc->bHwSec) {
-               SKCIPHER_REQUEST_ON_STACK(req, tkey->rx_tfm_arc4);
+               SYNC_SKCIPHER_REQUEST_ON_STACK(req, tkey->rx_tfm_arc4);
 
                if (iv32 < tkey->rx_iv32 ||
                (iv32 == tkey->rx_iv32 && iv16 <= tkey->rx_iv16)) {
 
                plen = skb->len - hdr_len - 12;
 
-               crypto_skcipher_setkey(tkey->rx_tfm_arc4, rc4key, 16);
+               crypto_sync_skcipher_setkey(tkey->rx_tfm_arc4, rc4key, 16);
                sg_init_one(&sg, pos, plen+4);
 
-               skcipher_request_set_tfm(req, tkey->rx_tfm_arc4);
+               skcipher_request_set_sync_tfm(req, tkey->rx_tfm_arc4);
                skcipher_request_set_callback(req, 0, NULL, NULL);
                skcipher_request_set_crypt(req, &sg, &sg, plen + 4, NULL);
 
        struct ieee80211_tkip_data *tkey = priv;
        int keyidx;
        struct crypto_shash *tfm = tkey->tx_tfm_michael;
-       struct crypto_skcipher *tfm2 = tkey->tx_tfm_arc4;
+       struct crypto_sync_skcipher *tfm2 = tkey->tx_tfm_arc4;
        struct crypto_shash *tfm3 = tkey->rx_tfm_michael;
-       struct crypto_skcipher *tfm4 = tkey->rx_tfm_arc4;
+       struct crypto_sync_skcipher *tfm4 = tkey->rx_tfm_arc4;
 
        keyidx = tkey->key_idx;
        memset(tkey, 0, sizeof(*tkey));
 
        u8 key[WEP_KEY_LEN + 1];
        u8 key_len;
        u8 key_idx;
-       struct crypto_skcipher *tx_tfm;
-       struct crypto_skcipher *rx_tfm;
+       struct crypto_sync_skcipher *tx_tfm;
+       struct crypto_sync_skcipher *rx_tfm;
 };
 
 
                return NULL;
        priv->key_idx = keyidx;
 
-       priv->tx_tfm = crypto_alloc_skcipher("ecb(arc4)", 0, CRYPTO_ALG_ASYNC);
+       priv->tx_tfm = crypto_alloc_sync_skcipher("ecb(arc4)", 0, 0);
        if (IS_ERR(priv->tx_tfm))
                goto free_priv;
-       priv->rx_tfm = crypto_alloc_skcipher("ecb(arc4)", 0, CRYPTO_ALG_ASYNC);
+       priv->rx_tfm = crypto_alloc_sync_skcipher("ecb(arc4)", 0, 0);
        if (IS_ERR(priv->rx_tfm))
                goto free_tx;
 
 
        return priv;
 free_tx:
-       crypto_free_skcipher(priv->tx_tfm);
+       crypto_free_sync_skcipher(priv->tx_tfm);
 free_priv:
        kfree(priv);
        return NULL;
        struct prism2_wep_data *_priv = priv;
 
        if (_priv) {
-               crypto_free_skcipher(_priv->tx_tfm);
-               crypto_free_skcipher(_priv->rx_tfm);
+               crypto_free_sync_skcipher(_priv->tx_tfm);
+               crypto_free_sync_skcipher(_priv->rx_tfm);
        }
        kfree(priv);
 }
        memcpy(key + 3, wep->key, wep->key_len);
 
        if (!tcb_desc->bHwSec) {
-               SKCIPHER_REQUEST_ON_STACK(req, wep->tx_tfm);
+               SYNC_SKCIPHER_REQUEST_ON_STACK(req, wep->tx_tfm);
 
                /* Append little-endian CRC32 and encrypt it to produce ICV */
                crc = ~crc32_le(~0, pos, len);
                icv[2] = crc >> 16;
                icv[3] = crc >> 24;
 
-               crypto_skcipher_setkey(wep->tx_tfm, key, klen);
+               crypto_sync_skcipher_setkey(wep->tx_tfm, key, klen);
                sg_init_one(&sg, pos, len+4);
 
-               skcipher_request_set_tfm(req, wep->tx_tfm);
+               skcipher_request_set_sync_tfm(req, wep->tx_tfm);
                skcipher_request_set_callback(req, 0, NULL, NULL);
                skcipher_request_set_crypt(req, &sg, &sg, len + 4, NULL);
 
        plen = skb->len - hdr_len - 8;
 
        if (!tcb_desc->bHwSec) {
-               SKCIPHER_REQUEST_ON_STACK(req, wep->rx_tfm);
+               SYNC_SKCIPHER_REQUEST_ON_STACK(req, wep->rx_tfm);
 
-               crypto_skcipher_setkey(wep->rx_tfm, key, klen);
+               crypto_sync_skcipher_setkey(wep->rx_tfm, key, klen);
                sg_init_one(&sg, pos, plen+4);
 
-               skcipher_request_set_tfm(req, wep->rx_tfm);
+               skcipher_request_set_sync_tfm(req, wep->rx_tfm);
                skcipher_request_set_callback(req, 0, NULL, NULL);
                skcipher_request_set_crypt(req, &sg, &sg, plen + 4, NULL);
 
 
 
        int key_idx;
 
-       struct crypto_skcipher *rx_tfm_arc4;
+       struct crypto_sync_skcipher *rx_tfm_arc4;
        struct crypto_shash *rx_tfm_michael;
-       struct crypto_skcipher *tx_tfm_arc4;
+       struct crypto_sync_skcipher *tx_tfm_arc4;
        struct crypto_shash *tx_tfm_michael;
 
        /* scratch buffers for virt_to_page() (crypto API) */
 
        priv->key_idx = key_idx;
 
-       priv->tx_tfm_arc4 = crypto_alloc_skcipher("ecb(arc4)", 0,
-                                                 CRYPTO_ALG_ASYNC);
+       priv->tx_tfm_arc4 = crypto_alloc_sync_skcipher("ecb(arc4)", 0, 0);
        if (IS_ERR(priv->tx_tfm_arc4)) {
                priv->tx_tfm_arc4 = NULL;
                goto fail;
                goto fail;
        }
 
-       priv->rx_tfm_arc4 = crypto_alloc_skcipher("ecb(arc4)", 0,
-                                                 CRYPTO_ALG_ASYNC);
+       priv->rx_tfm_arc4 = crypto_alloc_sync_skcipher("ecb(arc4)", 0, 0);
        if (IS_ERR(priv->rx_tfm_arc4)) {
                priv->rx_tfm_arc4 = NULL;
                goto fail;
       fail:
        if (priv) {
                crypto_free_shash(priv->tx_tfm_michael);
-               crypto_free_skcipher(priv->tx_tfm_arc4);
+               crypto_free_sync_skcipher(priv->tx_tfm_arc4);
                crypto_free_shash(priv->rx_tfm_michael);
-               crypto_free_skcipher(priv->rx_tfm_arc4);
+               crypto_free_sync_skcipher(priv->rx_tfm_arc4);
                kfree(priv);
        }
 
        struct lib80211_tkip_data *_priv = priv;
        if (_priv) {
                crypto_free_shash(_priv->tx_tfm_michael);
-               crypto_free_skcipher(_priv->tx_tfm_arc4);
+               crypto_free_sync_skcipher(_priv->tx_tfm_arc4);
                crypto_free_shash(_priv->rx_tfm_michael);
-               crypto_free_skcipher(_priv->rx_tfm_arc4);
+               crypto_free_sync_skcipher(_priv->rx_tfm_arc4);
        }
        kfree(priv);
 }
 static int lib80211_tkip_encrypt(struct sk_buff *skb, int hdr_len, void *priv)
 {
        struct lib80211_tkip_data *tkey = priv;
-       SKCIPHER_REQUEST_ON_STACK(req, tkey->tx_tfm_arc4);
+       SYNC_SKCIPHER_REQUEST_ON_STACK(req, tkey->tx_tfm_arc4);
        int len;
        u8 rc4key[16], *pos, *icv;
        u32 crc;
        icv[2] = crc >> 16;
        icv[3] = crc >> 24;
 
-       crypto_skcipher_setkey(tkey->tx_tfm_arc4, rc4key, 16);
+       crypto_sync_skcipher_setkey(tkey->tx_tfm_arc4, rc4key, 16);
        sg_init_one(&sg, pos, len + 4);
-       skcipher_request_set_tfm(req, tkey->tx_tfm_arc4);
+       skcipher_request_set_sync_tfm(req, tkey->tx_tfm_arc4);
        skcipher_request_set_callback(req, 0, NULL, NULL);
        skcipher_request_set_crypt(req, &sg, &sg, len + 4, NULL);
        err = crypto_skcipher_encrypt(req);
 static int lib80211_tkip_decrypt(struct sk_buff *skb, int hdr_len, void *priv)
 {
        struct lib80211_tkip_data *tkey = priv;
-       SKCIPHER_REQUEST_ON_STACK(req, tkey->rx_tfm_arc4);
+       SYNC_SKCIPHER_REQUEST_ON_STACK(req, tkey->rx_tfm_arc4);
        u8 rc4key[16];
        u8 keyidx, *pos;
        u32 iv32;
 
        plen = skb->len - hdr_len - 12;
 
-       crypto_skcipher_setkey(tkey->rx_tfm_arc4, rc4key, 16);
+       crypto_sync_skcipher_setkey(tkey->rx_tfm_arc4, rc4key, 16);
        sg_init_one(&sg, pos, plen + 4);
-       skcipher_request_set_tfm(req, tkey->rx_tfm_arc4);
+       skcipher_request_set_sync_tfm(req, tkey->rx_tfm_arc4);
        skcipher_request_set_callback(req, 0, NULL, NULL);
        skcipher_request_set_crypt(req, &sg, &sg, plen + 4, NULL);
        err = crypto_skcipher_decrypt(req);
        struct lib80211_tkip_data *tkey = priv;
        int keyidx;
        struct crypto_shash *tfm = tkey->tx_tfm_michael;
-       struct crypto_skcipher *tfm2 = tkey->tx_tfm_arc4;
+       struct crypto_sync_skcipher *tfm2 = tkey->tx_tfm_arc4;
        struct crypto_shash *tfm3 = tkey->rx_tfm_michael;
-       struct crypto_skcipher *tfm4 = tkey->rx_tfm_arc4;
+       struct crypto_sync_skcipher *tfm4 = tkey->rx_tfm_arc4;
 
        keyidx = tkey->key_idx;
        memset(tkey, 0, sizeof(*tkey));
 
        u8 key[WEP_KEY_LEN + 1];
        u8 key_len;
        u8 key_idx;
-       struct crypto_skcipher *tx_tfm;
-       struct crypto_skcipher *rx_tfm;
+       struct crypto_sync_skcipher *tx_tfm;
+       struct crypto_sync_skcipher *rx_tfm;
 };
 
 static void *lib80211_wep_init(int keyidx)
                goto fail;
        priv->key_idx = keyidx;
 
-       priv->tx_tfm = crypto_alloc_skcipher("ecb(arc4)", 0, CRYPTO_ALG_ASYNC);
+       priv->tx_tfm = crypto_alloc_sync_skcipher("ecb(arc4)", 0, 0);
        if (IS_ERR(priv->tx_tfm)) {
                priv->tx_tfm = NULL;
                goto fail;
        }
 
-       priv->rx_tfm = crypto_alloc_skcipher("ecb(arc4)", 0, CRYPTO_ALG_ASYNC);
+       priv->rx_tfm = crypto_alloc_sync_skcipher("ecb(arc4)", 0, 0);
        if (IS_ERR(priv->rx_tfm)) {
                priv->rx_tfm = NULL;
                goto fail;
 
       fail:
        if (priv) {
-               crypto_free_skcipher(priv->tx_tfm);
-               crypto_free_skcipher(priv->rx_tfm);
+               crypto_free_sync_skcipher(priv->tx_tfm);
+               crypto_free_sync_skcipher(priv->rx_tfm);
                kfree(priv);
        }
        return NULL;
 {
        struct lib80211_wep_data *_priv = priv;
        if (_priv) {
-               crypto_free_skcipher(_priv->tx_tfm);
-               crypto_free_skcipher(_priv->rx_tfm);
+               crypto_free_sync_skcipher(_priv->tx_tfm);
+               crypto_free_sync_skcipher(_priv->rx_tfm);
        }
        kfree(priv);
 }
 static int lib80211_wep_encrypt(struct sk_buff *skb, int hdr_len, void *priv)
 {
        struct lib80211_wep_data *wep = priv;
-       SKCIPHER_REQUEST_ON_STACK(req, wep->tx_tfm);
+       SYNC_SKCIPHER_REQUEST_ON_STACK(req, wep->tx_tfm);
        u32 crc, klen, len;
        u8 *pos, *icv;
        struct scatterlist sg;
        icv[2] = crc >> 16;
        icv[3] = crc >> 24;
 
-       crypto_skcipher_setkey(wep->tx_tfm, key, klen);
+       crypto_sync_skcipher_setkey(wep->tx_tfm, key, klen);
        sg_init_one(&sg, pos, len + 4);
-       skcipher_request_set_tfm(req, wep->tx_tfm);
+       skcipher_request_set_sync_tfm(req, wep->tx_tfm);
        skcipher_request_set_callback(req, 0, NULL, NULL);
        skcipher_request_set_crypt(req, &sg, &sg, len + 4, NULL);
        err = crypto_skcipher_encrypt(req);
 static int lib80211_wep_decrypt(struct sk_buff *skb, int hdr_len, void *priv)
 {
        struct lib80211_wep_data *wep = priv;
-       SKCIPHER_REQUEST_ON_STACK(req, wep->rx_tfm);
+       SYNC_SKCIPHER_REQUEST_ON_STACK(req, wep->rx_tfm);
        u32 crc, klen, plen;
        u8 key[WEP_KEY_LEN + 3];
        u8 keyidx, *pos, icv[4];
        /* Apply RC4 to data and compute CRC32 over decrypted data */
        plen = skb->len - hdr_len - 8;
 
-       crypto_skcipher_setkey(wep->rx_tfm, key, klen);
+       crypto_sync_skcipher_setkey(wep->rx_tfm, key, klen);
        sg_init_one(&sg, pos, plen + 4);
-       skcipher_request_set_tfm(req, wep->rx_tfm);
+       skcipher_request_set_sync_tfm(req, wep->rx_tfm);
        skcipher_request_set_callback(req, 0, NULL, NULL);
        skcipher_request_set_crypt(req, &sg, &sg, plen + 4, NULL);
        err = crypto_skcipher_decrypt(req);