Use kmem_cache_zalloc() directly.
With the exception of xlog_ticket_alloc() which will be dealt on the
next patch for readability.
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Carlos Maiolino <cmaiolino@redhat.com>
Reviewed-by: Darrick J. Wong <darrick.wong@oracle.com>
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
 
        ASSERT(btnum == XFS_BTNUM_BNO || btnum == XFS_BTNUM_CNT);
 
-       cur = kmem_zone_zalloc(xfs_btree_cur_zone, KM_NOFS);
+       cur = kmem_cache_zalloc(xfs_btree_cur_zone, GFP_NOFS | __GFP_NOFAIL);
 
        cur->bc_tp = tp;
        cur->bc_mp = mp;
 
        if (error)
                goto trans_cancel;
        ASSERT(ip->i_afp == NULL);
-       ip->i_afp = kmem_zone_zalloc(xfs_ifork_zone, 0);
+
+       ip->i_afp = kmem_cache_zalloc(xfs_ifork_zone,
+                                     GFP_KERNEL | __GFP_NOFAIL);
+
        ip->i_afp->if_format = XFS_DINODE_FMT_EXTENTS;
        ip->i_afp->if_flags = XFS_IFEXTENTS;
        logflags = 0;
 
        struct xfs_btree_cur    *cur;
        ASSERT(whichfork != XFS_COW_FORK);
 
-       cur = kmem_zone_zalloc(xfs_btree_cur_zone, KM_NOFS);
+       cur = kmem_cache_zalloc(xfs_btree_cur_zone, GFP_NOFS | __GFP_NOFAIL);
 
        cur->bc_tp = tp;
        cur->bc_mp = mp;
 
 xfs_da_state_t *
 xfs_da_state_alloc(void)
 {
-       return kmem_zone_zalloc(xfs_da_state_zone, KM_NOFS);
+       return kmem_cache_zalloc(xfs_da_state_zone, GFP_NOFS | __GFP_NOFAIL);
 }
 
 /*
 
 {
        struct xfs_btree_cur    *cur;
 
-       cur = kmem_zone_zalloc(xfs_btree_cur_zone, KM_NOFS);
+       cur = kmem_cache_zalloc(xfs_btree_cur_zone, GFP_NOFS | __GFP_NOFAIL);
        cur->bc_tp = tp;
        cur->bc_mp = mp;
        cur->bc_btnum = btnum;
 
         * Initialize the extent count early, as the per-format routines may
         * depend on it.
         */
-       ip->i_afp = kmem_zone_zalloc(xfs_ifork_zone, KM_NOFS);
+       ip->i_afp = kmem_cache_zalloc(xfs_ifork_zone, GFP_NOFS | __GFP_NOFAIL);
        ip->i_afp->if_format = dip->di_aformat;
        if (unlikely(ip->i_afp->if_format == 0)) /* pre IRIX 6.2 file system */
                ip->i_afp->if_format = XFS_DINODE_FMT_EXTENTS;
        if (ip->i_cowfp)
                return;
 
-       ip->i_cowfp = kmem_zone_zalloc(xfs_ifork_zone,
-                                      KM_NOFS);
+       ip->i_cowfp = kmem_cache_zalloc(xfs_ifork_zone,
+                                      GFP_NOFS | __GFP_NOFAIL);
        ip->i_cowfp->if_flags = XFS_IFEXTENTS;
        ip->i_cowfp->if_format = XFS_DINODE_FMT_EXTENTS;
 }
 
        ASSERT(agno != NULLAGNUMBER);
        ASSERT(agno < mp->m_sb.sb_agcount);
 
-       cur = kmem_zone_zalloc(xfs_btree_cur_zone, KM_NOFS);
+       cur = kmem_cache_zalloc(xfs_btree_cur_zone, GFP_NOFS | __GFP_NOFAIL);
        cur->bc_tp = tp;
        cur->bc_mp = mp;
        cur->bc_btnum = XFS_BTNUM_REFC;
 
 {
        struct xfs_btree_cur    *cur;
 
-       cur = kmem_zone_zalloc(xfs_btree_cur_zone, KM_NOFS);
+       cur = kmem_cache_zalloc(xfs_btree_cur_zone, GFP_NOFS | __GFP_NOFAIL);
        cur->bc_tp = tp;
        cur->bc_mp = mp;
        /* Overlapping btree; 2 keys per pointer. */
 
 {
        struct xfs_bui_log_item         *buip;
 
-       buip = kmem_zone_zalloc(xfs_bui_zone, 0);
+       buip = kmem_cache_zalloc(xfs_bui_zone, GFP_KERNEL | __GFP_NOFAIL);
 
        xfs_log_item_init(mp, &buip->bui_item, XFS_LI_BUI, &xfs_bui_item_ops);
        buip->bui_format.bui_nextents = XFS_BUI_MAX_FAST_EXTENTS;
 {
        struct xfs_bud_log_item         *budp;
 
-       budp = kmem_zone_zalloc(xfs_bud_zone, 0);
+       budp = kmem_cache_zalloc(xfs_bud_zone, GFP_KERNEL | __GFP_NOFAIL);
        xfs_log_item_init(tp->t_mountp, &budp->bud_item, XFS_LI_BUD,
                          &xfs_bud_item_ops);
        budp->bud_buip = buip;
 
        int                     i;
 
        *bpp = NULL;
-       bp = kmem_zone_zalloc(xfs_buf_zone, KM_NOFS);
-       if (unlikely(!bp))
-               return -ENOMEM;
+       bp = kmem_cache_zalloc(xfs_buf_zone, GFP_NOFS | __GFP_NOFAIL);
 
        /*
         * We don't want certain flags to appear in b_flags unless they are
 
                return 0;
        }
 
-       bip = kmem_zone_zalloc(xfs_buf_item_zone, 0);
+       bip = kmem_cache_zalloc(xfs_buf_item_zone, GFP_KERNEL | __GFP_NOFAIL);
        xfs_log_item_init(mp, &bip->bli_item, XFS_LI_BUF, &xfs_buf_item_ops);
        bip->bli_buf = bp;
 
 
 {
        struct xfs_dquot        *dqp;
 
-       dqp = kmem_zone_zalloc(xfs_qm_dqzone, 0);
+       dqp = kmem_cache_zalloc(xfs_qm_dqzone, GFP_KERNEL | __GFP_NOFAIL);
 
        dqp->q_type = type;
        dqp->q_id = id;
 
                        ((nextents - 1) * sizeof(xfs_extent_t)));
                efip = kmem_zalloc(size, 0);
        } else {
-               efip = kmem_zone_zalloc(xfs_efi_zone, 0);
+               efip = kmem_cache_zalloc(xfs_efi_zone,
+                                        GFP_KERNEL | __GFP_NOFAIL);
        }
 
        xfs_log_item_init(mp, &efip->efi_item, XFS_LI_EFI, &xfs_efi_item_ops);
                                (nextents - 1) * sizeof(struct xfs_extent),
                                0);
        } else {
-               efdp = kmem_zone_zalloc(xfs_efd_zone, 0);
+               efdp = kmem_cache_zalloc(xfs_efd_zone,
+                                       GFP_KERNEL | __GFP_NOFAIL);
        }
 
        xfs_log_item_init(tp->t_mountp, &efdp->efd_item, XFS_LI_EFD,
 
 {
        struct xfs_icreate_item *icp;
 
-       icp = kmem_zone_zalloc(xfs_icreate_zone, 0);
+       icp = kmem_cache_zalloc(xfs_icreate_zone, GFP_KERNEL | __GFP_NOFAIL);
 
        xfs_log_item_init(tp->t_mountp, &icp->ic_item, XFS_LI_ICREATE,
                          &xfs_icreate_item_ops);
 
        struct xfs_inode_log_item *iip;
 
        ASSERT(ip->i_itemp == NULL);
-       iip = ip->i_itemp = kmem_zone_zalloc(xfs_ili_zone, 0);
+       iip = ip->i_itemp = kmem_cache_zalloc(xfs_ili_zone,
+                                             GFP_KERNEL | __GFP_NOFAIL);
 
        iip->ili_inode = ip;
        spin_lock_init(&iip->ili_lock);
 
                cuip = kmem_zalloc(xfs_cui_log_item_sizeof(nextents),
                                0);
        else
-               cuip = kmem_zone_zalloc(xfs_cui_zone, 0);
+               cuip = kmem_cache_zalloc(xfs_cui_zone,
+                                        GFP_KERNEL | __GFP_NOFAIL);
 
        xfs_log_item_init(mp, &cuip->cui_item, XFS_LI_CUI, &xfs_cui_item_ops);
        cuip->cui_format.cui_nextents = nextents;
 {
        struct xfs_cud_log_item         *cudp;
 
-       cudp = kmem_zone_zalloc(xfs_cud_zone, 0);
+       cudp = kmem_cache_zalloc(xfs_cud_zone, GFP_KERNEL | __GFP_NOFAIL);
        xfs_log_item_init(tp->t_mountp, &cudp->cud_item, XFS_LI_CUD,
                          &xfs_cud_item_ops);
        cudp->cud_cuip = cuip;
 
        if (nextents > XFS_RUI_MAX_FAST_EXTENTS)
                ruip = kmem_zalloc(xfs_rui_log_item_sizeof(nextents), 0);
        else
-               ruip = kmem_zone_zalloc(xfs_rui_zone, 0);
+               ruip = kmem_cache_zalloc(xfs_rui_zone,
+                                        GFP_KERNEL | __GFP_NOFAIL);
 
        xfs_log_item_init(mp, &ruip->rui_item, XFS_LI_RUI, &xfs_rui_item_ops);
        ruip->rui_format.rui_nextents = nextents;
 {
        struct xfs_rud_log_item         *rudp;
 
-       rudp = kmem_zone_zalloc(xfs_rud_zone, 0);
+       rudp = kmem_cache_zalloc(xfs_rud_zone, GFP_KERNEL | __GFP_NOFAIL);
        xfs_log_item_init(tp->t_mountp, &rudp->rud_item, XFS_LI_RUD,
                          &xfs_rud_item_ops);
        rudp->rud_ruip = ruip;
 
 
        trace_xfs_trans_dup(tp, _RET_IP_);
 
-       ntp = kmem_zone_zalloc(xfs_trans_zone, 0);
+       ntp = kmem_cache_zalloc(xfs_trans_zone, GFP_KERNEL | __GFP_NOFAIL);
 
        /*
         * Initialize the new transaction structure.
         * GFP_NOFS allocation context so that we avoid lockdep false positives
         * by doing GFP_KERNEL allocations inside sb_start_intwrite().
         */
-       tp = kmem_zone_zalloc(xfs_trans_zone, 0);
+       tp = kmem_cache_zalloc(xfs_trans_zone, GFP_KERNEL | __GFP_NOFAIL);
        if (!(flags & XFS_TRANS_NO_WRITECOUNT))
                sb_start_intwrite(mp->m_super);
 
 
 xfs_trans_alloc_dqinfo(
        xfs_trans_t     *tp)
 {
-       tp->t_dqinfo = kmem_zone_zalloc(xfs_qm_dqtrxzone, 0);
+       tp->t_dqinfo = kmem_cache_zalloc(xfs_qm_dqtrxzone,
+                                        GFP_KERNEL | __GFP_NOFAIL);
 }
 
 void