}
        }
 
-       if (need_mm && mmget_not_zero(alloc->vma_vm_mm))
-               mm = alloc->vma_vm_mm;
+       if (need_mm && mmget_not_zero(alloc->mm))
+               mm = alloc->mm;
 
        if (mm) {
                mmap_read_lock(mm);
         */
        if (vma) {
                vm_start = vma->vm_start;
-               mmap_assert_write_locked(alloc->vma_vm_mm);
+               mmap_assert_write_locked(alloc->mm);
        } else {
-               mmap_assert_locked(alloc->vma_vm_mm);
+               mmap_assert_locked(alloc->mm);
        }
 
        alloc->vma_addr = vm_start;
        struct vm_area_struct *vma = NULL;
 
        if (alloc->vma_addr)
-               vma = vma_lookup(alloc->vma_vm_mm, alloc->vma_addr);
+               vma = vma_lookup(alloc->mm, alloc->vma_addr);
 
        return vma;
 }
        size_t size, data_offsets_size;
        int ret;
 
-       mmap_read_lock(alloc->vma_vm_mm);
+       mmap_read_lock(alloc->mm);
        if (!binder_alloc_get_vma(alloc)) {
-               mmap_read_unlock(alloc->vma_vm_mm);
+               mmap_read_unlock(alloc->mm);
                binder_alloc_debug(BINDER_DEBUG_USER_ERROR,
                                   "%d: binder_alloc_buf, no vma\n",
                                   alloc->pid);
                return ERR_PTR(-ESRCH);
        }
-       mmap_read_unlock(alloc->vma_vm_mm);
+       mmap_read_unlock(alloc->mm);
 
        data_offsets_size = ALIGN(data_size, sizeof(void *)) +
                ALIGN(offsets_size, sizeof(void *));
        buffers = 0;
        mutex_lock(&alloc->mutex);
        BUG_ON(alloc->vma_addr &&
-              vma_lookup(alloc->vma_vm_mm, alloc->vma_addr));
+              vma_lookup(alloc->mm, alloc->vma_addr));
 
        while ((n = rb_first(&alloc->allocated_buffers))) {
                buffer = rb_entry(n, struct binder_buffer, rb_node);
                kfree(alloc->pages);
        }
        mutex_unlock(&alloc->mutex);
-       if (alloc->vma_vm_mm)
-               mmdrop(alloc->vma_vm_mm);
+       if (alloc->mm)
+               mmdrop(alloc->mm);
 
        binder_alloc_debug(BINDER_DEBUG_OPEN_CLOSE,
                     "%s: %d buffers %d, pages %d\n",
         * read inconsistent state.
         */
 
-       mmap_read_lock(alloc->vma_vm_mm);
+       mmap_read_lock(alloc->mm);
        if (binder_alloc_get_vma(alloc) == NULL) {
-               mmap_read_unlock(alloc->vma_vm_mm);
+               mmap_read_unlock(alloc->mm);
                goto uninitialized;
        }
 
-       mmap_read_unlock(alloc->vma_vm_mm);
+       mmap_read_unlock(alloc->mm);
        for (i = 0; i < alloc->buffer_size / PAGE_SIZE; i++) {
                page = &alloc->pages[i];
                if (!page->page_ptr)
        index = page - alloc->pages;
        page_addr = (uintptr_t)alloc->buffer + index * PAGE_SIZE;
 
-       mm = alloc->vma_vm_mm;
+       mm = alloc->mm;
        if (!mmget_not_zero(mm))
                goto err_mmget;
        if (!mmap_read_trylock(mm))
 void binder_alloc_init(struct binder_alloc *alloc)
 {
        alloc->pid = current->group_leader->pid;
-       alloc->vma_vm_mm = current->mm;
-       mmgrab(alloc->vma_vm_mm);
+       alloc->mm = current->mm;
+       mmgrab(alloc->mm);
        mutex_init(&alloc->mutex);
        INIT_LIST_HEAD(&alloc->buffers);
 }