#define ctx_to_ppgtt(ctx) container_of((ctx)->vm, struct i915_hw_ppgtt, base)
 #define vm_to_ppgtt(vm) container_of(vm, struct i915_hw_ppgtt, base)
 int __must_check i915_gem_context_init(struct drm_device *dev);
-void ppgtt_release(struct kref *kref);
 void i915_gem_context_fini(struct drm_device *dev);
 void i915_gem_context_reset(struct drm_device *dev);
 int i915_gem_context_open(struct drm_device *dev, struct drm_file *file);
 
 void i915_gem_vma_destroy(struct i915_vma *vma)
 {
        struct i915_address_space *vm = NULL;
-       struct i915_hw_ppgtt *ppgtt = NULL;
        WARN_ON(vma->node.allocated);
 
        /* Keep the vma as a placeholder in the execbuffer reservation lists */
                return;
 
        vm = vma->vm;
-       ppgtt = vm_to_ppgtt(vm);
 
-       if (ppgtt)
-               kref_put(&ppgtt->ref, ppgtt_release);
+       i915_ppgtt_put(vm_to_ppgtt(vm));
 
        list_del(&vma->vma_link);
 
 
 #define GEN6_CONTEXT_ALIGN (64<<10)
 #define GEN7_CONTEXT_ALIGN 4096
 
-static void do_ppgtt_cleanup(struct i915_hw_ppgtt *ppgtt)
-{
-       struct drm_device *dev = ppgtt->base.dev;
-       struct drm_i915_private *dev_priv = dev->dev_private;
-       struct i915_address_space *vm = &ppgtt->base;
-
-       if (ppgtt == dev_priv->mm.aliasing_ppgtt ||
-           (list_empty(&vm->active_list) && list_empty(&vm->inactive_list))) {
-               ppgtt->base.cleanup(&ppgtt->base);
-               return;
-       }
-
-       /* vmas should already be unbound */
-       WARN_ON(!list_empty(&vm->active_list));
-
-       ppgtt->base.cleanup(&ppgtt->base);
-}
-
-void ppgtt_release(struct kref *kref)
-{
-       struct i915_hw_ppgtt *ppgtt =
-               container_of(kref, struct i915_hw_ppgtt, ref);
-
-       do_ppgtt_cleanup(ppgtt);
-       kfree(ppgtt);
-}
-
 static size_t get_context_alignment(struct drm_device *dev)
 {
        if (IS_GEN6(dev))
                        ppgtt = ctx_to_ppgtt(ctx);
        }
 
-       if (ppgtt)
-               kref_put(&ppgtt->ref, ppgtt_release);
+       i915_ppgtt_put(ppgtt);
        if (ctx->legacy_hw_ctx.rcs_state)
                drm_gem_object_unreference(&ctx->legacy_hw_ctx.rcs_state->base);
        list_del(&ctx->link);
        if (!ppgtt)
                return ERR_PTR(-ENOMEM);
 
-       ret = i915_gem_init_ppgtt(dev, ppgtt);
+       ret = i915_ppgtt_init(dev, ppgtt);
        if (ret) {
                kfree(ppgtt);
                return ERR_PTR(ret);
 
 static struct intel_context *
 __create_hw_context(struct drm_device *dev,
-                 struct drm_i915_file_private *file_priv)
+                   struct drm_i915_file_private *file_priv)
 {
        struct drm_i915_private *dev_priv = dev->dev_private;
        struct intel_context *ctx;
                /* For platforms which only have aliasing PPGTT, we fake the
                 * address space and refcounting. */
                ctx->vm = &dev_priv->mm.aliasing_ppgtt->base;
-               kref_get(&dev_priv->mm.aliasing_ppgtt->ref);
+               i915_ppgtt_get(dev_priv->mm.aliasing_ppgtt);
        } else
                ctx->vm = &dev_priv->gtt.base;
 
 
        return 0;
 }
 
-int i915_gem_init_ppgtt(struct drm_device *dev, struct i915_hw_ppgtt *ppgtt)
+int i915_ppgtt_init(struct drm_device *dev, struct i915_hw_ppgtt *ppgtt)
 {
        struct drm_i915_private *dev_priv = dev->dev_private;
        int ret = 0;
        return ret;
 }
 
+void  i915_ppgtt_release(struct kref *kref)
+{
+       struct i915_hw_ppgtt *ppgtt =
+               container_of(kref, struct i915_hw_ppgtt, ref);
+
+       /* vmas should already be unbound */
+       WARN_ON(!list_empty(&ppgtt->base.active_list));
+       WARN_ON(!list_empty(&ppgtt->base.inactive_list));
+
+       ppgtt->base.cleanup(&ppgtt->base);
+       kfree(ppgtt);
+}
+
 static void
 ppgtt_bind_vma(struct i915_vma *vma,
               enum i915_cache_level cache_level,
                                  struct i915_address_space *vm)
 {
        struct i915_vma *vma;
-       struct i915_hw_ppgtt *ppgtt = NULL;
 
        vma = i915_gem_obj_to_vma(obj, vm);
        if (!vma)
                vma = __i915_gem_vma_create(obj, vm);
 
-       ppgtt = vm_to_ppgtt(vm);
-       if (ppgtt)
-               kref_get(&ppgtt->ref);
+       i915_ppgtt_get(vm_to_ppgtt(vm));
 
        return vma;
 }
 
 void i915_gem_setup_global_gtt(struct drm_device *dev, unsigned long start,
                               unsigned long mappable_end, unsigned long end);
 
-int i915_gem_init_ppgtt(struct drm_device *dev, struct i915_hw_ppgtt *ppgtt);
+
+int i915_ppgtt_init(struct drm_device *dev, struct i915_hw_ppgtt *ppgtt);
+void i915_ppgtt_release(struct kref *kref);
+static inline void i915_ppgtt_get(struct i915_hw_ppgtt *ppgtt)
+{
+       if (ppgtt)
+               kref_get(&ppgtt->ref);
+}
+static inline void i915_ppgtt_put(struct i915_hw_ppgtt *ppgtt)
+{
+       if (ppgtt)
+               kref_put(&ppgtt->ref, i915_ppgtt_release);
+}
 
 void i915_check_and_clear_faults(struct drm_device *dev);
 void i915_gem_suspend_gtt_mappings(struct drm_device *dev);