* Enqueue the glock on the work queue.  Passes one glock reference on to the
  * work queue.
  */
-static void __gfs2_glock_queue_work(struct gfs2_glock *gl, unsigned long delay) {
+static void gfs2_glock_queue_work(struct gfs2_glock *gl, unsigned long delay) {
        if (!queue_delayed_work(glock_workqueue, &gl->gl_work, delay)) {
                /*
                 * We are holding the lockref spinlock, and the work was still
        }
 }
 
-static void gfs2_glock_queue_work(struct gfs2_glock *gl, unsigned long delay) {
-       spin_lock(&gl->gl_lockref.lock);
-       __gfs2_glock_queue_work(gl, delay);
-       spin_unlock(&gl->gl_lockref.lock);
-}
-
 static void __gfs2_glock_put(struct gfs2_glock *gl)
 {
        struct gfs2_sbd *sdp = gl->gl_name.ln_sbd;
        if (lockref_put_or_lock(&gl->gl_lockref))
                return;
 
-       __gfs2_glock_queue_work(gl, 0);
+       GLOCK_BUG_ON(gl, gl->gl_lockref.count != 1);
+       gfs2_glock_queue_work(gl, 0);
        spin_unlock(&gl->gl_lockref.lock);
 }
 
                         */
                        clear_bit(GLF_LOCK, &gl->gl_flags);
                        clear_bit(GLF_DEMOTE_IN_PROGRESS, &gl->gl_flags);
-                       __gfs2_glock_queue_work(gl, GL_GLOCK_DFT_HOLD);
+                       gfs2_glock_queue_work(gl, GL_GLOCK_DFT_HOLD);
                        return;
                } else {
                        clear_bit(GLF_INVALIDATE_IN_PROGRESS, &gl->gl_flags);
 
        /* Complete the operation now. */
        finish_xmote(gl, target);
-       __gfs2_glock_queue_work(gl, 0);
+       gfs2_glock_queue_work(gl, 0);
 }
 
 /**
        clear_bit(GLF_LOCK, &gl->gl_flags);
        smp_mb__after_atomic();
        gl->gl_lockref.count++;
-       __gfs2_glock_queue_work(gl, 0);
+       gfs2_glock_queue_work(gl, 0);
        return;
 
 out_unlock:
                drop_refs--;
                if (gl->gl_name.ln_type != LM_TYPE_INODE)
                        delay = 0;
-               __gfs2_glock_queue_work(gl, delay);
+               gfs2_glock_queue_work(gl, delay);
        }
 
        /*
         * Drop the remaining glock references manually here. (Mind that
-        * __gfs2_glock_queue_work depends on the lockref spinlock begin held
+        * gfs2_glock_queue_work depends on the lockref spinlock begin held
         * here as well.)
         */
        gl->gl_lockref.count -= drop_refs;
                     test_and_clear_bit(GLF_FROZEN, &gl->gl_flags))) {
                set_bit(GLF_REPLY_PENDING, &gl->gl_flags);
                gl->gl_lockref.count++;
-               __gfs2_glock_queue_work(gl, 0);
+               gfs2_glock_queue_work(gl, 0);
        }
        run_queue(gl, 1);
        spin_unlock(&gl->gl_lockref.lock);
                    !test_bit(GLF_DEMOTE, &gl->gl_flags) &&
                    gl->gl_name.ln_type == LM_TYPE_INODE)
                        delay = gl->gl_hold_time;
-               __gfs2_glock_queue_work(gl, delay);
+               gfs2_glock_queue_work(gl, delay);
        }
 }
 
                        delay = gl->gl_hold_time;
        }
        handle_callback(gl, state, delay, true);
-       __gfs2_glock_queue_work(gl, delay);
+       gfs2_glock_queue_work(gl, delay);
        spin_unlock(&gl->gl_lockref.lock);
 }
 
 
        gl->gl_lockref.count++;
        set_bit(GLF_REPLY_PENDING, &gl->gl_flags);
-       __gfs2_glock_queue_work(gl, 0);
+       gfs2_glock_queue_work(gl, 0);
        spin_unlock(&gl->gl_lockref.lock);
 }
 
                gl->gl_lockref.count++;
                if (demote_ok(gl))
                        handle_callback(gl, LM_ST_UNLOCKED, 0, false);
-               __gfs2_glock_queue_work(gl, 0);
+               gfs2_glock_queue_work(gl, 0);
                spin_unlock(&gl->gl_lockref.lock);
                cond_resched_lock(&lru_lock);
        }
 
        spin_lock(&gl->gl_lockref.lock);
        set_bit(GLF_REPLY_PENDING, &gl->gl_flags);
-       __gfs2_glock_queue_work(gl, 0);
+       gfs2_glock_queue_work(gl, 0);
        spin_unlock(&gl->gl_lockref.lock);
 }
 
                gl->gl_lockref.count++;
                if (gl->gl_state != LM_ST_UNLOCKED)
                        handle_callback(gl, LM_ST_UNLOCKED, 0, false);
-               __gfs2_glock_queue_work(gl, 0);
+               gfs2_glock_queue_work(gl, 0);
        }
        spin_unlock(&gl->gl_lockref.lock);
 }