href_node);
        }
 
-       head->processing = 1;
+       head->processing = true;
        WARN_ON(delayed_refs->num_heads_ready == 0);
        delayed_refs->num_heads_ready--;
        delayed_refs->run_delayed_start = head->bytenr +
        RB_CLEAR_NODE(&head->href_node);
        atomic_dec(&delayed_refs->num_entries);
        delayed_refs->num_heads--;
-       if (head->processing == 0)
+       if (!head->processing)
                delayed_refs->num_heads_ready--;
 }
 
                                  bool is_system)
 {
        int count_mod = 1;
-       int must_insert_reserved = 0;
+       bool must_insert_reserved = false;
 
        /* If reserved is provided, it must be a data extent. */
        BUG_ON(!is_data && reserved);
         * BTRFS_ADD_DELAYED_REF because other special casing is not required.
         */
        if (action == BTRFS_ADD_DELAYED_EXTENT)
-               must_insert_reserved = 1;
+               must_insert_reserved = true;
        else
-               must_insert_reserved = 0;
+               must_insert_reserved = false;
 
        refcount_set(&head_ref->refs, 1);
        head_ref->bytenr = bytenr;
        head_ref->ref_tree = RB_ROOT_CACHED;
        INIT_LIST_HEAD(&head_ref->ref_add_list);
        RB_CLEAR_NODE(&head_ref->href_node);
-       head_ref->processing = 0;
+       head_ref->processing = false;
        head_ref->total_ref_mod = count_mod;
        spin_lock_init(&head_ref->lock);
        mutex_init(&head_ref->mutex);
 
         * we need to update the in ram accounting to properly reflect
         * the free has happened.
         */
-       unsigned int must_insert_reserved:1;
-       unsigned int is_data:1;
-       unsigned int is_system:1;
-       unsigned int processing:1;
+       bool must_insert_reserved;
+       bool is_data;
+       bool is_system;
+       bool processing;
 };
 
 struct btrfs_delayed_tree_ref {
 
 static int run_delayed_data_ref(struct btrfs_trans_handle *trans,
                                struct btrfs_delayed_ref_node *node,
                                struct btrfs_delayed_extent_op *extent_op,
-                               int insert_reserved)
+                               bool insert_reserved)
 {
        int ret = 0;
        struct btrfs_delayed_data_ref *ref;
 static int run_delayed_tree_ref(struct btrfs_trans_handle *trans,
                                struct btrfs_delayed_ref_node *node,
                                struct btrfs_delayed_extent_op *extent_op,
-                               int insert_reserved)
+                               bool insert_reserved)
 {
        int ret = 0;
        struct btrfs_delayed_tree_ref *ref;
 static int run_one_delayed_ref(struct btrfs_trans_handle *trans,
                               struct btrfs_delayed_ref_node *node,
                               struct btrfs_delayed_extent_op *extent_op,
-                              int insert_reserved)
+                              bool insert_reserved)
 {
        int ret = 0;
 
                                      struct btrfs_delayed_ref_head *head)
 {
        spin_lock(&delayed_refs->lock);
-       head->processing = 0;
+       head->processing = false;
        delayed_refs->num_heads_ready++;
        spin_unlock(&delayed_refs->lock);
        btrfs_delayed_ref_unlock(head);
        struct btrfs_delayed_ref_root *delayed_refs;
        struct btrfs_delayed_extent_op *extent_op;
        struct btrfs_delayed_ref_node *ref;
-       int must_insert_reserved = 0;
+       bool must_insert_reserved;
        int ret;
 
        delayed_refs = &trans->transaction->delayed_refs;
                 * spin lock.
                 */
                must_insert_reserved = locked_ref->must_insert_reserved;
-               locked_ref->must_insert_reserved = 0;
+               locked_ref->must_insert_reserved = false;
 
                extent_op = locked_ref->extent_op;
                locked_ref->extent_op = NULL;
                goto out;
 
        btrfs_delete_ref_head(delayed_refs, head);
-       head->processing = 0;
+       head->processing = false;
 
        spin_unlock(&head->lock);
        spin_unlock(&delayed_refs->lock);