* boundary as appropriate.
  */
 static enum netfs_io_source cachefiles_prepare_read(struct netfs_io_subrequest *subreq,
-                                                   loff_t i_size)
+                                                   unsigned long long i_size)
 {
        return cachefiles_do_prepare_read(&subreq->rreq->cache_resources,
                                          subreq->start, &subreq->len, i_size,
 
         * block, but do not exceed the file size, unless the original
         * request already exceeds it.
         */
-       new_end = min(round_up(end, lo->stripe_unit), rreq->i_size);
+       new_end = umin(round_up(end, lo->stripe_unit), rreq->i_size);
        if (new_end > end && new_end <= rreq->start + max_len)
                rreq->len = new_end - rreq->start;
 
 
 }
 
 static void netfs_cache_expand_readahead(struct netfs_io_request *rreq,
-                                        loff_t *_start, size_t *_len, loff_t i_size)
+                                        unsigned long long *_start,
+                                        unsigned long long *_len,
+                                        unsigned long long i_size)
 {
        struct netfs_cache_resources *cres = &rreq->cache_resources;
 
 
        last = (wreq->start + wreq->len - 1) / PAGE_SIZE;
        xas_for_each(&xas, folio, last) {
                WARN(!folio_test_writeback(folio),
-                    "bad %zx @%llx page %lx %lx\n",
+                    "bad %llx @%llx page %lx %lx\n",
                     wreq->len, wreq->start, folio->index, last);
 
                if ((finfo = netfs_folio_info(folio))) {
 
 
 set:
        if (subreq->len > rreq->len)
-               pr_warn("R=%08x[%u] SREQ>RREQ %zx > %zx\n",
+               pr_warn("R=%08x[%u] SREQ>RREQ %zx > %llx\n",
                        rreq->debug_id, subreq->debug_index,
                        subreq->len, rreq->len);
 
        subreq->start           = rreq->start + rreq->submitted;
        subreq->len             = io_iter->count;
 
-       _debug("slice %llx,%zx,%zx", subreq->start, subreq->len, rreq->submitted);
+       _debug("slice %llx,%zx,%llx", subreq->start, subreq->len, rreq->submitted);
        list_add_tail(&subreq->rreq_link, &rreq->subrequests);
 
        /* Call out to the cache to find out what it can do with the remaining
        atomic_set(&rreq->nr_outstanding, 1);
        io_iter = rreq->io_iter;
        do {
-               _debug("submit %llx + %zx >= %llx",
+               _debug("submit %llx + %llx >= %llx",
                       rreq->start, rreq->submitted, rreq->i_size);
                if (rreq->origin == NETFS_DIO_READ &&
                    rreq->start + rreq->submitted >= rreq->i_size)
 
 
        rreq = list_entry(v, struct netfs_io_request, proc_link);
        seq_printf(m,
-                  "%08x %s %3d %2lx %4d %3d @%04llx %zx/%zx",
+                  "%08x %s %3d %2lx %4d %3d @%04llx %llx/%llx",
                   rreq->debug_id,
                   netfs_origins[rreq->origin],
                   refcount_read(&rreq->ref),
 
  */
 int netfs_advance_writethrough(struct netfs_io_request *wreq, size_t copied, bool to_page_end)
 {
-       _enter("ic=%zu sb=%zu ws=%u cp=%zu tp=%u",
+       _enter("ic=%zu sb=%llu ws=%u cp=%zu tp=%u",
               wreq->iter.count, wreq->submitted, wreq->wsize, copied, to_page_end);
 
        wreq->iter.count += copied;
 {
        int ret = -EIOCBQUEUED;
 
-       _enter("ic=%zu sb=%zu ws=%u",
+       _enter("ic=%zu sb=%llu ws=%u",
               wreq->iter.count, wreq->submitted, wreq->wsize);
 
        if (wreq->submitted < wreq->io_iter.count)
 
        struct work_struct      work;
        struct list_head        rreq_link;      /* Link in rreq->subrequests */
        struct iov_iter         io_iter;        /* Iterator for this subrequest */
-       loff_t                  start;          /* Where to start the I/O */
+       unsigned long long      start;          /* Where to start the I/O */
        size_t                  len;            /* Size of the I/O */
        size_t                  transferred;    /* Amount of data transferred */
        refcount_t              ref;
        atomic_t                subreq_counter; /* Next subreq->debug_index */
        atomic_t                nr_outstanding; /* Number of ops in progress */
        atomic_t                nr_copy_ops;    /* Number of copy-to-cache ops in progress */
-       size_t                  submitted;      /* Amount submitted for I/O so far */
-       size_t                  len;            /* Length of the request */
        size_t                  upper_len;      /* Length can be extended to here */
+       unsigned long long      submitted;      /* Amount submitted for I/O so far */
+       unsigned long long      len;            /* Length of the request */
        size_t                  transferred;    /* Amount to be indicated as transferred */
        short                   error;          /* 0 or error that occurred */
        enum netfs_io_origin    origin;         /* Origin of the request */
        bool                    direct_bv_unpin; /* T if direct_bv[] must be unpinned */
-       loff_t                  i_size;         /* Size of the file */
-       loff_t                  start;          /* Start position */
+       unsigned long long      i_size;         /* Size of the file */
+       unsigned long long      start;          /* Start position */
        pgoff_t                 no_unlock_folio; /* Don't unlock this folio after read */
        refcount_t              ref;
        unsigned long           flags;
 
        /* Expand readahead request */
        void (*expand_readahead)(struct netfs_cache_resources *cres,
-                                loff_t *_start, size_t *_len, loff_t i_size);
+                                unsigned long long *_start,
+                                unsigned long long *_len,
+                                unsigned long long i_size);
 
        /* Prepare a read operation, shortening it to a cached/uncached
         * boundary as appropriate.
         */
        enum netfs_io_source (*prepare_read)(struct netfs_io_subrequest *subreq,
-                                            loff_t i_size);
+                                            unsigned long long i_size);
 
        /* Prepare a write operation, working out what part of the write we can
         * actually do.
 
                    __entry->start      = sreq->start;
                           ),
 
-           TP_printk("R=%08x[%u] %s %s f=%02x s=%llx %zx/%zx e=%d",
+           TP_printk("R=%08x[%x] %s %s f=%02x s=%llx %zx/%zx e=%d",
                      __entry->rreq, __entry->index,
                      __print_symbolic(__entry->source, netfs_sreq_sources),
                      __print_symbolic(__entry->what, netfs_sreq_traces),
                    __entry->start      = sreq ? sreq->start : 0;
                           ),
 
-           TP_printk("R=%08x[%d] %s f=%02x s=%llx %zx/%zx %s e=%d",
+           TP_printk("R=%08x[%x] %s f=%02x s=%llx %zx/%zx %s e=%d",
                      __entry->rreq, __entry->index,
                      __print_symbolic(__entry->source, netfs_sreq_sources),
                      __entry->flags,
                    __field(unsigned int,               cookie          )
                    __field(enum netfs_write_trace,     what            )
                    __field(unsigned long long,         start           )
-                   __field(size_t,                     len             )
+                   __field(unsigned long long,         len             )
                             ),
 
            TP_fast_assign(