]> git.ipfire.org Git - thirdparty/kernel/linux.git/commitdiff
netfs: Fix wait/wake to be consistent about the waitqueue used
authorDavid Howells <dhowells@redhat.com>
Mon, 19 May 2025 09:07:04 +0000 (10:07 +0100)
committerChristian Brauner <brauner@kernel.org>
Wed, 21 May 2025 12:35:21 +0000 (14:35 +0200)
Fix further inconsistencies in the use of waitqueues
(clear_and_wake_up_bit() vs private waitqueue).

Move some of this stuff from the read and write sides into common code so
that it can be done in fewer places.

To make this work, async I/O needs to set NETFS_RREQ_OFFLOAD_COLLECTION to
indicate that a workqueue will do the collecting and places that call the
wait function need to deal with it returning the amount transferred.

Fixes: e2d46f2ec332 ("netfs: Change the read result collector to only use one work item")
Signed-off-by: David Howells <dhowells@redhat.com>
Link: https://lore.kernel.org/20250519090707.2848510-5-dhowells@redhat.com
cc: Marc Dionne <marc.dionne@auristor.com>
cc: Steve French <stfrench@microsoft.com>
cc: Ihor Solodrai <ihor.solodrai@pm.me>
cc: Eric Van Hensbergen <ericvh@kernel.org>
cc: Latchesar Ionkov <lucho@ionkov.net>
cc: Dominique Martinet <asmadeus@codewreck.org>
cc: Christian Schoenebeck <linux_oss@crudebyte.com>
cc: Paulo Alcantara <pc@manguebit.com>
cc: Jeff Layton <jlayton@kernel.org>
cc: v9fs@lists.linux.dev
cc: linux-cifs@vger.kernel.org
cc: netfs@lists.linux.dev
cc: linux-fsdevel@vger.kernel.org
Signed-off-by: Christian Brauner <brauner@kernel.org>
fs/netfs/buffered_read.c
fs/netfs/buffered_write.c
fs/netfs/direct_read.c
fs/netfs/direct_write.c
fs/netfs/internal.h
fs/netfs/misc.c
fs/netfs/read_collect.c
fs/netfs/read_retry.c
fs/netfs/write_collect.c
fs/netfs/write_issue.c
fs/netfs/write_retry.c

index cb6202efc4668ff45f38df08fd692d72773e525a..fd4619275801be98f6a936a84a4724bf600b0dfe 100644 (file)
@@ -312,7 +312,7 @@ static void netfs_read_to_pagecache(struct netfs_io_request *rreq)
        if (unlikely(size > 0)) {
                smp_wmb(); /* Write lists before ALL_QUEUED. */
                set_bit(NETFS_RREQ_ALL_QUEUED, &rreq->flags);
-               netfs_wake_read_collector(rreq);
+               netfs_wake_collector(rreq);
        }
 
        /* Defer error return as we may need to wait for outstanding I/O. */
index b4826360a4111270f5b750540112d8d8fec860d3..dbb544e183d13db603188cc6d658c97f07ebcf3a 100644 (file)
@@ -386,7 +386,7 @@ out:
                wbc_detach_inode(&wbc);
                if (ret2 == -EIOCBQUEUED)
                        return ret2;
-               if (ret == 0)
+               if (ret == 0 && ret2 < 0)
                        ret = ret2;
        }
 
index cb3c6dc0b1654c4156c1c7515372b7be4800b837..a24e63d2c8186a6059201ee850aa20bb52871e01 100644 (file)
@@ -103,7 +103,7 @@ static int netfs_dispatch_unbuffered_reads(struct netfs_io_request *rreq)
                rreq->netfs_ops->issue_read(subreq);
 
                if (test_bit(NETFS_RREQ_PAUSE, &rreq->flags))
-                       netfs_wait_for_pause(rreq);
+                       netfs_wait_for_paused_read(rreq);
                if (test_bit(NETFS_RREQ_FAILED, &rreq->flags))
                        break;
                if (test_bit(NETFS_RREQ_BLOCKED, &rreq->flags) &&
@@ -115,7 +115,7 @@ static int netfs_dispatch_unbuffered_reads(struct netfs_io_request *rreq)
        if (unlikely(size > 0)) {
                smp_wmb(); /* Write lists before ALL_QUEUED. */
                set_bit(NETFS_RREQ_ALL_QUEUED, &rreq->flags);
-               netfs_wake_read_collector(rreq);
+               netfs_wake_collector(rreq);
        }
 
        return ret;
index c98f1676f86dfc66a2657c3a3221d09912378bf3..fa9a5bf3c6d512d905178fb6c920f538b1f6ee73 100644 (file)
@@ -87,6 +87,8 @@ ssize_t netfs_unbuffered_write_iter_locked(struct kiocb *iocb, struct iov_iter *
        }
 
        __set_bit(NETFS_RREQ_USE_IO_ITER, &wreq->flags);
+       if (async)
+               __set_bit(NETFS_RREQ_OFFLOAD_COLLECTION, &wreq->flags);
 
        /* Copy the data into the bounce buffer and encrypt it. */
        // TODO
@@ -105,13 +107,9 @@ ssize_t netfs_unbuffered_write_iter_locked(struct kiocb *iocb, struct iov_iter *
 
        if (!async) {
                trace_netfs_rreq(wreq, netfs_rreq_trace_wait_ip);
-               wait_on_bit(&wreq->flags, NETFS_RREQ_IN_PROGRESS,
-                           TASK_UNINTERRUPTIBLE);
-               ret = wreq->error;
-               if (ret == 0) {
-                       ret = wreq->transferred;
+               ret = netfs_wait_for_write(wreq);
+               if (ret > 0)
                        iocb->ki_pos += ret;
-               }
        } else {
                ret = -EIOCBQUEUED;
        }
index b6500a7cda81df16f9c80b8cbd92df054fa163e8..e2ee9183392b932ec5a05b6b7a59a3ac360870cc 100644 (file)
@@ -62,6 +62,14 @@ static inline void netfs_proc_del_rreq(struct netfs_io_request *rreq) {}
 struct folio_queue *netfs_buffer_make_space(struct netfs_io_request *rreq,
                                            enum netfs_folioq_trace trace);
 void netfs_reset_iter(struct netfs_io_subrequest *subreq);
+void netfs_wake_collector(struct netfs_io_request *rreq);
+void netfs_subreq_clear_in_progress(struct netfs_io_subrequest *subreq);
+void netfs_wait_for_in_progress_stream(struct netfs_io_request *rreq,
+                                      struct netfs_io_stream *stream);
+ssize_t netfs_wait_for_read(struct netfs_io_request *rreq);
+ssize_t netfs_wait_for_write(struct netfs_io_request *rreq);
+void netfs_wait_for_paused_read(struct netfs_io_request *rreq);
+void netfs_wait_for_paused_write(struct netfs_io_request *rreq);
 
 /*
  * objects.c
@@ -91,11 +99,9 @@ static inline void netfs_see_subrequest(struct netfs_io_subrequest *subreq,
 /*
  * read_collect.c
  */
+bool netfs_read_collection(struct netfs_io_request *rreq);
 void netfs_read_collection_worker(struct work_struct *work);
-void netfs_wake_read_collector(struct netfs_io_request *rreq);
 void netfs_cache_read_terminated(void *priv, ssize_t transferred_or_error);
-ssize_t netfs_wait_for_read(struct netfs_io_request *rreq);
-void netfs_wait_for_pause(struct netfs_io_request *rreq);
 
 /*
  * read_pgpriv2.c
@@ -175,8 +181,8 @@ static inline void netfs_stat_d(atomic_t *stat)
  * write_collect.c
  */
 int netfs_folio_written_back(struct folio *folio);
+bool netfs_write_collection(struct netfs_io_request *wreq);
 void netfs_write_collection_worker(struct work_struct *work);
-void netfs_wake_write_collector(struct netfs_io_request *wreq);
 
 /*
  * write_issue.c
@@ -197,8 +203,8 @@ struct netfs_io_request *netfs_begin_writethrough(struct kiocb *iocb, size_t len
 int netfs_advance_writethrough(struct netfs_io_request *wreq, struct writeback_control *wbc,
                               struct folio *folio, size_t copied, bool to_page_end,
                               struct folio **writethrough_cache);
-int netfs_end_writethrough(struct netfs_io_request *wreq, struct writeback_control *wbc,
-                          struct folio *writethrough_cache);
+ssize_t netfs_end_writethrough(struct netfs_io_request *wreq, struct writeback_control *wbc,
+                              struct folio *writethrough_cache);
 int netfs_unbuffered_write(struct netfs_io_request *wreq, bool may_wait, size_t len);
 
 /*
@@ -253,6 +259,21 @@ static inline void netfs_put_group_many(struct netfs_group *netfs_group, int nr)
                netfs_group->free(netfs_group);
 }
 
+/*
+ * Clear and wake up a NETFS_RREQ_* flag bit on a request.
+ */
+static inline void netfs_wake_rreq_flag(struct netfs_io_request *rreq,
+                                       unsigned int rreq_flag,
+                                       enum netfs_rreq_trace trace)
+{
+       if (test_bit(rreq_flag, &rreq->flags)) {
+               trace_netfs_rreq(rreq, trace);
+               clear_bit_unlock(rreq_flag, &rreq->flags);
+               smp_mb__after_atomic(); /* Set flag before task state */
+               wake_up(&rreq->waitq);
+       }
+}
+
 /*
  * fscache-cache.c
  */
index 7099aa07737ac07179c179c4b370120d2895e917..77e7f7c79d27c2a69758d8c86c745bce53cfb643 100644 (file)
@@ -313,3 +313,221 @@ bool netfs_release_folio(struct folio *folio, gfp_t gfp)
        return true;
 }
 EXPORT_SYMBOL(netfs_release_folio);
+
+/*
+ * Wake the collection work item.
+ */
+void netfs_wake_collector(struct netfs_io_request *rreq)
+{
+       if (test_bit(NETFS_RREQ_OFFLOAD_COLLECTION, &rreq->flags) &&
+           !test_bit(NETFS_RREQ_RETRYING, &rreq->flags)) {
+               queue_work(system_unbound_wq, &rreq->work);
+       } else {
+               trace_netfs_rreq(rreq, netfs_rreq_trace_wake_queue);
+               wake_up(&rreq->waitq);
+       }
+}
+
+/*
+ * Mark a subrequest as no longer being in progress and, if need be, wake the
+ * collector.
+ */
+void netfs_subreq_clear_in_progress(struct netfs_io_subrequest *subreq)
+{
+       struct netfs_io_request *rreq = subreq->rreq;
+       struct netfs_io_stream *stream = &rreq->io_streams[subreq->stream_nr];
+
+       clear_bit_unlock(NETFS_SREQ_IN_PROGRESS, &subreq->flags);
+       smp_mb__after_atomic(); /* Clear IN_PROGRESS before task state */
+
+       /* If we are at the head of the queue, wake up the collector. */
+       if (list_is_first(&subreq->rreq_link, &stream->subrequests) ||
+           test_bit(NETFS_RREQ_RETRYING, &rreq->flags))
+               netfs_wake_collector(rreq);
+}
+
+/*
+ * Wait for all outstanding I/O in a stream to quiesce.
+ */
+void netfs_wait_for_in_progress_stream(struct netfs_io_request *rreq,
+                                      struct netfs_io_stream *stream)
+{
+       struct netfs_io_subrequest *subreq;
+       DEFINE_WAIT(myself);
+
+       list_for_each_entry(subreq, &stream->subrequests, rreq_link) {
+               if (!test_bit(NETFS_SREQ_IN_PROGRESS, &subreq->flags))
+                       continue;
+
+               trace_netfs_rreq(rreq, netfs_rreq_trace_wait_queue);
+               for (;;) {
+                       prepare_to_wait(&rreq->waitq, &myself, TASK_UNINTERRUPTIBLE);
+
+                       if (!test_bit(NETFS_SREQ_IN_PROGRESS, &subreq->flags))
+                               break;
+
+                       trace_netfs_sreq(subreq, netfs_sreq_trace_wait_for);
+                       schedule();
+                       trace_netfs_rreq(rreq, netfs_rreq_trace_woke_queue);
+               }
+       }
+
+       finish_wait(&rreq->waitq, &myself);
+}
+
+/*
+ * Perform collection in app thread if not offloaded to workqueue.
+ */
+static int netfs_collect_in_app(struct netfs_io_request *rreq,
+                               bool (*collector)(struct netfs_io_request *rreq))
+{
+       bool need_collect = false, inactive = true;
+
+       for (int i = 0; i < NR_IO_STREAMS; i++) {
+               struct netfs_io_subrequest *subreq;
+               struct netfs_io_stream *stream = &rreq->io_streams[i];
+
+               if (!stream->active)
+                       continue;
+               inactive = false;
+               trace_netfs_collect_stream(rreq, stream);
+               subreq = list_first_entry_or_null(&stream->subrequests,
+                                                 struct netfs_io_subrequest,
+                                                 rreq_link);
+               if (subreq &&
+                   (!test_bit(NETFS_SREQ_IN_PROGRESS, &subreq->flags) ||
+                    test_bit(NETFS_SREQ_MADE_PROGRESS, &subreq->flags))) {
+                       need_collect = true;
+                       break;
+               }
+       }
+
+       if (!need_collect && !inactive)
+               return 0; /* Sleep */
+
+       __set_current_state(TASK_RUNNING);
+       if (collector(rreq)) {
+               /* Drop the ref from the NETFS_RREQ_IN_PROGRESS flag. */
+               netfs_put_request(rreq, netfs_rreq_trace_put_work_ip);
+               return 1; /* Done */
+       }
+
+       if (inactive) {
+               WARN(true, "Failed to collect inactive req R=%08x\n",
+                    rreq->debug_id);
+               cond_resched();
+       }
+       return 2; /* Again */
+}
+
+/*
+ * Wait for a request to complete, successfully or otherwise.
+ */
+static ssize_t netfs_wait_for_request(struct netfs_io_request *rreq,
+                                     bool (*collector)(struct netfs_io_request *rreq))
+{
+       DEFINE_WAIT(myself);
+       ssize_t ret;
+
+       for (;;) {
+               trace_netfs_rreq(rreq, netfs_rreq_trace_wait_queue);
+               prepare_to_wait(&rreq->waitq, &myself, TASK_UNINTERRUPTIBLE);
+
+               if (!test_bit(NETFS_RREQ_OFFLOAD_COLLECTION, &rreq->flags)) {
+                       switch (netfs_collect_in_app(rreq, collector)) {
+                       case 0:
+                               break;
+                       case 1:
+                               goto all_collected;
+                       case 2:
+                               continue;
+                       }
+               }
+
+               if (!test_bit(NETFS_RREQ_IN_PROGRESS, &rreq->flags))
+                       break;
+
+               schedule();
+               trace_netfs_rreq(rreq, netfs_rreq_trace_woke_queue);
+       }
+
+all_collected:
+       finish_wait(&rreq->waitq, &myself);
+
+       ret = rreq->error;
+       if (ret == 0) {
+               ret = rreq->transferred;
+               switch (rreq->origin) {
+               case NETFS_DIO_READ:
+               case NETFS_DIO_WRITE:
+               case NETFS_READ_SINGLE:
+               case NETFS_UNBUFFERED_WRITE:
+                       break;
+               default:
+                       if (rreq->submitted < rreq->len) {
+                               trace_netfs_failure(rreq, NULL, ret, netfs_fail_short_read);
+                               ret = -EIO;
+                       }
+                       break;
+               }
+       }
+
+       return ret;
+}
+
+ssize_t netfs_wait_for_read(struct netfs_io_request *rreq)
+{
+       return netfs_wait_for_request(rreq, netfs_read_collection);
+}
+
+ssize_t netfs_wait_for_write(struct netfs_io_request *rreq)
+{
+       return netfs_wait_for_request(rreq, netfs_write_collection);
+}
+
+/*
+ * Wait for a paused operation to unpause or complete in some manner.
+ */
+static void netfs_wait_for_pause(struct netfs_io_request *rreq,
+                                bool (*collector)(struct netfs_io_request *rreq))
+{
+       DEFINE_WAIT(myself);
+
+       trace_netfs_rreq(rreq, netfs_rreq_trace_wait_pause);
+
+       for (;;) {
+               trace_netfs_rreq(rreq, netfs_rreq_trace_wait_queue);
+               prepare_to_wait(&rreq->waitq, &myself, TASK_UNINTERRUPTIBLE);
+
+               if (!test_bit(NETFS_RREQ_OFFLOAD_COLLECTION, &rreq->flags)) {
+                       switch (netfs_collect_in_app(rreq, collector)) {
+                       case 0:
+                               break;
+                       case 1:
+                               goto all_collected;
+                       case 2:
+                               continue;
+                       }
+               }
+
+               if (!test_bit(NETFS_RREQ_IN_PROGRESS, &rreq->flags) ||
+                   !test_bit(NETFS_RREQ_PAUSE, &rreq->flags))
+                       break;
+
+               schedule();
+               trace_netfs_rreq(rreq, netfs_rreq_trace_woke_queue);
+       }
+
+all_collected:
+       finish_wait(&rreq->waitq, &myself);
+}
+
+void netfs_wait_for_paused_read(struct netfs_io_request *rreq)
+{
+       return netfs_wait_for_pause(rreq, netfs_read_collection);
+}
+
+void netfs_wait_for_paused_write(struct netfs_io_request *rreq)
+{
+       return netfs_wait_for_pause(rreq, netfs_write_collection);
+}
index 1197ebce56757dffcaaf0074d6a22ac003f7b8f5..900dd51c3b941cbf7d629a0c306335658444f0e6 100644 (file)
@@ -315,14 +315,8 @@ reassess:
 
        if (notes & NEED_RETRY)
                goto need_retry;
-       if ((notes & MADE_PROGRESS) && test_bit(NETFS_RREQ_PAUSE, &rreq->flags)) {
-               trace_netfs_rreq(rreq, netfs_rreq_trace_unpause);
-               clear_bit_unlock(NETFS_RREQ_PAUSE, &rreq->flags);
-               smp_mb__after_atomic(); /* Set PAUSE before task state */
-               wake_up(&rreq->waitq);
-       }
-
        if (notes & MADE_PROGRESS) {
+               netfs_wake_rreq_flag(rreq, NETFS_RREQ_PAUSE, netfs_rreq_trace_unpause);
                //cond_resched();
                goto reassess;
        }
@@ -399,7 +393,7 @@ static void netfs_rreq_assess_single(struct netfs_io_request *rreq)
  * Note that we're in normal kernel thread context at this point, possibly
  * running on a workqueue.
  */
-static bool netfs_read_collection(struct netfs_io_request *rreq)
+bool netfs_read_collection(struct netfs_io_request *rreq)
 {
        struct netfs_io_stream *stream = &rreq->io_streams[0];
 
@@ -434,8 +428,7 @@ static bool netfs_read_collection(struct netfs_io_request *rreq)
        }
        task_io_account_read(rreq->transferred);
 
-       trace_netfs_rreq(rreq, netfs_rreq_trace_wake_ip);
-       clear_and_wake_up_bit(NETFS_RREQ_IN_PROGRESS, &rreq->flags);
+       netfs_wake_rreq_flag(rreq, NETFS_RREQ_IN_PROGRESS, netfs_rreq_trace_wake_ip);
        /* As we cleared NETFS_RREQ_IN_PROGRESS, we acquired its ref. */
 
        trace_netfs_rreq(rreq, netfs_rreq_trace_done);
@@ -460,20 +453,6 @@ void netfs_read_collection_worker(struct work_struct *work)
        }
 }
 
-/*
- * Wake the collection work item.
- */
-void netfs_wake_read_collector(struct netfs_io_request *rreq)
-{
-       if (test_bit(NETFS_RREQ_OFFLOAD_COLLECTION, &rreq->flags) &&
-           !test_bit(NETFS_RREQ_RETRYING, &rreq->flags)) {
-               queue_work(system_unbound_wq, &rreq->work);
-       } else {
-               trace_netfs_rreq(rreq, netfs_rreq_trace_wake_queue);
-               wake_up(&rreq->waitq);
-       }
-}
-
 /**
  * netfs_read_subreq_progress - Note progress of a read operation.
  * @subreq: The read request that has terminated.
@@ -502,7 +481,7 @@ void netfs_read_subreq_progress(struct netfs_io_subrequest *subreq)
            list_is_first(&subreq->rreq_link, &stream->subrequests)
            ) {
                __set_bit(NETFS_SREQ_MADE_PROGRESS, &subreq->flags);
-               netfs_wake_read_collector(rreq);
+               netfs_wake_collector(rreq);
        }
 }
 EXPORT_SYMBOL(netfs_read_subreq_progress);
@@ -526,7 +505,6 @@ EXPORT_SYMBOL(netfs_read_subreq_progress);
 void netfs_read_subreq_terminated(struct netfs_io_subrequest *subreq)
 {
        struct netfs_io_request *rreq = subreq->rreq;
-       struct netfs_io_stream *stream = &rreq->io_streams[0];
 
        switch (subreq->source) {
        case NETFS_READ_FROM_CACHE:
@@ -573,15 +551,7 @@ void netfs_read_subreq_terminated(struct netfs_io_subrequest *subreq)
        }
 
        trace_netfs_sreq(subreq, netfs_sreq_trace_terminated);
-
-       clear_bit_unlock(NETFS_SREQ_IN_PROGRESS, &subreq->flags);
-       smp_mb__after_atomic(); /* Clear IN_PROGRESS before task state */
-
-       /* If we are at the head of the queue, wake up the collector. */
-       if (list_is_first(&subreq->rreq_link, &stream->subrequests) ||
-           test_bit(NETFS_RREQ_RETRYING, &rreq->flags))
-               netfs_wake_read_collector(rreq);
-
+       netfs_subreq_clear_in_progress(subreq);
        netfs_put_subrequest(subreq, netfs_sreq_trace_put_terminated);
 }
 EXPORT_SYMBOL(netfs_read_subreq_terminated);
@@ -604,102 +574,3 @@ void netfs_cache_read_terminated(void *priv, ssize_t transferred_or_error)
        }
        netfs_read_subreq_terminated(subreq);
 }
-
-/*
- * Wait for the read operation to complete, successfully or otherwise.
- */
-ssize_t netfs_wait_for_read(struct netfs_io_request *rreq)
-{
-       struct netfs_io_subrequest *subreq;
-       struct netfs_io_stream *stream = &rreq->io_streams[0];
-       DEFINE_WAIT(myself);
-       ssize_t ret;
-
-       for (;;) {
-               trace_netfs_rreq(rreq, netfs_rreq_trace_wait_queue);
-               prepare_to_wait(&rreq->waitq, &myself, TASK_UNINTERRUPTIBLE);
-
-               subreq = list_first_entry_or_null(&stream->subrequests,
-                                                 struct netfs_io_subrequest, rreq_link);
-               if (subreq &&
-                   (!test_bit(NETFS_SREQ_IN_PROGRESS, &subreq->flags) ||
-                    test_bit(NETFS_SREQ_MADE_PROGRESS, &subreq->flags))) {
-                       __set_current_state(TASK_RUNNING);
-                       if (netfs_read_collection(rreq)) {
-                               /* Drop the ref from the NETFS_RREQ_IN_PROGRESS flag. */
-                               netfs_put_request(rreq, netfs_rreq_trace_put_work_ip);
-                               break;
-                       }
-                       continue;
-               }
-
-               if (!test_bit(NETFS_RREQ_IN_PROGRESS, &rreq->flags))
-                       break;
-
-               schedule();
-               trace_netfs_rreq(rreq, netfs_rreq_trace_woke_queue);
-       }
-
-       finish_wait(&rreq->waitq, &myself);
-
-       ret = rreq->error;
-       if (ret == 0) {
-               ret = rreq->transferred;
-               switch (rreq->origin) {
-               case NETFS_DIO_READ:
-               case NETFS_READ_SINGLE:
-                       ret = rreq->transferred;
-                       break;
-               default:
-                       if (rreq->submitted < rreq->len) {
-                               trace_netfs_failure(rreq, NULL, ret, netfs_fail_short_read);
-                               ret = -EIO;
-                       }
-                       break;
-               }
-       }
-
-       return ret;
-}
-
-/*
- * Wait for a paused read operation to unpause or complete in some manner.
- */
-void netfs_wait_for_pause(struct netfs_io_request *rreq)
-{
-       struct netfs_io_subrequest *subreq;
-       struct netfs_io_stream *stream = &rreq->io_streams[0];
-       DEFINE_WAIT(myself);
-
-       trace_netfs_rreq(rreq, netfs_rreq_trace_wait_pause);
-
-       for (;;) {
-               trace_netfs_rreq(rreq, netfs_rreq_trace_wait_queue);
-               prepare_to_wait(&rreq->waitq, &myself, TASK_UNINTERRUPTIBLE);
-
-               if (!test_bit(NETFS_RREQ_OFFLOAD_COLLECTION, &rreq->flags)) {
-                       subreq = list_first_entry_or_null(&stream->subrequests,
-                                                         struct netfs_io_subrequest, rreq_link);
-                       if (subreq &&
-                           (!test_bit(NETFS_SREQ_IN_PROGRESS, &subreq->flags) ||
-                            test_bit(NETFS_SREQ_MADE_PROGRESS, &subreq->flags))) {
-                               __set_current_state(TASK_RUNNING);
-                               if (netfs_read_collection(rreq)) {
-                                       /* Drop the ref from the NETFS_RREQ_IN_PROGRESS flag. */
-                                       netfs_put_request(rreq, netfs_rreq_trace_put_work_ip);
-                                       break;
-                               }
-                               continue;
-                       }
-               }
-
-               if (!test_bit(NETFS_RREQ_IN_PROGRESS, &rreq->flags) ||
-                   !test_bit(NETFS_RREQ_PAUSE, &rreq->flags))
-                       break;
-
-               schedule();
-               trace_netfs_rreq(rreq, netfs_rreq_trace_woke_queue);
-       }
-
-       finish_wait(&rreq->waitq, &myself);
-}
index 1378dc7fa2ccdb37c365368002d720174bdb9b90..b99e84a8170af2791d325979088430984485a5a1 100644 (file)
@@ -257,35 +257,15 @@ abandon:
  */
 void netfs_retry_reads(struct netfs_io_request *rreq)
 {
-       struct netfs_io_subrequest *subreq;
        struct netfs_io_stream *stream = &rreq->io_streams[0];
-       DEFINE_WAIT(myself);
 
        netfs_stat(&netfs_n_rh_retry_read_req);
 
-       set_bit(NETFS_RREQ_RETRYING, &rreq->flags);
-
        /* Wait for all outstanding I/O to quiesce before performing retries as
         * we may need to renegotiate the I/O sizes.
         */
-       list_for_each_entry(subreq, &stream->subrequests, rreq_link) {
-               if (!test_bit(NETFS_SREQ_IN_PROGRESS, &subreq->flags))
-                       continue;
-
-               trace_netfs_rreq(rreq, netfs_rreq_trace_wait_queue);
-               for (;;) {
-                       prepare_to_wait(&rreq->waitq, &myself, TASK_UNINTERRUPTIBLE);
-
-                       if (!test_bit(NETFS_SREQ_IN_PROGRESS, &subreq->flags))
-                               break;
-
-                       trace_netfs_sreq(subreq, netfs_sreq_trace_wait_for);
-                       schedule();
-                       trace_netfs_rreq(rreq, netfs_rreq_trace_woke_queue);
-               }
-
-               finish_wait(&rreq->waitq, &myself);
-       }
+       set_bit(NETFS_RREQ_RETRYING, &rreq->flags);
+       netfs_wait_for_in_progress_stream(rreq, stream);
        clear_bit(NETFS_RREQ_RETRYING, &rreq->flags);
 
        trace_netfs_rreq(rreq, netfs_rreq_trace_resubmit);
index 7241d1fd2c14af2a7225ca669d7d44ad5db580f5..0ce7b53e7fe83f37f953da82c24ce9ee7e5922f6 100644 (file)
@@ -321,18 +321,14 @@ reassess_streams:
 
        if (notes & NEED_RETRY)
                goto need_retry;
-       if ((notes & MADE_PROGRESS) && test_bit(NETFS_RREQ_PAUSE, &wreq->flags)) {
-               trace_netfs_rreq(wreq, netfs_rreq_trace_unpause);
-               clear_bit_unlock(NETFS_RREQ_PAUSE, &wreq->flags);
-               smp_mb__after_atomic(); /* Set PAUSE before task state */
-               wake_up(&wreq->waitq);
-       }
 
-       if (notes & NEED_REASSESS) {
+       if (notes & MADE_PROGRESS) {
+               netfs_wake_rreq_flag(wreq, NETFS_RREQ_PAUSE, netfs_rreq_trace_unpause);
                //cond_resched();
                goto reassess_streams;
        }
-       if (notes & MADE_PROGRESS) {
+
+       if (notes & NEED_REASSESS) {
                //cond_resched();
                goto reassess_streams;
        }
@@ -356,7 +352,7 @@ need_retry:
 /*
  * Perform the collection of subrequests, folios and encryption buffers.
  */
-static bool netfs_write_collection(struct netfs_io_request *wreq)
+bool netfs_write_collection(struct netfs_io_request *wreq)
 {
        struct netfs_inode *ictx = netfs_inode(wreq->inode);
        size_t transferred;
@@ -417,8 +413,7 @@ static bool netfs_write_collection(struct netfs_io_request *wreq)
                inode_dio_end(wreq->inode);
 
        _debug("finished");
-       trace_netfs_rreq(wreq, netfs_rreq_trace_wake_ip);
-       clear_and_wake_up_bit(NETFS_RREQ_IN_PROGRESS, &wreq->flags);
+       netfs_wake_rreq_flag(wreq, NETFS_RREQ_IN_PROGRESS, netfs_rreq_trace_wake_ip);
        /* As we cleared NETFS_RREQ_IN_PROGRESS, we acquired its ref. */
 
        if (wreq->iocb) {
@@ -448,14 +443,6 @@ void netfs_write_collection_worker(struct work_struct *work)
        }
 }
 
-/*
- * Wake the collection work item.
- */
-void netfs_wake_write_collector(struct netfs_io_request *wreq)
-{
-       queue_work(system_unbound_wq, &wreq->work);
-}
-
 /**
  * netfs_write_subrequest_terminated - Note the termination of a write operation.
  * @_op: The I/O request that has terminated.
@@ -479,7 +466,6 @@ void netfs_write_subrequest_terminated(void *_op, ssize_t transferred_or_error)
 {
        struct netfs_io_subrequest *subreq = _op;
        struct netfs_io_request *wreq = subreq->rreq;
-       struct netfs_io_stream *stream = &wreq->io_streams[subreq->stream_nr];
 
        _enter("%x[%x] %zd", wreq->debug_id, subreq->debug_index, transferred_or_error);
 
@@ -531,15 +517,7 @@ void netfs_write_subrequest_terminated(void *_op, ssize_t transferred_or_error)
        }
 
        trace_netfs_sreq(subreq, netfs_sreq_trace_terminated);
-
-       clear_and_wake_up_bit(NETFS_SREQ_IN_PROGRESS, &subreq->flags);
-
-       /* If we are at the head of the queue, wake up the collector,
-        * transferring a ref to it if we were the ones to do so.
-        */
-       if (list_is_first(&subreq->rreq_link, &stream->subrequests))
-               netfs_wake_write_collector(wreq);
-
+       netfs_subreq_clear_in_progress(subreq);
        netfs_put_subrequest(subreq, netfs_sreq_trace_put_terminated);
 }
 EXPORT_SYMBOL(netfs_write_subrequest_terminated);
index 8744ed3faf29b646629d56fdfbb8b828e6c2b7b1..50bee2c4130d1eacf9899317a41abd836aa27a42 100644 (file)
@@ -542,7 +542,7 @@ static void netfs_end_issue_write(struct netfs_io_request *wreq)
        }
 
        if (needs_poke)
-               netfs_wake_write_collector(wreq);
+               netfs_wake_collector(wreq);
 }
 
 /*
@@ -576,6 +576,7 @@ int netfs_writepages(struct address_space *mapping,
                goto couldnt_start;
        }
 
+       __set_bit(NETFS_RREQ_OFFLOAD_COLLECTION, &wreq->flags);
        trace_netfs_write(wreq, netfs_write_trace_writeback);
        netfs_stat(&netfs_n_wh_writepages);
 
@@ -599,7 +600,7 @@ int netfs_writepages(struct address_space *mapping,
        netfs_end_issue_write(wreq);
 
        mutex_unlock(&ictx->wb_lock);
-       netfs_wake_write_collector(wreq);
+       netfs_wake_collector(wreq);
 
        netfs_put_request(wreq, netfs_rreq_trace_put_return);
        _leave(" = %d", error);
@@ -674,11 +675,11 @@ int netfs_advance_writethrough(struct netfs_io_request *wreq, struct writeback_c
 /*
  * End a write operation used when writing through the pagecache.
  */
-int netfs_end_writethrough(struct netfs_io_request *wreq, struct writeback_control *wbc,
-                          struct folio *writethrough_cache)
+ssize_t netfs_end_writethrough(struct netfs_io_request *wreq, struct writeback_control *wbc,
+                              struct folio *writethrough_cache)
 {
        struct netfs_inode *ictx = netfs_inode(wreq->inode);
-       int ret;
+       ssize_t ret;
 
        _enter("R=%x", wreq->debug_id);
 
@@ -689,12 +690,10 @@ int netfs_end_writethrough(struct netfs_io_request *wreq, struct writeback_contr
 
        mutex_unlock(&ictx->wb_lock);
 
-       if (wreq->iocb) {
+       if (wreq->iocb)
                ret = -EIOCBQUEUED;
-       } else {
-               wait_on_bit(&wreq->flags, NETFS_RREQ_IN_PROGRESS, TASK_UNINTERRUPTIBLE);
-               ret = wreq->error;
-       }
+       else
+               ret = netfs_wait_for_write(wreq);
        netfs_put_request(wreq, netfs_rreq_trace_put_return);
        return ret;
 }
@@ -723,10 +722,8 @@ int netfs_unbuffered_write(struct netfs_io_request *wreq, bool may_wait, size_t
                start += part;
                len -= part;
                rolling_buffer_advance(&wreq->buffer, part);
-               if (test_bit(NETFS_RREQ_PAUSE, &wreq->flags)) {
-                       trace_netfs_rreq(wreq, netfs_rreq_trace_wait_pause);
-                       wait_event(wreq->waitq, !test_bit(NETFS_RREQ_PAUSE, &wreq->flags));
-               }
+               if (test_bit(NETFS_RREQ_PAUSE, &wreq->flags))
+                       netfs_wait_for_paused_write(wreq);
                if (test_bit(NETFS_RREQ_FAILED, &wreq->flags))
                        break;
        }
@@ -886,6 +883,7 @@ int netfs_writeback_single(struct address_space *mapping,
                goto couldnt_start;
        }
 
+       __set_bit(NETFS_RREQ_OFFLOAD_COLLECTION, &wreq->flags);
        trace_netfs_write(wreq, netfs_write_trace_writeback_single);
        netfs_stat(&netfs_n_wh_writepages);
 
@@ -915,7 +913,7 @@ stop:
        set_bit(NETFS_RREQ_ALL_QUEUED, &wreq->flags);
 
        mutex_unlock(&ictx->wb_lock);
-       netfs_wake_write_collector(wreq);
+       netfs_wake_collector(wreq);
 
        netfs_put_request(wreq, netfs_rreq_trace_put_return);
        _leave(" = %d", ret);
index 7408f6bb8e42e5f27c07fad6dbdb63d45fd00d93..9d1d8a8bab726149d150c4588a106f15106384ea 100644 (file)
@@ -200,7 +200,6 @@ static void netfs_retry_write_stream(struct netfs_io_request *wreq,
  */
 void netfs_retry_writes(struct netfs_io_request *wreq)
 {
-       struct netfs_io_subrequest *subreq;
        struct netfs_io_stream *stream;
        int s;
 
@@ -209,16 +208,13 @@ void netfs_retry_writes(struct netfs_io_request *wreq)
        /* Wait for all outstanding I/O to quiesce before performing retries as
         * we may need to renegotiate the I/O sizes.
         */
+       set_bit(NETFS_RREQ_RETRYING, &wreq->flags);
        for (s = 0; s < NR_IO_STREAMS; s++) {
                stream = &wreq->io_streams[s];
-               if (!stream->active)
-                       continue;
-
-               list_for_each_entry(subreq, &stream->subrequests, rreq_link) {
-                       wait_on_bit(&subreq->flags, NETFS_SREQ_IN_PROGRESS,
-                                   TASK_UNINTERRUPTIBLE);
-               }
+               if (stream->active)
+                       netfs_wait_for_in_progress_stream(wreq, stream);
        }
+       clear_bit(NETFS_RREQ_RETRYING, &wreq->flags);
 
        // TODO: Enc: Fetch changed partial pages
        // TODO: Enc: Reencrypt content if needed.