mirror of
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-08-05 16:54:27 +00:00

Make a number of updates to the netfs tracepoints: (1) Remove a duplicate trace from netfs_unbuffered_write_iter_locked(). (2) Move the trace in netfs_wake_rreq_flag() to after the flag is cleared so that the change appears in the trace. (3) Differentiate the use of netfs_rreq_trace_wait/woke_queue symbols. (4) Don't do so many trace emissions in the wait functions as some of them are redundant. (5) In netfs_collect_read_results(), differentiate a subreq that's being abandoned vs one that has been consumed in a regular way. (6) Add a tracepoint to indicate the call to ->ki_complete(). (7) Don't double-increment the subreq_counter when retrying a write. (8) Move the netfs_sreq_trace_io_progress tracepoint within cifs code to just MID_RESPONSE_RECEIVED and add different tracepoints for other MID states and note check failure. Signed-off-by: David Howells <dhowells@redhat.com> Co-developed-by: Paulo Alcantara <pc@manguebit.org> Signed-off-by: Paulo Alcantara <pc@manguebit.org> Link: https://lore.kernel.org/20250701163852.2171681-14-dhowells@redhat.com cc: Steve French <sfrench@samba.org> cc: netfs@lists.linux.dev cc: linux-fsdevel@vger.kernel.org cc: linux-cifs@vger.kernel.org Signed-off-by: Christian Brauner <brauner@kernel.org>
583 lines
18 KiB
C
583 lines
18 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/* Network filesystem read subrequest result collection, assessment and
|
|
* retrying.
|
|
*
|
|
* Copyright (C) 2024 Red Hat, Inc. All Rights Reserved.
|
|
* Written by David Howells (dhowells@redhat.com)
|
|
*/
|
|
|
|
#include <linux/export.h>
|
|
#include <linux/fs.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/pagemap.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/task_io_accounting_ops.h>
|
|
#include "internal.h"
|
|
|
|
/* Notes made in the collector */
|
|
#define HIT_PENDING 0x01 /* A front op was still pending */
|
|
#define MADE_PROGRESS 0x04 /* Made progress cleaning up a stream or the folio set */
|
|
#define BUFFERED 0x08 /* The pagecache needs cleaning up */
|
|
#define NEED_RETRY 0x10 /* A front op requests retrying */
|
|
#define COPY_TO_CACHE 0x40 /* Need to copy subrequest to cache */
|
|
#define ABANDON_SREQ 0x80 /* Need to abandon untransferred part of subrequest */
|
|
|
|
/*
|
|
* Clear the unread part of an I/O request.
|
|
*/
|
|
static void netfs_clear_unread(struct netfs_io_subrequest *subreq)
|
|
{
|
|
netfs_reset_iter(subreq);
|
|
WARN_ON_ONCE(subreq->len - subreq->transferred != iov_iter_count(&subreq->io_iter));
|
|
iov_iter_zero(iov_iter_count(&subreq->io_iter), &subreq->io_iter);
|
|
if (subreq->start + subreq->transferred >= subreq->rreq->i_size)
|
|
__set_bit(NETFS_SREQ_HIT_EOF, &subreq->flags);
|
|
}
|
|
|
|
/*
|
|
* Flush, mark and unlock a folio that's now completely read. If we want to
|
|
* cache the folio, we set the group to NETFS_FOLIO_COPY_TO_CACHE, mark it
|
|
* dirty and let writeback handle it.
|
|
*/
|
|
static void netfs_unlock_read_folio(struct netfs_io_request *rreq,
|
|
struct folio_queue *folioq,
|
|
int slot)
|
|
{
|
|
struct netfs_folio *finfo;
|
|
struct folio *folio = folioq_folio(folioq, slot);
|
|
|
|
if (unlikely(folio_pos(folio) < rreq->abandon_to)) {
|
|
trace_netfs_folio(folio, netfs_folio_trace_abandon);
|
|
goto just_unlock;
|
|
}
|
|
|
|
flush_dcache_folio(folio);
|
|
folio_mark_uptodate(folio);
|
|
|
|
if (!test_bit(NETFS_RREQ_USE_PGPRIV2, &rreq->flags)) {
|
|
finfo = netfs_folio_info(folio);
|
|
if (finfo) {
|
|
trace_netfs_folio(folio, netfs_folio_trace_filled_gaps);
|
|
if (finfo->netfs_group)
|
|
folio_change_private(folio, finfo->netfs_group);
|
|
else
|
|
folio_detach_private(folio);
|
|
kfree(finfo);
|
|
}
|
|
|
|
if (test_bit(NETFS_RREQ_FOLIO_COPY_TO_CACHE, &rreq->flags)) {
|
|
if (!WARN_ON_ONCE(folio_get_private(folio) != NULL)) {
|
|
trace_netfs_folio(folio, netfs_folio_trace_copy_to_cache);
|
|
folio_attach_private(folio, NETFS_FOLIO_COPY_TO_CACHE);
|
|
folio_mark_dirty(folio);
|
|
}
|
|
} else {
|
|
trace_netfs_folio(folio, netfs_folio_trace_read_done);
|
|
}
|
|
|
|
folioq_clear(folioq, slot);
|
|
} else {
|
|
// TODO: Use of PG_private_2 is deprecated.
|
|
if (test_bit(NETFS_RREQ_FOLIO_COPY_TO_CACHE, &rreq->flags))
|
|
netfs_pgpriv2_copy_to_cache(rreq, folio);
|
|
}
|
|
|
|
just_unlock:
|
|
if (folio->index == rreq->no_unlock_folio &&
|
|
test_bit(NETFS_RREQ_NO_UNLOCK_FOLIO, &rreq->flags)) {
|
|
_debug("no unlock");
|
|
} else {
|
|
trace_netfs_folio(folio, netfs_folio_trace_read_unlock);
|
|
folio_unlock(folio);
|
|
}
|
|
|
|
folioq_clear(folioq, slot);
|
|
}
|
|
|
|
/*
|
|
* Unlock any folios we've finished with.
|
|
*/
|
|
static void netfs_read_unlock_folios(struct netfs_io_request *rreq,
|
|
unsigned int *notes)
|
|
{
|
|
struct folio_queue *folioq = rreq->buffer.tail;
|
|
unsigned long long collected_to = rreq->collected_to;
|
|
unsigned int slot = rreq->buffer.first_tail_slot;
|
|
|
|
if (rreq->cleaned_to >= rreq->collected_to)
|
|
return;
|
|
|
|
// TODO: Begin decryption
|
|
|
|
if (slot >= folioq_nr_slots(folioq)) {
|
|
folioq = rolling_buffer_delete_spent(&rreq->buffer);
|
|
if (!folioq) {
|
|
rreq->front_folio_order = 0;
|
|
return;
|
|
}
|
|
slot = 0;
|
|
}
|
|
|
|
for (;;) {
|
|
struct folio *folio;
|
|
unsigned long long fpos, fend;
|
|
unsigned int order;
|
|
size_t fsize;
|
|
|
|
if (*notes & COPY_TO_CACHE)
|
|
set_bit(NETFS_RREQ_FOLIO_COPY_TO_CACHE, &rreq->flags);
|
|
|
|
folio = folioq_folio(folioq, slot);
|
|
if (WARN_ONCE(!folio_test_locked(folio),
|
|
"R=%08x: folio %lx is not locked\n",
|
|
rreq->debug_id, folio->index))
|
|
trace_netfs_folio(folio, netfs_folio_trace_not_locked);
|
|
|
|
order = folioq_folio_order(folioq, slot);
|
|
rreq->front_folio_order = order;
|
|
fsize = PAGE_SIZE << order;
|
|
fpos = folio_pos(folio);
|
|
fend = umin(fpos + fsize, rreq->i_size);
|
|
|
|
trace_netfs_collect_folio(rreq, folio, fend, collected_to);
|
|
|
|
/* Unlock any folio we've transferred all of. */
|
|
if (collected_to < fend)
|
|
break;
|
|
|
|
netfs_unlock_read_folio(rreq, folioq, slot);
|
|
WRITE_ONCE(rreq->cleaned_to, fpos + fsize);
|
|
*notes |= MADE_PROGRESS;
|
|
|
|
clear_bit(NETFS_RREQ_FOLIO_COPY_TO_CACHE, &rreq->flags);
|
|
|
|
/* Clean up the head folioq. If we clear an entire folioq, then
|
|
* we can get rid of it provided it's not also the tail folioq
|
|
* being filled by the issuer.
|
|
*/
|
|
folioq_clear(folioq, slot);
|
|
slot++;
|
|
if (slot >= folioq_nr_slots(folioq)) {
|
|
folioq = rolling_buffer_delete_spent(&rreq->buffer);
|
|
if (!folioq)
|
|
goto done;
|
|
slot = 0;
|
|
trace_netfs_folioq(folioq, netfs_trace_folioq_read_progress);
|
|
}
|
|
|
|
if (fpos + fsize >= collected_to)
|
|
break;
|
|
}
|
|
|
|
rreq->buffer.tail = folioq;
|
|
done:
|
|
rreq->buffer.first_tail_slot = slot;
|
|
}
|
|
|
|
/*
|
|
* Collect and assess the results of various read subrequests. We may need to
|
|
* retry some of the results.
|
|
*
|
|
* Note that we have a sequence of subrequests, which may be drawing on
|
|
* different sources and may or may not be the same size or starting position
|
|
* and may not even correspond in boundary alignment.
|
|
*/
|
|
static void netfs_collect_read_results(struct netfs_io_request *rreq)
|
|
{
|
|
struct netfs_io_subrequest *front, *remove;
|
|
struct netfs_io_stream *stream = &rreq->io_streams[0];
|
|
unsigned int notes;
|
|
|
|
_enter("%llx-%llx", rreq->start, rreq->start + rreq->len);
|
|
trace_netfs_rreq(rreq, netfs_rreq_trace_collect);
|
|
trace_netfs_collect(rreq);
|
|
|
|
reassess:
|
|
if (rreq->origin == NETFS_READAHEAD ||
|
|
rreq->origin == NETFS_READPAGE ||
|
|
rreq->origin == NETFS_READ_FOR_WRITE)
|
|
notes = BUFFERED;
|
|
else
|
|
notes = 0;
|
|
|
|
/* Remove completed subrequests from the front of the stream and
|
|
* advance the completion point. We stop when we hit something that's
|
|
* in progress. The issuer thread may be adding stuff to the tail
|
|
* whilst we're doing this.
|
|
*/
|
|
front = READ_ONCE(stream->front);
|
|
while (front) {
|
|
size_t transferred;
|
|
|
|
trace_netfs_collect_sreq(rreq, front);
|
|
_debug("sreq [%x] %llx %zx/%zx",
|
|
front->debug_index, front->start, front->transferred, front->len);
|
|
|
|
if (stream->collected_to < front->start) {
|
|
trace_netfs_collect_gap(rreq, stream, front->start, 'F');
|
|
stream->collected_to = front->start;
|
|
}
|
|
|
|
if (netfs_check_subreq_in_progress(front))
|
|
notes |= HIT_PENDING;
|
|
smp_rmb(); /* Read counters after IN_PROGRESS flag. */
|
|
transferred = READ_ONCE(front->transferred);
|
|
|
|
/* If we can now collect the next folio, do so. We don't want
|
|
* to defer this as we have to decide whether we need to copy
|
|
* to the cache or not, and that may differ between adjacent
|
|
* subreqs.
|
|
*/
|
|
if (notes & BUFFERED) {
|
|
size_t fsize = PAGE_SIZE << rreq->front_folio_order;
|
|
|
|
/* Clear the tail of a short read. */
|
|
if (!(notes & HIT_PENDING) &&
|
|
front->error == 0 &&
|
|
transferred < front->len &&
|
|
(test_bit(NETFS_SREQ_HIT_EOF, &front->flags) ||
|
|
test_bit(NETFS_SREQ_CLEAR_TAIL, &front->flags))) {
|
|
netfs_clear_unread(front);
|
|
transferred = front->transferred = front->len;
|
|
trace_netfs_sreq(front, netfs_sreq_trace_clear);
|
|
}
|
|
|
|
stream->collected_to = front->start + transferred;
|
|
rreq->collected_to = stream->collected_to;
|
|
|
|
if (test_bit(NETFS_SREQ_COPY_TO_CACHE, &front->flags))
|
|
notes |= COPY_TO_CACHE;
|
|
|
|
if (test_bit(NETFS_SREQ_FAILED, &front->flags)) {
|
|
rreq->abandon_to = front->start + front->len;
|
|
front->transferred = front->len;
|
|
transferred = front->len;
|
|
trace_netfs_rreq(rreq, netfs_rreq_trace_set_abandon);
|
|
}
|
|
if (front->start + transferred >= rreq->cleaned_to + fsize ||
|
|
test_bit(NETFS_SREQ_HIT_EOF, &front->flags))
|
|
netfs_read_unlock_folios(rreq, ¬es);
|
|
} else {
|
|
stream->collected_to = front->start + transferred;
|
|
rreq->collected_to = stream->collected_to;
|
|
}
|
|
|
|
/* Stall if the front is still undergoing I/O. */
|
|
if (notes & HIT_PENDING)
|
|
break;
|
|
|
|
if (test_bit(NETFS_SREQ_FAILED, &front->flags)) {
|
|
if (!stream->failed) {
|
|
stream->error = front->error;
|
|
rreq->error = front->error;
|
|
set_bit(NETFS_RREQ_FAILED, &rreq->flags);
|
|
stream->failed = true;
|
|
}
|
|
notes |= MADE_PROGRESS | ABANDON_SREQ;
|
|
} else if (test_bit(NETFS_SREQ_NEED_RETRY, &front->flags)) {
|
|
stream->need_retry = true;
|
|
notes |= NEED_RETRY | MADE_PROGRESS;
|
|
break;
|
|
} else if (test_bit(NETFS_RREQ_SHORT_TRANSFER, &rreq->flags)) {
|
|
notes |= MADE_PROGRESS;
|
|
} else {
|
|
if (!stream->failed)
|
|
stream->transferred += transferred;
|
|
if (front->transferred < front->len)
|
|
set_bit(NETFS_RREQ_SHORT_TRANSFER, &rreq->flags);
|
|
notes |= MADE_PROGRESS;
|
|
}
|
|
|
|
/* Remove if completely consumed. */
|
|
stream->source = front->source;
|
|
spin_lock(&rreq->lock);
|
|
|
|
remove = front;
|
|
trace_netfs_sreq(front,
|
|
notes & ABANDON_SREQ ?
|
|
netfs_sreq_trace_abandoned : netfs_sreq_trace_consumed);
|
|
list_del_init(&front->rreq_link);
|
|
front = list_first_entry_or_null(&stream->subrequests,
|
|
struct netfs_io_subrequest, rreq_link);
|
|
stream->front = front;
|
|
spin_unlock(&rreq->lock);
|
|
netfs_put_subrequest(remove,
|
|
notes & ABANDON_SREQ ?
|
|
netfs_sreq_trace_put_abandon :
|
|
netfs_sreq_trace_put_done);
|
|
}
|
|
|
|
trace_netfs_collect_stream(rreq, stream);
|
|
trace_netfs_collect_state(rreq, rreq->collected_to, notes);
|
|
|
|
if (!(notes & BUFFERED))
|
|
rreq->cleaned_to = rreq->collected_to;
|
|
|
|
if (notes & NEED_RETRY)
|
|
goto need_retry;
|
|
if (notes & MADE_PROGRESS) {
|
|
netfs_wake_rreq_flag(rreq, NETFS_RREQ_PAUSE, netfs_rreq_trace_unpause);
|
|
//cond_resched();
|
|
goto reassess;
|
|
}
|
|
|
|
out:
|
|
_leave(" = %x", notes);
|
|
return;
|
|
|
|
need_retry:
|
|
/* Okay... We're going to have to retry parts of the stream. Note
|
|
* that any partially completed op will have had any wholly transferred
|
|
* folios removed from it.
|
|
*/
|
|
_debug("retry");
|
|
netfs_retry_reads(rreq);
|
|
goto out;
|
|
}
|
|
|
|
/*
|
|
* Do page flushing and suchlike after DIO.
|
|
*/
|
|
static void netfs_rreq_assess_dio(struct netfs_io_request *rreq)
|
|
{
|
|
unsigned int i;
|
|
|
|
if (rreq->origin == NETFS_UNBUFFERED_READ ||
|
|
rreq->origin == NETFS_DIO_READ) {
|
|
for (i = 0; i < rreq->direct_bv_count; i++) {
|
|
flush_dcache_page(rreq->direct_bv[i].bv_page);
|
|
// TODO: cifs marks pages in the destination buffer
|
|
// dirty under some circumstances after a read. Do we
|
|
// need to do that too?
|
|
set_page_dirty(rreq->direct_bv[i].bv_page);
|
|
}
|
|
}
|
|
|
|
if (rreq->iocb) {
|
|
rreq->iocb->ki_pos += rreq->transferred;
|
|
if (rreq->iocb->ki_complete) {
|
|
trace_netfs_rreq(rreq, netfs_rreq_trace_ki_complete);
|
|
rreq->iocb->ki_complete(
|
|
rreq->iocb, rreq->error ? rreq->error : rreq->transferred);
|
|
}
|
|
}
|
|
if (rreq->netfs_ops->done)
|
|
rreq->netfs_ops->done(rreq);
|
|
if (rreq->origin == NETFS_UNBUFFERED_READ ||
|
|
rreq->origin == NETFS_DIO_READ)
|
|
inode_dio_end(rreq->inode);
|
|
}
|
|
|
|
/*
|
|
* Do processing after reading a monolithic single object.
|
|
*/
|
|
static void netfs_rreq_assess_single(struct netfs_io_request *rreq)
|
|
{
|
|
struct netfs_io_stream *stream = &rreq->io_streams[0];
|
|
|
|
if (!rreq->error && stream->source == NETFS_DOWNLOAD_FROM_SERVER &&
|
|
fscache_resources_valid(&rreq->cache_resources)) {
|
|
trace_netfs_rreq(rreq, netfs_rreq_trace_dirty);
|
|
netfs_single_mark_inode_dirty(rreq->inode);
|
|
}
|
|
|
|
if (rreq->iocb) {
|
|
rreq->iocb->ki_pos += rreq->transferred;
|
|
if (rreq->iocb->ki_complete) {
|
|
trace_netfs_rreq(rreq, netfs_rreq_trace_ki_complete);
|
|
rreq->iocb->ki_complete(
|
|
rreq->iocb, rreq->error ? rreq->error : rreq->transferred);
|
|
}
|
|
}
|
|
if (rreq->netfs_ops->done)
|
|
rreq->netfs_ops->done(rreq);
|
|
}
|
|
|
|
/*
|
|
* Perform the collection of subrequests and folios.
|
|
*
|
|
* Note that we're in normal kernel thread context at this point, possibly
|
|
* running on a workqueue.
|
|
*/
|
|
bool netfs_read_collection(struct netfs_io_request *rreq)
|
|
{
|
|
struct netfs_io_stream *stream = &rreq->io_streams[0];
|
|
|
|
netfs_collect_read_results(rreq);
|
|
|
|
/* We're done when the app thread has finished posting subreqs and the
|
|
* queue is empty.
|
|
*/
|
|
if (!test_bit(NETFS_RREQ_ALL_QUEUED, &rreq->flags))
|
|
return false;
|
|
smp_rmb(); /* Read ALL_QUEUED before subreq lists. */
|
|
|
|
if (!list_empty(&stream->subrequests))
|
|
return false;
|
|
|
|
/* Okay, declare that all I/O is complete. */
|
|
rreq->transferred = stream->transferred;
|
|
trace_netfs_rreq(rreq, netfs_rreq_trace_complete);
|
|
|
|
//netfs_rreq_is_still_valid(rreq);
|
|
|
|
switch (rreq->origin) {
|
|
case NETFS_UNBUFFERED_READ:
|
|
case NETFS_DIO_READ:
|
|
case NETFS_READ_GAPS:
|
|
netfs_rreq_assess_dio(rreq);
|
|
break;
|
|
case NETFS_READ_SINGLE:
|
|
netfs_rreq_assess_single(rreq);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
task_io_account_read(rreq->transferred);
|
|
|
|
netfs_wake_rreq_flag(rreq, NETFS_RREQ_IN_PROGRESS, netfs_rreq_trace_wake_ip);
|
|
/* As we cleared NETFS_RREQ_IN_PROGRESS, we acquired its ref. */
|
|
|
|
trace_netfs_rreq(rreq, netfs_rreq_trace_done);
|
|
netfs_clear_subrequests(rreq);
|
|
netfs_unlock_abandoned_read_pages(rreq);
|
|
if (unlikely(rreq->copy_to_cache))
|
|
netfs_pgpriv2_end_copy_to_cache(rreq);
|
|
return true;
|
|
}
|
|
|
|
void netfs_read_collection_worker(struct work_struct *work)
|
|
{
|
|
struct netfs_io_request *rreq = container_of(work, struct netfs_io_request, work);
|
|
|
|
netfs_see_request(rreq, netfs_rreq_trace_see_work);
|
|
if (netfs_check_rreq_in_progress(rreq)) {
|
|
if (netfs_read_collection(rreq))
|
|
/* Drop the ref from the IN_PROGRESS flag. */
|
|
netfs_put_request(rreq, netfs_rreq_trace_put_work_ip);
|
|
else
|
|
netfs_see_request(rreq, netfs_rreq_trace_see_work_complete);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* netfs_read_subreq_progress - Note progress of a read operation.
|
|
* @subreq: The read request that has terminated.
|
|
*
|
|
* This tells the read side of netfs lib that a contributory I/O operation has
|
|
* made some progress and that it may be possible to unlock some folios.
|
|
*
|
|
* Before calling, the filesystem should update subreq->transferred to track
|
|
* the amount of data copied into the output buffer.
|
|
*/
|
|
void netfs_read_subreq_progress(struct netfs_io_subrequest *subreq)
|
|
{
|
|
struct netfs_io_request *rreq = subreq->rreq;
|
|
struct netfs_io_stream *stream = &rreq->io_streams[0];
|
|
size_t fsize = PAGE_SIZE << rreq->front_folio_order;
|
|
|
|
trace_netfs_sreq(subreq, netfs_sreq_trace_progress);
|
|
|
|
/* If we are at the head of the queue, wake up the collector,
|
|
* getting a ref to it if we were the ones to do so.
|
|
*/
|
|
if (subreq->start + subreq->transferred > rreq->cleaned_to + fsize &&
|
|
(rreq->origin == NETFS_READAHEAD ||
|
|
rreq->origin == NETFS_READPAGE ||
|
|
rreq->origin == NETFS_READ_FOR_WRITE) &&
|
|
list_is_first(&subreq->rreq_link, &stream->subrequests)
|
|
) {
|
|
__set_bit(NETFS_SREQ_MADE_PROGRESS, &subreq->flags);
|
|
netfs_wake_collector(rreq);
|
|
}
|
|
}
|
|
EXPORT_SYMBOL(netfs_read_subreq_progress);
|
|
|
|
/**
|
|
* netfs_read_subreq_terminated - Note the termination of an I/O operation.
|
|
* @subreq: The I/O request that has terminated.
|
|
*
|
|
* This tells the read helper that a contributory I/O operation has terminated,
|
|
* one way or another, and that it should integrate the results.
|
|
*
|
|
* The caller indicates the outcome of the operation through @subreq->error,
|
|
* supplying 0 to indicate a successful or retryable transfer (if
|
|
* NETFS_SREQ_NEED_RETRY is set) or a negative error code. The helper will
|
|
* look after reissuing I/O operations as appropriate and writing downloaded
|
|
* data to the cache.
|
|
*
|
|
* Before calling, the filesystem should update subreq->transferred to track
|
|
* the amount of data copied into the output buffer.
|
|
*/
|
|
void netfs_read_subreq_terminated(struct netfs_io_subrequest *subreq)
|
|
{
|
|
struct netfs_io_request *rreq = subreq->rreq;
|
|
|
|
switch (subreq->source) {
|
|
case NETFS_READ_FROM_CACHE:
|
|
netfs_stat(&netfs_n_rh_read_done);
|
|
break;
|
|
case NETFS_DOWNLOAD_FROM_SERVER:
|
|
netfs_stat(&netfs_n_rh_download_done);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
/* Deal with retry requests, short reads and errors. If we retry
|
|
* but don't make progress, we abandon the attempt.
|
|
*/
|
|
if (!subreq->error && subreq->transferred < subreq->len) {
|
|
if (test_bit(NETFS_SREQ_HIT_EOF, &subreq->flags)) {
|
|
trace_netfs_sreq(subreq, netfs_sreq_trace_hit_eof);
|
|
} else if (test_bit(NETFS_SREQ_CLEAR_TAIL, &subreq->flags)) {
|
|
trace_netfs_sreq(subreq, netfs_sreq_trace_need_clear);
|
|
} else if (test_bit(NETFS_SREQ_NEED_RETRY, &subreq->flags)) {
|
|
trace_netfs_sreq(subreq, netfs_sreq_trace_need_retry);
|
|
} else if (test_bit(NETFS_SREQ_MADE_PROGRESS, &subreq->flags)) {
|
|
__set_bit(NETFS_SREQ_NEED_RETRY, &subreq->flags);
|
|
trace_netfs_sreq(subreq, netfs_sreq_trace_partial_read);
|
|
} else {
|
|
__set_bit(NETFS_SREQ_FAILED, &subreq->flags);
|
|
subreq->error = -ENODATA;
|
|
trace_netfs_sreq(subreq, netfs_sreq_trace_short);
|
|
}
|
|
}
|
|
|
|
if (unlikely(subreq->error < 0)) {
|
|
trace_netfs_failure(rreq, subreq, subreq->error, netfs_fail_read);
|
|
if (subreq->source == NETFS_READ_FROM_CACHE) {
|
|
netfs_stat(&netfs_n_rh_read_failed);
|
|
__set_bit(NETFS_SREQ_NEED_RETRY, &subreq->flags);
|
|
} else {
|
|
netfs_stat(&netfs_n_rh_download_failed);
|
|
__set_bit(NETFS_SREQ_FAILED, &subreq->flags);
|
|
}
|
|
trace_netfs_rreq(rreq, netfs_rreq_trace_set_pause);
|
|
set_bit(NETFS_RREQ_PAUSE, &rreq->flags);
|
|
}
|
|
|
|
trace_netfs_sreq(subreq, netfs_sreq_trace_terminated);
|
|
netfs_subreq_clear_in_progress(subreq);
|
|
netfs_put_subrequest(subreq, netfs_sreq_trace_put_terminated);
|
|
}
|
|
EXPORT_SYMBOL(netfs_read_subreq_terminated);
|
|
|
|
/*
|
|
* Handle termination of a read from the cache.
|
|
*/
|
|
void netfs_cache_read_terminated(void *priv, ssize_t transferred_or_error)
|
|
{
|
|
struct netfs_io_subrequest *subreq = priv;
|
|
|
|
if (transferred_or_error > 0) {
|
|
subreq->error = 0;
|
|
if (transferred_or_error > 0) {
|
|
subreq->transferred += transferred_or_error;
|
|
__set_bit(NETFS_SREQ_MADE_PROGRESS, &subreq->flags);
|
|
}
|
|
} else {
|
|
subreq->error = transferred_or_error;
|
|
}
|
|
netfs_read_subreq_terminated(subreq);
|
|
}
|