Skip to content

Commit

Permalink
---
Browse files Browse the repository at this point in the history
yaml
---
r: 63040
b: refs/heads/master
c: 43b58b3
h: refs/heads/master
v: v3
  • Loading branch information
Pierre Ossman committed Jul 25, 2007
1 parent c9230b1 commit 576112e
Show file tree
Hide file tree
Showing 485 changed files with 5,051 additions and 10,785 deletions.
2 changes: 1 addition & 1 deletion [refs]
Original file line number Diff line number Diff line change
@@ -1,2 +1,2 @@
---
refs/heads/master: 9f5577d8158d8190174d95cbf21713251cc8a044
refs/heads/master: 43b58b36b7e6554b8a96be6b9f63542c583c06e5
5 changes: 2 additions & 3 deletions trunk/CREDITS
Original file line number Diff line number Diff line change
Expand Up @@ -966,7 +966,6 @@ N: Pekka Enberg
E: penberg@cs.helsinki.fi
W: http://www.cs.helsinki.fi/u/penberg/
D: Various kernel hacks, fixes, and cleanups.
D: Slab allocators
S: Finland

N: David Engebretsen
Expand Down Expand Up @@ -1940,8 +1939,8 @@ D: for Menuconfig's lxdialog.
N: Christoph Lameter
E: christoph@lameter.com
D: Digiboard PC/Xe and PC/Xi, Digiboard EPCA
D: NUMA support, Slab allocators, Page migration
D: Scalability, Time subsystem
D: Early protocol filter for bridging code
D: Bug fixes

N: Paul Laufer
E: paul@laufernet.com
Expand Down
6 changes: 3 additions & 3 deletions trunk/Documentation/block/barrier.txt
Original file line number Diff line number Diff line change
Expand Up @@ -79,9 +79,9 @@ and how to prepare flush requests. Note that the term 'ordered' is
used to indicate the whole sequence of performing barrier requests
including draining and flushing.

typedef void (prepare_flush_fn)(struct request_queue *q, struct request *rq);
typedef void (prepare_flush_fn)(request_queue_t *q, struct request *rq);

int blk_queue_ordered(struct request_queue *q, unsigned ordered,
int blk_queue_ordered(request_queue_t *q, unsigned ordered,
prepare_flush_fn *prepare_flush_fn);

@q : the queue in question
Expand All @@ -92,7 +92,7 @@ int blk_queue_ordered(struct request_queue *q, unsigned ordered,
For example, SCSI disk driver's prepare_flush_fn looks like the
following.

static void sd_prepare_flush(struct request_queue *q, struct request *rq)
static void sd_prepare_flush(request_queue_t *q, struct request *rq)
{
memset(rq->cmd, 0, sizeof(rq->cmd));
rq->cmd_type = REQ_TYPE_BLOCK_PC;
Expand Down
10 changes: 5 additions & 5 deletions trunk/Documentation/block/biodoc.txt
Original file line number Diff line number Diff line change
Expand Up @@ -740,12 +740,12 @@ Block now offers some simple generic functionality to help support command
queueing (typically known as tagged command queueing), ie manage more than
one outstanding command on a queue at any given time.

blk_queue_init_tags(struct request_queue *q, int depth)
blk_queue_init_tags(request_queue_t *q, int depth)

Initialize internal command tagging structures for a maximum
depth of 'depth'.

blk_queue_free_tags((struct request_queue *q)
blk_queue_free_tags((request_queue_t *q)

Teardown tag info associated with the queue. This will be done
automatically by block if blk_queue_cleanup() is called on a queue
Expand All @@ -754,15 +754,15 @@ one outstanding command on a queue at any given time.
The above are initialization and exit management, the main helpers during
normal operations are:

blk_queue_start_tag(struct request_queue *q, struct request *rq)
blk_queue_start_tag(request_queue_t *q, struct request *rq)

Start tagged operation for this request. A free tag number between
0 and 'depth' is assigned to the request (rq->tag holds this number),
and 'rq' is added to the internal tag management. If the maximum depth
for this queue is already achieved (or if the tag wasn't started for
some other reason), 1 is returned. Otherwise 0 is returned.

blk_queue_end_tag(struct request_queue *q, struct request *rq)
blk_queue_end_tag(request_queue_t *q, struct request *rq)

End tagged operation on this request. 'rq' is removed from the internal
book keeping structures.
Expand All @@ -781,7 +781,7 @@ queue. For instance, on IDE any tagged request error needs to clear both
the hardware and software block queue and enable the driver to sanely restart
all the outstanding requests. There's a third helper to do that:

blk_queue_invalidate_tags(struct request_queue *q)
blk_queue_invalidate_tags(request_queue_t *q)

Clear the internal block tag queue and re-add all the pending requests
to the request queue. The driver will receive them again on the
Expand Down
2 changes: 1 addition & 1 deletion trunk/Documentation/block/request.txt
Original file line number Diff line number Diff line change
Expand Up @@ -83,6 +83,6 @@ struct bio *bio DBI First bio in request

struct bio *biotail DBI Last bio in request

struct request_queue *q DB Request queue this request belongs to
request_queue_t *q DB Request queue this request belongs to

struct request_list *rl B Request list this request came from
2 changes: 1 addition & 1 deletion trunk/Documentation/iostats.txt
Original file line number Diff line number Diff line change
Expand Up @@ -79,7 +79,7 @@ Field 8 -- # of milliseconds spent writing
measured from __make_request() to end_that_request_last()).
Field 9 -- # of I/Os currently in progress
The only field that should go to zero. Incremented as requests are
given to appropriate struct request_queue and decremented as they finish.
given to appropriate request_queue_t and decremented as they finish.
Field 10 -- # of milliseconds spent doing I/Os
This field is increases so long as field 9 is nonzero.
Field 11 -- weighted # of milliseconds spent doing I/Os
Expand Down
58 changes: 0 additions & 58 deletions trunk/Documentation/lguest/extract

This file was deleted.

Loading

0 comments on commit 576112e

Please sign in to comment.