Skip to content

Commit

Permalink
---
Browse files Browse the repository at this point in the history
yaml
---
r: 116675
b: refs/heads/master
c: 6f807ac
h: refs/heads/master
i:
  116673: 0a5d972
  116671: 227651a
v: v3
  • Loading branch information
Steven Rostedt authored and Ingo Molnar committed Oct 14, 2008
1 parent 996de36 commit 046eb66
Show file tree
Hide file tree
Showing 2 changed files with 42 additions and 35 deletions.
2 changes: 1 addition & 1 deletion [refs]
Original file line number Diff line number Diff line change
@@ -1,2 +1,2 @@
---
refs/heads/master: 097d036a2f25eecc42435c57e010aaf4a2eed2d9
refs/heads/master: 6f807acd27734197b11d42829d3cbb9c0937b572
75 changes: 41 additions & 34 deletions trunk/kernel/trace/ring_buffer.c
Original file line number Diff line number Diff line change
Expand Up @@ -117,6 +117,8 @@ void *ring_buffer_event_data(struct ring_buffer_event *event)
struct buffer_page {
u64 time_stamp; /* page time stamp */
unsigned size; /* size of page data */
unsigned write; /* index for next write */
unsigned read; /* index for next read */
struct list_head list; /* list of free pages */
void *page; /* Actual data page */
};
Expand Down Expand Up @@ -153,11 +155,8 @@ struct ring_buffer_per_cpu {
spinlock_t lock;
struct lock_class_key lock_key;
struct list_head pages;
unsigned long head; /* read from head */
unsigned long tail; /* write to tail */
unsigned long reader;
struct buffer_page *head_page;
struct buffer_page *tail_page;
struct buffer_page *head_page; /* read from head */
struct buffer_page *tail_page; /* write to tail */
struct buffer_page *reader_page;
unsigned long overrun;
unsigned long entries;
Expand Down Expand Up @@ -566,34 +565,41 @@ int ring_buffer_resize(struct ring_buffer *buffer, unsigned long size)

static inline int rb_per_cpu_empty(struct ring_buffer_per_cpu *cpu_buffer)
{
return (cpu_buffer->reader == cpu_buffer->reader_page->size &&
return cpu_buffer->reader_page->read == cpu_buffer->reader_page->size &&
(cpu_buffer->tail_page == cpu_buffer->reader_page ||
(cpu_buffer->tail_page == cpu_buffer->head_page &&
cpu_buffer->head == cpu_buffer->tail)));
cpu_buffer->head_page->read ==
cpu_buffer->tail_page->write));
}

static inline int rb_null_event(struct ring_buffer_event *event)
{
return event->type == RINGBUF_TYPE_PADDING;
}

static inline void *rb_page_index(struct buffer_page *page, unsigned index)
static inline void *__rb_page_index(struct buffer_page *page, unsigned index)
{
return page->page + index;
}

static inline struct ring_buffer_event *
rb_reader_event(struct ring_buffer_per_cpu *cpu_buffer)
{
return rb_page_index(cpu_buffer->reader_page,
cpu_buffer->reader);
return __rb_page_index(cpu_buffer->reader_page,
cpu_buffer->reader_page->read);
}

static inline struct ring_buffer_event *
rb_head_event(struct ring_buffer_per_cpu *cpu_buffer)
{
return __rb_page_index(cpu_buffer->head_page,
cpu_buffer->head_page->read);
}

static inline struct ring_buffer_event *
rb_iter_head_event(struct ring_buffer_iter *iter)
{
return rb_page_index(iter->head_page,
iter->head);
return __rb_page_index(iter->head_page, iter->head);
}

/*
Expand All @@ -610,7 +616,7 @@ static void rb_update_overflow(struct ring_buffer_per_cpu *cpu_buffer)
for (head = 0; head < rb_head_size(cpu_buffer);
head += rb_event_length(event)) {

event = rb_page_index(cpu_buffer->head_page, head);
event = __rb_page_index(cpu_buffer->head_page, head);
BUG_ON(rb_null_event(event));
/* Only count data entries */
if (event->type != RINGBUF_TYPE_DATA)
Expand Down Expand Up @@ -640,13 +646,13 @@ rb_add_stamp(struct ring_buffer_per_cpu *cpu_buffer, u64 *ts)

static void rb_reset_head_page(struct ring_buffer_per_cpu *cpu_buffer)
{
cpu_buffer->head = 0;
cpu_buffer->head_page->read = 0;
}

static void rb_reset_reader_page(struct ring_buffer_per_cpu *cpu_buffer)
{
cpu_buffer->read_stamp = cpu_buffer->reader_page->time_stamp;
cpu_buffer->reader = 0;
cpu_buffer->reader_page->read = 0;
}

static inline void rb_inc_iter(struct ring_buffer_iter *iter)
Expand Down Expand Up @@ -743,9 +749,8 @@ __rb_reserve_next(struct ring_buffer_per_cpu *cpu_buffer,
struct ring_buffer *buffer = cpu_buffer->buffer;
struct ring_buffer_event *event;

/* No locking needed for tail page */
tail_page = cpu_buffer->tail_page;
tail = cpu_buffer->tail;
tail = cpu_buffer->tail_page->write;

if (tail + length > BUF_PAGE_SIZE) {
struct buffer_page *next_page = tail_page;
Expand Down Expand Up @@ -774,7 +779,7 @@ __rb_reserve_next(struct ring_buffer_per_cpu *cpu_buffer,
}

if (tail != BUF_PAGE_SIZE) {
event = rb_page_index(tail_page, tail);
event = __rb_page_index(tail_page, tail);
/* page padding */
event->type = RINGBUF_TYPE_PADDING;
}
Expand All @@ -784,14 +789,14 @@ __rb_reserve_next(struct ring_buffer_per_cpu *cpu_buffer,
tail_page->size = 0;
tail = 0;
cpu_buffer->tail_page = tail_page;
cpu_buffer->tail = tail;
cpu_buffer->tail_page->write = tail;
rb_add_stamp(cpu_buffer, ts);
spin_unlock(&cpu_buffer->lock);
}

BUG_ON(tail + length > BUF_PAGE_SIZE);

event = rb_page_index(tail_page, tail);
event = __rb_page_index(tail_page, tail);
rb_update_event(event, type, length);

return event;
Expand Down Expand Up @@ -823,12 +828,12 @@ rb_add_time_stamp(struct ring_buffer_per_cpu *cpu_buffer,
return -1;

/* check to see if we went to the next page */
if (cpu_buffer->tail) {
if (cpu_buffer->tail_page->write) {
/* Still on same page, update timestamp */
event->time_delta = *delta & TS_MASK;
event->array[0] = *delta >> TS_SHIFT;
/* commit the time event */
cpu_buffer->tail +=
cpu_buffer->tail_page->write +=
rb_event_length(event);
cpu_buffer->write_stamp = *ts;
*delta = 0;
Expand All @@ -846,7 +851,7 @@ rb_reserve_next_event(struct ring_buffer_per_cpu *cpu_buffer,

ts = ring_buffer_time_stamp(cpu_buffer->cpu);

if (cpu_buffer->tail) {
if (cpu_buffer->tail_page->write) {
delta = ts - cpu_buffer->write_stamp;

if (test_time_stamp(delta)) {
Expand All @@ -868,7 +873,7 @@ rb_reserve_next_event(struct ring_buffer_per_cpu *cpu_buffer,
return NULL;

/* If the reserve went to the next page, our delta is zero */
if (!cpu_buffer->tail)
if (!cpu_buffer->tail_page->write)
delta = 0;

event->time_delta = delta;
Expand Down Expand Up @@ -933,8 +938,8 @@ ring_buffer_lock_reserve(struct ring_buffer *buffer,
static void rb_commit(struct ring_buffer_per_cpu *cpu_buffer,
struct ring_buffer_event *event)
{
cpu_buffer->tail += rb_event_length(event);
cpu_buffer->tail_page->size = cpu_buffer->tail;
cpu_buffer->tail_page->write += rb_event_length(event);
cpu_buffer->tail_page->size = cpu_buffer->tail_page->write;
cpu_buffer->write_stamp += event->time_delta;
cpu_buffer->entries++;
}
Expand Down Expand Up @@ -1178,10 +1183,10 @@ void ring_buffer_iter_reset(struct ring_buffer_iter *iter)
/* Iterator usage is expected to have record disabled */
if (list_empty(&cpu_buffer->reader_page->list)) {
iter->head_page = cpu_buffer->head_page;
iter->head = cpu_buffer->head;
iter->head = cpu_buffer->head_page->read;
} else {
iter->head_page = cpu_buffer->reader_page;
iter->head = cpu_buffer->reader;
iter->head = cpu_buffer->reader_page->read;
}
if (iter->head)
iter->read_stamp = cpu_buffer->read_stamp;
Expand All @@ -1200,7 +1205,7 @@ int ring_buffer_iter_empty(struct ring_buffer_iter *iter)
cpu_buffer = iter->cpu_buffer;

return iter->head_page == cpu_buffer->tail_page &&
iter->head == cpu_buffer->tail;
iter->head == cpu_buffer->tail_page->write;
}

static void
Expand Down Expand Up @@ -1277,11 +1282,11 @@ rb_get_reader_page(struct ring_buffer_per_cpu *cpu_buffer)
reader = cpu_buffer->reader_page;

/* If there's more to read, return this page */
if (cpu_buffer->reader < reader->size)
if (cpu_buffer->reader_page->read < reader->size)
goto out;

/* Never should we have an index greater than the size */
WARN_ON(cpu_buffer->reader > reader->size);
WARN_ON(cpu_buffer->reader_page->read > reader->size);

/* check if we caught up to the tail */
reader = NULL;
Expand Down Expand Up @@ -1342,7 +1347,7 @@ static void rb_advance_reader(struct ring_buffer_per_cpu *cpu_buffer)
rb_update_read_stamp(cpu_buffer, event);

length = rb_event_length(event);
cpu_buffer->reader += length;
cpu_buffer->reader_page->read += length;
}

static void rb_advance_iter(struct ring_buffer_iter *iter)
Expand Down Expand Up @@ -1373,7 +1378,7 @@ static void rb_advance_iter(struct ring_buffer_iter *iter)
* at the tail of the buffer.
*/
BUG_ON((iter->head_page == cpu_buffer->tail_page) &&
(iter->head + length > cpu_buffer->tail));
(iter->head + length > cpu_buffer->tail_page->write));

rb_update_iter_read_stamp(iter, event);

Expand Down Expand Up @@ -1623,7 +1628,9 @@ rb_reset_cpu(struct ring_buffer_per_cpu *cpu_buffer)
INIT_LIST_HEAD(&cpu_buffer->reader_page->list);
cpu_buffer->reader_page->size = 0;

cpu_buffer->head = cpu_buffer->tail = cpu_buffer->reader = 0;
cpu_buffer->head_page->read = 0;
cpu_buffer->tail_page->write = 0;
cpu_buffer->reader_page->read = 0;

cpu_buffer->overrun = 0;
cpu_buffer->entries = 0;
Expand Down

0 comments on commit 046eb66

Please sign in to comment.