From 2b44c84ec34b0e29a8f5741ce6c4209a601cb53c Mon Sep 17 00:00:00 2001 From: Jerome Glisse Date: Fri, 13 Nov 2009 20:56:58 +0100 Subject: [PATCH] --- yaml --- r: 168871 b: refs/heads/master c: a698cf34ea867efef12fc29dd63d443f0c71a53c h: refs/heads/master i: 168869: 0a2ee02b521ecc0453eb36440e5033a6552b9449 168867: ba46655ca9bf767c1f03ac1411ab115769b9ac82 168863: 86852339954ffeb287687826fcb58403a1152825 v: v3 --- [refs] | 2 +- trunk/drivers/gpu/drm/drm_mm.c | 9 +++++++++ 2 files changed, 10 insertions(+), 1 deletion(-) diff --git a/[refs] b/[refs] index d8e93326331a..7e62d6defd30 100644 --- a/[refs] +++ b/[refs] @@ -1,2 +1,2 @@ --- -refs/heads/master: 0beb81ab45c67de4b3aa85faad604cff8ed133a8 +refs/heads/master: a698cf34ea867efef12fc29dd63d443f0c71a53c diff --git a/trunk/drivers/gpu/drm/drm_mm.c b/trunk/drivers/gpu/drm/drm_mm.c index c861d80fd779..97dc5a4f0de4 100644 --- a/trunk/drivers/gpu/drm/drm_mm.c +++ b/trunk/drivers/gpu/drm/drm_mm.c @@ -103,6 +103,11 @@ static struct drm_mm_node *drm_mm_kmalloc(struct drm_mm *mm, int atomic) return child; } +/* drm_mm_pre_get() - pre allocate drm_mm_node structure + * drm_mm: memory manager struct we are pre-allocating for + * + * Returns 0 on success or -ENOMEM if allocation fails. + */ int drm_mm_pre_get(struct drm_mm *mm) { struct drm_mm_node *node; @@ -253,12 +258,14 @@ void drm_mm_put_block(struct drm_mm_node *cur) prev_node->size += next_node->size; list_del(&next_node->ml_entry); list_del(&next_node->fl_entry); + spin_lock(&mm->unused_lock); if (mm->num_unused < MM_UNUSED_TARGET) { list_add(&next_node->fl_entry, &mm->unused_nodes); ++mm->num_unused; } else kfree(next_node); + spin_unlock(&mm->unused_lock); } else { next_node->size += cur->size; next_node->start = cur->start; @@ -271,11 +278,13 @@ void drm_mm_put_block(struct drm_mm_node *cur) list_add(&cur->fl_entry, &mm->fl_entry); } else { list_del(&cur->ml_entry); + spin_lock(&mm->unused_lock); if (mm->num_unused < MM_UNUSED_TARGET) { list_add(&cur->fl_entry, &mm->unused_nodes); ++mm->num_unused; } else kfree(cur); + spin_unlock(&mm->unused_lock); } }