Re: Lockless page cache test results

From: Wu Fengguang
Date: Fri Apr 28 2006 - 07:27:51 EST


On Wed, Apr 26, 2006 at 01:12:00PM -0700, Andrew Morton wrote:
> Jens Axboe <axboe@xxxxxxx> wrote:
> >
> > With a 16-page gang lookup in splice, the top profile for the 4-client
> > case (which is now at 4GiB/sec instead of 3) are:
> >
> > samples % symbol name
> > 30396 36.7217 __do_page_cache_readahead
> > 25843 31.2212 find_get_pages_contig
> > 9699 11.7174 default_idle
>
> __do_page_cache_readahead() should use gang lookup. We never got around to
> that, mainly because nothing really demonstrated a need.

I have been testing a patch for this for a while. The new function
looks like

static int
__do_page_cache_readahead(struct address_space *mapping, struct file *filp,
pgoff_t offset, unsigned long nr_to_read)
{
struct inode *inode = mapping->host;
struct page *page;
LIST_HEAD(page_pool);
pgoff_t last_index; /* The last page we want to read */
pgoff_t hole_index;
int ret = 0;
loff_t isize = i_size_read(inode);

last_index = ((isize - 1) >> PAGE_CACHE_SHIFT);

if (unlikely(!isize || !nr_to_read))
goto out;
if (unlikely(last_index < offset))
goto out;

if (last_index > offset + nr_to_read - 1 &&
offset < offset + nr_to_read)
last_index = offset + nr_to_read - 1;

/*
* Go through ranges of holes and preallocate all the absent pages.
*/
next_hole_range:
cond_resched();

read_lock_irq(&mapping->tree_lock);
hole_index = radix_tree_scan_hole(&mapping->page_tree,
offset, last_index - offset + 1);

if (hole_index > last_index) { /* no more holes? */
read_unlock_irq(&mapping->tree_lock);
goto submit_io;
}

offset = radix_tree_scan_data(&mapping->page_tree, (void **)&page,
hole_index, last_index);
read_unlock_irq(&mapping->tree_lock);

ddprintk("ra range %lu-%lu(%p)-%lu\n", hole_index, offset, page, last_index);

for (;;) {
page = page_cache_alloc_cold(mapping);
if (!page)
break;

page->index = hole_index;
list_add(&page->lru, &page_pool);
ret++;
BUG_ON(ret > nr_to_read);

if (hole_index >= last_index)
break;

if (++hole_index >= offset)
goto next_hole_range;
}

submit_io:
/*
* Now start the IO. We ignore I/O errors - if the page is not
* uptodate then the caller will launch readpage again, and
* will then handle the error.
*/
if (ret)
read_pages(mapping, filp, &page_pool, ret);
BUG_ON(!list_empty(&page_pool));
out:
return ret;
}

The radix_tree_scan_data()/radix_tree_scan_hole() functions called
above are more flexible than the original __lookup(). Perhaps we can
rebase radix_tree_gang_lookup() and find_get_pages_contig() on them.

If it is deemed ok, I'll clean it up and submit the patch asap.

Thanks,
Wu
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/