From mboxrd@z Thu Jan 1 00:00:00 1970 Date: Wed, 16 Aug 2006 09:35:46 +0400 From: Evgeniy Polyakov Subject: Re: [PATCH 1/1] network memory allocator. Message-ID: <20060816053545.GB22921@2ka.mipt.ru> References: <20060814110359.GA27704@2ka.mipt.ru> <200608152221.22883.arnd@arndb.de> Mime-Version: 1.0 Content-Type: text/plain; charset=koi8-r Content-Disposition: inline Content-Transfer-Encoding: 8bit In-Reply-To: <200608152221.22883.arnd@arndb.de> Sender: owner-linux-mm@kvack.org Return-Path: To: Arnd Bergmann Cc: David Miller , netdev@vger.kernel.org, linux-kernel@vger.kernel.org, linux-mm@kvack.org List-ID: On Tue, Aug 15, 2006 at 10:21:22PM +0200, Arnd Bergmann (arnd@arndb.de) wrote: > Am Monday 14 August 2006 13:04 schrieb Evgeniy Polyakov: > > ?* full per CPU allocation and freeing (objects are never freed on > > ????????different CPU) > > Many of your data structures are per cpu, but your underlying allocations > are all using regular kzalloc/__get_free_page/__get_free_pages functions. > Shouldn't these be converted to calls to kmalloc_node and alloc_pages_node > in order to get better locality on NUMA systems? > > OTOH, we have recently experimented with doing the dev_alloc_skb calls > with affinity to the NUMA node that holds the actual network adapter, and > got significant improvements on the Cell blade server. That of course > may be a conflicting goal since it would mean having per-cpu per-node > page pools if any CPU is supposed to be able to allocate pages for use > as DMA buffers on any node. Doesn't alloc_pages() automatically switches to alloc_pages_node() or alloc_pages_current()? > Arnd <>< -- Evgeniy Polyakov -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: email@kvack.org