From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail-pd0-f172.google.com (mail-pd0-f172.google.com [209.85.192.172]) by kanga.kvack.org (Postfix) with ESMTP id 1FC3C6B0072 for ; Fri, 26 Dec 2014 09:40:35 -0500 (EST) Received: by mail-pd0-f172.google.com with SMTP id y13so13219262pdi.3 for ; Fri, 26 Dec 2014 06:40:34 -0800 (PST) Received: from mailout2.w1.samsung.com (mailout2.w1.samsung.com. [210.118.77.12]) by mx.google.com with ESMTPS id si2si14355504pac.97.2014.12.26.06.40.32 for (version=TLSv1 cipher=RC4-MD5 bits=128/128); Fri, 26 Dec 2014 06:40:33 -0800 (PST) Received: from eucpsbgm1.samsung.com (unknown [203.254.199.244]) by mailout2.w1.samsung.com (Oracle Communications Messaging Server 7u4-24.01(7.0.4.24.0) 64bit (built Nov 17 2011)) with ESMTP id <0NH700C4O2AAQW80@mailout2.w1.samsung.com> for linux-mm@kvack.org; Fri, 26 Dec 2014 14:44:34 +0000 (GMT) From: "Stefan I. Strogin" Subject: [PATCH 3/3] cma: add functions to get region pages counters Date: Fri, 26 Dec 2014 17:39:04 +0300 Message-id: In-reply-to: References: In-reply-to: References: Sender: owner-linux-mm@kvack.org List-ID: To: linux-mm@kvack.org, linux-kernel@vger.kernel.org Cc: Dmitry Safonov , s.strogin@partner.samsung.com, Joonsoo Kim , Andrew Morton , Marek Szyprowski , Michal Nazarewicz , aneesh.kumar@linux.vnet.ibm.com, Laurent Pinchart , Pintu Kumar , Weijie Yang , Laura Abbott , SeongJae Park , Hui Zhu , Minchan Kim , Dyasly Sergey , Vyacheslav Tyrtov From: Dmitry Safonov Here are two functions that provide interface to compute/get used size and size of biggest free chunk in cma region. Added that information in cmainfo. Signed-off-by: Dmitry Safonov --- include/linux/cma.h | 2 ++ mm/cma.c | 34 ++++++++++++++++++++++++++++++++++ 2 files changed, 36 insertions(+) diff --git a/include/linux/cma.h b/include/linux/cma.h index 9384ba6..855e6f2 100644 --- a/include/linux/cma.h +++ b/include/linux/cma.h @@ -18,6 +18,8 @@ struct cma; extern unsigned long totalcma_pages; extern phys_addr_t cma_get_base(struct cma *cma); extern unsigned long cma_get_size(struct cma *cma); +extern unsigned long cma_get_used(struct cma *cma); +extern unsigned long cma_get_maxchunk(struct cma *cma); extern int __init cma_declare_contiguous(phys_addr_t base, phys_addr_t size, phys_addr_t limit, diff --git a/mm/cma.c b/mm/cma.c index ffaea26..5e560ed 100644 --- a/mm/cma.c +++ b/mm/cma.c @@ -78,6 +78,36 @@ unsigned long cma_get_size(struct cma *cma) return cma->count << PAGE_SHIFT; } +unsigned long cma_get_used(struct cma *cma) +{ + unsigned long ret = 0; + + mutex_lock(&cma->lock); + /* pages counter is smaller than sizeof(int) */ + ret = bitmap_weight(cma->bitmap, (int)cma->count); + mutex_unlock(&cma->lock); + + return ret << (PAGE_SHIFT + cma->order_per_bit); +} + +unsigned long cma_get_maxchunk(struct cma *cma) +{ + unsigned long maxchunk = 0; + unsigned long start, end = 0; + + mutex_lock(&cma->lock); + for (;;) { + start = find_next_zero_bit(cma->bitmap, cma->count, end); + if (start >= cma->count) + break; + end = find_next_bit(cma->bitmap, cma->count, start); + maxchunk = max(end - start, maxchunk); + } + mutex_unlock(&cma->lock); + + return maxchunk << (PAGE_SHIFT + cma->order_per_bit); +} + static unsigned long cma_bitmap_aligned_mask(struct cma *cma, int align_order) { if (align_order <= cma->order_per_bit) @@ -591,6 +621,10 @@ static int s_show(struct seq_file *m, void *p) struct cma_buffer *cmabuf; struct stack_trace trace; + seq_printf(m, "CMARegion stat: %8lu kB total, %8lu kB used, %8lu kB max contiguous chunk\n\n", + cma_get_size(cma) >> 10, + cma_get_used(cma) >> 10, + cma_get_maxchunk(cma) >> 10); mutex_lock(&cma->list_lock); list_for_each_entry(cmabuf, &cma->buffers_list, list) { -- 2.1.0 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: email@kvack.org