From: "Thomas Hellström (VMware)" <thomas_os@shipmail.org>
To: "Christian König" <christian.koenig@amd.com>,
linux-mm@kvack.org, dri-devel@lists.freedesktop.org,
linux-kernel@vger.kernel.org
Cc: pv-drivers@vmware.com, linux-graphics-maintainer@vmware.com,
"Thomas Hellstrom" <thellstrom@vmware.com>,
"Andrew Morton" <akpm@linux-foundation.org>,
"Michal Hocko" <mhocko@suse.com>,
"Matthew Wilcox (Oracle)" <willy@infradead.org>,
"Kirill A. Shutemov" <kirill.shutemov@linux.intel.com>,
"Ralph Campbell" <rcampbell@nvidia.com>,
"Jérôme Glisse" <jglisse@redhat.com>,
"Dan Williams" <dan.j.williams@intel.com>,
"Roland Scheidegger" <sroland@vmware.com>
Subject: Re: [PATCH 5/9] mm, drm/ttm, drm/vmwgfx: Support huge TTM pagefaults
Date: Thu, 30 Jan 2020 14:29:43 +0100 [thread overview]
Message-ID: <615c9c1f-2e53-2c0e-7722-661bcf81554d@shipmail.org> (raw)
In-Reply-To: <bc30bf7e-5c96-0272-6e7e-64d22490d6a2@amd.com>
On 1/29/20 3:55 PM, Christian König wrote:
> Am 24.01.20 um 10:09 schrieb Thomas Hellström (VMware):
>> From: Thomas Hellstrom <thellstrom@vmware.com>
>>
>> Support huge (PMD-size and PUD-size) page-table entries by providing a
>> huge_fault() callback.
>> We still support private mappings and write-notify by splitting the huge
>> page-table entries on write-access.
>>
>> Note that for huge page-faults to occur, either the kernel needs to be
>> compiled with trans-huge-pages always enabled, or the kernel needs to be
>> compiled with trans-huge-pages enabled using madvise, and the user-space
>> app needs to call madvise() to enable trans-huge pages on a per-mapping
>> basis.
>>
>> Furthermore huge page-faults will not succeed unless buffer objects and
>> user-space addresses are aligned on huge page size boundaries.
>>
>> Cc: Andrew Morton <akpm@linux-foundation.org>
>> Cc: Michal Hocko <mhocko@suse.com>
>> Cc: "Matthew Wilcox (Oracle)" <willy@infradead.org>
>> Cc: "Kirill A. Shutemov" <kirill.shutemov@linux.intel.com>
>> Cc: Ralph Campbell <rcampbell@nvidia.com>
>> Cc: "Jérôme Glisse" <jglisse@redhat.com>
>> Cc: "Christian König" <christian.koenig@amd.com>
>> Cc: Dan Williams <dan.j.williams@intel.com>
>> Signed-off-by: Thomas Hellstrom <thellstrom@vmware.com>
>> Reviewed-by: Roland Scheidegger <sroland@vmware.com>
>> ---
>> drivers/gpu/drm/ttm/ttm_bo_vm.c | 145 ++++++++++++++++++++-
>> drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c | 2 +-
>> include/drm/ttm/ttm_bo_api.h | 3 +-
>> 3 files changed, 145 insertions(+), 5 deletions(-)
>>
>> diff --git a/drivers/gpu/drm/ttm/ttm_bo_vm.c
>> b/drivers/gpu/drm/ttm/ttm_bo_vm.c
>> index 389128b8c4dd..49704261a00d 100644
>> --- a/drivers/gpu/drm/ttm/ttm_bo_vm.c
>> +++ b/drivers/gpu/drm/ttm/ttm_bo_vm.c
>> @@ -156,6 +156,89 @@ vm_fault_t ttm_bo_vm_reserve(struct
>> ttm_buffer_object *bo,
>> }
>> EXPORT_SYMBOL(ttm_bo_vm_reserve);
>> +#ifdef CONFIG_TRANSPARENT_HUGEPAGE
>> +/**
>> + * ttm_bo_vm_insert_huge - Insert a pfn for PUD or PMD faults
>> + * @vmf: Fault data
>> + * @bo: The buffer object
>> + * @page_offset: Page offset from bo start
>> + * @fault_page_size: The size of the fault in pages.
>> + * @pgprot: The page protections.
>> + * Does additional checking whether it's possible to insert a PUD or
>> PMD
>> + * pfn and performs the insertion.
>> + *
>> + * Return: VM_FAULT_NOPAGE on successful insertion,
>> VM_FAULT_FALLBACK if
>> + * a huge fault was not possible, and a VM_FAULT_ERROR code otherwise.
>> + */
>> +static vm_fault_t ttm_bo_vm_insert_huge(struct vm_fault *vmf,
>> + struct ttm_buffer_object *bo,
>> + pgoff_t page_offset,
>> + pgoff_t fault_page_size,
>> + pgprot_t pgprot)
>> +{
>> + pgoff_t i;
>> + vm_fault_t ret;
>> + unsigned long pfn;
>> + pfn_t pfnt;
>> + struct ttm_tt *ttm = bo->ttm;
>> + bool write = vmf->flags & FAULT_FLAG_WRITE;
>> +
>> + /* Fault should not cross bo boundary. */
>> + page_offset &= ~(fault_page_size - 1);
>> + if (page_offset + fault_page_size > bo->num_pages)
>> + goto out_fallback;
>> +
>> + if (bo->mem.bus.is_iomem)
>> + pfn = ttm_bo_io_mem_pfn(bo, page_offset);
>> + else
>> + pfn = page_to_pfn(ttm->pages[page_offset]);
>> +
>> + /* pfn must be fault_page_size aligned. */
>> + if ((pfn & (fault_page_size - 1)) != 0)
>> + goto out_fallback;
>> +
>> + /* Check that memory is contiguous. */
>> + if (!bo->mem.bus.is_iomem)
>> + for (i = 1; i < fault_page_size; ++i) {
>> + if (page_to_pfn(ttm->pages[page_offset + i]) != pfn + i)
>> + goto out_fallback;
>> + }
>> + /* IO mem without the io_mem_pfn callback is always contiguous. */
>> + else if (bo->bdev->driver->io_mem_pfn)
>> + for (i = 1; i < fault_page_size; ++i) {
>> + if (ttm_bo_io_mem_pfn(bo, page_offset + i) != pfn + i)
>> + goto out_fallback;
>> + }
>
> Maybe add {} to the if to make clear where things start/end.
>
>> +
>> + pfnt = __pfn_to_pfn_t(pfn, PFN_DEV);
>> + if (fault_page_size == (HPAGE_PMD_SIZE >> PAGE_SHIFT))
>> + ret = vmf_insert_pfn_pmd_prot(vmf, pfnt, pgprot, write);
>> +#ifdef CONFIG_HAVE_ARCH_TRANSPARENT_HUGEPAGE_PUD
>> + else if (fault_page_size == (HPAGE_PUD_SIZE >> PAGE_SHIFT))
>> + ret = vmf_insert_pfn_pud_prot(vmf, pfnt, pgprot, write);
>> +#endif
>> + else
>> + WARN_ON_ONCE(ret = VM_FAULT_FALLBACK);
>> +
>> + if (ret != VM_FAULT_NOPAGE)
>> + goto out_fallback;
>> +
>> + return VM_FAULT_NOPAGE;
>> +out_fallback:
>> + count_vm_event(THP_FAULT_FALLBACK);
>> + return VM_FAULT_FALLBACK;
>
> This doesn't seem to match the function documentation since we never
> return ret here as far as I can see.
>
> Apart from those comments it looks like that should work,
> Christian.
Thanks for reviewing, Christian. I'll update the next version with your
feedback.
/Thomas
next prev parent reply other threads:[~2020-01-30 13:29 UTC|newest]
Thread overview: 12+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-01-24 9:09 [PATCH v2 0/9] Huge page-table entries for TTM Thomas Hellström (VMware)
2020-01-24 9:09 ` [PATCH 1/9] fs: Constify vma argument to vma_is_dax Thomas Hellström (VMware)
2020-01-24 9:09 ` [PATCH 2/9] mm: Introduce vma_is_special_huge Thomas Hellström (VMware)
2020-01-24 9:09 ` [PATCH 3/9] mm: Split huge pages on write-notify or COW Thomas Hellström (VMware)
2020-01-24 9:09 ` [PATCH 4/9] mm: Add vmf_insert_pfn_xxx_prot() for huge page-table entries Thomas Hellström (VMware)
2020-01-24 9:09 ` [PATCH 5/9] mm, drm/ttm, drm/vmwgfx: Support huge TTM pagefaults Thomas Hellström (VMware)
2020-01-29 14:55 ` Christian König
2020-01-30 13:29 ` Thomas Hellström (VMware) [this message]
2020-01-24 9:09 ` [PATCH 6/9] drm/vmwgfx: Support huge page faults Thomas Hellström (VMware)
2020-01-24 9:09 ` [PATCH 7/9] drm: Add a drm_get_unmapped_area() helper Thomas Hellström (VMware)
2020-01-24 9:09 ` [PATCH 8/9] drm/vmwgfx: Introduce a huge page aligning TTM range manager Thomas Hellström (VMware)
2020-01-24 9:09 ` [PATCH 9/9] drm/vmwgfx: Hook up the helpers to align buffer objects Thomas Hellström (VMware)
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=615c9c1f-2e53-2c0e-7722-661bcf81554d@shipmail.org \
--to=thomas_os@shipmail.org \
--cc=akpm@linux-foundation.org \
--cc=christian.koenig@amd.com \
--cc=dan.j.williams@intel.com \
--cc=dri-devel@lists.freedesktop.org \
--cc=jglisse@redhat.com \
--cc=kirill.shutemov@linux.intel.com \
--cc=linux-graphics-maintainer@vmware.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=mhocko@suse.com \
--cc=pv-drivers@vmware.com \
--cc=rcampbell@nvidia.com \
--cc=sroland@vmware.com \
--cc=thellstrom@vmware.com \
--cc=willy@infradead.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox