[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [PATCH RFC PKS/PMEM 09/58] drivers/gpu: Utilize new kmap_thread()



On Sat, Oct 10, 2020 at 12:03:49AM +0200, Daniel Vetter wrote:
> On Fri, Oct 09, 2020 at 12:49:44PM -0700, ira.weiny@xxxxxxxxx wrote:
> > From: Ira Weiny <ira.weiny@xxxxxxxxx>
> > 
> > These kmap() calls in the gpu stack are localized to a single thread.
> > To avoid the over head of global PKRS updates use the new kmap_thread()
> > call.
> > 
> > Cc: David Airlie <airlied@xxxxxxxx>
> > Cc: Daniel Vetter <daniel@xxxxxxxx>
> > Cc: Patrik Jakobsson <patrik.r.jakobsson@xxxxxxxxx>
> > Signed-off-by: Ira Weiny <ira.weiny@xxxxxxxxx>
> 
> I'm guessing the entire pile goes in through some other tree.
>

Apologies for not realizing there were multiple maintainers here.

But, I was thinking it would land together through the mm tree once the core
support lands.  I've tried to split these out in a way they can be easily
reviewed/acked by the correct developers.

> If so:
> 
> Acked-by: Daniel Vetter <daniel.vetter@xxxxxxxx>
> 
> If you want this to land through maintainer trees, then we need a
> per-driver split (since aside from amdgpu and radeon they're all different
> subtrees).

It is just RFC for the moment.  I need to get the core support accepted first
then this can land.

> 
> btw the two kmap calls in drm you highlight in the cover letter should
> also be convertible to kmap_thread. We only hold vmalloc mappings for a
> longer time (or it'd be quite a driver bug). So if you want maybe throw
> those two as two additional patches on top, and we can do some careful
> review & testing for them.

Cool.  I'll add them in.

Ira

> -Daniel
> 
> > ---
> >  drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c              | 12 ++++++------
> >  drivers/gpu/drm/gma500/gma_display.c                 |  4 ++--
> >  drivers/gpu/drm/gma500/mmu.c                         | 10 +++++-----
> >  drivers/gpu/drm/i915/gem/i915_gem_shmem.c            |  4 ++--
> >  .../gpu/drm/i915/gem/selftests/i915_gem_context.c    |  4 ++--
> >  drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c   |  8 ++++----
> >  drivers/gpu/drm/i915/gt/intel_ggtt_fencing.c         |  4 ++--
> >  drivers/gpu/drm/i915/gt/intel_gtt.c                  |  4 ++--
> >  drivers/gpu/drm/i915/gt/shmem_utils.c                |  4 ++--
> >  drivers/gpu/drm/i915/i915_gem.c                      |  8 ++++----
> >  drivers/gpu/drm/i915/i915_gpu_error.c                |  4 ++--
> >  drivers/gpu/drm/i915/selftests/i915_perf.c           |  4 ++--
> >  drivers/gpu/drm/radeon/radeon_ttm.c                  |  4 ++--
> >  13 files changed, 37 insertions(+), 37 deletions(-)
> > 
> > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 
> > b/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
> > index 978bae731398..bd564bccb7a3 100644
> > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
> > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
> > @@ -2437,11 +2437,11 @@ static ssize_t amdgpu_ttm_gtt_read(struct file *f, 
> > char __user *buf,
> >  
> >             page = adev->gart.pages[p];
> >             if (page) {
> > -                   ptr = kmap(page);
> > +                   ptr = kmap_thread(page);
> >                     ptr += off;
> >  
> >                     r = copy_to_user(buf, ptr, cur_size);
> > -                   kunmap(adev->gart.pages[p]);
> > +                   kunmap_thread(adev->gart.pages[p]);
> >             } else
> >                     r = clear_user(buf, cur_size);
> >  
> > @@ -2507,9 +2507,9 @@ static ssize_t amdgpu_iomem_read(struct file *f, char 
> > __user *buf,
> >             if (p->mapping != adev->mman.bdev.dev_mapping)
> >                     return -EPERM;
> >  
> > -           ptr = kmap(p);
> > +           ptr = kmap_thread(p);
> >             r = copy_to_user(buf, ptr + off, bytes);
> > -           kunmap(p);
> > +           kunmap_thread(p);
> >             if (r)
> >                     return -EFAULT;
> >  
> > @@ -2558,9 +2558,9 @@ static ssize_t amdgpu_iomem_write(struct file *f, 
> > const char __user *buf,
> >             if (p->mapping != adev->mman.bdev.dev_mapping)
> >                     return -EPERM;
> >  
> > -           ptr = kmap(p);
> > +           ptr = kmap_thread(p);
> >             r = copy_from_user(ptr + off, buf, bytes);
> > -           kunmap(p);
> > +           kunmap_thread(p);
> >             if (r)
> >                     return -EFAULT;
> >  
> > diff --git a/drivers/gpu/drm/gma500/gma_display.c 
> > b/drivers/gpu/drm/gma500/gma_display.c
> > index 3df6d6e850f5..35f4e55c941f 100644
> > --- a/drivers/gpu/drm/gma500/gma_display.c
> > +++ b/drivers/gpu/drm/gma500/gma_display.c
> > @@ -400,9 +400,9 @@ int gma_crtc_cursor_set(struct drm_crtc *crtc,
> >             /* Copy the cursor to cursor mem */
> >             tmp_dst = dev_priv->vram_addr + cursor_gt->offset;
> >             for (i = 0; i < cursor_pages; i++) {
> > -                   tmp_src = kmap(gt->pages[i]);
> > +                   tmp_src = kmap_thread(gt->pages[i]);
> >                     memcpy(tmp_dst, tmp_src, PAGE_SIZE);
> > -                   kunmap(gt->pages[i]);
> > +                   kunmap_thread(gt->pages[i]);
> >                     tmp_dst += PAGE_SIZE;
> >             }
> >  
> > diff --git a/drivers/gpu/drm/gma500/mmu.c b/drivers/gpu/drm/gma500/mmu.c
> > index 505044c9a673..fba7a3a461fd 100644
> > --- a/drivers/gpu/drm/gma500/mmu.c
> > +++ b/drivers/gpu/drm/gma500/mmu.c
> > @@ -192,20 +192,20 @@ struct psb_mmu_pd *psb_mmu_alloc_pd(struct 
> > psb_mmu_driver *driver,
> >             pd->invalid_pte = 0;
> >     }
> >  
> > -   v = kmap(pd->dummy_pt);
> > +   v = kmap_thread(pd->dummy_pt);
> >     for (i = 0; i < (PAGE_SIZE / sizeof(uint32_t)); ++i)
> >             v[i] = pd->invalid_pte;
> >  
> > -   kunmap(pd->dummy_pt);
> > +   kunmap_thread(pd->dummy_pt);
> >  
> > -   v = kmap(pd->p);
> > +   v = kmap_thread(pd->p);
> >     for (i = 0; i < (PAGE_SIZE / sizeof(uint32_t)); ++i)
> >             v[i] = pd->invalid_pde;
> >  
> > -   kunmap(pd->p);
> > +   kunmap_thread(pd->p);
> >  
> >     clear_page(kmap(pd->dummy_page));
> > -   kunmap(pd->dummy_page);
> > +   kunmap_thread(pd->dummy_page);
> >  
> >     pd->tables = vmalloc_user(sizeof(struct psb_mmu_pt *) * 1024);
> >     if (!pd->tables)
> > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_shmem.c 
> > b/drivers/gpu/drm/i915/gem/i915_gem_shmem.c
> > index 38113d3c0138..274424795fb7 100644
> > --- a/drivers/gpu/drm/i915/gem/i915_gem_shmem.c
> > +++ b/drivers/gpu/drm/i915/gem/i915_gem_shmem.c
> > @@ -566,9 +566,9 @@ i915_gem_object_create_shmem_from_data(struct 
> > drm_i915_private *dev_priv,
> >             if (err < 0)
> >                     goto fail;
> >  
> > -           vaddr = kmap(page);
> > +           vaddr = kmap_thread(page);
> >             memcpy(vaddr, data, len);
> > -           kunmap(page);
> > +           kunmap_thread(page);
> >  
> >             err = pagecache_write_end(file, file->f_mapping,
> >                                       offset, len, len,
> > diff --git a/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c 
> > b/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
> > index 7ffc3c751432..b466c677d007 100644
> > --- a/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
> > +++ b/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c
> > @@ -1754,7 +1754,7 @@ static int check_scratch_page(struct i915_gem_context 
> > *ctx, u32 *out)
> >             return -EINVAL;
> >     }
> >  
> > -   vaddr = kmap(page);
> > +   vaddr = kmap_thread(page);
> >     if (!vaddr) {
> >             pr_err("No (mappable) scratch page!\n");
> >             return -EINVAL;
> > @@ -1765,7 +1765,7 @@ static int check_scratch_page(struct i915_gem_context 
> > *ctx, u32 *out)
> >             pr_err("Inconsistent initial state of scratch page!\n");
> >             err = -EINVAL;
> >     }
> > -   kunmap(page);
> > +   kunmap_thread(page);
> >  
> >     return err;
> >  }
> > diff --git a/drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c 
> > b/drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
> > index 9c7402ce5bf9..447df22e2e06 100644
> > --- a/drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
> > +++ b/drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
> > @@ -143,7 +143,7 @@ static int check_partial_mapping(struct 
> > drm_i915_gem_object *obj,
> >     intel_gt_flush_ggtt_writes(&to_i915(obj->base.dev)->gt);
> >  
> >     p = i915_gem_object_get_page(obj, offset >> PAGE_SHIFT);
> > -   cpu = kmap(p) + offset_in_page(offset);
> > +   cpu = kmap_thread(p) + offset_in_page(offset);
> >     drm_clflush_virt_range(cpu, sizeof(*cpu));
> >     if (*cpu != (u32)page) {
> >             pr_err("Partial view for %lu [%u] (offset=%llu, size=%u [%llu, 
> > row size %u], fence=%d, tiling=%d, stride=%d) misalignment, expected write 
> > to page (%llu + %u [0x%llx]) of 0x%x, found 0x%x\n",
> > @@ -161,7 +161,7 @@ static int check_partial_mapping(struct 
> > drm_i915_gem_object *obj,
> >     }
> >     *cpu = 0;
> >     drm_clflush_virt_range(cpu, sizeof(*cpu));
> > -   kunmap(p);
> > +   kunmap_thread(p);
> >  
> >  out:
> >     __i915_vma_put(vma);
> > @@ -236,7 +236,7 @@ static int check_partial_mappings(struct 
> > drm_i915_gem_object *obj,
> >             intel_gt_flush_ggtt_writes(&to_i915(obj->base.dev)->gt);
> >  
> >             p = i915_gem_object_get_page(obj, offset >> PAGE_SHIFT);
> > -           cpu = kmap(p) + offset_in_page(offset);
> > +           cpu = kmap_thread(p) + offset_in_page(offset);
> >             drm_clflush_virt_range(cpu, sizeof(*cpu));
> >             if (*cpu != (u32)page) {
> >                     pr_err("Partial view for %lu [%u] (offset=%llu, size=%u 
> > [%llu, row size %u], fence=%d, tiling=%d, stride=%d) misalignment, expected 
> > write to page (%llu + %u [0x%llx]) of 0x%x, found 0x%x\n",
> > @@ -254,7 +254,7 @@ static int check_partial_mappings(struct 
> > drm_i915_gem_object *obj,
> >             }
> >             *cpu = 0;
> >             drm_clflush_virt_range(cpu, sizeof(*cpu));
> > -           kunmap(p);
> > +           kunmap_thread(p);
> >             if (err)
> >                     return err;
> >  
> > diff --git a/drivers/gpu/drm/i915/gt/intel_ggtt_fencing.c 
> > b/drivers/gpu/drm/i915/gt/intel_ggtt_fencing.c
> > index 7fb36b12fe7a..38da348282f1 100644
> > --- a/drivers/gpu/drm/i915/gt/intel_ggtt_fencing.c
> > +++ b/drivers/gpu/drm/i915/gt/intel_ggtt_fencing.c
> > @@ -731,7 +731,7 @@ static void swizzle_page(struct page *page)
> >     char *vaddr;
> >     int i;
> >  
> > -   vaddr = kmap(page);
> > +   vaddr = kmap_thread(page);
> >  
> >     for (i = 0; i < PAGE_SIZE; i += 128) {
> >             memcpy(temp, &vaddr[i], 64);
> > @@ -739,7 +739,7 @@ static void swizzle_page(struct page *page)
> >             memcpy(&vaddr[i + 64], temp, 64);
> >     }
> >  
> > -   kunmap(page);
> > +   kunmap_thread(page);
> >  }
> >  
> >  /**
> > diff --git a/drivers/gpu/drm/i915/gt/intel_gtt.c 
> > b/drivers/gpu/drm/i915/gt/intel_gtt.c
> > index 2a72cce63fd9..4cfb24e9ed62 100644
> > --- a/drivers/gpu/drm/i915/gt/intel_gtt.c
> > +++ b/drivers/gpu/drm/i915/gt/intel_gtt.c
> > @@ -312,9 +312,9 @@ static void poison_scratch_page(struct page *page, 
> > unsigned long size)
> >     do {
> >             void *vaddr;
> >  
> > -           vaddr = kmap(page);
> > +           vaddr = kmap_thread(page);
> >             memset(vaddr, POISON_FREE, PAGE_SIZE);
> > -           kunmap(page);
> > +           kunmap_thread(page);
> >  
> >             page = pfn_to_page(page_to_pfn(page) + 1);
> >             size -= PAGE_SIZE;
> > diff --git a/drivers/gpu/drm/i915/gt/shmem_utils.c 
> > b/drivers/gpu/drm/i915/gt/shmem_utils.c
> > index 43c7acbdc79d..a40d3130cebf 100644
> > --- a/drivers/gpu/drm/i915/gt/shmem_utils.c
> > +++ b/drivers/gpu/drm/i915/gt/shmem_utils.c
> > @@ -142,12 +142,12 @@ static int __shmem_rw(struct file *file, loff_t off,
> >             if (IS_ERR(page))
> >                     return PTR_ERR(page);
> >  
> > -           vaddr = kmap(page);
> > +           vaddr = kmap_thread(page);
> >             if (write)
> >                     memcpy(vaddr + offset_in_page(off), ptr, this);
> >             else
> >                     memcpy(ptr, vaddr + offset_in_page(off), this);
> > -           kunmap(page);
> > +           kunmap_thread(page);
> >             put_page(page);
> >  
> >             len -= this;
> > diff --git a/drivers/gpu/drm/i915/i915_gem.c 
> > b/drivers/gpu/drm/i915/i915_gem.c
> > index 9aa3066cb75d..cae8300fd224 100644
> > --- a/drivers/gpu/drm/i915/i915_gem.c
> > +++ b/drivers/gpu/drm/i915/i915_gem.c
> > @@ -312,14 +312,14 @@ shmem_pread(struct page *page, int offset, int len, 
> > char __user *user_data,
> >     char *vaddr;
> >     int ret;
> >  
> > -   vaddr = kmap(page);
> > +   vaddr = kmap_thread(page);
> >  
> >     if (needs_clflush)
> >             drm_clflush_virt_range(vaddr + offset, len);
> >  
> >     ret = __copy_to_user(user_data, vaddr + offset, len);
> >  
> > -   kunmap(page);
> > +   kunmap_thread(page);
> >  
> >     return ret ? -EFAULT : 0;
> >  }
> > @@ -708,7 +708,7 @@ shmem_pwrite(struct page *page, int offset, int len, 
> > char __user *user_data,
> >     char *vaddr;
> >     int ret;
> >  
> > -   vaddr = kmap(page);
> > +   vaddr = kmap_thread(page);
> >  
> >     if (needs_clflush_before)
> >             drm_clflush_virt_range(vaddr + offset, len);
> > @@ -717,7 +717,7 @@ shmem_pwrite(struct page *page, int offset, int len, 
> > char __user *user_data,
> >     if (!ret && needs_clflush_after)
> >             drm_clflush_virt_range(vaddr + offset, len);
> >  
> > -   kunmap(page);
> > +   kunmap_thread(page);
> >  
> >     return ret ? -EFAULT : 0;
> >  }
> > diff --git a/drivers/gpu/drm/i915/i915_gpu_error.c 
> > b/drivers/gpu/drm/i915/i915_gpu_error.c
> > index 3e6cbb0d1150..aecd469b6b6e 100644
> > --- a/drivers/gpu/drm/i915/i915_gpu_error.c
> > +++ b/drivers/gpu/drm/i915/i915_gpu_error.c
> > @@ -1058,9 +1058,9 @@ i915_vma_coredump_create(const struct intel_gt *gt,
> >  
> >                     drm_clflush_pages(&page, 1);
> >  
> > -                   s = kmap(page);
> > +                   s = kmap_thread(page);
> >                     ret = compress_page(compress, s, dst, false);
> > -                   kunmap(page);
> > +                   kunmap_thread(page);
> >  
> >                     drm_clflush_pages(&page, 1);
> >  
> > diff --git a/drivers/gpu/drm/i915/selftests/i915_perf.c 
> > b/drivers/gpu/drm/i915/selftests/i915_perf.c
> > index c2d001d9c0ec..7f7ef2d056f4 100644
> > --- a/drivers/gpu/drm/i915/selftests/i915_perf.c
> > +++ b/drivers/gpu/drm/i915/selftests/i915_perf.c
> > @@ -307,7 +307,7 @@ static int live_noa_gpr(void *arg)
> >     }
> >  
> >     /* Poison the ce->vm so we detect writes not to the GGTT gt->scratch */
> > -   scratch = kmap(ce->vm->scratch[0].base.page);
> > +   scratch = kmap_thread(ce->vm->scratch[0].base.page);
> >     memset(scratch, POISON_FREE, PAGE_SIZE);
> >  
> >     rq = intel_context_create_request(ce);
> > @@ -405,7 +405,7 @@ static int live_noa_gpr(void *arg)
> >  out_rq:
> >     i915_request_put(rq);
> >  out_ce:
> > -   kunmap(ce->vm->scratch[0].base.page);
> > +   kunmap_thread(ce->vm->scratch[0].base.page);
> >     intel_context_put(ce);
> >  out:
> >     stream_destroy(stream);
> > diff --git a/drivers/gpu/drm/radeon/radeon_ttm.c 
> > b/drivers/gpu/drm/radeon/radeon_ttm.c
> > index 004344dce140..0aba0cac51e1 100644
> > --- a/drivers/gpu/drm/radeon/radeon_ttm.c
> > +++ b/drivers/gpu/drm/radeon/radeon_ttm.c
> > @@ -1013,11 +1013,11 @@ static ssize_t radeon_ttm_gtt_read(struct file *f, 
> > char __user *buf,
> >  
> >             page = rdev->gart.pages[p];
> >             if (page) {
> > -                   ptr = kmap(page);
> > +                   ptr = kmap_thread(page);
> >                     ptr += off;
> >  
> >                     r = copy_to_user(buf, ptr, cur_size);
> > -                   kunmap(rdev->gart.pages[p]);
> > +                   kunmap_thread(rdev->gart.pages[p]);
> >             } else
> >                     r = clear_user(buf, cur_size);
> >  
> > -- 
> > 2.28.0.rc0.12.gb6a658bd00c9
> > 
> 
> -- 
> Daniel Vetter
> Software Engineer, Intel Corporation
> http://blog.ffwll.ch



 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.