You've already forked linux-apfs
mirror of
https://github.com/linux-apfs/linux-apfs.git
synced 2026-05-01 15:00:59 -07:00
Merge branch 'mm-pat-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip
Pull dma_*_writecombine rename from Ingo Molnar: "Rename dma_*_writecombine() to dma_*_wc() This is a tree-wide API rename, to move the dma_*() write-combining APIs closer in name to their usual API families. (The old API names are kept as compatibility wrappers to not introduce extra breakage.) The patch was Coccinelle generated" * 'mm-pat-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip: dma, mm/pat: Rename dma_*_writecombine() to dma_*_wc()
This commit is contained in:
@@ -86,8 +86,8 @@ static int lpc32xx_clcd_setup(struct clcd_fb *fb)
|
||||
{
|
||||
dma_addr_t dma;
|
||||
|
||||
fb->fb.screen_base = dma_alloc_writecombine(&fb->dev->dev,
|
||||
PANEL_SIZE, &dma, GFP_KERNEL);
|
||||
fb->fb.screen_base = dma_alloc_wc(&fb->dev->dev, PANEL_SIZE, &dma,
|
||||
GFP_KERNEL);
|
||||
if (!fb->fb.screen_base) {
|
||||
printk(KERN_ERR "CLCD: unable to map framebuffer\n");
|
||||
return -ENOMEM;
|
||||
@@ -116,15 +116,14 @@ static int lpc32xx_clcd_setup(struct clcd_fb *fb)
|
||||
|
||||
static int lpc32xx_clcd_mmap(struct clcd_fb *fb, struct vm_area_struct *vma)
|
||||
{
|
||||
return dma_mmap_writecombine(&fb->dev->dev, vma,
|
||||
fb->fb.screen_base, fb->fb.fix.smem_start,
|
||||
fb->fb.fix.smem_len);
|
||||
return dma_mmap_wc(&fb->dev->dev, vma, fb->fb.screen_base,
|
||||
fb->fb.fix.smem_start, fb->fb.fix.smem_len);
|
||||
}
|
||||
|
||||
static void lpc32xx_clcd_remove(struct clcd_fb *fb)
|
||||
{
|
||||
dma_free_writecombine(&fb->dev->dev, fb->fb.fix.smem_len,
|
||||
fb->fb.screen_base, fb->fb.fix.smem_start);
|
||||
dma_free_wc(&fb->dev->dev, fb->fb.fix.smem_len, fb->fb.screen_base,
|
||||
fb->fb.fix.smem_start);
|
||||
}
|
||||
|
||||
/*
|
||||
|
||||
@@ -42,8 +42,8 @@ int netx_clcd_setup(struct clcd_fb *fb)
|
||||
|
||||
fb->panel = netx_panel;
|
||||
|
||||
fb->fb.screen_base = dma_alloc_writecombine(&fb->dev->dev, 1024*1024,
|
||||
&dma, GFP_KERNEL);
|
||||
fb->fb.screen_base = dma_alloc_wc(&fb->dev->dev, 1024 * 1024, &dma,
|
||||
GFP_KERNEL);
|
||||
if (!fb->fb.screen_base) {
|
||||
printk(KERN_ERR "CLCD: unable to map framebuffer\n");
|
||||
return -ENOMEM;
|
||||
@@ -57,16 +57,14 @@ int netx_clcd_setup(struct clcd_fb *fb)
|
||||
|
||||
int netx_clcd_mmap(struct clcd_fb *fb, struct vm_area_struct *vma)
|
||||
{
|
||||
return dma_mmap_writecombine(&fb->dev->dev, vma,
|
||||
fb->fb.screen_base,
|
||||
fb->fb.fix.smem_start,
|
||||
fb->fb.fix.smem_len);
|
||||
return dma_mmap_wc(&fb->dev->dev, vma, fb->fb.screen_base,
|
||||
fb->fb.fix.smem_start, fb->fb.fix.smem_len);
|
||||
}
|
||||
|
||||
void netx_clcd_remove(struct clcd_fb *fb)
|
||||
{
|
||||
dma_free_writecombine(&fb->dev->dev, fb->fb.fix.smem_len,
|
||||
fb->fb.screen_base, fb->fb.fix.smem_start);
|
||||
dma_free_wc(&fb->dev->dev, fb->fb.fix.smem_len, fb->fb.screen_base,
|
||||
fb->fb.fix.smem_start);
|
||||
}
|
||||
|
||||
static AMBA_AHB_DEVICE(fb, "fb", 0, 0x00104000, { NETX_IRQ_LCD }, NULL);
|
||||
|
||||
@@ -90,8 +90,8 @@ int nspire_clcd_setup(struct clcd_fb *fb)
|
||||
panel_size = ((panel->mode.xres * panel->mode.yres) * panel->bpp) / 8;
|
||||
panel_size = ALIGN(panel_size, PAGE_SIZE);
|
||||
|
||||
fb->fb.screen_base = dma_alloc_writecombine(&fb->dev->dev,
|
||||
panel_size, &dma, GFP_KERNEL);
|
||||
fb->fb.screen_base = dma_alloc_wc(&fb->dev->dev, panel_size, &dma,
|
||||
GFP_KERNEL);
|
||||
|
||||
if (!fb->fb.screen_base) {
|
||||
pr_err("CLCD: unable to map framebuffer\n");
|
||||
@@ -107,13 +107,12 @@ int nspire_clcd_setup(struct clcd_fb *fb)
|
||||
|
||||
int nspire_clcd_mmap(struct clcd_fb *fb, struct vm_area_struct *vma)
|
||||
{
|
||||
return dma_mmap_writecombine(&fb->dev->dev, vma,
|
||||
fb->fb.screen_base, fb->fb.fix.smem_start,
|
||||
fb->fb.fix.smem_len);
|
||||
return dma_mmap_wc(&fb->dev->dev, vma, fb->fb.screen_base,
|
||||
fb->fb.fix.smem_start, fb->fb.fix.smem_len);
|
||||
}
|
||||
|
||||
void nspire_clcd_remove(struct clcd_fb *fb)
|
||||
{
|
||||
dma_free_writecombine(&fb->dev->dev, fb->fb.fix.smem_len,
|
||||
fb->fb.screen_base, fb->fb.fix.smem_start);
|
||||
dma_free_wc(&fb->dev->dev, fb->fb.fix.smem_len, fb->fb.screen_base,
|
||||
fb->fb.fix.smem_start);
|
||||
}
|
||||
|
||||
@@ -1300,10 +1300,10 @@ static int iop_adma_probe(struct platform_device *pdev)
|
||||
* note: writecombine gives slightly better performance, but
|
||||
* requires that we explicitly flush the writes
|
||||
*/
|
||||
adev->dma_desc_pool_virt = dma_alloc_writecombine(&pdev->dev,
|
||||
plat_data->pool_size,
|
||||
&adev->dma_desc_pool,
|
||||
GFP_KERNEL);
|
||||
adev->dma_desc_pool_virt = dma_alloc_wc(&pdev->dev,
|
||||
plat_data->pool_size,
|
||||
&adev->dma_desc_pool,
|
||||
GFP_KERNEL);
|
||||
if (!adev->dma_desc_pool_virt) {
|
||||
ret = -ENOMEM;
|
||||
goto err_free_adev;
|
||||
|
||||
@@ -964,8 +964,8 @@ mv_xor_channel_add(struct mv_xor_device *xordev,
|
||||
* requires that we explicitly flush the writes
|
||||
*/
|
||||
mv_chan->dma_desc_pool_virt =
|
||||
dma_alloc_writecombine(&pdev->dev, MV_XOR_POOL_SIZE,
|
||||
&mv_chan->dma_desc_pool, GFP_KERNEL);
|
||||
dma_alloc_wc(&pdev->dev, MV_XOR_POOL_SIZE, &mv_chan->dma_desc_pool,
|
||||
GFP_KERNEL);
|
||||
if (!mv_chan->dma_desc_pool_virt)
|
||||
return ERR_PTR(-ENOMEM);
|
||||
|
||||
|
||||
@@ -502,8 +502,8 @@ static int bam_alloc_chan(struct dma_chan *chan)
|
||||
return 0;
|
||||
|
||||
/* allocate FIFO descriptor space, but only if necessary */
|
||||
bchan->fifo_virt = dma_alloc_writecombine(bdev->dev, BAM_DESC_FIFO_SIZE,
|
||||
&bchan->fifo_phys, GFP_KERNEL);
|
||||
bchan->fifo_virt = dma_alloc_wc(bdev->dev, BAM_DESC_FIFO_SIZE,
|
||||
&bchan->fifo_phys, GFP_KERNEL);
|
||||
|
||||
if (!bchan->fifo_virt) {
|
||||
dev_err(bdev->dev, "Failed to allocate desc fifo\n");
|
||||
@@ -538,8 +538,8 @@ static void bam_free_chan(struct dma_chan *chan)
|
||||
bam_reset_channel(bchan);
|
||||
spin_unlock_irqrestore(&bchan->vc.lock, flags);
|
||||
|
||||
dma_free_writecombine(bdev->dev, BAM_DESC_FIFO_SIZE, bchan->fifo_virt,
|
||||
bchan->fifo_phys);
|
||||
dma_free_wc(bdev->dev, BAM_DESC_FIFO_SIZE, bchan->fifo_virt,
|
||||
bchan->fifo_phys);
|
||||
bchan->fifo_virt = NULL;
|
||||
|
||||
/* mask irq for pipe/channel */
|
||||
@@ -1231,9 +1231,9 @@ static int bam_dma_remove(struct platform_device *pdev)
|
||||
bam_dma_terminate_all(&bdev->channels[i].vc.chan);
|
||||
tasklet_kill(&bdev->channels[i].vc.task);
|
||||
|
||||
dma_free_writecombine(bdev->dev, BAM_DESC_FIFO_SIZE,
|
||||
bdev->channels[i].fifo_virt,
|
||||
bdev->channels[i].fifo_phys);
|
||||
dma_free_wc(bdev->dev, BAM_DESC_FIFO_SIZE,
|
||||
bdev->channels[i].fifo_virt,
|
||||
bdev->channels[i].fifo_phys);
|
||||
}
|
||||
|
||||
tasklet_kill(&bdev->task);
|
||||
|
||||
@@ -109,8 +109,8 @@ struct drm_gem_cma_object *drm_gem_cma_create(struct drm_device *drm,
|
||||
if (IS_ERR(cma_obj))
|
||||
return cma_obj;
|
||||
|
||||
cma_obj->vaddr = dma_alloc_writecombine(drm->dev, size,
|
||||
&cma_obj->paddr, GFP_KERNEL | __GFP_NOWARN);
|
||||
cma_obj->vaddr = dma_alloc_wc(drm->dev, size, &cma_obj->paddr,
|
||||
GFP_KERNEL | __GFP_NOWARN);
|
||||
if (!cma_obj->vaddr) {
|
||||
dev_err(drm->dev, "failed to allocate buffer with size %zu\n",
|
||||
size);
|
||||
@@ -192,8 +192,8 @@ void drm_gem_cma_free_object(struct drm_gem_object *gem_obj)
|
||||
cma_obj = to_drm_gem_cma_obj(gem_obj);
|
||||
|
||||
if (cma_obj->vaddr) {
|
||||
dma_free_writecombine(gem_obj->dev->dev, cma_obj->base.size,
|
||||
cma_obj->vaddr, cma_obj->paddr);
|
||||
dma_free_wc(gem_obj->dev->dev, cma_obj->base.size,
|
||||
cma_obj->vaddr, cma_obj->paddr);
|
||||
} else if (gem_obj->import_attach) {
|
||||
drm_prime_gem_destroy(gem_obj, cma_obj->sgt);
|
||||
}
|
||||
@@ -324,9 +324,8 @@ static int drm_gem_cma_mmap_obj(struct drm_gem_cma_object *cma_obj,
|
||||
vma->vm_flags &= ~VM_PFNMAP;
|
||||
vma->vm_pgoff = 0;
|
||||
|
||||
ret = dma_mmap_writecombine(cma_obj->base.dev->dev, vma,
|
||||
cma_obj->vaddr, cma_obj->paddr,
|
||||
vma->vm_end - vma->vm_start);
|
||||
ret = dma_mmap_wc(cma_obj->base.dev->dev, vma, cma_obj->vaddr,
|
||||
cma_obj->paddr, vma->vm_end - vma->vm_start);
|
||||
if (ret)
|
||||
drm_gem_vm_close(vma);
|
||||
|
||||
|
||||
@@ -1113,8 +1113,8 @@ struct etnaviv_cmdbuf *etnaviv_gpu_cmdbuf_new(struct etnaviv_gpu *gpu, u32 size,
|
||||
if (!cmdbuf)
|
||||
return NULL;
|
||||
|
||||
cmdbuf->vaddr = dma_alloc_writecombine(gpu->dev, size, &cmdbuf->paddr,
|
||||
GFP_KERNEL);
|
||||
cmdbuf->vaddr = dma_alloc_wc(gpu->dev, size, &cmdbuf->paddr,
|
||||
GFP_KERNEL);
|
||||
if (!cmdbuf->vaddr) {
|
||||
kfree(cmdbuf);
|
||||
return NULL;
|
||||
@@ -1128,8 +1128,8 @@ struct etnaviv_cmdbuf *etnaviv_gpu_cmdbuf_new(struct etnaviv_gpu *gpu, u32 size,
|
||||
|
||||
void etnaviv_gpu_cmdbuf_free(struct etnaviv_cmdbuf *cmdbuf)
|
||||
{
|
||||
dma_free_writecombine(cmdbuf->gpu->dev, cmdbuf->size,
|
||||
cmdbuf->vaddr, cmdbuf->paddr);
|
||||
dma_free_wc(cmdbuf->gpu->dev, cmdbuf->size, cmdbuf->vaddr,
|
||||
cmdbuf->paddr);
|
||||
kfree(cmdbuf);
|
||||
}
|
||||
|
||||
|
||||
@@ -573,10 +573,9 @@ static int omap_dmm_remove(struct platform_device *dev)
|
||||
|
||||
kfree(omap_dmm->engines);
|
||||
if (omap_dmm->refill_va)
|
||||
dma_free_writecombine(omap_dmm->dev,
|
||||
REFILL_BUFFER_SIZE * omap_dmm->num_engines,
|
||||
omap_dmm->refill_va,
|
||||
omap_dmm->refill_pa);
|
||||
dma_free_wc(omap_dmm->dev,
|
||||
REFILL_BUFFER_SIZE * omap_dmm->num_engines,
|
||||
omap_dmm->refill_va, omap_dmm->refill_pa);
|
||||
if (omap_dmm->dummy_page)
|
||||
__free_page(omap_dmm->dummy_page);
|
||||
|
||||
@@ -701,9 +700,9 @@ static int omap_dmm_probe(struct platform_device *dev)
|
||||
omap_dmm->dummy_pa = page_to_phys(omap_dmm->dummy_page);
|
||||
|
||||
/* alloc refill memory */
|
||||
omap_dmm->refill_va = dma_alloc_writecombine(&dev->dev,
|
||||
REFILL_BUFFER_SIZE * omap_dmm->num_engines,
|
||||
&omap_dmm->refill_pa, GFP_KERNEL);
|
||||
omap_dmm->refill_va = dma_alloc_wc(&dev->dev,
|
||||
REFILL_BUFFER_SIZE * omap_dmm->num_engines,
|
||||
&omap_dmm->refill_pa, GFP_KERNEL);
|
||||
if (!omap_dmm->refill_va) {
|
||||
dev_err(&dev->dev, "could not allocate refill memory\n");
|
||||
goto fail;
|
||||
|
||||
@@ -1330,8 +1330,8 @@ void omap_gem_free_object(struct drm_gem_object *obj)
|
||||
omap_gem_detach_pages(obj);
|
||||
|
||||
if (!is_shmem(obj)) {
|
||||
dma_free_writecombine(dev->dev, obj->size,
|
||||
omap_obj->vaddr, omap_obj->paddr);
|
||||
dma_free_wc(dev->dev, obj->size, omap_obj->vaddr,
|
||||
omap_obj->paddr);
|
||||
} else if (omap_obj->vaddr) {
|
||||
vunmap(omap_obj->vaddr);
|
||||
}
|
||||
@@ -1395,8 +1395,8 @@ struct drm_gem_object *omap_gem_new(struct drm_device *dev,
|
||||
/* attempt to allocate contiguous memory if we don't
|
||||
* have DMM for remappign discontiguous buffers
|
||||
*/
|
||||
omap_obj->vaddr = dma_alloc_writecombine(dev->dev, size,
|
||||
&omap_obj->paddr, GFP_KERNEL);
|
||||
omap_obj->vaddr = dma_alloc_wc(dev->dev, size,
|
||||
&omap_obj->paddr, GFP_KERNEL);
|
||||
if (!omap_obj->vaddr) {
|
||||
kfree(omap_obj);
|
||||
|
||||
|
||||
@@ -157,17 +157,15 @@ static void sti_cursor_atomic_update(struct drm_plane *drm_plane,
|
||||
cursor->height = src_h;
|
||||
|
||||
if (cursor->pixmap.base)
|
||||
dma_free_writecombine(cursor->dev,
|
||||
cursor->pixmap.size,
|
||||
cursor->pixmap.base,
|
||||
cursor->pixmap.paddr);
|
||||
dma_free_wc(cursor->dev, cursor->pixmap.size,
|
||||
cursor->pixmap.base, cursor->pixmap.paddr);
|
||||
|
||||
cursor->pixmap.size = cursor->width * cursor->height;
|
||||
|
||||
cursor->pixmap.base = dma_alloc_writecombine(cursor->dev,
|
||||
cursor->pixmap.size,
|
||||
&cursor->pixmap.paddr,
|
||||
GFP_KERNEL | GFP_DMA);
|
||||
cursor->pixmap.base = dma_alloc_wc(cursor->dev,
|
||||
cursor->pixmap.size,
|
||||
&cursor->pixmap.paddr,
|
||||
GFP_KERNEL | GFP_DMA);
|
||||
if (!cursor->pixmap.base) {
|
||||
DRM_ERROR("Failed to allocate memory for pixmap\n");
|
||||
return;
|
||||
@@ -252,8 +250,8 @@ struct drm_plane *sti_cursor_create(struct drm_device *drm_dev,
|
||||
|
||||
/* Allocate clut buffer */
|
||||
size = 0x100 * sizeof(unsigned short);
|
||||
cursor->clut = dma_alloc_writecombine(dev, size, &cursor->clut_paddr,
|
||||
GFP_KERNEL | GFP_DMA);
|
||||
cursor->clut = dma_alloc_wc(dev, size, &cursor->clut_paddr,
|
||||
GFP_KERNEL | GFP_DMA);
|
||||
|
||||
if (!cursor->clut) {
|
||||
DRM_ERROR("Failed to allocate memory for cursor clut\n");
|
||||
@@ -286,7 +284,7 @@ struct drm_plane *sti_cursor_create(struct drm_device *drm_dev,
|
||||
return &cursor->plane.drm_plane;
|
||||
|
||||
err_plane:
|
||||
dma_free_writecombine(dev, size, cursor->clut, cursor->clut_paddr);
|
||||
dma_free_wc(dev, size, cursor->clut, cursor->clut_paddr);
|
||||
err_clut:
|
||||
devm_kfree(dev, cursor);
|
||||
return NULL;
|
||||
|
||||
@@ -312,8 +312,7 @@ static void sti_gdp_init(struct sti_gdp *gdp)
|
||||
/* Allocate all the nodes within a single memory page */
|
||||
size = sizeof(struct sti_gdp_node) *
|
||||
GDP_NODE_PER_FIELD * GDP_NODE_NB_BANK;
|
||||
base = dma_alloc_writecombine(gdp->dev,
|
||||
size, &dma_addr, GFP_KERNEL | GFP_DMA);
|
||||
base = dma_alloc_wc(gdp->dev, size, &dma_addr, GFP_KERNEL | GFP_DMA);
|
||||
|
||||
if (!base) {
|
||||
DRM_ERROR("Failed to allocate memory for GDP node\n");
|
||||
|
||||
@@ -617,9 +617,9 @@ static void sti_hqvdp_init(struct sti_hqvdp *hqvdp)
|
||||
|
||||
/* Allocate memory for the VDP commands */
|
||||
size = NB_VDP_CMD * sizeof(struct sti_hqvdp_cmd);
|
||||
hqvdp->hqvdp_cmd = dma_alloc_writecombine(hqvdp->dev, size,
|
||||
&hqvdp->hqvdp_cmd_paddr,
|
||||
GFP_KERNEL | GFP_DMA);
|
||||
hqvdp->hqvdp_cmd = dma_alloc_wc(hqvdp->dev, size,
|
||||
&hqvdp->hqvdp_cmd_paddr,
|
||||
GFP_KERNEL | GFP_DMA);
|
||||
if (!hqvdp->hqvdp_cmd) {
|
||||
DRM_ERROR("Failed to allocate memory for VDP cmd\n");
|
||||
return;
|
||||
|
||||
@@ -175,8 +175,7 @@ static void tegra_bo_free(struct drm_device *drm, struct tegra_bo *bo)
|
||||
sg_free_table(bo->sgt);
|
||||
kfree(bo->sgt);
|
||||
} else if (bo->vaddr) {
|
||||
dma_free_writecombine(drm->dev, bo->gem.size, bo->vaddr,
|
||||
bo->paddr);
|
||||
dma_free_wc(drm->dev, bo->gem.size, bo->vaddr, bo->paddr);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -233,8 +232,8 @@ static int tegra_bo_alloc(struct drm_device *drm, struct tegra_bo *bo)
|
||||
} else {
|
||||
size_t size = bo->gem.size;
|
||||
|
||||
bo->vaddr = dma_alloc_writecombine(drm->dev, size, &bo->paddr,
|
||||
GFP_KERNEL | __GFP_NOWARN);
|
||||
bo->vaddr = dma_alloc_wc(drm->dev, size, &bo->paddr,
|
||||
GFP_KERNEL | __GFP_NOWARN);
|
||||
if (!bo->vaddr) {
|
||||
dev_err(drm->dev,
|
||||
"failed to allocate buffer of size %zu\n",
|
||||
@@ -472,8 +471,8 @@ int tegra_drm_mmap(struct file *file, struct vm_area_struct *vma)
|
||||
vma->vm_flags &= ~VM_PFNMAP;
|
||||
vma->vm_pgoff = 0;
|
||||
|
||||
ret = dma_mmap_writecombine(gem->dev->dev, vma, bo->vaddr,
|
||||
bo->paddr, gem->size);
|
||||
ret = dma_mmap_wc(gem->dev->dev, vma, bo->vaddr, bo->paddr,
|
||||
gem->size);
|
||||
if (ret) {
|
||||
drm_gem_vm_close(vma);
|
||||
return ret;
|
||||
|
||||
@@ -398,9 +398,8 @@ int vc4_mmap(struct file *filp, struct vm_area_struct *vma)
|
||||
vma->vm_flags &= ~VM_PFNMAP;
|
||||
vma->vm_pgoff = 0;
|
||||
|
||||
ret = dma_mmap_writecombine(bo->base.base.dev->dev, vma,
|
||||
bo->base.vaddr, bo->base.paddr,
|
||||
vma->vm_end - vma->vm_start);
|
||||
ret = dma_mmap_wc(bo->base.base.dev->dev, vma, bo->base.vaddr,
|
||||
bo->base.paddr, vma->vm_end - vma->vm_start);
|
||||
if (ret)
|
||||
drm_gem_vm_close(vma);
|
||||
|
||||
|
||||
@@ -52,8 +52,8 @@ static void host1x_pushbuffer_destroy(struct push_buffer *pb)
|
||||
struct host1x *host1x = cdma_to_host1x(cdma);
|
||||
|
||||
if (pb->phys != 0)
|
||||
dma_free_writecombine(host1x->dev, pb->size_bytes + 4,
|
||||
pb->mapped, pb->phys);
|
||||
dma_free_wc(host1x->dev, pb->size_bytes + 4, pb->mapped,
|
||||
pb->phys);
|
||||
|
||||
pb->mapped = NULL;
|
||||
pb->phys = 0;
|
||||
@@ -76,8 +76,8 @@ static int host1x_pushbuffer_init(struct push_buffer *pb)
|
||||
pb->pos = 0;
|
||||
|
||||
/* allocate and map pushbuffer memory */
|
||||
pb->mapped = dma_alloc_writecombine(host1x->dev, pb->size_bytes + 4,
|
||||
&pb->phys, GFP_KERNEL);
|
||||
pb->mapped = dma_alloc_wc(host1x->dev, pb->size_bytes + 4, &pb->phys,
|
||||
GFP_KERNEL);
|
||||
if (!pb->mapped)
|
||||
goto fail;
|
||||
|
||||
|
||||
@@ -467,9 +467,8 @@ static inline int copy_gathers(struct host1x_job *job, struct device *dev)
|
||||
size += g->words * sizeof(u32);
|
||||
}
|
||||
|
||||
job->gather_copy_mapped = dma_alloc_writecombine(dev, size,
|
||||
&job->gather_copy,
|
||||
GFP_KERNEL);
|
||||
job->gather_copy_mapped = dma_alloc_wc(dev, size, &job->gather_copy,
|
||||
GFP_KERNEL);
|
||||
if (!job->gather_copy_mapped) {
|
||||
job->gather_copy_mapped = NULL;
|
||||
return -ENOMEM;
|
||||
@@ -578,9 +577,8 @@ void host1x_job_unpin(struct host1x_job *job)
|
||||
job->num_unpins = 0;
|
||||
|
||||
if (job->gather_copy_size)
|
||||
dma_free_writecombine(job->channel->dev, job->gather_copy_size,
|
||||
job->gather_copy_mapped,
|
||||
job->gather_copy);
|
||||
dma_free_wc(job->channel->dev, job->gather_copy_size,
|
||||
job->gather_copy_mapped, job->gather_copy);
|
||||
}
|
||||
EXPORT_SYMBOL(host1x_job_unpin);
|
||||
|
||||
|
||||
@@ -1455,9 +1455,9 @@ static int coda_alloc_bitstream_buffer(struct coda_ctx *ctx,
|
||||
return 0;
|
||||
|
||||
ctx->bitstream.size = roundup_pow_of_two(q_data->sizeimage * 2);
|
||||
ctx->bitstream.vaddr = dma_alloc_writecombine(
|
||||
&ctx->dev->plat_dev->dev, ctx->bitstream.size,
|
||||
&ctx->bitstream.paddr, GFP_KERNEL);
|
||||
ctx->bitstream.vaddr = dma_alloc_wc(&ctx->dev->plat_dev->dev,
|
||||
ctx->bitstream.size,
|
||||
&ctx->bitstream.paddr, GFP_KERNEL);
|
||||
if (!ctx->bitstream.vaddr) {
|
||||
v4l2_err(&ctx->dev->v4l2_dev,
|
||||
"failed to allocate bitstream ringbuffer");
|
||||
@@ -1474,8 +1474,8 @@ static void coda_free_bitstream_buffer(struct coda_ctx *ctx)
|
||||
if (ctx->bitstream.vaddr == NULL)
|
||||
return;
|
||||
|
||||
dma_free_writecombine(&ctx->dev->plat_dev->dev, ctx->bitstream.size,
|
||||
ctx->bitstream.vaddr, ctx->bitstream.paddr);
|
||||
dma_free_wc(&ctx->dev->plat_dev->dev, ctx->bitstream.size,
|
||||
ctx->bitstream.vaddr, ctx->bitstream.paddr);
|
||||
ctx->bitstream.vaddr = NULL;
|
||||
kfifo_init(&ctx->bitstream_fifo, NULL, 0);
|
||||
}
|
||||
|
||||
@@ -1040,8 +1040,8 @@ static int acornfb_probe(struct platform_device *dev)
|
||||
* for the framebuffer if we are not using
|
||||
* VRAM.
|
||||
*/
|
||||
base = dma_alloc_writecombine(current_par.dev, size, &handle,
|
||||
GFP_KERNEL);
|
||||
base = dma_alloc_wc(current_par.dev, size, &handle,
|
||||
GFP_KERNEL);
|
||||
if (base == NULL) {
|
||||
printk(KERN_ERR "acornfb: unable to allocate screen "
|
||||
"memory\n");
|
||||
|
||||
@@ -154,8 +154,8 @@ int versatile_clcd_setup_dma(struct clcd_fb *fb, unsigned long framesize)
|
||||
{
|
||||
dma_addr_t dma;
|
||||
|
||||
fb->fb.screen_base = dma_alloc_writecombine(&fb->dev->dev, framesize,
|
||||
&dma, GFP_KERNEL);
|
||||
fb->fb.screen_base = dma_alloc_wc(&fb->dev->dev, framesize, &dma,
|
||||
GFP_KERNEL);
|
||||
if (!fb->fb.screen_base) {
|
||||
pr_err("CLCD: unable to map framebuffer\n");
|
||||
return -ENOMEM;
|
||||
@@ -169,14 +169,12 @@ int versatile_clcd_setup_dma(struct clcd_fb *fb, unsigned long framesize)
|
||||
|
||||
int versatile_clcd_mmap_dma(struct clcd_fb *fb, struct vm_area_struct *vma)
|
||||
{
|
||||
return dma_mmap_writecombine(&fb->dev->dev, vma,
|
||||
fb->fb.screen_base,
|
||||
fb->fb.fix.smem_start,
|
||||
fb->fb.fix.smem_len);
|
||||
return dma_mmap_wc(&fb->dev->dev, vma, fb->fb.screen_base,
|
||||
fb->fb.fix.smem_start, fb->fb.fix.smem_len);
|
||||
}
|
||||
|
||||
void versatile_clcd_remove_dma(struct clcd_fb *fb)
|
||||
{
|
||||
dma_free_writecombine(&fb->dev->dev, fb->fb.fix.smem_len,
|
||||
fb->fb.screen_base, fb->fb.fix.smem_start);
|
||||
dma_free_wc(&fb->dev->dev, fb->fb.fix.smem_len, fb->fb.screen_base,
|
||||
fb->fb.fix.smem_start);
|
||||
}
|
||||
|
||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user