Merge branch 'stable/ttm.pci-api.v5' of git://git.kernel.org/pub/scm/linux/kernel/git/konrad/xen into drm-next
* 'stable/ttm.pci-api.v5' of git://git.kernel.org/pub/scm/linux/kernel/git/konrad/xen: ttm: Include the 'struct dev' when using the DMA API. nouveau/ttm/PCIe: Use dma_addr if TTM has set it. radeon/ttm/PCIe: Use dma_addr if TTM has set it. ttm: Expand (*populate) to support an array of DMA addresses. ttm: Utilize the DMA API for pages that have TTM_PAGE_FLAG_DMA32 set. ttm: Introduce a placeholder for DMA (bus) addresses.
This commit is contained in:
@@ -47,7 +47,8 @@ struct ttm_agp_backend {
|
||||
|
||||
static int ttm_agp_populate(struct ttm_backend *backend,
|
||||
unsigned long num_pages, struct page **pages,
|
||||
struct page *dummy_read_page)
|
||||
struct page *dummy_read_page,
|
||||
dma_addr_t *dma_addrs)
|
||||
{
|
||||
struct ttm_agp_backend *agp_be =
|
||||
container_of(backend, struct ttm_agp_backend, backend);
|
||||
|
@@ -38,6 +38,7 @@
|
||||
#include <linux/mm.h>
|
||||
#include <linux/seq_file.h> /* for seq_printf */
|
||||
#include <linux/slab.h>
|
||||
#include <linux/dma-mapping.h>
|
||||
|
||||
#include <asm/atomic.h>
|
||||
|
||||
@@ -662,7 +663,8 @@ out:
|
||||
* cached pages.
|
||||
*/
|
||||
int ttm_get_pages(struct list_head *pages, int flags,
|
||||
enum ttm_caching_state cstate, unsigned count)
|
||||
enum ttm_caching_state cstate, unsigned count,
|
||||
dma_addr_t *dma_address, struct device *dev)
|
||||
{
|
||||
struct ttm_page_pool *pool = ttm_get_pool(flags, cstate);
|
||||
struct page *p = NULL;
|
||||
@@ -681,14 +683,22 @@ int ttm_get_pages(struct list_head *pages, int flags,
|
||||
gfp_flags |= GFP_HIGHUSER;
|
||||
|
||||
for (r = 0; r < count; ++r) {
|
||||
p = alloc_page(gfp_flags);
|
||||
if ((flags & TTM_PAGE_FLAG_DMA32) && dma_address) {
|
||||
void *addr;
|
||||
addr = dma_alloc_coherent(dev, PAGE_SIZE,
|
||||
&dma_address[r],
|
||||
gfp_flags);
|
||||
if (addr == NULL)
|
||||
return -ENOMEM;
|
||||
p = virt_to_page(addr);
|
||||
} else
|
||||
p = alloc_page(gfp_flags);
|
||||
if (!p) {
|
||||
|
||||
printk(KERN_ERR TTM_PFX
|
||||
"Unable to allocate page.");
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
||||
list_add(&p->lru, pages);
|
||||
}
|
||||
return 0;
|
||||
@@ -720,7 +730,7 @@ int ttm_get_pages(struct list_head *pages, int flags,
|
||||
printk(KERN_ERR TTM_PFX
|
||||
"Failed to allocate extra pages "
|
||||
"for large request.");
|
||||
ttm_put_pages(pages, 0, flags, cstate);
|
||||
ttm_put_pages(pages, 0, flags, cstate, NULL, NULL);
|
||||
return r;
|
||||
}
|
||||
}
|
||||
@@ -731,17 +741,30 @@ int ttm_get_pages(struct list_head *pages, int flags,
|
||||
|
||||
/* Put all pages in pages list to correct pool to wait for reuse */
|
||||
void ttm_put_pages(struct list_head *pages, unsigned page_count, int flags,
|
||||
enum ttm_caching_state cstate)
|
||||
enum ttm_caching_state cstate, dma_addr_t *dma_address,
|
||||
struct device *dev)
|
||||
{
|
||||
unsigned long irq_flags;
|
||||
struct ttm_page_pool *pool = ttm_get_pool(flags, cstate);
|
||||
struct page *p, *tmp;
|
||||
unsigned r;
|
||||
|
||||
if (pool == NULL) {
|
||||
/* No pool for this memory type so free the pages */
|
||||
|
||||
r = page_count-1;
|
||||
list_for_each_entry_safe(p, tmp, pages, lru) {
|
||||
__free_page(p);
|
||||
if ((flags & TTM_PAGE_FLAG_DMA32) && dma_address) {
|
||||
void *addr = page_address(p);
|
||||
WARN_ON(!addr || !dma_address[r]);
|
||||
if (addr)
|
||||
dma_free_coherent(dev, PAGE_SIZE,
|
||||
addr,
|
||||
dma_address[r]);
|
||||
dma_address[r] = 0;
|
||||
} else
|
||||
__free_page(p);
|
||||
r--;
|
||||
}
|
||||
/* Make the pages list empty */
|
||||
INIT_LIST_HEAD(pages);
|
||||
|
@@ -49,12 +49,16 @@ static int ttm_tt_swapin(struct ttm_tt *ttm);
|
||||
static void ttm_tt_alloc_page_directory(struct ttm_tt *ttm)
|
||||
{
|
||||
ttm->pages = drm_calloc_large(ttm->num_pages, sizeof(*ttm->pages));
|
||||
ttm->dma_address = drm_calloc_large(ttm->num_pages,
|
||||
sizeof(*ttm->dma_address));
|
||||
}
|
||||
|
||||
static void ttm_tt_free_page_directory(struct ttm_tt *ttm)
|
||||
{
|
||||
drm_free_large(ttm->pages);
|
||||
ttm->pages = NULL;
|
||||
drm_free_large(ttm->dma_address);
|
||||
ttm->dma_address = NULL;
|
||||
}
|
||||
|
||||
static void ttm_tt_free_user_pages(struct ttm_tt *ttm)
|
||||
@@ -105,7 +109,8 @@ static struct page *__ttm_tt_get_page(struct ttm_tt *ttm, int index)
|
||||
|
||||
INIT_LIST_HEAD(&h);
|
||||
|
||||
ret = ttm_get_pages(&h, ttm->page_flags, ttm->caching_state, 1);
|
||||
ret = ttm_get_pages(&h, ttm->page_flags, ttm->caching_state, 1,
|
||||
&ttm->dma_address[index], ttm->be->bdev->dev);
|
||||
|
||||
if (ret != 0)
|
||||
return NULL;
|
||||
@@ -164,7 +169,7 @@ int ttm_tt_populate(struct ttm_tt *ttm)
|
||||
}
|
||||
|
||||
be->func->populate(be, ttm->num_pages, ttm->pages,
|
||||
ttm->dummy_read_page);
|
||||
ttm->dummy_read_page, ttm->dma_address);
|
||||
ttm->state = tt_unbound;
|
||||
return 0;
|
||||
}
|
||||
@@ -298,7 +303,8 @@ static void ttm_tt_free_alloced_pages(struct ttm_tt *ttm)
|
||||
count++;
|
||||
}
|
||||
}
|
||||
ttm_put_pages(&h, count, ttm->page_flags, ttm->caching_state);
|
||||
ttm_put_pages(&h, count, ttm->page_flags, ttm->caching_state,
|
||||
ttm->dma_address, ttm->be->bdev->dev);
|
||||
ttm->state = tt_unpopulated;
|
||||
ttm->first_himem_page = ttm->num_pages;
|
||||
ttm->last_lomem_page = -1;
|
||||
|
Reference in New Issue
Block a user