123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405 |
- // SPDX-License-Identifier: GPL-2.0
- /*
- * Driver for FPGA Accelerated Function Unit (AFU) DMA Region Management
- *
- * Copyright (C) 2017-2018 Intel Corporation, Inc.
- *
- * Authors:
- * Wu Hao <[email protected]>
- * Xiao Guangrong <[email protected]>
- */
- #include <linux/dma-mapping.h>
- #include <linux/sched/signal.h>
- #include <linux/uaccess.h>
- #include <linux/mm.h>
- #include "dfl-afu.h"
- void afu_dma_region_init(struct dfl_feature_platform_data *pdata)
- {
- struct dfl_afu *afu = dfl_fpga_pdata_get_private(pdata);
- afu->dma_regions = RB_ROOT;
- }
- /**
- * afu_dma_pin_pages - pin pages of given dma memory region
- * @pdata: feature device platform data
- * @region: dma memory region to be pinned
- *
- * Pin all the pages of given dfl_afu_dma_region.
- * Return 0 for success or negative error code.
- */
- static int afu_dma_pin_pages(struct dfl_feature_platform_data *pdata,
- struct dfl_afu_dma_region *region)
- {
- int npages = region->length >> PAGE_SHIFT;
- struct device *dev = &pdata->dev->dev;
- int ret, pinned;
- ret = account_locked_vm(current->mm, npages, true);
- if (ret)
- return ret;
- region->pages = kcalloc(npages, sizeof(struct page *), GFP_KERNEL);
- if (!region->pages) {
- ret = -ENOMEM;
- goto unlock_vm;
- }
- pinned = pin_user_pages_fast(region->user_addr, npages, FOLL_WRITE,
- region->pages);
- if (pinned < 0) {
- ret = pinned;
- goto free_pages;
- } else if (pinned != npages) {
- ret = -EFAULT;
- goto unpin_pages;
- }
- dev_dbg(dev, "%d pages pinned\n", pinned);
- return 0;
- unpin_pages:
- unpin_user_pages(region->pages, pinned);
- free_pages:
- kfree(region->pages);
- unlock_vm:
- account_locked_vm(current->mm, npages, false);
- return ret;
- }
- /**
- * afu_dma_unpin_pages - unpin pages of given dma memory region
- * @pdata: feature device platform data
- * @region: dma memory region to be unpinned
- *
- * Unpin all the pages of given dfl_afu_dma_region.
- * Return 0 for success or negative error code.
- */
- static void afu_dma_unpin_pages(struct dfl_feature_platform_data *pdata,
- struct dfl_afu_dma_region *region)
- {
- long npages = region->length >> PAGE_SHIFT;
- struct device *dev = &pdata->dev->dev;
- unpin_user_pages(region->pages, npages);
- kfree(region->pages);
- account_locked_vm(current->mm, npages, false);
- dev_dbg(dev, "%ld pages unpinned\n", npages);
- }
- /**
- * afu_dma_check_continuous_pages - check if pages are continuous
- * @region: dma memory region
- *
- * Return true if pages of given dma memory region have continuous physical
- * address, otherwise return false.
- */
- static bool afu_dma_check_continuous_pages(struct dfl_afu_dma_region *region)
- {
- int npages = region->length >> PAGE_SHIFT;
- int i;
- for (i = 0; i < npages - 1; i++)
- if (page_to_pfn(region->pages[i]) + 1 !=
- page_to_pfn(region->pages[i + 1]))
- return false;
- return true;
- }
- /**
- * dma_region_check_iova - check if memory area is fully contained in the region
- * @region: dma memory region
- * @iova: address of the dma memory area
- * @size: size of the dma memory area
- *
- * Compare the dma memory area defined by @iova and @size with given dma region.
- * Return true if memory area is fully contained in the region, otherwise false.
- */
- static bool dma_region_check_iova(struct dfl_afu_dma_region *region,
- u64 iova, u64 size)
- {
- if (!size && region->iova != iova)
- return false;
- return (region->iova <= iova) &&
- (region->length + region->iova >= iova + size);
- }
- /**
- * afu_dma_region_add - add given dma region to rbtree
- * @pdata: feature device platform data
- * @region: dma region to be added
- *
- * Return 0 for success, -EEXIST if dma region has already been added.
- *
- * Needs to be called with pdata->lock heold.
- */
- static int afu_dma_region_add(struct dfl_feature_platform_data *pdata,
- struct dfl_afu_dma_region *region)
- {
- struct dfl_afu *afu = dfl_fpga_pdata_get_private(pdata);
- struct rb_node **new, *parent = NULL;
- dev_dbg(&pdata->dev->dev, "add region (iova = %llx)\n",
- (unsigned long long)region->iova);
- new = &afu->dma_regions.rb_node;
- while (*new) {
- struct dfl_afu_dma_region *this;
- this = container_of(*new, struct dfl_afu_dma_region, node);
- parent = *new;
- if (dma_region_check_iova(this, region->iova, region->length))
- return -EEXIST;
- if (region->iova < this->iova)
- new = &((*new)->rb_left);
- else if (region->iova > this->iova)
- new = &((*new)->rb_right);
- else
- return -EEXIST;
- }
- rb_link_node(®ion->node, parent, new);
- rb_insert_color(®ion->node, &afu->dma_regions);
- return 0;
- }
- /**
- * afu_dma_region_remove - remove given dma region from rbtree
- * @pdata: feature device platform data
- * @region: dma region to be removed
- *
- * Needs to be called with pdata->lock heold.
- */
- static void afu_dma_region_remove(struct dfl_feature_platform_data *pdata,
- struct dfl_afu_dma_region *region)
- {
- struct dfl_afu *afu;
- dev_dbg(&pdata->dev->dev, "del region (iova = %llx)\n",
- (unsigned long long)region->iova);
- afu = dfl_fpga_pdata_get_private(pdata);
- rb_erase(®ion->node, &afu->dma_regions);
- }
- /**
- * afu_dma_region_destroy - destroy all regions in rbtree
- * @pdata: feature device platform data
- *
- * Needs to be called with pdata->lock heold.
- */
- void afu_dma_region_destroy(struct dfl_feature_platform_data *pdata)
- {
- struct dfl_afu *afu = dfl_fpga_pdata_get_private(pdata);
- struct rb_node *node = rb_first(&afu->dma_regions);
- struct dfl_afu_dma_region *region;
- while (node) {
- region = container_of(node, struct dfl_afu_dma_region, node);
- dev_dbg(&pdata->dev->dev, "del region (iova = %llx)\n",
- (unsigned long long)region->iova);
- rb_erase(node, &afu->dma_regions);
- if (region->iova)
- dma_unmap_page(dfl_fpga_pdata_to_parent(pdata),
- region->iova, region->length,
- DMA_BIDIRECTIONAL);
- if (region->pages)
- afu_dma_unpin_pages(pdata, region);
- node = rb_next(node);
- kfree(region);
- }
- }
- /**
- * afu_dma_region_find - find the dma region from rbtree based on iova and size
- * @pdata: feature device platform data
- * @iova: address of the dma memory area
- * @size: size of the dma memory area
- *
- * It finds the dma region from the rbtree based on @iova and @size:
- * - if @size == 0, it finds the dma region which starts from @iova
- * - otherwise, it finds the dma region which fully contains
- * [@iova, @iova+size)
- * If nothing is matched returns NULL.
- *
- * Needs to be called with pdata->lock held.
- */
- struct dfl_afu_dma_region *
- afu_dma_region_find(struct dfl_feature_platform_data *pdata, u64 iova, u64 size)
- {
- struct dfl_afu *afu = dfl_fpga_pdata_get_private(pdata);
- struct rb_node *node = afu->dma_regions.rb_node;
- struct device *dev = &pdata->dev->dev;
- while (node) {
- struct dfl_afu_dma_region *region;
- region = container_of(node, struct dfl_afu_dma_region, node);
- if (dma_region_check_iova(region, iova, size)) {
- dev_dbg(dev, "find region (iova = %llx)\n",
- (unsigned long long)region->iova);
- return region;
- }
- if (iova < region->iova)
- node = node->rb_left;
- else if (iova > region->iova)
- node = node->rb_right;
- else
- /* the iova region is not fully covered. */
- break;
- }
- dev_dbg(dev, "region with iova %llx and size %llx is not found\n",
- (unsigned long long)iova, (unsigned long long)size);
- return NULL;
- }
- /**
- * afu_dma_region_find_iova - find the dma region from rbtree by iova
- * @pdata: feature device platform data
- * @iova: address of the dma region
- *
- * Needs to be called with pdata->lock held.
- */
- static struct dfl_afu_dma_region *
- afu_dma_region_find_iova(struct dfl_feature_platform_data *pdata, u64 iova)
- {
- return afu_dma_region_find(pdata, iova, 0);
- }
- /**
- * afu_dma_map_region - map memory region for dma
- * @pdata: feature device platform data
- * @user_addr: address of the memory region
- * @length: size of the memory region
- * @iova: pointer of iova address
- *
- * Map memory region defined by @user_addr and @length, and return dma address
- * of the memory region via @iova.
- * Return 0 for success, otherwise error code.
- */
- int afu_dma_map_region(struct dfl_feature_platform_data *pdata,
- u64 user_addr, u64 length, u64 *iova)
- {
- struct dfl_afu_dma_region *region;
- int ret;
- /*
- * Check Inputs, only accept page-aligned user memory region with
- * valid length.
- */
- if (!PAGE_ALIGNED(user_addr) || !PAGE_ALIGNED(length) || !length)
- return -EINVAL;
- /* Check overflow */
- if (user_addr + length < user_addr)
- return -EINVAL;
- region = kzalloc(sizeof(*region), GFP_KERNEL);
- if (!region)
- return -ENOMEM;
- region->user_addr = user_addr;
- region->length = length;
- /* Pin the user memory region */
- ret = afu_dma_pin_pages(pdata, region);
- if (ret) {
- dev_err(&pdata->dev->dev, "failed to pin memory region\n");
- goto free_region;
- }
- /* Only accept continuous pages, return error else */
- if (!afu_dma_check_continuous_pages(region)) {
- dev_err(&pdata->dev->dev, "pages are not continuous\n");
- ret = -EINVAL;
- goto unpin_pages;
- }
- /* As pages are continuous then start to do DMA mapping */
- region->iova = dma_map_page(dfl_fpga_pdata_to_parent(pdata),
- region->pages[0], 0,
- region->length,
- DMA_BIDIRECTIONAL);
- if (dma_mapping_error(dfl_fpga_pdata_to_parent(pdata), region->iova)) {
- dev_err(&pdata->dev->dev, "failed to map for dma\n");
- ret = -EFAULT;
- goto unpin_pages;
- }
- *iova = region->iova;
- mutex_lock(&pdata->lock);
- ret = afu_dma_region_add(pdata, region);
- mutex_unlock(&pdata->lock);
- if (ret) {
- dev_err(&pdata->dev->dev, "failed to add dma region\n");
- goto unmap_dma;
- }
- return 0;
- unmap_dma:
- dma_unmap_page(dfl_fpga_pdata_to_parent(pdata),
- region->iova, region->length, DMA_BIDIRECTIONAL);
- unpin_pages:
- afu_dma_unpin_pages(pdata, region);
- free_region:
- kfree(region);
- return ret;
- }
- /**
- * afu_dma_unmap_region - unmap dma memory region
- * @pdata: feature device platform data
- * @iova: dma address of the region
- *
- * Unmap dma memory region based on @iova.
- * Return 0 for success, otherwise error code.
- */
- int afu_dma_unmap_region(struct dfl_feature_platform_data *pdata, u64 iova)
- {
- struct dfl_afu_dma_region *region;
- mutex_lock(&pdata->lock);
- region = afu_dma_region_find_iova(pdata, iova);
- if (!region) {
- mutex_unlock(&pdata->lock);
- return -EINVAL;
- }
- if (region->in_use) {
- mutex_unlock(&pdata->lock);
- return -EBUSY;
- }
- afu_dma_region_remove(pdata, region);
- mutex_unlock(&pdata->lock);
- dma_unmap_page(dfl_fpga_pdata_to_parent(pdata),
- region->iova, region->length, DMA_BIDIRECTIONAL);
- afu_dma_unpin_pages(pdata, region);
- kfree(region);
- return 0;
- }
|