0001
0002
0003
0004
0005
0006
0007
0008
0009
0010
0011
0012 #include <linux/dma-mapping.h>
0013 #include <linux/sched/signal.h>
0014 #include <linux/uaccess.h>
0015 #include <linux/mm.h>
0016
0017 #include "dfl-afu.h"
0018
0019 void afu_dma_region_init(struct dfl_feature_platform_data *pdata)
0020 {
0021 struct dfl_afu *afu = dfl_fpga_pdata_get_private(pdata);
0022
0023 afu->dma_regions = RB_ROOT;
0024 }
0025
0026
0027
0028
0029
0030
0031
0032
0033
0034 static int afu_dma_pin_pages(struct dfl_feature_platform_data *pdata,
0035 struct dfl_afu_dma_region *region)
0036 {
0037 int npages = region->length >> PAGE_SHIFT;
0038 struct device *dev = &pdata->dev->dev;
0039 int ret, pinned;
0040
0041 ret = account_locked_vm(current->mm, npages, true);
0042 if (ret)
0043 return ret;
0044
0045 region->pages = kcalloc(npages, sizeof(struct page *), GFP_KERNEL);
0046 if (!region->pages) {
0047 ret = -ENOMEM;
0048 goto unlock_vm;
0049 }
0050
0051 pinned = pin_user_pages_fast(region->user_addr, npages, FOLL_WRITE,
0052 region->pages);
0053 if (pinned < 0) {
0054 ret = pinned;
0055 goto free_pages;
0056 } else if (pinned != npages) {
0057 ret = -EFAULT;
0058 goto unpin_pages;
0059 }
0060
0061 dev_dbg(dev, "%d pages pinned\n", pinned);
0062
0063 return 0;
0064
0065 unpin_pages:
0066 unpin_user_pages(region->pages, pinned);
0067 free_pages:
0068 kfree(region->pages);
0069 unlock_vm:
0070 account_locked_vm(current->mm, npages, false);
0071 return ret;
0072 }
0073
0074
0075
0076
0077
0078
0079
0080
0081
0082 static void afu_dma_unpin_pages(struct dfl_feature_platform_data *pdata,
0083 struct dfl_afu_dma_region *region)
0084 {
0085 long npages = region->length >> PAGE_SHIFT;
0086 struct device *dev = &pdata->dev->dev;
0087
0088 unpin_user_pages(region->pages, npages);
0089 kfree(region->pages);
0090 account_locked_vm(current->mm, npages, false);
0091
0092 dev_dbg(dev, "%ld pages unpinned\n", npages);
0093 }
0094
0095
0096
0097
0098
0099
0100
0101
0102 static bool afu_dma_check_continuous_pages(struct dfl_afu_dma_region *region)
0103 {
0104 int npages = region->length >> PAGE_SHIFT;
0105 int i;
0106
0107 for (i = 0; i < npages - 1; i++)
0108 if (page_to_pfn(region->pages[i]) + 1 !=
0109 page_to_pfn(region->pages[i + 1]))
0110 return false;
0111
0112 return true;
0113 }
0114
0115
0116
0117
0118
0119
0120
0121
0122
0123
0124 static bool dma_region_check_iova(struct dfl_afu_dma_region *region,
0125 u64 iova, u64 size)
0126 {
0127 if (!size && region->iova != iova)
0128 return false;
0129
0130 return (region->iova <= iova) &&
0131 (region->length + region->iova >= iova + size);
0132 }
0133
0134
0135
0136
0137
0138
0139
0140
0141
0142
0143 static int afu_dma_region_add(struct dfl_feature_platform_data *pdata,
0144 struct dfl_afu_dma_region *region)
0145 {
0146 struct dfl_afu *afu = dfl_fpga_pdata_get_private(pdata);
0147 struct rb_node **new, *parent = NULL;
0148
0149 dev_dbg(&pdata->dev->dev, "add region (iova = %llx)\n",
0150 (unsigned long long)region->iova);
0151
0152 new = &afu->dma_regions.rb_node;
0153
0154 while (*new) {
0155 struct dfl_afu_dma_region *this;
0156
0157 this = container_of(*new, struct dfl_afu_dma_region, node);
0158
0159 parent = *new;
0160
0161 if (dma_region_check_iova(this, region->iova, region->length))
0162 return -EEXIST;
0163
0164 if (region->iova < this->iova)
0165 new = &((*new)->rb_left);
0166 else if (region->iova > this->iova)
0167 new = &((*new)->rb_right);
0168 else
0169 return -EEXIST;
0170 }
0171
0172 rb_link_node(®ion->node, parent, new);
0173 rb_insert_color(®ion->node, &afu->dma_regions);
0174
0175 return 0;
0176 }
0177
0178
0179
0180
0181
0182
0183
0184
0185 static void afu_dma_region_remove(struct dfl_feature_platform_data *pdata,
0186 struct dfl_afu_dma_region *region)
0187 {
0188 struct dfl_afu *afu;
0189
0190 dev_dbg(&pdata->dev->dev, "del region (iova = %llx)\n",
0191 (unsigned long long)region->iova);
0192
0193 afu = dfl_fpga_pdata_get_private(pdata);
0194 rb_erase(®ion->node, &afu->dma_regions);
0195 }
0196
0197
0198
0199
0200
0201
0202
0203 void afu_dma_region_destroy(struct dfl_feature_platform_data *pdata)
0204 {
0205 struct dfl_afu *afu = dfl_fpga_pdata_get_private(pdata);
0206 struct rb_node *node = rb_first(&afu->dma_regions);
0207 struct dfl_afu_dma_region *region;
0208
0209 while (node) {
0210 region = container_of(node, struct dfl_afu_dma_region, node);
0211
0212 dev_dbg(&pdata->dev->dev, "del region (iova = %llx)\n",
0213 (unsigned long long)region->iova);
0214
0215 rb_erase(node, &afu->dma_regions);
0216
0217 if (region->iova)
0218 dma_unmap_page(dfl_fpga_pdata_to_parent(pdata),
0219 region->iova, region->length,
0220 DMA_BIDIRECTIONAL);
0221
0222 if (region->pages)
0223 afu_dma_unpin_pages(pdata, region);
0224
0225 node = rb_next(node);
0226 kfree(region);
0227 }
0228 }
0229
0230
0231
0232
0233
0234
0235
0236
0237
0238
0239
0240
0241
0242
0243
0244 struct dfl_afu_dma_region *
0245 afu_dma_region_find(struct dfl_feature_platform_data *pdata, u64 iova, u64 size)
0246 {
0247 struct dfl_afu *afu = dfl_fpga_pdata_get_private(pdata);
0248 struct rb_node *node = afu->dma_regions.rb_node;
0249 struct device *dev = &pdata->dev->dev;
0250
0251 while (node) {
0252 struct dfl_afu_dma_region *region;
0253
0254 region = container_of(node, struct dfl_afu_dma_region, node);
0255
0256 if (dma_region_check_iova(region, iova, size)) {
0257 dev_dbg(dev, "find region (iova = %llx)\n",
0258 (unsigned long long)region->iova);
0259 return region;
0260 }
0261
0262 if (iova < region->iova)
0263 node = node->rb_left;
0264 else if (iova > region->iova)
0265 node = node->rb_right;
0266 else
0267
0268 break;
0269 }
0270
0271 dev_dbg(dev, "region with iova %llx and size %llx is not found\n",
0272 (unsigned long long)iova, (unsigned long long)size);
0273
0274 return NULL;
0275 }
0276
0277
0278
0279
0280
0281
0282
0283
0284 static struct dfl_afu_dma_region *
0285 afu_dma_region_find_iova(struct dfl_feature_platform_data *pdata, u64 iova)
0286 {
0287 return afu_dma_region_find(pdata, iova, 0);
0288 }
0289
0290
0291
0292
0293
0294
0295
0296
0297
0298
0299
0300
0301 int afu_dma_map_region(struct dfl_feature_platform_data *pdata,
0302 u64 user_addr, u64 length, u64 *iova)
0303 {
0304 struct dfl_afu_dma_region *region;
0305 int ret;
0306
0307
0308
0309
0310
0311 if (!PAGE_ALIGNED(user_addr) || !PAGE_ALIGNED(length) || !length)
0312 return -EINVAL;
0313
0314
0315 if (user_addr + length < user_addr)
0316 return -EINVAL;
0317
0318 region = kzalloc(sizeof(*region), GFP_KERNEL);
0319 if (!region)
0320 return -ENOMEM;
0321
0322 region->user_addr = user_addr;
0323 region->length = length;
0324
0325
0326 ret = afu_dma_pin_pages(pdata, region);
0327 if (ret) {
0328 dev_err(&pdata->dev->dev, "failed to pin memory region\n");
0329 goto free_region;
0330 }
0331
0332
0333 if (!afu_dma_check_continuous_pages(region)) {
0334 dev_err(&pdata->dev->dev, "pages are not continuous\n");
0335 ret = -EINVAL;
0336 goto unpin_pages;
0337 }
0338
0339
0340 region->iova = dma_map_page(dfl_fpga_pdata_to_parent(pdata),
0341 region->pages[0], 0,
0342 region->length,
0343 DMA_BIDIRECTIONAL);
0344 if (dma_mapping_error(dfl_fpga_pdata_to_parent(pdata), region->iova)) {
0345 dev_err(&pdata->dev->dev, "failed to map for dma\n");
0346 ret = -EFAULT;
0347 goto unpin_pages;
0348 }
0349
0350 *iova = region->iova;
0351
0352 mutex_lock(&pdata->lock);
0353 ret = afu_dma_region_add(pdata, region);
0354 mutex_unlock(&pdata->lock);
0355 if (ret) {
0356 dev_err(&pdata->dev->dev, "failed to add dma region\n");
0357 goto unmap_dma;
0358 }
0359
0360 return 0;
0361
0362 unmap_dma:
0363 dma_unmap_page(dfl_fpga_pdata_to_parent(pdata),
0364 region->iova, region->length, DMA_BIDIRECTIONAL);
0365 unpin_pages:
0366 afu_dma_unpin_pages(pdata, region);
0367 free_region:
0368 kfree(region);
0369 return ret;
0370 }
0371
0372
0373
0374
0375
0376
0377
0378
0379
0380 int afu_dma_unmap_region(struct dfl_feature_platform_data *pdata, u64 iova)
0381 {
0382 struct dfl_afu_dma_region *region;
0383
0384 mutex_lock(&pdata->lock);
0385 region = afu_dma_region_find_iova(pdata, iova);
0386 if (!region) {
0387 mutex_unlock(&pdata->lock);
0388 return -EINVAL;
0389 }
0390
0391 if (region->in_use) {
0392 mutex_unlock(&pdata->lock);
0393 return -EBUSY;
0394 }
0395
0396 afu_dma_region_remove(pdata, region);
0397 mutex_unlock(&pdata->lock);
0398
0399 dma_unmap_page(dfl_fpga_pdata_to_parent(pdata),
0400 region->iova, region->length, DMA_BIDIRECTIONAL);
0401 afu_dma_unpin_pages(pdata, region);
0402 kfree(region);
0403
0404 return 0;
0405 }