0001
0002
0003
0004
0005
0006
0007
0008
0009
0010
0011
0012
0013 #include <linux/fs.h>
0014 #include <linux/pci.h>
0015 #include <linux/uaccess.h>
0016 #include <linux/io.h>
0017 #include <linux/vfio.h>
0018 #include <linux/vgaarb.h>
0019
0020 #include <linux/vfio_pci_core.h>
0021
0022 #ifdef __LITTLE_ENDIAN
0023 #define vfio_ioread64 ioread64
0024 #define vfio_iowrite64 iowrite64
0025 #define vfio_ioread32 ioread32
0026 #define vfio_iowrite32 iowrite32
0027 #define vfio_ioread16 ioread16
0028 #define vfio_iowrite16 iowrite16
0029 #else
0030 #define vfio_ioread64 ioread64be
0031 #define vfio_iowrite64 iowrite64be
0032 #define vfio_ioread32 ioread32be
0033 #define vfio_iowrite32 iowrite32be
0034 #define vfio_ioread16 ioread16be
0035 #define vfio_iowrite16 iowrite16be
0036 #endif
0037 #define vfio_ioread8 ioread8
0038 #define vfio_iowrite8 iowrite8
0039
0040 #define VFIO_IOWRITE(size) \
0041 static int vfio_pci_iowrite##size(struct vfio_pci_core_device *vdev, \
0042 bool test_mem, u##size val, void __iomem *io) \
0043 { \
0044 if (test_mem) { \
0045 down_read(&vdev->memory_lock); \
0046 if (!__vfio_pci_memory_enabled(vdev)) { \
0047 up_read(&vdev->memory_lock); \
0048 return -EIO; \
0049 } \
0050 } \
0051 \
0052 vfio_iowrite##size(val, io); \
0053 \
0054 if (test_mem) \
0055 up_read(&vdev->memory_lock); \
0056 \
0057 return 0; \
0058 }
0059
0060 VFIO_IOWRITE(8)
0061 VFIO_IOWRITE(16)
0062 VFIO_IOWRITE(32)
0063 #ifdef iowrite64
0064 VFIO_IOWRITE(64)
0065 #endif
0066
0067 #define VFIO_IOREAD(size) \
0068 static int vfio_pci_ioread##size(struct vfio_pci_core_device *vdev, \
0069 bool test_mem, u##size *val, void __iomem *io) \
0070 { \
0071 if (test_mem) { \
0072 down_read(&vdev->memory_lock); \
0073 if (!__vfio_pci_memory_enabled(vdev)) { \
0074 up_read(&vdev->memory_lock); \
0075 return -EIO; \
0076 } \
0077 } \
0078 \
0079 *val = vfio_ioread##size(io); \
0080 \
0081 if (test_mem) \
0082 up_read(&vdev->memory_lock); \
0083 \
0084 return 0; \
0085 }
0086
0087 VFIO_IOREAD(8)
0088 VFIO_IOREAD(16)
0089 VFIO_IOREAD(32)
0090
0091
0092
0093
0094
0095
0096
0097 static ssize_t do_io_rw(struct vfio_pci_core_device *vdev, bool test_mem,
0098 void __iomem *io, char __user *buf,
0099 loff_t off, size_t count, size_t x_start,
0100 size_t x_end, bool iswrite)
0101 {
0102 ssize_t done = 0;
0103 int ret;
0104
0105 while (count) {
0106 size_t fillable, filled;
0107
0108 if (off < x_start)
0109 fillable = min(count, (size_t)(x_start - off));
0110 else if (off >= x_end)
0111 fillable = count;
0112 else
0113 fillable = 0;
0114
0115 if (fillable >= 4 && !(off % 4)) {
0116 u32 val;
0117
0118 if (iswrite) {
0119 if (copy_from_user(&val, buf, 4))
0120 return -EFAULT;
0121
0122 ret = vfio_pci_iowrite32(vdev, test_mem,
0123 val, io + off);
0124 if (ret)
0125 return ret;
0126 } else {
0127 ret = vfio_pci_ioread32(vdev, test_mem,
0128 &val, io + off);
0129 if (ret)
0130 return ret;
0131
0132 if (copy_to_user(buf, &val, 4))
0133 return -EFAULT;
0134 }
0135
0136 filled = 4;
0137 } else if (fillable >= 2 && !(off % 2)) {
0138 u16 val;
0139
0140 if (iswrite) {
0141 if (copy_from_user(&val, buf, 2))
0142 return -EFAULT;
0143
0144 ret = vfio_pci_iowrite16(vdev, test_mem,
0145 val, io + off);
0146 if (ret)
0147 return ret;
0148 } else {
0149 ret = vfio_pci_ioread16(vdev, test_mem,
0150 &val, io + off);
0151 if (ret)
0152 return ret;
0153
0154 if (copy_to_user(buf, &val, 2))
0155 return -EFAULT;
0156 }
0157
0158 filled = 2;
0159 } else if (fillable) {
0160 u8 val;
0161
0162 if (iswrite) {
0163 if (copy_from_user(&val, buf, 1))
0164 return -EFAULT;
0165
0166 ret = vfio_pci_iowrite8(vdev, test_mem,
0167 val, io + off);
0168 if (ret)
0169 return ret;
0170 } else {
0171 ret = vfio_pci_ioread8(vdev, test_mem,
0172 &val, io + off);
0173 if (ret)
0174 return ret;
0175
0176 if (copy_to_user(buf, &val, 1))
0177 return -EFAULT;
0178 }
0179
0180 filled = 1;
0181 } else {
0182
0183 filled = min(count, (size_t)(x_end - off));
0184 if (!iswrite) {
0185 u8 val = 0xFF;
0186 size_t i;
0187
0188 for (i = 0; i < filled; i++)
0189 if (copy_to_user(buf + i, &val, 1))
0190 return -EFAULT;
0191 }
0192 }
0193
0194 count -= filled;
0195 done += filled;
0196 off += filled;
0197 buf += filled;
0198 }
0199
0200 return done;
0201 }
0202
0203 static int vfio_pci_setup_barmap(struct vfio_pci_core_device *vdev, int bar)
0204 {
0205 struct pci_dev *pdev = vdev->pdev;
0206 int ret;
0207 void __iomem *io;
0208
0209 if (vdev->barmap[bar])
0210 return 0;
0211
0212 ret = pci_request_selected_regions(pdev, 1 << bar, "vfio");
0213 if (ret)
0214 return ret;
0215
0216 io = pci_iomap(pdev, bar, 0);
0217 if (!io) {
0218 pci_release_selected_regions(pdev, 1 << bar);
0219 return -ENOMEM;
0220 }
0221
0222 vdev->barmap[bar] = io;
0223
0224 return 0;
0225 }
0226
0227 ssize_t vfio_pci_bar_rw(struct vfio_pci_core_device *vdev, char __user *buf,
0228 size_t count, loff_t *ppos, bool iswrite)
0229 {
0230 struct pci_dev *pdev = vdev->pdev;
0231 loff_t pos = *ppos & VFIO_PCI_OFFSET_MASK;
0232 int bar = VFIO_PCI_OFFSET_TO_INDEX(*ppos);
0233 size_t x_start = 0, x_end = 0;
0234 resource_size_t end;
0235 void __iomem *io;
0236 struct resource *res = &vdev->pdev->resource[bar];
0237 ssize_t done;
0238
0239 if (pci_resource_start(pdev, bar))
0240 end = pci_resource_len(pdev, bar);
0241 else if (bar == PCI_ROM_RESOURCE &&
0242 pdev->resource[bar].flags & IORESOURCE_ROM_SHADOW)
0243 end = 0x20000;
0244 else
0245 return -EINVAL;
0246
0247 if (pos >= end)
0248 return -EINVAL;
0249
0250 count = min(count, (size_t)(end - pos));
0251
0252 if (bar == PCI_ROM_RESOURCE) {
0253
0254
0255
0256
0257
0258 io = pci_map_rom(pdev, &x_start);
0259 if (!io) {
0260 done = -ENOMEM;
0261 goto out;
0262 }
0263 x_end = end;
0264 } else {
0265 int ret = vfio_pci_setup_barmap(vdev, bar);
0266 if (ret) {
0267 done = ret;
0268 goto out;
0269 }
0270
0271 io = vdev->barmap[bar];
0272 }
0273
0274 if (bar == vdev->msix_bar) {
0275 x_start = vdev->msix_offset;
0276 x_end = vdev->msix_offset + vdev->msix_size;
0277 }
0278
0279 done = do_io_rw(vdev, res->flags & IORESOURCE_MEM, io, buf, pos,
0280 count, x_start, x_end, iswrite);
0281
0282 if (done >= 0)
0283 *ppos += done;
0284
0285 if (bar == PCI_ROM_RESOURCE)
0286 pci_unmap_rom(pdev, io);
0287 out:
0288 return done;
0289 }
0290
0291 #ifdef CONFIG_VFIO_PCI_VGA
0292 ssize_t vfio_pci_vga_rw(struct vfio_pci_core_device *vdev, char __user *buf,
0293 size_t count, loff_t *ppos, bool iswrite)
0294 {
0295 int ret;
0296 loff_t off, pos = *ppos & VFIO_PCI_OFFSET_MASK;
0297 void __iomem *iomem = NULL;
0298 unsigned int rsrc;
0299 bool is_ioport;
0300 ssize_t done;
0301
0302 if (!vdev->has_vga)
0303 return -EINVAL;
0304
0305 if (pos > 0xbfffful)
0306 return -EINVAL;
0307
0308 switch ((u32)pos) {
0309 case 0xa0000 ... 0xbffff:
0310 count = min(count, (size_t)(0xc0000 - pos));
0311 iomem = ioremap(0xa0000, 0xbffff - 0xa0000 + 1);
0312 off = pos - 0xa0000;
0313 rsrc = VGA_RSRC_LEGACY_MEM;
0314 is_ioport = false;
0315 break;
0316 case 0x3b0 ... 0x3bb:
0317 count = min(count, (size_t)(0x3bc - pos));
0318 iomem = ioport_map(0x3b0, 0x3bb - 0x3b0 + 1);
0319 off = pos - 0x3b0;
0320 rsrc = VGA_RSRC_LEGACY_IO;
0321 is_ioport = true;
0322 break;
0323 case 0x3c0 ... 0x3df:
0324 count = min(count, (size_t)(0x3e0 - pos));
0325 iomem = ioport_map(0x3c0, 0x3df - 0x3c0 + 1);
0326 off = pos - 0x3c0;
0327 rsrc = VGA_RSRC_LEGACY_IO;
0328 is_ioport = true;
0329 break;
0330 default:
0331 return -EINVAL;
0332 }
0333
0334 if (!iomem)
0335 return -ENOMEM;
0336
0337 ret = vga_get_interruptible(vdev->pdev, rsrc);
0338 if (ret) {
0339 is_ioport ? ioport_unmap(iomem) : iounmap(iomem);
0340 return ret;
0341 }
0342
0343
0344
0345
0346
0347
0348 done = do_io_rw(vdev, false, iomem, buf, off, count, 0, 0, iswrite);
0349
0350 vga_put(vdev->pdev, rsrc);
0351
0352 is_ioport ? ioport_unmap(iomem) : iounmap(iomem);
0353
0354 if (done >= 0)
0355 *ppos += done;
0356
0357 return done;
0358 }
0359 #endif
0360
0361 static void vfio_pci_ioeventfd_do_write(struct vfio_pci_ioeventfd *ioeventfd,
0362 bool test_mem)
0363 {
0364 switch (ioeventfd->count) {
0365 case 1:
0366 vfio_pci_iowrite8(ioeventfd->vdev, test_mem,
0367 ioeventfd->data, ioeventfd->addr);
0368 break;
0369 case 2:
0370 vfio_pci_iowrite16(ioeventfd->vdev, test_mem,
0371 ioeventfd->data, ioeventfd->addr);
0372 break;
0373 case 4:
0374 vfio_pci_iowrite32(ioeventfd->vdev, test_mem,
0375 ioeventfd->data, ioeventfd->addr);
0376 break;
0377 #ifdef iowrite64
0378 case 8:
0379 vfio_pci_iowrite64(ioeventfd->vdev, test_mem,
0380 ioeventfd->data, ioeventfd->addr);
0381 break;
0382 #endif
0383 }
0384 }
0385
0386 static int vfio_pci_ioeventfd_handler(void *opaque, void *unused)
0387 {
0388 struct vfio_pci_ioeventfd *ioeventfd = opaque;
0389 struct vfio_pci_core_device *vdev = ioeventfd->vdev;
0390
0391 if (ioeventfd->test_mem) {
0392 if (!down_read_trylock(&vdev->memory_lock))
0393 return 1;
0394 if (!__vfio_pci_memory_enabled(vdev)) {
0395 up_read(&vdev->memory_lock);
0396 return 0;
0397 }
0398 }
0399
0400 vfio_pci_ioeventfd_do_write(ioeventfd, false);
0401
0402 if (ioeventfd->test_mem)
0403 up_read(&vdev->memory_lock);
0404
0405 return 0;
0406 }
0407
0408 static void vfio_pci_ioeventfd_thread(void *opaque, void *unused)
0409 {
0410 struct vfio_pci_ioeventfd *ioeventfd = opaque;
0411
0412 vfio_pci_ioeventfd_do_write(ioeventfd, ioeventfd->test_mem);
0413 }
0414
0415 long vfio_pci_ioeventfd(struct vfio_pci_core_device *vdev, loff_t offset,
0416 uint64_t data, int count, int fd)
0417 {
0418 struct pci_dev *pdev = vdev->pdev;
0419 loff_t pos = offset & VFIO_PCI_OFFSET_MASK;
0420 int ret, bar = VFIO_PCI_OFFSET_TO_INDEX(offset);
0421 struct vfio_pci_ioeventfd *ioeventfd;
0422
0423
0424 if (bar > VFIO_PCI_BAR5_REGION_INDEX)
0425 return -EINVAL;
0426
0427 if (pos + count > pci_resource_len(pdev, bar))
0428 return -EINVAL;
0429
0430
0431 if (bar == vdev->msix_bar &&
0432 !(pos + count <= vdev->msix_offset ||
0433 pos >= vdev->msix_offset + vdev->msix_size))
0434 return -EINVAL;
0435
0436 #ifndef iowrite64
0437 if (count == 8)
0438 return -EINVAL;
0439 #endif
0440
0441 ret = vfio_pci_setup_barmap(vdev, bar);
0442 if (ret)
0443 return ret;
0444
0445 mutex_lock(&vdev->ioeventfds_lock);
0446
0447 list_for_each_entry(ioeventfd, &vdev->ioeventfds_list, next) {
0448 if (ioeventfd->pos == pos && ioeventfd->bar == bar &&
0449 ioeventfd->data == data && ioeventfd->count == count) {
0450 if (fd == -1) {
0451 vfio_virqfd_disable(&ioeventfd->virqfd);
0452 list_del(&ioeventfd->next);
0453 vdev->ioeventfds_nr--;
0454 kfree(ioeventfd);
0455 ret = 0;
0456 } else
0457 ret = -EEXIST;
0458
0459 goto out_unlock;
0460 }
0461 }
0462
0463 if (fd < 0) {
0464 ret = -ENODEV;
0465 goto out_unlock;
0466 }
0467
0468 if (vdev->ioeventfds_nr >= VFIO_PCI_IOEVENTFD_MAX) {
0469 ret = -ENOSPC;
0470 goto out_unlock;
0471 }
0472
0473 ioeventfd = kzalloc(sizeof(*ioeventfd), GFP_KERNEL);
0474 if (!ioeventfd) {
0475 ret = -ENOMEM;
0476 goto out_unlock;
0477 }
0478
0479 ioeventfd->vdev = vdev;
0480 ioeventfd->addr = vdev->barmap[bar] + pos;
0481 ioeventfd->data = data;
0482 ioeventfd->pos = pos;
0483 ioeventfd->bar = bar;
0484 ioeventfd->count = count;
0485 ioeventfd->test_mem = vdev->pdev->resource[bar].flags & IORESOURCE_MEM;
0486
0487 ret = vfio_virqfd_enable(ioeventfd, vfio_pci_ioeventfd_handler,
0488 vfio_pci_ioeventfd_thread, NULL,
0489 &ioeventfd->virqfd, fd);
0490 if (ret) {
0491 kfree(ioeventfd);
0492 goto out_unlock;
0493 }
0494
0495 list_add(&ioeventfd->next, &vdev->ioeventfds_list);
0496 vdev->ioeventfds_nr++;
0497
0498 out_unlock:
0499 mutex_unlock(&vdev->ioeventfds_lock);
0500
0501 return ret;
0502 }