0001
0002
0003
0004
0005
0006
0007
0008
0009
0010
0011
0012
0013
0014
0015
0016
0017
0018
0019
0020
0021
0022 #include "vmm.h"
0023
0024 #include <subdev/fb.h>
0025 #include <subdev/ltc.h>
0026 #include <subdev/timer.h>
0027
0028 #include <nvif/if900d.h>
0029 #include <nvif/unpack.h>
0030
0031 static inline void
0032 gf100_vmm_pgt_pte(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
0033 u32 ptei, u32 ptes, struct nvkm_vmm_map *map, u64 addr)
0034 {
0035 u64 base = (addr >> 8) | map->type;
0036 u64 data = base;
0037
0038 if (map->ctag && !(map->next & (1ULL << 44))) {
0039 while (ptes--) {
0040 data = base | ((map->ctag >> 1) << 44);
0041 if (!(map->ctag++ & 1))
0042 data |= BIT_ULL(60);
0043
0044 VMM_WO064(pt, vmm, ptei++ * 8, data);
0045 base += map->next;
0046 }
0047 } else {
0048 map->type += ptes * map->ctag;
0049
0050 while (ptes--) {
0051 VMM_WO064(pt, vmm, ptei++ * 8, data);
0052 data += map->next;
0053 }
0054 }
0055 }
0056
0057 void
0058 gf100_vmm_pgt_sgl(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
0059 u32 ptei, u32 ptes, struct nvkm_vmm_map *map)
0060 {
0061 VMM_MAP_ITER_SGL(vmm, pt, ptei, ptes, map, gf100_vmm_pgt_pte);
0062 }
0063
0064 void
0065 gf100_vmm_pgt_dma(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
0066 u32 ptei, u32 ptes, struct nvkm_vmm_map *map)
0067 {
0068 if (map->page->shift == PAGE_SHIFT) {
0069 VMM_SPAM(vmm, "DMAA %08x %08x PTE(s)", ptei, ptes);
0070 nvkm_kmap(pt->memory);
0071 while (ptes--) {
0072 const u64 data = (*map->dma++ >> 8) | map->type;
0073 VMM_WO064(pt, vmm, ptei++ * 8, data);
0074 map->type += map->ctag;
0075 }
0076 nvkm_done(pt->memory);
0077 return;
0078 }
0079
0080 VMM_MAP_ITER_DMA(vmm, pt, ptei, ptes, map, gf100_vmm_pgt_pte);
0081 }
0082
0083 void
0084 gf100_vmm_pgt_mem(struct nvkm_vmm *vmm, struct nvkm_mmu_pt *pt,
0085 u32 ptei, u32 ptes, struct nvkm_vmm_map *map)
0086 {
0087 VMM_MAP_ITER_MEM(vmm, pt, ptei, ptes, map, gf100_vmm_pgt_pte);
0088 }
0089
0090 void
0091 gf100_vmm_pgt_unmap(struct nvkm_vmm *vmm,
0092 struct nvkm_mmu_pt *pt, u32 ptei, u32 ptes)
0093 {
0094 VMM_FO064(pt, vmm, ptei * 8, 0ULL, ptes);
0095 }
0096
0097 const struct nvkm_vmm_desc_func
0098 gf100_vmm_pgt = {
0099 .unmap = gf100_vmm_pgt_unmap,
0100 .mem = gf100_vmm_pgt_mem,
0101 .dma = gf100_vmm_pgt_dma,
0102 .sgl = gf100_vmm_pgt_sgl,
0103 };
0104
0105 void
0106 gf100_vmm_pgd_pde(struct nvkm_vmm *vmm, struct nvkm_vmm_pt *pgd, u32 pdei)
0107 {
0108 struct nvkm_vmm_pt *pgt = pgd->pde[pdei];
0109 struct nvkm_mmu_pt *pd = pgd->pt[0];
0110 struct nvkm_mmu_pt *pt;
0111 u64 data = 0;
0112
0113 if ((pt = pgt->pt[0])) {
0114 switch (nvkm_memory_target(pt->memory)) {
0115 case NVKM_MEM_TARGET_VRAM: data |= 1ULL << 0; break;
0116 case NVKM_MEM_TARGET_HOST: data |= 2ULL << 0;
0117 data |= BIT_ULL(35);
0118 break;
0119 case NVKM_MEM_TARGET_NCOH: data |= 3ULL << 0; break;
0120 default:
0121 WARN_ON(1);
0122 return;
0123 }
0124 data |= pt->addr >> 8;
0125 }
0126
0127 if ((pt = pgt->pt[1])) {
0128 switch (nvkm_memory_target(pt->memory)) {
0129 case NVKM_MEM_TARGET_VRAM: data |= 1ULL << 32; break;
0130 case NVKM_MEM_TARGET_HOST: data |= 2ULL << 32;
0131 data |= BIT_ULL(34);
0132 break;
0133 case NVKM_MEM_TARGET_NCOH: data |= 3ULL << 32; break;
0134 default:
0135 WARN_ON(1);
0136 return;
0137 }
0138 data |= pt->addr << 24;
0139 }
0140
0141 nvkm_kmap(pd->memory);
0142 VMM_WO064(pd, vmm, pdei * 8, data);
0143 nvkm_done(pd->memory);
0144 }
0145
0146 const struct nvkm_vmm_desc_func
0147 gf100_vmm_pgd = {
0148 .unmap = gf100_vmm_pgt_unmap,
0149 .pde = gf100_vmm_pgd_pde,
0150 };
0151
0152 static const struct nvkm_vmm_desc
0153 gf100_vmm_desc_17_12[] = {
0154 { SPT, 15, 8, 0x1000, &gf100_vmm_pgt },
0155 { PGD, 13, 8, 0x1000, &gf100_vmm_pgd },
0156 {}
0157 };
0158
0159 static const struct nvkm_vmm_desc
0160 gf100_vmm_desc_17_17[] = {
0161 { LPT, 10, 8, 0x1000, &gf100_vmm_pgt },
0162 { PGD, 13, 8, 0x1000, &gf100_vmm_pgd },
0163 {}
0164 };
0165
0166 static const struct nvkm_vmm_desc
0167 gf100_vmm_desc_16_12[] = {
0168 { SPT, 14, 8, 0x1000, &gf100_vmm_pgt },
0169 { PGD, 14, 8, 0x1000, &gf100_vmm_pgd },
0170 {}
0171 };
0172
0173 static const struct nvkm_vmm_desc
0174 gf100_vmm_desc_16_16[] = {
0175 { LPT, 10, 8, 0x1000, &gf100_vmm_pgt },
0176 { PGD, 14, 8, 0x1000, &gf100_vmm_pgd },
0177 {}
0178 };
0179
0180 void
0181 gf100_vmm_invalidate_pdb(struct nvkm_vmm *vmm, u64 addr)
0182 {
0183 struct nvkm_device *device = vmm->mmu->subdev.device;
0184 nvkm_wr32(device, 0x100cb8, addr);
0185 }
0186
0187 void
0188 gf100_vmm_invalidate(struct nvkm_vmm *vmm, u32 type)
0189 {
0190 struct nvkm_device *device = vmm->mmu->subdev.device;
0191 struct nvkm_mmu_pt *pd = vmm->pd->pt[0];
0192 u64 addr = 0;
0193
0194 mutex_lock(&vmm->mmu->mutex);
0195
0196
0197
0198 nvkm_msec(device, 2000,
0199 if (nvkm_rd32(device, 0x100c80) & 0x00ff0000)
0200 break;
0201 );
0202
0203 if (!(type & 0x00000002) ) {
0204 switch (nvkm_memory_target(pd->memory)) {
0205 case NVKM_MEM_TARGET_VRAM: addr |= 0x00000000; break;
0206 case NVKM_MEM_TARGET_HOST: addr |= 0x00000002; break;
0207 case NVKM_MEM_TARGET_NCOH: addr |= 0x00000003; break;
0208 default:
0209 WARN_ON(1);
0210 break;
0211 }
0212 addr |= (vmm->pd->pt[0]->addr >> 12) << 4;
0213
0214 vmm->func->invalidate_pdb(vmm, addr);
0215 }
0216
0217 nvkm_wr32(device, 0x100cbc, 0x80000000 | type);
0218
0219
0220 nvkm_msec(device, 2000,
0221 if (nvkm_rd32(device, 0x100c80) & 0x00008000)
0222 break;
0223 );
0224 mutex_unlock(&vmm->mmu->mutex);
0225 }
0226
0227 void
0228 gf100_vmm_flush(struct nvkm_vmm *vmm, int depth)
0229 {
0230 u32 type = 0x00000001;
0231 if (atomic_read(&vmm->engref[NVKM_SUBDEV_BAR]))
0232 type |= 0x00000004;
0233 gf100_vmm_invalidate(vmm, type);
0234 }
0235
0236 int
0237 gf100_vmm_valid(struct nvkm_vmm *vmm, void *argv, u32 argc,
0238 struct nvkm_vmm_map *map)
0239 {
0240 const enum nvkm_memory_target target = nvkm_memory_target(map->memory);
0241 const struct nvkm_vmm_page *page = map->page;
0242 const bool gm20x = page->desc->func->sparse != NULL;
0243 union {
0244 struct gf100_vmm_map_vn vn;
0245 struct gf100_vmm_map_v0 v0;
0246 } *args = argv;
0247 struct nvkm_device *device = vmm->mmu->subdev.device;
0248 struct nvkm_memory *memory = map->memory;
0249 u8 kind, kind_inv, priv, ro, vol;
0250 int kindn, aper, ret = -ENOSYS;
0251 const u8 *kindm;
0252
0253 map->next = (1 << page->shift) >> 8;
0254 map->type = map->ctag = 0;
0255
0256 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) {
0257 vol = !!args->v0.vol;
0258 ro = !!args->v0.ro;
0259 priv = !!args->v0.priv;
0260 kind = args->v0.kind;
0261 } else
0262 if (!(ret = nvif_unvers(ret, &argv, &argc, args->vn))) {
0263 vol = target == NVKM_MEM_TARGET_HOST;
0264 ro = 0;
0265 priv = 0;
0266 kind = 0x00;
0267 } else {
0268 VMM_DEBUG(vmm, "args");
0269 return ret;
0270 }
0271
0272 aper = vmm->func->aper(target);
0273 if (WARN_ON(aper < 0))
0274 return aper;
0275
0276 kindm = vmm->mmu->func->kind(vmm->mmu, &kindn, &kind_inv);
0277 if (kind >= kindn || kindm[kind] == kind_inv) {
0278 VMM_DEBUG(vmm, "kind %02x", kind);
0279 return -EINVAL;
0280 }
0281
0282 if (kindm[kind] != kind) {
0283 u32 comp = (page->shift == 16 && !gm20x) ? 16 : 17;
0284 u32 tags = ALIGN(nvkm_memory_size(memory), 1 << 17) >> comp;
0285 if (aper != 0 || !(page->type & NVKM_VMM_PAGE_COMP)) {
0286 VMM_DEBUG(vmm, "comp %d %02x", aper, page->type);
0287 return -EINVAL;
0288 }
0289
0290 ret = nvkm_memory_tags_get(memory, device, tags,
0291 nvkm_ltc_tags_clear,
0292 &map->tags);
0293 if (ret) {
0294 VMM_DEBUG(vmm, "comp %d", ret);
0295 return ret;
0296 }
0297
0298 if (map->tags->mn) {
0299 u64 tags = map->tags->mn->offset + (map->offset >> 17);
0300 if (page->shift == 17 || !gm20x) {
0301 map->type |= tags << 44;
0302 map->ctag |= 1ULL << 44;
0303 map->next |= 1ULL << 44;
0304 } else {
0305 map->ctag |= tags << 1 | 1;
0306 }
0307 } else {
0308 kind = kindm[kind];
0309 }
0310 }
0311
0312 map->type |= BIT(0);
0313 map->type |= (u64)priv << 1;
0314 map->type |= (u64) ro << 2;
0315 map->type |= (u64) vol << 32;
0316 map->type |= (u64)aper << 33;
0317 map->type |= (u64)kind << 36;
0318 return 0;
0319 }
0320
0321 int
0322 gf100_vmm_aper(enum nvkm_memory_target target)
0323 {
0324 switch (target) {
0325 case NVKM_MEM_TARGET_VRAM: return 0;
0326 case NVKM_MEM_TARGET_HOST: return 2;
0327 case NVKM_MEM_TARGET_NCOH: return 3;
0328 default:
0329 return -EINVAL;
0330 }
0331 }
0332
0333 void
0334 gf100_vmm_part(struct nvkm_vmm *vmm, struct nvkm_memory *inst)
0335 {
0336 nvkm_fo64(inst, 0x0200, 0x00000000, 2);
0337 }
0338
0339 int
0340 gf100_vmm_join_(struct nvkm_vmm *vmm, struct nvkm_memory *inst, u64 base)
0341 {
0342 struct nvkm_mmu_pt *pd = vmm->pd->pt[0];
0343
0344 switch (nvkm_memory_target(pd->memory)) {
0345 case NVKM_MEM_TARGET_VRAM: base |= 0ULL << 0; break;
0346 case NVKM_MEM_TARGET_HOST: base |= 2ULL << 0;
0347 base |= BIT_ULL(2) ;
0348 break;
0349 case NVKM_MEM_TARGET_NCOH: base |= 3ULL << 0; break;
0350 default:
0351 WARN_ON(1);
0352 return -EINVAL;
0353 }
0354 base |= pd->addr;
0355
0356 nvkm_kmap(inst);
0357 nvkm_wo64(inst, 0x0200, base);
0358 nvkm_wo64(inst, 0x0208, vmm->limit - 1);
0359 nvkm_done(inst);
0360 return 0;
0361 }
0362
0363 int
0364 gf100_vmm_join(struct nvkm_vmm *vmm, struct nvkm_memory *inst)
0365 {
0366 return gf100_vmm_join_(vmm, inst, 0);
0367 }
0368
0369 static const struct nvkm_vmm_func
0370 gf100_vmm_17 = {
0371 .join = gf100_vmm_join,
0372 .part = gf100_vmm_part,
0373 .aper = gf100_vmm_aper,
0374 .valid = gf100_vmm_valid,
0375 .flush = gf100_vmm_flush,
0376 .invalidate_pdb = gf100_vmm_invalidate_pdb,
0377 .page = {
0378 { 17, &gf100_vmm_desc_17_17[0], NVKM_VMM_PAGE_xVxC },
0379 { 12, &gf100_vmm_desc_17_12[0], NVKM_VMM_PAGE_xVHx },
0380 {}
0381 }
0382 };
0383
0384 static const struct nvkm_vmm_func
0385 gf100_vmm_16 = {
0386 .join = gf100_vmm_join,
0387 .part = gf100_vmm_part,
0388 .aper = gf100_vmm_aper,
0389 .valid = gf100_vmm_valid,
0390 .flush = gf100_vmm_flush,
0391 .invalidate_pdb = gf100_vmm_invalidate_pdb,
0392 .page = {
0393 { 16, &gf100_vmm_desc_16_16[0], NVKM_VMM_PAGE_xVxC },
0394 { 12, &gf100_vmm_desc_16_12[0], NVKM_VMM_PAGE_xVHx },
0395 {}
0396 }
0397 };
0398
0399 int
0400 gf100_vmm_new_(const struct nvkm_vmm_func *func_16,
0401 const struct nvkm_vmm_func *func_17,
0402 struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size,
0403 void *argv, u32 argc, struct lock_class_key *key,
0404 const char *name, struct nvkm_vmm **pvmm)
0405 {
0406 switch (mmu->subdev.device->fb->page) {
0407 case 16: return nv04_vmm_new_(func_16, mmu, 0, managed, addr, size,
0408 argv, argc, key, name, pvmm);
0409 case 17: return nv04_vmm_new_(func_17, mmu, 0, managed, addr, size,
0410 argv, argc, key, name, pvmm);
0411 default:
0412 WARN_ON(1);
0413 return -EINVAL;
0414 }
0415 }
0416
0417 int
0418 gf100_vmm_new(struct nvkm_mmu *mmu, bool managed, u64 addr, u64 size,
0419 void *argv, u32 argc, struct lock_class_key *key,
0420 const char *name, struct nvkm_vmm **pvmm)
0421 {
0422 return gf100_vmm_new_(&gf100_vmm_16, &gf100_vmm_17, mmu, managed, addr,
0423 size, argv, argc, key, name, pvmm);
0424 }