/dports/multimedia/v4l-utils/linux-5.13-rc2/arch/arm64/kvm/hyp/ |
H A D | pgtable.c | 296 struct kvm_pgtable *pgt = data->pgt; in _kvm_pgtable_walk() local 302 if (!pgt->pgd) in _kvm_pgtable_walk() 320 .pgt = pgt, in kvm_pgtable_walk() 434 if (!pgt->pgd) in kvm_pgtable_hyp_init() 461 WARN_ON(kvm_pgtable_walk(pgt, 0, BIT(pgt->ia_bits), &walker)); in kvm_pgtable_hyp_destroy() 462 pgt->mm_ops->put_page(pgt->pgd); in kvm_pgtable_hyp_destroy() 463 pgt->pgd = NULL; in kvm_pgtable_hyp_destroy() 856 .arg = pgt, in kvm_pgtable_stage2_unmap() 1004 .arg = pgt, in kvm_pgtable_stage2_flush() 1025 if (!pgt->pgd) in kvm_pgtable_stage2_init_flags() [all …]
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/arch/arm64/kvm/hyp/ |
H A D | pgtable.c | 296 struct kvm_pgtable *pgt = data->pgt; in _kvm_pgtable_walk() local 302 if (!pgt->pgd) in _kvm_pgtable_walk() 320 .pgt = pgt, in kvm_pgtable_walk() 434 if (!pgt->pgd) in kvm_pgtable_hyp_init() 461 WARN_ON(kvm_pgtable_walk(pgt, 0, BIT(pgt->ia_bits), &walker)); in kvm_pgtable_hyp_destroy() 462 pgt->mm_ops->put_page(pgt->pgd); in kvm_pgtable_hyp_destroy() 463 pgt->pgd = NULL; in kvm_pgtable_hyp_destroy() 856 .arg = pgt, in kvm_pgtable_stage2_unmap() 1004 .arg = pgt, in kvm_pgtable_stage2_flush() 1025 if (!pgt->pgd) in kvm_pgtable_stage2_init_flags() [all …]
|
/dports/multimedia/libv4l/linux-5.13-rc2/arch/arm64/kvm/hyp/ |
H A D | pgtable.c | 296 struct kvm_pgtable *pgt = data->pgt; in _kvm_pgtable_walk() local 302 if (!pgt->pgd) in _kvm_pgtable_walk() 320 .pgt = pgt, in kvm_pgtable_walk() 434 if (!pgt->pgd) in kvm_pgtable_hyp_init() 461 WARN_ON(kvm_pgtable_walk(pgt, 0, BIT(pgt->ia_bits), &walker)); in kvm_pgtable_hyp_destroy() 462 pgt->mm_ops->put_page(pgt->pgd); in kvm_pgtable_hyp_destroy() 463 pgt->pgd = NULL; in kvm_pgtable_hyp_destroy() 856 .arg = pgt, in kvm_pgtable_stage2_unmap() 1004 .arg = pgt, in kvm_pgtable_stage2_flush() 1025 if (!pgt->pgd) in kvm_pgtable_stage2_init_flags() [all …]
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/arch/arm64/include/asm/ |
H A D | kvm_pgtable.h | 160 int kvm_pgtable_hyp_init(struct kvm_pgtable *pgt, u32 va_bits, 170 void kvm_pgtable_hyp_destroy(struct kvm_pgtable *pgt); 224 #define kvm_pgtable_stage2_init(pgt, arch, mm_ops) \ argument 225 kvm_pgtable_stage2_init_flags(pgt, arch, mm_ops, 0) 234 void kvm_pgtable_stage2_destroy(struct kvm_pgtable *pgt); 264 int kvm_pgtable_stage2_map(struct kvm_pgtable *pgt, u64 addr, u64 size, 355 kvm_pte_t kvm_pgtable_stage2_mkold(struct kvm_pgtable *pgt, u64 addr); 373 int kvm_pgtable_stage2_relax_perms(struct kvm_pgtable *pgt, u64 addr, 386 bool kvm_pgtable_stage2_is_young(struct kvm_pgtable *pgt, u64 addr); 424 int kvm_pgtable_walk(struct kvm_pgtable *pgt, u64 addr, u64 size, [all …]
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/arch/arm64/include/asm/ |
H A D | kvm_pgtable.h | 160 int kvm_pgtable_hyp_init(struct kvm_pgtable *pgt, u32 va_bits, 170 void kvm_pgtable_hyp_destroy(struct kvm_pgtable *pgt); 224 #define kvm_pgtable_stage2_init(pgt, arch, mm_ops) \ argument 225 kvm_pgtable_stage2_init_flags(pgt, arch, mm_ops, 0) 234 void kvm_pgtable_stage2_destroy(struct kvm_pgtable *pgt); 264 int kvm_pgtable_stage2_map(struct kvm_pgtable *pgt, u64 addr, u64 size, 355 kvm_pte_t kvm_pgtable_stage2_mkold(struct kvm_pgtable *pgt, u64 addr); 373 int kvm_pgtable_stage2_relax_perms(struct kvm_pgtable *pgt, u64 addr, 386 bool kvm_pgtable_stage2_is_young(struct kvm_pgtable *pgt, u64 addr); 424 int kvm_pgtable_walk(struct kvm_pgtable *pgt, u64 addr, u64 size, [all …]
|
/dports/multimedia/libv4l/linux-5.13-rc2/arch/arm64/include/asm/ |
H A D | kvm_pgtable.h | 160 int kvm_pgtable_hyp_init(struct kvm_pgtable *pgt, u32 va_bits, 170 void kvm_pgtable_hyp_destroy(struct kvm_pgtable *pgt); 224 #define kvm_pgtable_stage2_init(pgt, arch, mm_ops) \ argument 225 kvm_pgtable_stage2_init_flags(pgt, arch, mm_ops, 0) 234 void kvm_pgtable_stage2_destroy(struct kvm_pgtable *pgt); 264 int kvm_pgtable_stage2_map(struct kvm_pgtable *pgt, u64 addr, u64 size, 355 kvm_pte_t kvm_pgtable_stage2_mkold(struct kvm_pgtable *pgt, u64 addr); 373 int kvm_pgtable_stage2_relax_perms(struct kvm_pgtable *pgt, u64 addr, 386 bool kvm_pgtable_stage2_is_young(struct kvm_pgtable *pgt, u64 addr); 424 int kvm_pgtable_walk(struct kvm_pgtable *pgt, u64 addr, u64 size, [all …]
|
/dports/misc/rump/buildrump.sh-b914579/src/sys/external/bsd/drm2/dist/drm/nouveau/core/subdev/vm/ |
H A D | nouveau_subdev_vm_nv44.c | 52 tmp[0] = nv_ro32(pgt, base + 0x0); in nv44_vm_fill() 53 tmp[1] = nv_ro32(pgt, base + 0x4); in nv44_vm_fill() 54 tmp[2] = nv_ro32(pgt, base + 0x8); in nv44_vm_fill() 55 tmp[3] = nv_ro32(pgt, base + 0xc); in nv44_vm_fill() 85 nv_wo32(pgt, base + 0x0, tmp[0]); in nv44_vm_fill() 86 nv_wo32(pgt, base + 0x4, tmp[1]); in nv44_vm_fill() 87 nv_wo32(pgt, base + 0x8, tmp[2]); in nv44_vm_fill() 136 nv_wo32(pgt, pte++ * 4, 0x00000000); in nv44_vm_unmap() 137 nv_wo32(pgt, pte++ * 4, 0x00000000); in nv44_vm_unmap() 250 &priv->vm->pgt[0].obj[0]); in nv44_vmmgr_ctor() [all …]
|
H A D | nouveau_subdev_vm_nv50.c | 44 struct nouveau_gpuobj *pgt[2]) in nv50_vm_map_pgt() 49 if (pgt[0]) { in nv50_vm_map_pgt() 51 coverage = (pgt[0]->size >> 3) << 12; in nv50_vm_map_pgt() 53 if (pgt[1]) { in nv50_vm_map_pgt() 55 coverage = (pgt[1]->size >> 3) << 16; in nv50_vm_map_pgt() 123 nv_wo32(pgt, pte + 0, offset_l); in nv50_vm_map() 124 nv_wo32(pgt, pte + 4, offset_h); in nv50_vm_map() 139 nv_wo32(pgt, pte + 0, lower_32_bits(phys)); in nv50_vm_map_sg() 140 nv_wo32(pgt, pte + 4, upper_32_bits(phys)); in nv50_vm_map_sg() 150 nv_wo32(pgt, pte + 0, 0x00000000); in nv50_vm_unmap() [all …]
|
H A D | nouveau_subdev_vm_nvc0.c | 86 struct nouveau_gpuobj *pgt[2]) in nvc0_vm_map_pgt() 90 if (pgt[0]) in nvc0_vm_map_pgt() 91 pde[1] = 0x00000001 | (pgt[0]->addr >> 8); in nvc0_vm_map_pgt() 92 if (pgt[1]) in nvc0_vm_map_pgt() 93 pde[0] = 0x00000001 | (pgt[1]->addr >> 8); in nvc0_vm_map_pgt() 133 nv_wo32(pgt, pte + 0, lower_32_bits(phys)); in nvc0_vm_map() 134 nv_wo32(pgt, pte + 4, upper_32_bits(phys)); in nvc0_vm_map() 151 nv_wo32(pgt, pte + 0, lower_32_bits(phys)); in nvc0_vm_map_sg() 152 nv_wo32(pgt, pte + 4, upper_32_bits(phys)); in nvc0_vm_map_sg() 162 nv_wo32(pgt, pte + 0, 0x00000000); in nvc0_vm_unmap() [all …]
|
H A D | nouveau_subdev_vm_base.c | 56 struct nouveau_gpuobj *pgt = vm->pgt[pde].obj[big]; in nouveau_vm_map_at() local 98 struct nouveau_gpuobj *pgt = vm->pgt[pde].obj[big]; in nouveau_vm_map_dma() local 144 struct nouveau_gpuobj *pgt = vm->pgt[pde].obj[big]; in nouveau_vm_map_sg_table() local 200 struct nouveau_gpuobj *pgt = vm->pgt[pde].obj[big]; in nouveau_vm_map_sg() local 257 struct nouveau_gpuobj *pgt = vm->pgt[pde].obj[big]; in nouveau_vm_unmap_at() local 297 pgt = vpgt->obj[big]; in nouveau_vm_unmap_pgt() 339 vpgt->obj[big] = pgt; in nouveau_vm_map_pgt() 433 vm->pgt = vzalloc((vm->lpde - vm->fpde + 1) * sizeof(*vm->pgt)); in nouveau_vm_create() 434 if (!vm->pgt) { in nouveau_vm_create() 442 vfree(vm->pgt); in nouveau_vm_create() [all …]
|
/dports/print/pdflib-perl/PDFlib-Lite-7.0.5p3/libs/pdcore/ |
H A D | pc_encoding.c | 2256 return pgt; in pdc_new_pglyphtab() 2264 if (pgt) in pdc_delete_pglyphtab() 2293 if (pgt == NULL) in pdc_get_glyphtab() 2296 return pgt; in pdc_get_glyphtab() 2309 if (pgt->glyph_tab_size == pgt->glyph_tab_capacity) in pdc_register_glyphname() 2368 pgt->name2unicode[slot].code = pgt->name2unicode[slot-1].code; in pdc_register_glyphname() 2369 pgt->name2unicode[slot].name = pgt->name2unicode[slot-1].name; in pdc_register_glyphname() 2399 pgt->unicode2name[slotuv].name = pgt->name2unicode[slotname].name; in pdc_register_glyphname() 2411 if (pgt && pgt->glyph_tab_size) in pdc_privglyphname2unicode() 2511 if (pgt && pgt->glyph_tab_size) in pdc_glyphname2privglyphname() [all …]
|
/dports/print/pdflib/PDFlib-Lite-7.0.5p3/libs/pdcore/ |
H A D | pc_encoding.c | 2256 return pgt; in pdc_new_pglyphtab() 2264 if (pgt) in pdc_delete_pglyphtab() 2293 if (pgt == NULL) in pdc_get_glyphtab() 2296 return pgt; in pdc_get_glyphtab() 2309 if (pgt->glyph_tab_size == pgt->glyph_tab_capacity) in pdc_register_glyphname() 2368 pgt->name2unicode[slot].code = pgt->name2unicode[slot-1].code; in pdc_register_glyphname() 2369 pgt->name2unicode[slot].name = pgt->name2unicode[slot-1].name; in pdc_register_glyphname() 2399 pgt->unicode2name[slotuv].name = pgt->name2unicode[slotname].name; in pdc_register_glyphname() 2411 if (pgt && pgt->glyph_tab_size) in pdc_privglyphname2unicode() 2511 if (pgt && pgt->glyph_tab_size) in pdc_glyphname2privglyphname() [all …]
|
/dports/multimedia/libv4l/linux-5.13-rc2/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | vmm.c | 31 if (pgt) { in nvkm_vmm_pt_del() 32 kvfree(pgt->pde); in nvkm_vmm_pt_del() 33 kfree(pgt); in nvkm_vmm_pt_del() 56 if (!(pgt = kzalloc(sizeof(*pgt) + lpte, GFP_KERNEL))) in nvkm_vmm_pt_new() 62 pgt->pde = kvcalloc(pten, sizeof(*pgt->pde), GFP_KERNEL); in nvkm_vmm_pt_new() 63 if (!pgt->pde) { in nvkm_vmm_pt_new() 64 kfree(pgt); in nvkm_vmm_pt_new() 69 return pgt; in nvkm_vmm_pt_new() 216 if (!pgt->refs[0]) in nvkm_vmm_unref_sptes() 280 if (desc->type == SPT && (pgt->refs[0] || pgt->refs[1])) in nvkm_vmm_unref_ptes() [all …]
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | vmm.c | 31 if (pgt) { in nvkm_vmm_pt_del() 32 kvfree(pgt->pde); in nvkm_vmm_pt_del() 33 kfree(pgt); in nvkm_vmm_pt_del() 56 if (!(pgt = kzalloc(sizeof(*pgt) + lpte, GFP_KERNEL))) in nvkm_vmm_pt_new() 62 pgt->pde = kvcalloc(pten, sizeof(*pgt->pde), GFP_KERNEL); in nvkm_vmm_pt_new() 63 if (!pgt->pde) { in nvkm_vmm_pt_new() 64 kfree(pgt); in nvkm_vmm_pt_new() 69 return pgt; in nvkm_vmm_pt_new() 216 if (!pgt->refs[0]) in nvkm_vmm_unref_sptes() 280 if (desc->type == SPT && (pgt->refs[0] || pgt->refs[1])) in nvkm_vmm_unref_ptes() [all …]
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | vmm.c | 31 if (pgt) { in nvkm_vmm_pt_del() 32 kvfree(pgt->pde); in nvkm_vmm_pt_del() 33 kfree(pgt); in nvkm_vmm_pt_del() 56 if (!(pgt = kzalloc(sizeof(*pgt) + lpte, GFP_KERNEL))) in nvkm_vmm_pt_new() 62 pgt->pde = kvcalloc(pten, sizeof(*pgt->pde), GFP_KERNEL); in nvkm_vmm_pt_new() 63 if (!pgt->pde) { in nvkm_vmm_pt_new() 64 kfree(pgt); in nvkm_vmm_pt_new() 69 return pgt; in nvkm_vmm_pt_new() 216 if (!pgt->refs[0]) in nvkm_vmm_unref_sptes() 280 if (desc->type == SPT && (pgt->refs[0] || pgt->refs[1])) in nvkm_vmm_unref_ptes() [all …]
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/arch/arm64/kvm/ |
H A D | mmu.c | 51 struct kvm_pgtable *pgt = kvm->arch.mmu.pgt; in stage2_apply_range() local 52 if (!pgt) in stage2_apply_range() 456 pgt = kzalloc(sizeof(*pgt), GFP_KERNEL); in kvm_init_stage2_mmu() 457 if (!pgt) in kvm_init_stage2_mmu() 474 mmu->pgt = pgt; in kvm_init_stage2_mmu() 482 kfree(pgt); in kvm_init_stage2_mmu() 560 pgt = mmu->pgt; in kvm_free_stage2_pgd() 561 if (pgt) { in kvm_free_stage2_pgd() 568 if (pgt) { in kvm_free_stage2_pgd() 589 struct kvm_pgtable *pgt = kvm->arch.mmu.pgt; in kvm_phys_addr_ioremap() local [all …]
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/arch/arm64/kvm/ |
H A D | mmu.c | 51 struct kvm_pgtable *pgt = kvm->arch.mmu.pgt; in stage2_apply_range() local 52 if (!pgt) in stage2_apply_range() 456 pgt = kzalloc(sizeof(*pgt), GFP_KERNEL); in kvm_init_stage2_mmu() 457 if (!pgt) in kvm_init_stage2_mmu() 474 mmu->pgt = pgt; in kvm_init_stage2_mmu() 482 kfree(pgt); in kvm_init_stage2_mmu() 560 pgt = mmu->pgt; in kvm_free_stage2_pgd() 561 if (pgt) { in kvm_free_stage2_pgd() 568 if (pgt) { in kvm_free_stage2_pgd() 589 struct kvm_pgtable *pgt = kvm->arch.mmu.pgt; in kvm_phys_addr_ioremap() local [all …]
|
/dports/multimedia/libv4l/linux-5.13-rc2/arch/arm64/kvm/ |
H A D | mmu.c | 51 struct kvm_pgtable *pgt = kvm->arch.mmu.pgt; in stage2_apply_range() local 52 if (!pgt) in stage2_apply_range() 456 pgt = kzalloc(sizeof(*pgt), GFP_KERNEL); in kvm_init_stage2_mmu() 457 if (!pgt) in kvm_init_stage2_mmu() 474 mmu->pgt = pgt; in kvm_init_stage2_mmu() 482 kfree(pgt); in kvm_init_stage2_mmu() 560 pgt = mmu->pgt; in kvm_free_stage2_pgd() 561 if (pgt) { in kvm_free_stage2_pgd() 568 if (pgt) { in kvm_free_stage2_pgd() 589 struct kvm_pgtable *pgt = kvm->arch.mmu.pgt; in kvm_phys_addr_ioremap() local [all …]
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/arch/arm64/kvm/hyp/nvhe/ |
H A D | mem_protect.c | 101 ret = kvm_pgtable_stage2_init_flags(&host_kvm.pgt, &host_kvm.arch, in kvm_host_prepare_stage2() 106 mmu->pgd_phys = __hyp_pa(host_kvm.pgt.pgd); in kvm_host_prepare_stage2() 108 mmu->pgt = &host_kvm.pgt; in kvm_host_prepare_stage2() 144 struct kvm_pgtable *pgt = &host_kvm.pgt; in host_stage2_unmap_dev_all() local 152 ret = kvm_pgtable_stage2_unmap(pgt, addr, reg->base - addr); in host_stage2_unmap_dev_all() 156 return kvm_pgtable_stage2_unmap(pgt, addr, BIT(pgt->ia_bits) - addr); in host_stage2_unmap_dev_all() 205 return kvm_pgtable_stage2_map(&host_kvm.pgt, start, end - start, start, in __host_stage2_idmap() 221 ret = kvm_pgtable_stage2_find_range(&host_kvm.pgt, addr, prot, &range); in host_stage2_idmap() 260 ret = kvm_pgtable_stage2_set_owner(&host_kvm.pgt, start, end - start, in __pkvm_mark_hyp()
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/arch/arm64/kvm/hyp/nvhe/ |
H A D | mem_protect.c | 101 ret = kvm_pgtable_stage2_init_flags(&host_kvm.pgt, &host_kvm.arch, in kvm_host_prepare_stage2() 106 mmu->pgd_phys = __hyp_pa(host_kvm.pgt.pgd); in kvm_host_prepare_stage2() 108 mmu->pgt = &host_kvm.pgt; in kvm_host_prepare_stage2() 144 struct kvm_pgtable *pgt = &host_kvm.pgt; in host_stage2_unmap_dev_all() local 152 ret = kvm_pgtable_stage2_unmap(pgt, addr, reg->base - addr); in host_stage2_unmap_dev_all() 156 return kvm_pgtable_stage2_unmap(pgt, addr, BIT(pgt->ia_bits) - addr); in host_stage2_unmap_dev_all() 205 return kvm_pgtable_stage2_map(&host_kvm.pgt, start, end - start, start, in __host_stage2_idmap() 221 ret = kvm_pgtable_stage2_find_range(&host_kvm.pgt, addr, prot, &range); in host_stage2_idmap() 260 ret = kvm_pgtable_stage2_set_owner(&host_kvm.pgt, start, end - start, in __pkvm_mark_hyp()
|
/dports/multimedia/libv4l/linux-5.13-rc2/arch/arm64/kvm/hyp/nvhe/ |
H A D | mem_protect.c | 101 ret = kvm_pgtable_stage2_init_flags(&host_kvm.pgt, &host_kvm.arch, in kvm_host_prepare_stage2() 106 mmu->pgd_phys = __hyp_pa(host_kvm.pgt.pgd); in kvm_host_prepare_stage2() 108 mmu->pgt = &host_kvm.pgt; in kvm_host_prepare_stage2() 144 struct kvm_pgtable *pgt = &host_kvm.pgt; in host_stage2_unmap_dev_all() local 152 ret = kvm_pgtable_stage2_unmap(pgt, addr, reg->base - addr); in host_stage2_unmap_dev_all() 156 return kvm_pgtable_stage2_unmap(pgt, addr, BIT(pgt->ia_bits) - addr); in host_stage2_unmap_dev_all() 205 return kvm_pgtable_stage2_map(&host_kvm.pgt, start, end - start, start, in __host_stage2_idmap() 221 ret = kvm_pgtable_stage2_find_range(&host_kvm.pgt, addr, prot, &range); in host_stage2_idmap() 260 ret = kvm_pgtable_stage2_set_owner(&host_kvm.pgt, start, end - start, in __pkvm_mark_hyp()
|
/dports/net-mgmt/ocsinventory-ocsreports/ocsinventory-ocsreports/vendor/jasig/phpcas/source/CAS/PGTStorage/ |
H A D | File.php | 194 * @param string $pgt the PGT 201 function write($pgt,$pgt_iou) argument 210 if (fputs($f, $pgt) === false) { 237 $pgt = false; 243 if (($pgt=fgets($f)) === false) { 254 phpCAS::traceEnd($pgt); 255 return $pgt;
|
/dports/www/tikiwiki/tiki-21.2/vendor_bundled/vendor/jasig/phpcas/source/CAS/PGTStorage/ |
H A D | File.php | 194 * @param string $pgt the PGT 201 function write($pgt,$pgt_iou) argument 210 if (fputs($f, $pgt) === false) { 237 $pgt = false; 243 if (($pgt=fgets($f)) === false) { 254 phpCAS::traceEnd($pgt); 255 return $pgt;
|
/dports/www/moodle310/moodle/auth/cas/CAS/CAS/PGTStorage/ |
H A D | File.php | 194 * @param string $pgt the PGT 201 function write($pgt,$pgt_iou) argument 210 if (fputs($f, $pgt) === false) { 237 $pgt = false; 243 if (($pgt=fgets($f)) === false) { 254 phpCAS::traceEnd($pgt); 255 return $pgt;
|
/dports/www/moodle311/moodle/auth/cas/CAS/CAS/PGTStorage/ |
H A D | File.php | 194 * @param string $pgt the PGT 201 function write($pgt,$pgt_iou) argument 210 if (fputs($f, $pgt) === false) { 237 $pgt = false; 243 if (($pgt=fgets($f)) === false) { 254 phpCAS::traceEnd($pgt); 255 return $pgt;
|