va 148 kernel/exec.c loadseg(pagetable_t pagetable, uint64 va, struct inode *ip, uint offset, uint sz) va 154 kernel/exec.c pa = walkaddr(pagetable, va + i); va 41 kernel/proc.c uint64 va = KSTACK((int) (p - proc)); va 42 kernel/proc.c kvmmap(kpgtbl, va, (uint64)pa, PGSIZE, PTE_R | PTE_W); va 376 kernel/riscv.h #define PX(level, va) ((((uint64) (va)) >> PXSHIFT(level)) & PXMASK) va 86 kernel/vm.c walk(pagetable_t pagetable, uint64 va, int alloc) va 88 kernel/vm.c if(va >= MAXVA) va 92 kernel/vm.c pte_t *pte = &pagetable[PX(level, va)]; va 102 kernel/vm.c return &pagetable[PX(0, va)]; va 109 kernel/vm.c walkaddr(pagetable_t pagetable, uint64 va) va 114 kernel/vm.c if(va >= MAXVA) va 117 kernel/vm.c pte = walk(pagetable, va, 0); va 132 kernel/vm.c kvmmap(pagetable_t kpgtbl, uint64 va, uint64 pa, uint64 sz, int perm) va 134 kernel/vm.c if(mappages(kpgtbl, va, sz, pa, perm) != 0) va 144 kernel/vm.c mappages(pagetable_t pagetable, uint64 va, uint64 size, uint64 pa, int perm) va 149 kernel/vm.c if((va % PGSIZE) != 0) va 158 kernel/vm.c a = va; va 159 kernel/vm.c last = va + size - PGSIZE; va 178 kernel/vm.c uvmunmap(pagetable_t pagetable, uint64 va, uint64 npages, int do_free) va 183 kernel/vm.c if((va % PGSIZE) != 0) va 186 kernel/vm.c for(a = va; a < va + npages*PGSIZE; a += PGSIZE){ va 345 kernel/vm.c uvmclear(pagetable_t pagetable, uint64 va) va 349 kernel/vm.c pte = walk(pagetable, va, 0);