va                148 kernel/exec.c  loadseg(pagetable_t pagetable, uint64 va, struct inode *ip, uint offset, uint sz)
va                154 kernel/exec.c      pa = walkaddr(pagetable, va + i);
va                 41 kernel/proc.c      uint64 va = KSTACK((int) (p - proc));
va                 42 kernel/proc.c      kvmmap(kpgtbl, va, (uint64)pa, PGSIZE, PTE_R | PTE_W);
va                376 kernel/riscv.h #define PX(level, va) ((((uint64) (va)) >> PXSHIFT(level)) & PXMASK)
va                 86 kernel/vm.c    walk(pagetable_t pagetable, uint64 va, int alloc)
va                 88 kernel/vm.c      if(va >= MAXVA)
va                 92 kernel/vm.c        pte_t *pte = &pagetable[PX(level, va)];
va                102 kernel/vm.c      return &pagetable[PX(0, va)];
va                109 kernel/vm.c    walkaddr(pagetable_t pagetable, uint64 va)
va                114 kernel/vm.c      if(va >= MAXVA)
va                117 kernel/vm.c      pte = walk(pagetable, va, 0);
va                132 kernel/vm.c    kvmmap(pagetable_t kpgtbl, uint64 va, uint64 pa, uint64 sz, int perm)
va                134 kernel/vm.c      if(mappages(kpgtbl, va, sz, pa, perm) != 0)
va                144 kernel/vm.c    mappages(pagetable_t pagetable, uint64 va, uint64 size, uint64 pa, int perm)
va                149 kernel/vm.c      if((va % PGSIZE) != 0)
va                158 kernel/vm.c      a = va;
va                159 kernel/vm.c      last = va + size - PGSIZE;
va                178 kernel/vm.c    uvmunmap(pagetable_t pagetable, uint64 va, uint64 npages, int do_free)
va                183 kernel/vm.c      if((va % PGSIZE) != 0)
va                186 kernel/vm.c      for(a = va; a < va + npages*PGSIZE; a += PGSIZE){
va                345 kernel/vm.c    uvmclear(pagetable_t pagetable, uint64 va)
va                349 kernel/vm.c      pte = walk(pagetable, va, 0);