#include #include #include #include #include #include #include #include #include struct proc_mm* procvm_create(struct proc_info* proc) { struct proc_mm* mm = valloc(sizeof(struct proc_mm)); assert(mm); mm->heap = 0; mm->proc = proc; llist_init_head(&mm->regions); return mm; } static ptr_t __dup_vmspace(ptr_t mount_point, bool only_kernel) { ptr_t ptd_pp = pmm_alloc_page(PP_FGPERSIST); vmm_set_mapping(VMS_SELF, PG_MOUNT_1, ptd_pp, PG_PREM_RW, VMAP_NULL); x86_page_table* ptd = (x86_page_table*)PG_MOUNT_1; x86_page_table* pptd = (x86_page_table*)(mount_point | (0x3FF << 12)); size_t kspace_l1inx = L1_INDEX(KERNEL_EXEC); size_t i = 1; // skip first 4MiB, to avoid bring other thread's stack ptd->entry[0] = 0; if (only_kernel) { i = kspace_l1inx; memset(ptd, 0, PG_SIZE); } for (; i < PG_MAX_ENTRIES - 1; i++) { x86_pte_t ptde = pptd->entry[i]; // 空或者是未在内存中的L1页表项直接照搬过去。 // 内核地址空间直接共享过去。 if (!ptde || i >= kspace_l1inx || !(ptde & PG_PRESENT)) { ptd->entry[i] = ptde; continue; } // 复制L2页表 ptr_t pt_pp = pmm_alloc_page(PP_FGPERSIST); vmm_set_mapping(VMS_SELF, PG_MOUNT_2, pt_pp, PG_PREM_RW, VMAP_NULL); x86_page_table* ppt = (x86_page_table*)(mount_point | (i << 12)); x86_page_table* pt = (x86_page_table*)PG_MOUNT_2; for (size_t j = 0; j < PG_MAX_ENTRIES; j++) { x86_pte_t pte = ppt->entry[j]; pmm_ref_page(PG_ENTRY_ADDR(pte)); pt->entry[j] = pte; } ptd->entry[i] = (ptr_t)pt_pp | PG_ENTRY_FLAGS(ptde); } ptd->entry[PG_MAX_ENTRIES - 1] = NEW_L1_ENTRY(T_SELF_REF_PERM, ptd_pp); return ptd_pp; } void procvm_dup(struct proc_info* proc) { struct proc_mm* mm = vmspace(proc); struct proc_mm* mm_current = vmspace(__current); mm->heap = mm_current->heap; mm->vmroot = __dup_vmspace(VMS_SELF, false); region_copy_mm(mm_current, mm); } void procvm_init_clean(struct proc_info* proc) { struct proc_mm* mm = vmspace(proc); mm->vmroot = __dup_vmspace(VMS_SELF, true); } static void __delete_vmspace(ptr_t vm_mnt) { x86_page_table* pptd = (x86_page_table*)(vm_mnt | (0x3FF << 12)); // only remove user address space for (size_t i = 0; i < L1_INDEX(KERNEL_EXEC); i++) { x86_pte_t ptde = pptd->entry[i]; if (!ptde || !(ptde & PG_PRESENT)) { continue; } x86_page_table* ppt = (x86_page_table*)(vm_mnt | (i << 12)); for (size_t j = 0; j < PG_MAX_ENTRIES; j++) { x86_pte_t pte = ppt->entry[j]; // free the 4KB data page if ((pte & PG_PRESENT)) { pmm_free_page(PG_ENTRY_ADDR(pte)); } } // free the L2 page table pmm_free_page(PG_ENTRY_ADDR(ptde)); } // free the L1 directory pmm_free_page(PG_ENTRY_ADDR(pptd->entry[PG_MAX_ENTRIES - 1])); } void procvm_cleanup(ptr_t vm_mnt, struct proc_info* proc) { struct mm_region *pos, *n; llist_for_each(pos, n, vmregions(proc), head) { mem_sync_pages(vm_mnt, pos, pos->start, pos->end - pos->start, 0); region_release(pos); } vfree(proc->mm); __delete_vmspace(vm_mnt); } ptr_t procvm_enter_remote(struct remote_vmctx* rvmctx, struct proc_mm* mm, ptr_t vm_mnt, ptr_t remote_base, size_t size) { ptr_t size_pn = PN(size + MEM_PAGE); assert(size_pn < REMOTEVM_MAX_PAGES); struct mm_region* region = region_get(&mm->regions, remote_base); assert(region && region_contains(region, remote_base + size)); rvmctx->vms_mnt = vm_mnt; rvmctx->page_cnt = size_pn; remote_base = PG_ALIGN(remote_base); rvmctx->remote = remote_base; rvmctx->local_mnt = PG_MOUNT_4_END + 1; v_mapping m; unsigned int pattr = region_ptattr(region); ptr_t raddr = remote_base, lmnt = rvmctx->local_mnt; for (size_t i = 0; i < size_pn; i++, lmnt += MEM_PAGE, raddr += MEM_PAGE) { if (vmm_lookupat(vm_mnt, raddr, &m) && PG_IS_PRESENT(m.flags)) { vmm_set_mapping(VMS_SELF, lmnt, m.pa, PG_PREM_RW, 0); continue; } ptr_t pa = pmm_alloc_page(0); vmm_set_mapping(VMS_SELF, lmnt, pa, PG_PREM_RW, 0); vmm_set_mapping(vm_mnt, raddr, pa, pattr, 0); } return vm_mnt; } int procvm_copy_remote_transaction(struct remote_vmctx* rvmctx, ptr_t remote_dest, void* local_src, size_t sz) { if (remote_dest < rvmctx->remote) { return -1; } ptr_t offset = remote_dest - rvmctx->remote; if (PN(offset + sz) >= rvmctx->page_cnt) { return -1; } memcpy((void*)(rvmctx->local_mnt + offset), local_src, sz); return sz; } void procvm_exit_remote_transaction(struct remote_vmctx* rvmctx) { ptr_t lmnt = rvmctx->local_mnt; for (size_t i = 0; i < rvmctx->page_cnt; i++, lmnt += MEM_PAGE) { vmm_del_mapping(VMS_SELF, lmnt); } }