+#include <lunaix/mm/page.h>
#include <lunaix/mm/region.h>
#include <lunaix/mm/valloc.h>
+#include <lunaix/spike.h>
+#include <lunaix/process.h>
+
+#include <sys/mm/mempart.h>
#include <klibc/string.h>
struct mm_region*
region_create(ptr_t start, ptr_t end, u32_t attr)
{
- return valloc(sizeof(struct mm_region));
+ assert_msg(PG_ALIGNED(start), "not page aligned");
+ assert_msg(PG_ALIGNED(end), "not page aligned");
+ struct mm_region* region = valloc(sizeof(struct mm_region));
+ *region =
+ (struct mm_region){ .attr = attr, .start = start, .end = end - 1 };
+ return region;
+}
+
+struct mm_region*
+region_create_range(ptr_t start, size_t length, u32_t attr)
+{
+ assert_msg(PG_ALIGNED(start), "not page aligned");
+ assert_msg(PG_ALIGNED(length), "not page aligned");
+ struct mm_region* region = valloc(sizeof(struct mm_region));
+ *region = (struct mm_region){ .attr = attr,
+ .start = start,
+ .end = ROUNDUP(start + length, MEM_PAGE) };
+ return region;
+}
+
+struct mm_region*
+region_dup(struct mm_region* origin)
+{
+ struct mm_region* region = valloc(sizeof(struct mm_region));
+ *region = *origin;
+
+ if (region->mfile) {
+ vfs_ref_file(region->mfile);
+ }
+
+ llist_init_head(®ion->head);
+ return region;
}
void
{
if (llist_empty(lead)) {
llist_append(lead, &vmregion->head);
- return vmregion;
+ return;
}
+ struct mm_region *pos, *n;
ptr_t cur_end = 0;
- struct mm_region *pos = (struct mm_region*)lead,
- *n = list_entry(lead->next, struct mm_region, head);
- do {
- if (vmregion->start >= cur_end && vmregion->end <= n->start) {
+
+ llist_for_each(pos, n, lead, head)
+ {
+ if (vmregion->start >= cur_end && vmregion->end <= pos->start) {
break;
}
- cur_end = n->end;
- pos = n;
- n = list_entry(n->head.next, struct mm_region, head);
- } while ((ptr_t)&pos->head != (ptr_t)lead);
+ cur_end = pos->end;
+ }
// XXX caution. require mm_region::head to be the lead of struct
- llist_insert_after(&pos->head, &vmregion->head);
+ llist_append(&pos->head, &vmregion->head);
+}
+
+void
+region_release(struct mm_region* region)
+{
+ if (region->destruct_region) {
+ region->destruct_region(region);
+ }
+
+ if (region->mfile) {
+ struct proc_mm* mm = region->proc_vms;
+ vfs_pclose(region->mfile, mm->proc->pid);
+ }
+
+ if (region->index) {
+ *region->index = NULL;
+ }
+
+ vfree(region);
}
void
llist_for_each(pos, n, lead, head)
{
- vfree(pos);
+ region_release(pos);
}
}
void
-region_copy(vm_regions_t* src, vm_regions_t* dest)
+region_copy_mm(struct proc_mm* src, struct proc_mm* dest)
{
- if (!src) {
- return;
- }
-
struct mm_region *pos, *n, *dup;
- llist_for_each(pos, n, src, head)
+ llist_for_each(pos, n, &src->regions, head)
{
dup = valloc(sizeof(struct mm_region));
memcpy(dup, pos, sizeof(*pos));
- region_add(dest, dup);
+
+ dup->proc_vms = dest;
+
+ if (dup->mfile) {
+ vfs_ref_file(dup->mfile);
+ }
+
+ if (dup->region_copied) {
+ dup->region_copied(dup);
+ }
+
+ llist_append(&dest->regions, &dup->head);
}
}
struct mm_region *pos, *n;
+ vaddr = PG_ALIGN(vaddr);
+
llist_for_each(pos, n, lead, head)
{
if (pos->start <= vaddr && vaddr < pos->end) {