-#include <lunaix/mm/mmio.h>
#include <lunaix/mm/page.h>
+#include <lunaix/mm/vmm.h>
+#include <lunaix/process.h>
#include <lunaix/spike.h>
+#include <lunaix/syslog.h>
#include <lunaix/trace.h>
+#include <sys/cpu.h>
+#include <sys/mm/mempart.h>
+
#include <klibc/string.h>
+#define NB_TRACEBACK 16
+
+LOG_MODULE("TRACE")
+
static struct trace_context trace_ctx;
void
for (size_t i = 0; i < bhctx->mods.mods_num; i++) {
struct boot_modent* mod = &bhctx->mods.entries[i];
if (streq(mod->str, "modksyms")) {
- // In case boot loader does not place our ksyms on page boundary
- ptr_t start = PG_ALIGN(mod->start);
+ assert(PG_ALIGNED(mod->start));
+
ptr_t end = ROUNDUP(mod->end, PG_SIZE);
- ptr_t ksym_va = (ptr_t)ioremap(start, (end - start));
+ ptr_t ksym_va =
+ (ptr_t)vmap(mod->start, (end - mod->start), PG_PREM_R, 0);
- trace_ctx.ksym_table =
- (struct ksyms*)(ksym_va + (mod->start - start));
+ assert(ksym_va);
+ trace_ctx.ksym_table = (struct ksyms*)ksym_va;
}
}
}
struct ksym_entry*
-trace_sym_lookup(ptr_t pc)
+trace_sym_lookup(ptr_t addr)
+{
+ int c = trace_ctx.ksym_table->ksym_count;
+ struct ksym_entry* ksent = trace_ctx.ksym_table->syms;
+
+ int i = c - 1, j = 0, m = 0;
+
+ if (addr > ksent[i].pc || addr < ksent[j].pc || addr < KERNEL_EXEC) {
+ return NULL;
+ }
+
+ while (i - j != 1) {
+ m = (i + j) / 2;
+ if (ksent[m].pc > addr) {
+ i = m;
+ } else if (ksent[m].pc < addr) {
+ j = m;
+ } else {
+ break;
+ }
+ }
+
+ struct ksym_entry* result = &ksent[MIN(i, j)];
+ if (result->pc > addr) {
+ return NULL;
+ }
+
+ return result;
+}
+
+static char*
+ksym_getstr(struct ksym_entry* sym)
+{
+ if (!sym) {
+ return "???";
+ }
+
+ return (char*)((ptr_t)trace_ctx.ksym_table +
+ trace_ctx.ksym_table->ksym_label_off + sym->label_off);
+}
+
+static inline bool valid_fp(ptr_t ptr) {
+ return KERNEL_STACK < ptr && ptr < KERNEL_EXEC_END;
+}
+
+int
+trace_walkback(struct trace_record* tb_buffer,
+ ptr_t fp,
+ int limit,
+ ptr_t* last_fp)
+{
+ ptr_t* frame = (ptr_t*)fp;
+ struct ksym_entry* current = NULL;
+ int i = 0;
+
+ while (valid_fp((ptr_t)frame) && i < limit) {
+ ptr_t pc = *(frame + 1);
+
+ current = trace_sym_lookup(pc);
+ tb_buffer[i] =
+ (struct trace_record){ .pc = pc,
+ .sym_pc = current ? current->pc : 0,
+ .symbol = ksym_getstr(current) };
+
+ frame = (ptr_t*)*frame;
+ i++;
+ }
+
+ if (last_fp) {
+ *last_fp = (ptr_t)frame;
+ }
+
+ return i;
+}
+
+static inline void
+trace_print_code_entry(ptr_t sym_pc, ptr_t inst_pc, char* sym)
+{
+ DEBUG("%p+%p: %s", sym_pc, inst_pc - sym_pc, sym);
+}
+
+void
+trace_printstack_of(ptr_t fp)
+{
+ struct trace_record tbs[NB_TRACEBACK];
+
+ // Let's get our Stackwalker does his job ;)
+ int n = trace_walkback(tbs, fp, NB_TRACEBACK, &fp);
+
+ if (fp) {
+ DEBUG("...<truncated>");
+ }
+
+ for (int i = 0; i < n; i++) {
+ struct trace_record* tb = &tbs[i];
+ trace_print_code_entry(tb->sym_pc, tb->pc, tb->symbol);
+ }
+}
+
+void
+trace_printstack()
+{
+ trace_printstack_of(cpu_get_fp());
+}
+
+static void
+trace_printswctx(const isr_param* p, char* direction)
{
- return NULL;
+
+ struct ksym_entry* sym = trace_sym_lookup(p->execp->eip);
+
+ DEBUG(">> (sw:%s) iv:%d, errno:%p <<",
+ direction,
+ p->execp->vector,
+ p->execp->err_code);
+
+ ptr_t sym_pc = sym ? sym->pc : p->execp->eip;
+ trace_print_code_entry(sym_pc, p->execp->eip, ksym_getstr(sym));
}
void
-trace_walkback(ptr_t fp)
+trace_printstack_isr(const isr_param* isrm)
{
+ isr_param* p = isrm;
+ ptr_t fp = cpu_get_fp();
+ int prev_fromusr = 0;
+
+ DEBUG("stack trace (pid=%d)\n", __current->pid);
+
+ trace_printstack_of(fp);
+
+ while (p) {
+ if (!prev_fromusr) {
+ if (uspace_context(p)) {
+ trace_printswctx(p, "s/u");
+ } else {
+ trace_printswctx(p, "s/s");
+ }
+ } else {
+ trace_printswctx(p, "u/s");
+ }
+
+ fp = saved_fp(p);
+ trace_printstack_of(fp);
+
+ prev_fromusr = uspace_context(p);
+
+ p = p->execp->saved_prev_ctx;
+ }
}
\ No newline at end of file