/* * Copyright (C) 2001 - 2007 Jeff Dike (jdike@{addtoit,linux.intel}.com) * Licensed under the GPL */ #include <linux/mm.h> #include <linux/sched.h> #include <linux/slab.h> #include <asm/unistd.h> #include <os.h> #include <proc_mm.h> #include <skas.h> #include <skas_ptrace.h> #include <sysdep/tls.h> extern int modify_ldt(int func, void *ptr, unsigned long bytecount); static long write_ldt_entry(struct mm_id *mm_idp, int func, struct user_desc *desc, void **addr, int done) { long res; if (proc_mm) { /* * This is a special handling for the case, that the mm to * modify isn't current->active_mm. * If this is called directly by modify_ldt, * (current->active_mm->context.skas.u == mm_idp) * will be true. So no call to __switch_mm(mm_idp) is done. * If this is called in case of init_new_ldt or PTRACE_LDT, * mm_idp won't belong to current->active_mm, but child->mm. * So we need to switch child's mm into our userspace, then * later switch back. * * Note: I'm unsure: should interrupts be disabled here? */ if (!current->active_mm || current->active_mm == &init_mm || mm_idp != ¤t->active_mm->context.id) __switch_mm(mm_idp); } if (ptrace_ldt) { struct ptrace_ldt ldt_op = (struct ptrace_ldt) { .func = func, .ptr = desc, .bytecount = sizeof(*desc)}; u32 cpu; int pid; if (!proc_mm) pid = mm_idp->u.pid; else { cpu = get_cpu(); pid = userspace_pid[cpu]; } res = os_ptrace_ldt(pid, 0, (unsigned long) &ldt_op); if (proc_mm) put_cpu(); } else { void *stub_addr; res = syscall_stub_data(mm_idp, (unsigned long *)desc, (sizeof(*desc) + sizeof(long) - 1) & ~(sizeof(long) - 1), addr, &stub_addr); if (!res) { unsigned long args[] = { func, (unsigned long)stub_addr, sizeof(*desc), 0, 0, 0 }; res = run_syscall_stub(mm_idp, __NR_modify_ldt, args, 0, addr, done); } } if (proc_mm) { /* * This is the second part of special handling, that makes * PTRACE_LDT possible to implement. */ if (current->active_mm && current->active_mm != &init_mm && mm_idp != ¤t->active_mm->context.id) __switch_mm(¤t->active_mm->context.id); } return res; } static long read_ldt_from_host(void __user * ptr, unsigned long bytecount) { int res, n; struct ptrace_ldt ptrace_ldt = (struct ptrace_ldt) { .func = 0, .bytecount = bytecount, .ptr = kmalloc(bytecount, GFP_KERNEL)}; u32 cpu; if (ptrace_ldt.ptr == NULL) return -ENOMEM; /* * This is called from sys_modify_ldt only, so userspace_pid gives * us the right number */ cpu = get_cpu(); res = os_ptrace_ldt(userspace_pid[cpu], 0, (unsigned long) &ptrace_ldt); put_cpu(); if (res < 0) goto out; n = copy_to_user(ptr, ptrace_ldt.ptr, res); if (n != 0) res = -EFAULT; out: kfree(ptrace_ldt.ptr); return res; } /* * In skas mode, we hold our own ldt data in UML. * Thus, the code implementing sys_modify_ldt_skas * is very similar to (and mostly stolen from) sys_modify_ldt * for arch/i386/kernel/ldt.c * The routines copied and modified in part are: * - read_ldt * - read_default_ldt * - write_ldt * - sys_modify_ldt_skas */ static int read_ldt(void __user * ptr, unsigned long bytecount) { int i, err = 0; unsigned long size; uml_ldt_t *ldt = ¤t->mm->context.arch.ldt; if (!ldt->entry_count) goto out; if (bytecount > LDT_ENTRY_SIZE*LDT_ENTRIES) bytecount = LDT_ENTRY_SIZE*LDT_ENTRIES; err = bytecount; if (ptrace_ldt) return read_ldt_from_host(ptr, bytecount); mutex_lock(&ldt->lock); if (ldt->entry_count <= LDT_DIRECT_ENTRIES) { size = LDT_ENTRY_SIZE*LDT_DIRECT_ENTRIES; if (size > bytecount) size = bytecount; if (copy_to_user(ptr, ldt->u.entries, size)) err = -EFAULT; bytecount -= size; ptr += size; } else { for (i=0; i<ldt->entry_count/LDT_ENTRIES_PER_PAGE && bytecount; i++) { size = PAGE_SIZE; if (size > bytecount) size = bytecount; if (copy_to_user(ptr, ldt->u.pages[i], size)) { err = -EFAULT; break; } bytecount -= size; ptr += size; } } mutex_unlock(&ldt->lock); if (bytecount == 0 || err == -EFAULT) goto out; if (clear_user(ptr, bytecount)) err = -EFAULT; out: return err; } static int read_default_ldt(void __user * ptr, unsigned long bytecount) { int err; if (bytecount > 5*LDT_ENTRY_SIZE) bytecount = 5*LDT_ENTRY_SIZE; err = bytecount; /* * UML doesn't support lcall7 and lcall27. * So, we don't really have a default ldt, but emulate * an empty ldt of common host default ldt size. */ if (clear_user(ptr, bytecount)) err = -EFAULT; return err; } static int write_ldt(void __user * ptr, unsigned long bytecount, int func) { uml_ldt_t *ldt = ¤t->mm->context.arch.ldt; struct mm_id * mm_idp = ¤t->mm->context.id; int i, err; struct user_desc ldt_info; struct ldt_entry entry0, *ldt_p; void *addr = NULL; err = -EINVAL; if (bytecount != sizeof(ldt_info)) goto out; err = -EFAULT; if (copy_from_user(&ldt_info, ptr, sizeof(ldt_info))) goto out; err = -EINVAL; if (ldt_info.entry_number >= LDT_ENTRIES) goto out; if (ldt_info.contents == 3) { if (func == 1) goto out; if (ldt_info.seg_not_present == 0) goto out; } if (!ptrace_ldt) mutex_lock(&ldt->lock); err = write_ldt_entry(mm_idp, func, &ldt_info, &addr, 1); if (err) goto out_unlock; else if (ptrace_ldt) { /* With PTRACE_LDT available, this is used as a flag only */ ldt->entry_count = 1; goto out; } if (ldt_info.entry_number >= ldt->entry_count && ldt_info.entry_number >= LDT_DIRECT_ENTRIES) { for (i=ldt->entry_count/LDT_ENTRIES_PER_PAGE; i*LDT_ENTRIES_PER_PAGE <= ldt_info.entry_number; i++) { if (i == 0) memcpy(&entry0, ldt->u.entries, sizeof(entry0)); ldt->u.pages[i] = (struct ldt_entry *) __get_free_page(GFP_KERNEL|__GFP_ZERO); if (!ldt->u.pages[i]) { err = -ENOMEM; /* Undo the change in host */ memset(&ldt_info, 0, sizeof(ldt_info)); write_ldt_entry(mm_idp, 1, &ldt_info, &addr, 1); goto out_unlock; } if (i == 0) { memcpy(ldt->u.pages[0], &entry0, sizeof(entry0)); memcpy(ldt->u.pages[0]+1, ldt->u.entries+1, sizeof(entry0)*(LDT_DIRECT_ENTRIES-1)); } ldt->entry_count = (i + 1) * LDT_ENTRIES_PER_PAGE; } } if (ldt->entry_count <= ldt_info.entry_number) ldt->entry_count = ldt_info.entry_number + 1; if (ldt->entry_count <= LDT_DIRECT_ENTRIES) ldt_p = ldt->u.entries + ldt_info.entry_number; else ldt_p = ldt->u.pages[ldt_info.entry_number/LDT_ENTRIES_PER_PAGE] + ldt_info.entry_number%LDT_ENTRIES_PER_PAGE; if (ldt_info.base_addr == 0 && ldt_info.limit == 0 && (func == 1 || LDT_empty(&ldt_info))) { ldt_p->a = 0; ldt_p->b = 0; } else{ if (func == 1) ldt_info.useable = 0; ldt_p->a = LDT_entry_a(&ldt_info); ldt_p->b = LDT_entry_b(&ldt_info); } err = 0; out_unlock: mutex_unlock(&ldt->lock); out: return err; } static long do_modify_ldt_skas(int func, void __user *ptr, unsigned long bytecount) { int ret = -ENOSYS; switch (func) { case 0: ret = read_ldt(ptr, bytecount); break; case 1: case 0x11: ret = write_ldt(ptr, bytecount, func); break; case 2: ret = read_default_ldt(ptr, bytecount); break; } return ret; } static DEFINE_SPINLOCK(host_ldt_lock); static short dummy_list[9] = {0, -1}; static short * host_ldt_entries = NULL; static void ldt_get_host_info(void) { long ret; struct ldt_entry * ldt; short *tmp; int i, size, k, order; spin_lock(&host_ldt_lock); if (host_ldt_entries != NULL) { spin_unlock(&host_ldt_lock); return; } host_ldt_entries = dummy_list+1; spin_unlock(&host_ldt_lock); for (i = LDT_PAGES_MAX-1, order=0; i; i>>=1, order++) ; ldt = (struct ldt_entry *) __get_free_pages(GFP_KERNEL|__GFP_ZERO, order); if (ldt == NULL) { printk(KERN_ERR "ldt_get_host_info: couldn't allocate buffer " "for host ldt\n"); return; } ret = modify_ldt(0, ldt, (1<<order)*PAGE_SIZE); if (ret < 0) { printk(KERN_ERR "ldt_get_host_info: couldn't read host ldt\n"); goto out_free; } if (ret == 0) { /* default_ldt is active, simply write an empty entry 0 */ host_ldt_entries = dummy_list; goto out_free; } for (i=0, size=0; i<ret/LDT_ENTRY_SIZE; i++) { if (ldt[i].a != 0 || ldt[i].b != 0) size++; } if (size < ARRAY_SIZE(dummy_list)) host_ldt_entries = dummy_list; else { size = (size + 1) * sizeof(dummy_list[0]); tmp = kmalloc(size, GFP_KERNEL); if (tmp == NULL) { printk(KERN_ERR "ldt_get_host_info: couldn't allocate " "host ldt list\n"); goto out_free; } host_ldt_entries = tmp; } for (i=0, k=0; i<ret/LDT_ENTRY_SIZE; i++) { if (ldt[i].a != 0 || ldt[i].b != 0) host_ldt_entries[k++] = i; } host_ldt_entries[k] = -1; out_free: free_pages((unsigned long)ldt, order); } long init_new_ldt(struct mm_context *new_mm, struct mm_context *from_mm) { struct user_desc desc; short * num_p; int i; long page, err=0; void *addr = NULL; struct proc_mm_op copy; if (!ptrace_ldt) mutex_init(&new_mm->arch.ldt.lock); if (!from_mm) { memset(&desc, 0, sizeof(desc)); /* * We have to initialize a clean ldt. */ if (proc_mm) { /* * If the new mm was created using proc_mm, host's * default-ldt currently is assigned, which normally * contains the call-gates for lcall7 and lcall27. * To remove these gates, we simply write an empty * entry as number 0 to the host. */ err = write_ldt_entry(&new_mm->id, 1, &desc, &addr, 1); } else{ /* * Now we try to retrieve info about the ldt, we * inherited from the host. All ldt-entries found * will be reset in the following loop */ ldt_get_host_info(); for (num_p=host_ldt_entries; *num_p != -1; num_p++) { desc.entry_number = *num_p; err = write_ldt_entry(&new_mm->id, 1, &desc, &addr, *(num_p + 1) == -1); if (err) break; } } new_mm->arch.ldt.entry_count = 0; goto out; } if (proc_mm) { /* * We have a valid from_mm, so we now have to copy the LDT of * from_mm to new_mm, because using proc_mm an new mm with * an empty/default LDT was created in new_mm() */ copy = ((struct proc_mm_op) { .op = MM_COPY_SEGMENTS, .u = { .copy_segments = from_mm->id.u.mm_fd } } ); i = os_write_file(new_mm->id.u.mm_fd, ©, sizeof(copy)); if (i != sizeof(copy)) printk(KERN_ERR "new_mm : /proc/mm copy_segments " "failed, err = %d\n", -i); } if (!ptrace_ldt) { /* * Our local LDT is used to supply the data for * modify_ldt(READLDT), if PTRACE_LDT isn't available, * i.e., we have to use the stub for modify_ldt, which * can't handle the big read buffer of up to 64kB. */ mutex_lock(&from_mm->arch.ldt.lock); if (from_mm->arch.ldt.entry_count <= LDT_DIRECT_ENTRIES) memcpy(new_mm->arch.ldt.u.entries, from_mm->arch.ldt.u.entries, sizeof(new_mm->arch.ldt.u.entries)); else { i = from_mm->arch.ldt.entry_count / LDT_ENTRIES_PER_PAGE; while (i-->0) { page = __get_free_page(GFP_KERNEL|__GFP_ZERO); if (!page) { err = -ENOMEM; break; } new_mm->arch.ldt.u.pages[i] = (struct ldt_entry *) page; memcpy(new_mm->arch.ldt.u.pages[i], from_mm->arch.ldt.u.pages[i], PAGE_SIZE); } } new_mm->arch.ldt.entry_count = from_mm->arch.ldt.entry_count; mutex_unlock(&from_mm->arch.ldt.lock); } out: return err; } void free_ldt(struct mm_context *mm) { int i; if (!ptrace_ldt && mm->arch.ldt.entry_count > LDT_DIRECT_ENTRIES) { i = mm->arch.ldt.entry_count / LDT_ENTRIES_PER_PAGE; while (i-- > 0) free_page((long) mm->arch.ldt.u.pages[i]); } mm->arch.ldt.entry_count = 0; } int sys_modify_ldt(int func, void __user *ptr, unsigned long bytecount) { return do_modify_ldt_skas(func, ptr, bytecount); }