uml: remove code made redundant by CHOOSE_MODE removal
This patch makes a number of simplifications enabled by the removal of CHOOSE_MODE. There were lots of functions that looked like int foo(args){ foo_skas(args); } The bodies of foo_skas are now folded into foo, and their declarations (and sometimes entire header files) are deleted. In addition, the union uml_pt_regs, which was a union between the tt and skas register formats, is now a struct, with the tt-mode arm of the union being removed. It turns out that usr2_handler was unused, so it is gone. Signed-off-by: Jeff Dike <jdike@linux.intel.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:

committad av
Linus Torvalds

förälder
ae2587e412
incheckning
77bf440031
@@ -3,7 +3,7 @@
|
||||
# Licensed under the GPL
|
||||
#
|
||||
|
||||
obj-y := clone.o exec.o mem.o mmu.o process.o syscall.o tlb.o uaccess.o
|
||||
obj-y := clone.o mmu.o process.o syscall.o uaccess.o
|
||||
|
||||
# clone.o is in the stub, so it can't be built with profiling
|
||||
# GCC hardened also auto-enables -fpic, but we need %ebx so it can't work ->
|
||||
|
@@ -1,40 +0,0 @@
|
||||
/*
|
||||
* Copyright (C) 2002 Jeff Dike (jdike@karaya.com)
|
||||
* Licensed under the GPL
|
||||
*/
|
||||
|
||||
#include "linux/kernel.h"
|
||||
#include "asm/current.h"
|
||||
#include "asm/page.h"
|
||||
#include "asm/signal.h"
|
||||
#include "asm/ptrace.h"
|
||||
#include "asm/uaccess.h"
|
||||
#include "asm/mmu_context.h"
|
||||
#include "tlb.h"
|
||||
#include "skas.h"
|
||||
#include "um_mmu.h"
|
||||
#include "os.h"
|
||||
|
||||
void flush_thread_skas(void)
|
||||
{
|
||||
void *data = NULL;
|
||||
unsigned long end = proc_mm ? task_size : CONFIG_STUB_START;
|
||||
int ret;
|
||||
|
||||
ret = unmap(¤t->mm->context.skas.id, 0, end, 1, &data);
|
||||
if(ret){
|
||||
printk("flush_thread_skas - clearing address space failed, "
|
||||
"err = %d\n", ret);
|
||||
force_sig(SIGKILL, current);
|
||||
}
|
||||
|
||||
switch_mm_skas(¤t->mm->context.skas.id);
|
||||
}
|
||||
|
||||
void start_thread_skas(struct pt_regs *regs, unsigned long eip,
|
||||
unsigned long esp)
|
||||
{
|
||||
set_fs(USER_DS);
|
||||
PT_REGS_IP(regs) = eip;
|
||||
PT_REGS_SP(regs) = esp;
|
||||
}
|
@@ -1,22 +0,0 @@
|
||||
/*
|
||||
* Copyright (C) 2002 Jeff Dike (jdike@karaya.com)
|
||||
* Licensed under the GPL
|
||||
*/
|
||||
|
||||
#include "linux/mm.h"
|
||||
#include "asm/pgtable.h"
|
||||
#include "mem_user.h"
|
||||
#include "skas.h"
|
||||
|
||||
unsigned long set_task_sizes_skas(unsigned long *task_size_out)
|
||||
{
|
||||
/* Round up to the nearest 4M */
|
||||
unsigned long host_task_size = ROUND_4M((unsigned long)
|
||||
&host_task_size);
|
||||
|
||||
if (!skas_needs_stub)
|
||||
*task_size_out = host_task_size;
|
||||
else *task_size_out = CONFIG_STUB_START & PGDIR_MASK;
|
||||
|
||||
return host_task_size;
|
||||
}
|
@@ -71,7 +71,7 @@ static int init_stub_pte(struct mm_struct *mm, unsigned long proc,
|
||||
return(-ENOMEM);
|
||||
}
|
||||
|
||||
int init_new_context_skas(struct task_struct *task, struct mm_struct *mm)
|
||||
int init_new_context(struct task_struct *task, struct mm_struct *mm)
|
||||
{
|
||||
struct mmu_context_skas *from_mm = NULL;
|
||||
struct mmu_context_skas *to_mm = &mm->context.skas;
|
||||
@@ -137,7 +137,7 @@ int init_new_context_skas(struct task_struct *task, struct mm_struct *mm)
|
||||
return ret;
|
||||
}
|
||||
|
||||
void destroy_context_skas(struct mm_struct *mm)
|
||||
void destroy_context(struct mm_struct *mm)
|
||||
{
|
||||
struct mmu_context_skas *mmu = &mm->context.skas;
|
||||
|
||||
|
@@ -18,129 +18,22 @@
|
||||
#include "os.h"
|
||||
#include "tlb.h"
|
||||
#include "kern.h"
|
||||
#include "mode.h"
|
||||
#include "registers.h"
|
||||
|
||||
void switch_to_skas(void *prev, void *next)
|
||||
{
|
||||
struct task_struct *from, *to;
|
||||
|
||||
from = prev;
|
||||
to = next;
|
||||
|
||||
/* XXX need to check runqueues[cpu].idle */
|
||||
if(current->pid == 0)
|
||||
switch_timers(0);
|
||||
|
||||
switch_threads(&from->thread.mode.skas.switch_buf,
|
||||
&to->thread.mode.skas.switch_buf);
|
||||
|
||||
arch_switch_to_skas(current->thread.prev_sched, current);
|
||||
|
||||
if(current->pid == 0)
|
||||
switch_timers(1);
|
||||
}
|
||||
|
||||
extern void schedule_tail(struct task_struct *prev);
|
||||
|
||||
/* This is called magically, by its address being stuffed in a jmp_buf
|
||||
* and being longjmp-d to.
|
||||
*/
|
||||
void new_thread_handler(void)
|
||||
{
|
||||
int (*fn)(void *), n;
|
||||
void *arg;
|
||||
|
||||
if(current->thread.prev_sched != NULL)
|
||||
schedule_tail(current->thread.prev_sched);
|
||||
current->thread.prev_sched = NULL;
|
||||
|
||||
fn = current->thread.request.u.thread.proc;
|
||||
arg = current->thread.request.u.thread.arg;
|
||||
|
||||
/* The return value is 1 if the kernel thread execs a process,
|
||||
* 0 if it just exits
|
||||
*/
|
||||
n = run_kernel_thread(fn, arg, ¤t->thread.exec_buf);
|
||||
if(n == 1){
|
||||
/* Handle any immediate reschedules or signals */
|
||||
interrupt_end();
|
||||
userspace(¤t->thread.regs.regs);
|
||||
}
|
||||
else do_exit(0);
|
||||
}
|
||||
|
||||
void release_thread_skas(struct task_struct *task)
|
||||
{
|
||||
}
|
||||
|
||||
/* Called magically, see new_thread_handler above */
|
||||
void fork_handler(void)
|
||||
{
|
||||
force_flush_all();
|
||||
if(current->thread.prev_sched == NULL)
|
||||
panic("blech");
|
||||
|
||||
schedule_tail(current->thread.prev_sched);
|
||||
|
||||
/* XXX: if interrupt_end() calls schedule, this call to
|
||||
* arch_switch_to_skas isn't needed. We could want to apply this to
|
||||
* improve performance. -bb */
|
||||
arch_switch_to_skas(current->thread.prev_sched, current);
|
||||
|
||||
current->thread.prev_sched = NULL;
|
||||
|
||||
/* Handle any immediate reschedules or signals */
|
||||
interrupt_end();
|
||||
|
||||
userspace(¤t->thread.regs.regs);
|
||||
}
|
||||
|
||||
int copy_thread_skas(int nr, unsigned long clone_flags, unsigned long sp,
|
||||
unsigned long stack_top, struct task_struct * p,
|
||||
struct pt_regs *regs)
|
||||
{
|
||||
void (*handler)(void);
|
||||
|
||||
if(current->thread.forking){
|
||||
memcpy(&p->thread.regs.regs.skas, ®s->regs.skas,
|
||||
sizeof(p->thread.regs.regs.skas));
|
||||
REGS_SET_SYSCALL_RETURN(p->thread.regs.regs.skas.regs, 0);
|
||||
if(sp != 0) REGS_SP(p->thread.regs.regs.skas.regs) = sp;
|
||||
|
||||
handler = fork_handler;
|
||||
|
||||
arch_copy_thread(¤t->thread.arch, &p->thread.arch);
|
||||
}
|
||||
else {
|
||||
init_thread_registers(&p->thread.regs.regs);
|
||||
p->thread.request.u.thread = current->thread.request.u.thread;
|
||||
handler = new_thread_handler;
|
||||
}
|
||||
|
||||
new_thread(task_stack_page(p), &p->thread.mode.skas.switch_buf,
|
||||
handler);
|
||||
return(0);
|
||||
}
|
||||
|
||||
int new_mm(unsigned long stack)
|
||||
{
|
||||
int fd;
|
||||
|
||||
fd = os_open_file("/proc/mm", of_cloexec(of_write(OPENFLAGS())), 0);
|
||||
if(fd < 0)
|
||||
return(fd);
|
||||
return fd;
|
||||
|
||||
if(skas_needs_stub)
|
||||
map_stub_pages(fd, CONFIG_STUB_CODE, CONFIG_STUB_DATA, stack);
|
||||
|
||||
return(fd);
|
||||
}
|
||||
|
||||
void init_idle_skas(void)
|
||||
{
|
||||
cpu_tasks[current_thread->cpu].pid = os_getpid();
|
||||
default_idle();
|
||||
return fd;
|
||||
}
|
||||
|
||||
extern void start_kernel(void);
|
||||
@@ -158,14 +51,14 @@ static int __init start_kernel_proc(void *unused)
|
||||
cpu_online_map = cpumask_of_cpu(0);
|
||||
#endif
|
||||
start_kernel();
|
||||
return(0);
|
||||
return 0;
|
||||
}
|
||||
|
||||
extern int userspace_pid[];
|
||||
|
||||
extern char cpu0_irqstack[];
|
||||
|
||||
int __init start_uml_skas(void)
|
||||
int __init start_uml(void)
|
||||
{
|
||||
stack_protections((unsigned long) &cpu0_irqstack);
|
||||
set_sigstack(cpu0_irqstack, THREAD_SIZE);
|
||||
@@ -176,49 +69,14 @@ int __init start_uml_skas(void)
|
||||
|
||||
init_task.thread.request.u.thread.proc = start_kernel_proc;
|
||||
init_task.thread.request.u.thread.arg = NULL;
|
||||
return(start_idle_thread(task_stack_page(&init_task),
|
||||
&init_task.thread.mode.skas.switch_buf));
|
||||
}
|
||||
|
||||
int external_pid_skas(struct task_struct *task)
|
||||
{
|
||||
/* FIXME: Need to look up userspace_pid by cpu */
|
||||
return(userspace_pid[0]);
|
||||
}
|
||||
|
||||
int thread_pid_skas(struct task_struct *task)
|
||||
{
|
||||
/* FIXME: Need to look up userspace_pid by cpu */
|
||||
return(userspace_pid[0]);
|
||||
}
|
||||
|
||||
void kill_off_processes_skas(void)
|
||||
{
|
||||
if(proc_mm)
|
||||
/*
|
||||
* FIXME: need to loop over userspace_pids in
|
||||
* kill_off_processes_skas
|
||||
*/
|
||||
os_kill_ptraced_process(userspace_pid[0], 1);
|
||||
else {
|
||||
struct task_struct *p;
|
||||
int pid, me;
|
||||
|
||||
me = os_getpid();
|
||||
for_each_process(p){
|
||||
if(p->mm == NULL)
|
||||
continue;
|
||||
|
||||
pid = p->mm->context.skas.id.u.pid;
|
||||
os_kill_ptraced_process(pid, 1);
|
||||
}
|
||||
}
|
||||
return start_idle_thread(task_stack_page(&init_task),
|
||||
&init_task.thread.switch_buf);
|
||||
}
|
||||
|
||||
unsigned long current_stub_stack(void)
|
||||
{
|
||||
if(current->mm == NULL)
|
||||
return(0);
|
||||
return 0;
|
||||
|
||||
return(current->mm->context.skas.id.stack);
|
||||
return current->mm->context.skas.id.stack;
|
||||
}
|
||||
|
@@ -13,7 +13,7 @@
|
||||
#include "kern_util.h"
|
||||
#include "syscall.h"
|
||||
|
||||
void handle_syscall(union uml_pt_regs *r)
|
||||
void handle_syscall(struct uml_pt_regs *r)
|
||||
{
|
||||
struct pt_regs *regs = container_of(r, struct pt_regs, regs);
|
||||
long result;
|
||||
@@ -37,7 +37,7 @@ void handle_syscall(union uml_pt_regs *r)
|
||||
result = -ENOSYS;
|
||||
else result = EXECUTE_SYSCALL(syscall, regs);
|
||||
|
||||
REGS_SET_SYSCALL_RETURN(r->skas.regs, result);
|
||||
REGS_SET_SYSCALL_RETURN(r->regs, result);
|
||||
|
||||
syscall_trace(r, 1);
|
||||
}
|
||||
|
@@ -1,164 +0,0 @@
|
||||
/*
|
||||
* Copyright (C) 2002 Jeff Dike (jdike@karaya.com)
|
||||
* Copyright 2003 PathScale, Inc.
|
||||
* Licensed under the GPL
|
||||
*/
|
||||
|
||||
#include "linux/stddef.h"
|
||||
#include "linux/sched.h"
|
||||
#include "linux/mm.h"
|
||||
#include "asm/page.h"
|
||||
#include "asm/pgtable.h"
|
||||
#include "asm/mmu.h"
|
||||
#include "mem_user.h"
|
||||
#include "mem.h"
|
||||
#include "skas.h"
|
||||
#include "os.h"
|
||||
#include "tlb.h"
|
||||
|
||||
static int do_ops(union mm_context *mmu, struct host_vm_op *ops, int last,
|
||||
int finished, void **flush)
|
||||
{
|
||||
struct host_vm_op *op;
|
||||
int i, ret = 0;
|
||||
|
||||
for(i = 0; i <= last && !ret; i++){
|
||||
op = &ops[i];
|
||||
switch(op->type){
|
||||
case MMAP:
|
||||
ret = map(&mmu->skas.id, op->u.mmap.addr,
|
||||
op->u.mmap.len, op->u.mmap.prot,
|
||||
op->u.mmap.fd, op->u.mmap.offset, finished,
|
||||
flush);
|
||||
break;
|
||||
case MUNMAP:
|
||||
ret = unmap(&mmu->skas.id, op->u.munmap.addr,
|
||||
op->u.munmap.len, finished, flush);
|
||||
break;
|
||||
case MPROTECT:
|
||||
ret = protect(&mmu->skas.id, op->u.mprotect.addr,
|
||||
op->u.mprotect.len, op->u.mprotect.prot,
|
||||
finished, flush);
|
||||
break;
|
||||
default:
|
||||
printk("Unknown op type %d in do_ops\n", op->type);
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
extern int proc_mm;
|
||||
|
||||
static void fix_range(struct mm_struct *mm, unsigned long start_addr,
|
||||
unsigned long end_addr, int force)
|
||||
{
|
||||
if(!proc_mm && (end_addr > CONFIG_STUB_START))
|
||||
end_addr = CONFIG_STUB_START;
|
||||
|
||||
fix_range_common(mm, start_addr, end_addr, force, do_ops);
|
||||
}
|
||||
|
||||
void __flush_tlb_one_skas(unsigned long addr)
|
||||
{
|
||||
flush_tlb_kernel_range_common(addr, addr + PAGE_SIZE);
|
||||
}
|
||||
|
||||
void flush_tlb_range_skas(struct vm_area_struct *vma, unsigned long start,
|
||||
unsigned long end)
|
||||
{
|
||||
if(vma->vm_mm == NULL)
|
||||
flush_tlb_kernel_range_common(start, end);
|
||||
else fix_range(vma->vm_mm, start, end, 0);
|
||||
}
|
||||
|
||||
void flush_tlb_mm_skas(struct mm_struct *mm)
|
||||
{
|
||||
unsigned long end;
|
||||
|
||||
/* Don't bother flushing if this address space is about to be
|
||||
* destroyed.
|
||||
*/
|
||||
if(atomic_read(&mm->mm_users) == 0)
|
||||
return;
|
||||
|
||||
end = proc_mm ? task_size : CONFIG_STUB_START;
|
||||
fix_range(mm, 0, end, 0);
|
||||
}
|
||||
|
||||
void force_flush_all_skas(void)
|
||||
{
|
||||
struct mm_struct *mm = current->mm;
|
||||
struct vm_area_struct *vma = mm->mmap;
|
||||
|
||||
while(vma != NULL) {
|
||||
fix_range(mm, vma->vm_start, vma->vm_end, 1);
|
||||
vma = vma->vm_next;
|
||||
}
|
||||
}
|
||||
|
||||
void flush_tlb_page_skas(struct vm_area_struct *vma, unsigned long address)
|
||||
{
|
||||
pgd_t *pgd;
|
||||
pud_t *pud;
|
||||
pmd_t *pmd;
|
||||
pte_t *pte;
|
||||
struct mm_struct *mm = vma->vm_mm;
|
||||
void *flush = NULL;
|
||||
int r, w, x, prot, err = 0;
|
||||
struct mm_id *mm_id;
|
||||
|
||||
pgd = pgd_offset(mm, address);
|
||||
if(!pgd_present(*pgd))
|
||||
goto kill;
|
||||
|
||||
pud = pud_offset(pgd, address);
|
||||
if(!pud_present(*pud))
|
||||
goto kill;
|
||||
|
||||
pmd = pmd_offset(pud, address);
|
||||
if(!pmd_present(*pmd))
|
||||
goto kill;
|
||||
|
||||
pte = pte_offset_kernel(pmd, address);
|
||||
|
||||
r = pte_read(*pte);
|
||||
w = pte_write(*pte);
|
||||
x = pte_exec(*pte);
|
||||
if (!pte_young(*pte)) {
|
||||
r = 0;
|
||||
w = 0;
|
||||
} else if (!pte_dirty(*pte)) {
|
||||
w = 0;
|
||||
}
|
||||
|
||||
mm_id = &mm->context.skas.id;
|
||||
prot = ((r ? UM_PROT_READ : 0) | (w ? UM_PROT_WRITE : 0) |
|
||||
(x ? UM_PROT_EXEC : 0));
|
||||
if(pte_newpage(*pte)){
|
||||
if(pte_present(*pte)){
|
||||
unsigned long long offset;
|
||||
int fd;
|
||||
|
||||
fd = phys_mapping(pte_val(*pte) & PAGE_MASK, &offset);
|
||||
err = map(mm_id, address, PAGE_SIZE, prot, fd, offset,
|
||||
1, &flush);
|
||||
}
|
||||
else err = unmap(mm_id, address, PAGE_SIZE, 1, &flush);
|
||||
}
|
||||
else if(pte_newprot(*pte))
|
||||
err = protect(mm_id, address, PAGE_SIZE, prot, 1, &flush);
|
||||
|
||||
if(err)
|
||||
goto kill;
|
||||
|
||||
*pte = pte_mkuptodate(*pte);
|
||||
|
||||
return;
|
||||
|
||||
kill:
|
||||
printk("Failed to flush page for address 0x%lx\n", address);
|
||||
force_sig(SIGKILL, current);
|
||||
}
|
||||
|
Referens i nytt ärende
Block a user