2006-06-12 15:22:12 +00:00
|
|
|
#include "types.h"
|
2007-08-27 23:26:33 +00:00
|
|
|
#include "defs.h"
|
|
|
|
#include "param.h"
|
2006-06-12 15:22:12 +00:00
|
|
|
#include "mmu.h"
|
|
|
|
#include "x86.h"
|
2006-06-22 20:47:23 +00:00
|
|
|
#include "proc.h"
|
2006-07-12 01:48:35 +00:00
|
|
|
#include "spinlock.h"
|
|
|
|
|
2006-08-10 22:08:14 +00:00
|
|
|
struct spinlock proc_table_lock;
|
2006-06-12 15:22:12 +00:00
|
|
|
|
|
|
|
struct proc proc[NPROC];
|
2007-08-23 14:35:28 +00:00
|
|
|
static struct proc *initproc;
|
|
|
|
|
2007-08-22 06:01:32 +00:00
|
|
|
int nextpid = 1;
|
2006-07-16 01:15:28 +00:00
|
|
|
extern void forkret(void);
|
2006-07-17 01:58:13 +00:00
|
|
|
extern void forkret1(struct trapframe*);
|
2006-06-12 15:22:12 +00:00
|
|
|
|
2006-08-10 22:08:14 +00:00
|
|
|
void
|
|
|
|
pinit(void)
|
|
|
|
{
|
|
|
|
initlock(&proc_table_lock, "proc_table");
|
|
|
|
}
|
|
|
|
|
2007-08-22 06:01:32 +00:00
|
|
|
// Look in the process table for an UNUSED proc.
|
|
|
|
// If found, change state to EMBRYO and return it.
|
|
|
|
// Otherwise return 0.
|
|
|
|
static struct proc*
|
|
|
|
allocproc(void)
|
2006-06-12 15:22:12 +00:00
|
|
|
{
|
2007-08-22 06:01:32 +00:00
|
|
|
int i;
|
|
|
|
struct proc *p;
|
2006-06-12 15:22:12 +00:00
|
|
|
|
2007-08-22 06:01:32 +00:00
|
|
|
acquire(&proc_table_lock);
|
|
|
|
for(i = 0; i < NPROC; i++){
|
|
|
|
p = &proc[i];
|
|
|
|
if(p->state == UNUSED){
|
|
|
|
p->state = EMBRYO;
|
|
|
|
p->pid = nextpid++;
|
2009-05-31 00:28:45 +00:00
|
|
|
goto found;
|
2007-08-22 06:01:32 +00:00
|
|
|
}
|
2006-08-15 22:18:20 +00:00
|
|
|
}
|
2007-08-22 06:01:32 +00:00
|
|
|
release(&proc_table_lock);
|
|
|
|
return 0;
|
2009-05-31 00:28:45 +00:00
|
|
|
|
|
|
|
found:
|
|
|
|
release(&proc_table_lock);
|
|
|
|
|
|
|
|
// Allocate kernel stack if necessary.
|
|
|
|
if((p->kstack = kalloc(KSTACKSIZE)) == 0){
|
|
|
|
p->state = UNUSED;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
p->tf = (struct trapframe*)(p->kstack + KSTACKSIZE) - 1;
|
|
|
|
|
|
|
|
// Set up new context to start executing at forkret (see below).
|
|
|
|
p->context = (struct context *)p->tf - 1;
|
|
|
|
memset(p->context, 0, sizeof(*p->context));
|
|
|
|
p->context->eip = (uint)forkret;
|
|
|
|
return p;
|
2006-06-12 15:22:12 +00:00
|
|
|
}
|
|
|
|
|
2006-09-08 14:26:51 +00:00
|
|
|
// Grow current process's memory by n bytes.
|
2009-05-31 00:28:45 +00:00
|
|
|
// Return 0 on success, -1 on failure.
|
2006-09-08 14:26:51 +00:00
|
|
|
int
|
|
|
|
growproc(int n)
|
|
|
|
{
|
2008-08-28 17:57:47 +00:00
|
|
|
char *newmem;
|
2006-09-08 14:26:51 +00:00
|
|
|
|
|
|
|
newmem = kalloc(cp->sz + n);
|
|
|
|
if(newmem == 0)
|
2007-08-24 20:22:55 +00:00
|
|
|
return -1;
|
2006-09-08 14:26:51 +00:00
|
|
|
memmove(newmem, cp->mem, cp->sz);
|
|
|
|
memset(newmem + cp->sz, 0, n);
|
2008-08-28 17:57:47 +00:00
|
|
|
kfree(cp->mem, cp->sz);
|
2006-09-08 14:26:51 +00:00
|
|
|
cp->mem = newmem;
|
|
|
|
cp->sz += n;
|
2009-05-31 00:28:45 +00:00
|
|
|
usegment();
|
|
|
|
return 0;
|
2006-09-08 14:26:51 +00:00
|
|
|
}
|
|
|
|
|
2009-05-31 00:28:45 +00:00
|
|
|
// Set up CPU's kernel segment descriptors.
|
2007-08-22 06:01:32 +00:00
|
|
|
void
|
2009-05-31 00:28:45 +00:00
|
|
|
ksegment(void)
|
2006-07-16 01:47:40 +00:00
|
|
|
{
|
2009-05-31 00:28:45 +00:00
|
|
|
struct cpu *c1;
|
|
|
|
|
|
|
|
c1 = &cpus[cpu()];
|
|
|
|
c1->gdt[0] = SEG_NULL;
|
|
|
|
c1->gdt[SEG_KCODE] = SEG(STA_X|STA_R, 0, 0x100000 + 64*1024-1, 0);
|
|
|
|
c1->gdt[SEG_KDATA] = SEG(STA_W, 0, 0xffffffff, 0);
|
|
|
|
c1->gdt[SEG_KCPU] = SEG(STA_W, (uint)&c1->tls+sizeof(c1->tls), 0xffffffff, 0);
|
|
|
|
c1->gdt[SEG_UCODE] = SEG_NULL;
|
|
|
|
c1->gdt[SEG_UDATA] = SEG_NULL;
|
|
|
|
c1->gdt[SEG_TSS] = SEG_NULL;
|
|
|
|
lgdt(c1->gdt, sizeof(c1->gdt));
|
2009-05-31 01:12:08 +00:00
|
|
|
loadfsgs(SEG_KCPU << 3);
|
2007-08-24 20:28:08 +00:00
|
|
|
|
2009-05-31 00:28:45 +00:00
|
|
|
// Initialize cpu-local variables.
|
|
|
|
c = c1;
|
|
|
|
cp = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Set up CPU's segment descriptors and task state for the current process.
|
|
|
|
// If cp==0, set up for "idle" state for when scheduler() is running.
|
|
|
|
void
|
|
|
|
usegment(void)
|
|
|
|
{
|
2007-09-27 20:09:40 +00:00
|
|
|
pushcli();
|
2007-09-27 20:38:53 +00:00
|
|
|
c->ts.ss0 = SEG_KDATA << 3;
|
2009-05-31 00:28:45 +00:00
|
|
|
if(cp)
|
|
|
|
c->ts.esp0 = (uint)(cp->kstack + KSTACKSIZE);
|
2007-08-22 06:01:32 +00:00
|
|
|
else
|
|
|
|
c->ts.esp0 = 0xffffffff;
|
|
|
|
|
2009-05-31 00:28:45 +00:00
|
|
|
if(cp){
|
|
|
|
c->gdt[SEG_UCODE] = SEG(STA_X|STA_R, (uint)cp->mem, cp->sz-1, DPL_USER);
|
|
|
|
c->gdt[SEG_UDATA] = SEG(STA_W, (uint)cp->mem, cp->sz-1, DPL_USER);
|
2007-08-28 18:37:41 +00:00
|
|
|
} else {
|
2007-08-22 06:01:32 +00:00
|
|
|
c->gdt[SEG_UCODE] = SEG_NULL;
|
|
|
|
c->gdt[SEG_UDATA] = SEG_NULL;
|
2006-07-16 01:47:40 +00:00
|
|
|
}
|
2009-05-31 00:28:45 +00:00
|
|
|
c->gdt[SEG_TSS] = SEG16(STS_T32A, (uint)&c->ts, sizeof(c->ts)-1, 0);
|
|
|
|
c->gdt[SEG_TSS].s = 0;
|
2007-08-22 06:01:32 +00:00
|
|
|
|
|
|
|
lgdt(c->gdt, sizeof(c->gdt));
|
|
|
|
ltr(SEG_TSS << 3);
|
2007-09-27 20:09:40 +00:00
|
|
|
popcli();
|
2006-07-16 01:47:40 +00:00
|
|
|
}
|
2006-06-12 15:22:12 +00:00
|
|
|
|
2006-07-16 01:47:40 +00:00
|
|
|
// Create a new process copying p as the parent.
|
2007-08-22 06:01:32 +00:00
|
|
|
// Sets up stack to return as if from system call.
|
|
|
|
// Caller must set state of returned proc to RUNNABLE.
|
2009-05-31 00:38:51 +00:00
|
|
|
int
|
|
|
|
fork(void)
|
2006-06-12 15:22:12 +00:00
|
|
|
{
|
2009-05-31 00:38:51 +00:00
|
|
|
int i, pid;
|
2006-06-12 15:22:12 +00:00
|
|
|
struct proc *np;
|
|
|
|
|
2006-07-16 01:47:40 +00:00
|
|
|
// Allocate process.
|
2007-08-22 06:01:32 +00:00
|
|
|
if((np = allocproc()) == 0)
|
2009-05-31 00:38:51 +00:00
|
|
|
return -1;
|
2006-07-12 01:48:35 +00:00
|
|
|
|
2009-05-31 00:28:45 +00:00
|
|
|
// Copy process state from p.
|
2009-05-31 00:38:51 +00:00
|
|
|
np->sz = cp->sz;
|
2009-05-31 00:28:45 +00:00
|
|
|
if((np->mem = kalloc(np->sz)) == 0){
|
|
|
|
kfree(np->kstack, KSTACKSIZE);
|
|
|
|
np->kstack = 0;
|
2006-07-12 01:48:35 +00:00
|
|
|
np->state = UNUSED;
|
2009-05-31 00:38:51 +00:00
|
|
|
return -1;
|
2006-06-12 15:22:12 +00:00
|
|
|
}
|
2009-05-31 00:38:51 +00:00
|
|
|
memmove(np->mem, cp->mem, np->sz);
|
|
|
|
np->parent = cp;
|
|
|
|
*np->tf = *cp->tf;
|
2006-09-06 17:27:19 +00:00
|
|
|
|
2009-05-31 00:38:51 +00:00
|
|
|
// Clear %eax so that fork returns 0 in the child.
|
|
|
|
np->tf->eax = 0;
|
2007-08-21 19:22:08 +00:00
|
|
|
|
2009-05-31 00:38:51 +00:00
|
|
|
for(i = 0; i < NOFILE; i++)
|
|
|
|
if(cp->ofile[i])
|
|
|
|
np->ofile[i] = filedup(cp->ofile[i]);
|
|
|
|
np->cwd = idup(cp->cwd);
|
|
|
|
|
|
|
|
pid = np->pid;
|
|
|
|
np->state = RUNNABLE;
|
|
|
|
|
|
|
|
return pid;
|
2006-06-12 15:22:12 +00:00
|
|
|
}
|
|
|
|
|
2007-08-22 06:01:32 +00:00
|
|
|
// Set up first user process.
|
|
|
|
void
|
|
|
|
userinit(void)
|
|
|
|
{
|
|
|
|
struct proc *p;
|
|
|
|
extern uchar _binary_initcode_start[], _binary_initcode_size[];
|
|
|
|
|
2009-05-31 00:28:45 +00:00
|
|
|
p = allocproc();
|
|
|
|
initproc = p;
|
|
|
|
|
|
|
|
// Initialize memory from initcode.S
|
2007-08-22 06:01:32 +00:00
|
|
|
p->sz = PAGE;
|
|
|
|
p->mem = kalloc(p->sz);
|
2009-05-31 00:28:45 +00:00
|
|
|
memmove(p->mem, _binary_initcode_start, (int)_binary_initcode_size);
|
|
|
|
|
2007-08-22 17:45:52 +00:00
|
|
|
memset(p->tf, 0, sizeof(*p->tf));
|
2007-08-22 06:01:32 +00:00
|
|
|
p->tf->cs = (SEG_UCODE << 3) | DPL_USER;
|
2007-08-24 20:22:55 +00:00
|
|
|
p->tf->ds = (SEG_UDATA << 3) | DPL_USER;
|
|
|
|
p->tf->es = p->tf->ds;
|
|
|
|
p->tf->ss = p->tf->ds;
|
2007-08-22 06:01:32 +00:00
|
|
|
p->tf->eflags = FL_IF;
|
|
|
|
p->tf->esp = p->sz;
|
2009-05-31 00:28:45 +00:00
|
|
|
p->tf->eip = 0; // beginning of initcode.S
|
2007-08-22 06:01:32 +00:00
|
|
|
|
|
|
|
safestrcpy(p->name, "initcode", sizeof(p->name));
|
2009-05-31 00:28:45 +00:00
|
|
|
p->cwd = namei("/");
|
kernel SMP interruptibility fixes.
Last year, right before I sent xv6 to the printer, I changed the
SETGATE calls so that interrupts would be disabled on entry to
interrupt handlers, and I added the nlock++ / nlock-- in trap()
so that interrupts would stay disabled while the hw handlers
(but not the syscall handler) did their work. I did this because
the kernel was otherwise causing Bochs to triple-fault in SMP
mode, and time was short.
Robert observed yesterday that something was keeping the SMP
preemption user test from working. It turned out that when I
simplified the lapic code I swapped the order of two register
writes that I didn't realize were order dependent. I fixed that
and then since I had everything paged in kept going and tried
to figure out why you can't leave interrupts on during interrupt
handlers. There are a few issues.
First, there must be some way to keep interrupts from "stacking
up" and overflowing the stack. Keeping interrupts off the whole
time solves this problem -- even if the clock tick handler runs
long enough that the next clock tick is waiting when it finishes,
keeping interrupts off means that the handler runs all the way
through the "iret" before the next handler begins. This is not
really a problem unless you are putting too many prints in trap
-- if the OS is doing its job right, the handlers should run
quickly and not stack up.
Second, if xv6 had page faults, then it would be important to
keep interrupts disabled between the start of the interrupt and
the time that cr2 was read, to avoid a scenario like:
p1 page faults [cr2 set to faulting address]
p1 starts executing trapasm.S
clock interrupt, p1 preempted, p2 starts executing
p2 page faults [cr2 set to another faulting address]
p2 starts, finishes fault handler
p1 rescheduled, reads cr2, sees wrong fault address
Alternately p1 could be rescheduled on the other cpu, in which
case it would still see the wrong cr2. That said, I think cr2
is the only interrupt state that isn't pushed onto the interrupt
stack atomically at fault time, and xv6 doesn't care. (This isn't
entirely hypothetical -- I debugged this problem on Plan 9.)
Third, and this is the big one, it is not safe to call cpu()
unless interrupts are disabled. If interrupts are enabled then
there is no guarantee that, between the time cpu() looks up the
cpu id and the time that it the result gets used, the process
has not been rescheduled to the other cpu. For example, the
very commonly-used expression curproc[cpu()] (aka the macro cp)
can end up referring to the wrong proc: the code stores the
result of cpu() in %eax, gets rescheduled to the other cpu at
just the wrong instant, and then reads curproc[%eax].
We use curproc[cpu()] to get the current process a LOT. In that
particular case, if we arranged for the current curproc entry
to be addressed by %fs:0 and just use a different %fs on each
CPU, then we could safely get at curproc even with interrupts
disabled, since the read of %fs would be atomic with the read
of %fs:0. Alternately, we could have a curproc() function that
disables interrupts while computing curproc[cpu()]. I've done
that last one.
Even in the current kernel, with interrupts off on entry to trap,
interrupts are enabled inside release if there are no locks held.
Also, the scheduler's idle loop must be interruptible at times
so that the clock and disk interrupts (which might make processes
runnable) can be handled.
In addition to the rampant use of curproc[cpu()], this little
snippet from acquire is wrong on smp:
if(cpus[cpu()].nlock == 0)
cli();
cpus[cpu()].nlock++;
because if interrupts are off then we might call cpu(), get
rescheduled to a different cpu, look at cpus[oldcpu].nlock, and
wrongly decide not to disable interrupts on the new cpu. The
fix is to always call cli(). But this is wrong too:
if(holding(lock))
panic("acquire");
cli();
cpus[cpu()].nlock++;
because holding looks at cpu(). The fix is:
cli();
if(holding(lock))
panic("acquire");
cpus[cpu()].nlock++;
I've done that, and I changed cpu() to complain the first time
it gets called with interrupts disabled. (It gets called too
much to complain every time.)
I added new functions splhi and spllo that are like acquire and
release but without the locking:
void
splhi(void)
{
cli();
cpus[cpu()].nsplhi++;
}
void
spllo(void)
{
if(--cpus[cpu()].nsplhi == 0)
sti();
}
and I've used those to protect other sections of code that refer
to cpu() when interrupts would otherwise be disabled (basically
just curproc and setupsegs). I also use them in acquire/release
and got rid of nlock.
I'm not thrilled with the names, but I think the concept -- a
counted cli/sti -- is sound. Having them also replaces the
nlock++/nlock-- in trap.c and main.c, which is nice.
Final note: it's still not safe to enable interrupts in
the middle of trap() between lapic_eoi and returning
to user space. I don't understand why, but we get a
fault on pop %es because 0x10 is a bad segment
descriptor (!) and then the fault faults trying to go into
a new interrupt because 0x8 is a bad segment descriptor too!
Triple fault. I haven't debugged this yet.
2007-09-27 12:58:42 +00:00
|
|
|
|
2009-05-31 00:28:45 +00:00
|
|
|
p->state = RUNNABLE;
|
kernel SMP interruptibility fixes.
Last year, right before I sent xv6 to the printer, I changed the
SETGATE calls so that interrupts would be disabled on entry to
interrupt handlers, and I added the nlock++ / nlock-- in trap()
so that interrupts would stay disabled while the hw handlers
(but not the syscall handler) did their work. I did this because
the kernel was otherwise causing Bochs to triple-fault in SMP
mode, and time was short.
Robert observed yesterday that something was keeping the SMP
preemption user test from working. It turned out that when I
simplified the lapic code I swapped the order of two register
writes that I didn't realize were order dependent. I fixed that
and then since I had everything paged in kept going and tried
to figure out why you can't leave interrupts on during interrupt
handlers. There are a few issues.
First, there must be some way to keep interrupts from "stacking
up" and overflowing the stack. Keeping interrupts off the whole
time solves this problem -- even if the clock tick handler runs
long enough that the next clock tick is waiting when it finishes,
keeping interrupts off means that the handler runs all the way
through the "iret" before the next handler begins. This is not
really a problem unless you are putting too many prints in trap
-- if the OS is doing its job right, the handlers should run
quickly and not stack up.
Second, if xv6 had page faults, then it would be important to
keep interrupts disabled between the start of the interrupt and
the time that cr2 was read, to avoid a scenario like:
p1 page faults [cr2 set to faulting address]
p1 starts executing trapasm.S
clock interrupt, p1 preempted, p2 starts executing
p2 page faults [cr2 set to another faulting address]
p2 starts, finishes fault handler
p1 rescheduled, reads cr2, sees wrong fault address
Alternately p1 could be rescheduled on the other cpu, in which
case it would still see the wrong cr2. That said, I think cr2
is the only interrupt state that isn't pushed onto the interrupt
stack atomically at fault time, and xv6 doesn't care. (This isn't
entirely hypothetical -- I debugged this problem on Plan 9.)
Third, and this is the big one, it is not safe to call cpu()
unless interrupts are disabled. If interrupts are enabled then
there is no guarantee that, between the time cpu() looks up the
cpu id and the time that it the result gets used, the process
has not been rescheduled to the other cpu. For example, the
very commonly-used expression curproc[cpu()] (aka the macro cp)
can end up referring to the wrong proc: the code stores the
result of cpu() in %eax, gets rescheduled to the other cpu at
just the wrong instant, and then reads curproc[%eax].
We use curproc[cpu()] to get the current process a LOT. In that
particular case, if we arranged for the current curproc entry
to be addressed by %fs:0 and just use a different %fs on each
CPU, then we could safely get at curproc even with interrupts
disabled, since the read of %fs would be atomic with the read
of %fs:0. Alternately, we could have a curproc() function that
disables interrupts while computing curproc[cpu()]. I've done
that last one.
Even in the current kernel, with interrupts off on entry to trap,
interrupts are enabled inside release if there are no locks held.
Also, the scheduler's idle loop must be interruptible at times
so that the clock and disk interrupts (which might make processes
runnable) can be handled.
In addition to the rampant use of curproc[cpu()], this little
snippet from acquire is wrong on smp:
if(cpus[cpu()].nlock == 0)
cli();
cpus[cpu()].nlock++;
because if interrupts are off then we might call cpu(), get
rescheduled to a different cpu, look at cpus[oldcpu].nlock, and
wrongly decide not to disable interrupts on the new cpu. The
fix is to always call cli(). But this is wrong too:
if(holding(lock))
panic("acquire");
cli();
cpus[cpu()].nlock++;
because holding looks at cpu(). The fix is:
cli();
if(holding(lock))
panic("acquire");
cpus[cpu()].nlock++;
I've done that, and I changed cpu() to complain the first time
it gets called with interrupts disabled. (It gets called too
much to complain every time.)
I added new functions splhi and spllo that are like acquire and
release but without the locking:
void
splhi(void)
{
cli();
cpus[cpu()].nsplhi++;
}
void
spllo(void)
{
if(--cpus[cpu()].nsplhi == 0)
sti();
}
and I've used those to protect other sections of code that refer
to cpu() when interrupts would otherwise be disabled (basically
just curproc and setupsegs). I also use them in acquire/release
and got rid of nlock.
I'm not thrilled with the names, but I think the concept -- a
counted cli/sti -- is sound. Having them also replaces the
nlock++/nlock-- in trap.c and main.c, which is nice.
Final note: it's still not safe to enable interrupts in
the middle of trap() between lapic_eoi and returning
to user space. I don't understand why, but we get a
fault on pop %es because 0x10 is a bad segment
descriptor (!) and then the fault faults trying to go into
a new interrupt because 0x8 is a bad segment descriptor too!
Triple fault. I haven't debugged this yet.
2007-09-27 12:58:42 +00:00
|
|
|
}
|
|
|
|
|
2006-09-07 14:12:30 +00:00
|
|
|
//PAGEBREAK: 42
|
2006-09-06 17:27:19 +00:00
|
|
|
// Per-CPU process scheduler.
|
2006-07-16 01:15:28 +00:00
|
|
|
// Each CPU calls scheduler() after setting itself up.
|
|
|
|
// Scheduler never returns. It loops, doing:
|
|
|
|
// - choose a process to run
|
2007-08-30 17:39:56 +00:00
|
|
|
// - swtch to start running that process
|
|
|
|
// - eventually that process transfers control
|
|
|
|
// via swtch back to the scheduler.
|
2006-06-12 15:22:12 +00:00
|
|
|
void
|
Changes to allow use of native x86 ELF compilers, which on my
Linux 2.4 box using gcc 3.4.6 don't seem to follow the same
conventions as the i386-jos-elf-gcc compilers.
Can run make 'TOOLPREFIX=' or edit the Makefile.
curproc[cpu()] can now be NULL, indicating that no proc is running.
This seemed safer to me than having curproc[0] and curproc[1]
both pointing at proc[0] potentially.
The old implementation of swtch depended on the stack frame layout
used inside swtch being okay to return from on the other stack
(exactly the V6 you are not expected to understand this).
It also could be called in two contexts: at boot time, to schedule
the very first process, and later, on behalf of a process, to sleep
or schedule some other process.
I split this into two functions: scheduler and swtch.
The scheduler is now a separate never-returning function, invoked
by each cpu once set up. The scheduler looks like:
scheduler() {
setjmp(cpu.context);
pick proc to schedule
blah blah blah
longjmp(proc.context)
}
The new swtch is intended to be called only when curproc[cpu()] is not NULL,
that is, only on behalf of a user proc. It does:
swtch() {
if(setjmp(proc.context) == 0)
longjmp(cpu.context)
}
to save the current proc context and then jump over to the scheduler,
running on the cpu stack.
Similarly the system call stubs are now in assembly in usys.S to avoid
needing to know the details of stack frame layout used by the compiler.
Also various changes in the debugging prints.
2006-07-11 01:07:40 +00:00
|
|
|
scheduler(void)
|
2006-06-12 15:22:12 +00:00
|
|
|
{
|
2006-07-16 01:15:28 +00:00
|
|
|
struct proc *p;
|
2006-07-01 21:26:01 +00:00
|
|
|
int i;
|
|
|
|
|
2006-07-16 01:15:28 +00:00
|
|
|
for(;;){
|
2008-10-15 05:01:39 +00:00
|
|
|
// Enable interrupts on this processor, in lieu of saving intena.
|
2007-09-27 21:25:37 +00:00
|
|
|
sti();
|
|
|
|
|
2006-07-16 01:15:28 +00:00
|
|
|
// Loop over process table looking for process to run.
|
|
|
|
acquire(&proc_table_lock);
|
2006-07-01 21:26:01 +00:00
|
|
|
for(i = 0; i < NPROC; i++){
|
2006-07-16 01:15:28 +00:00
|
|
|
p = &proc[i];
|
|
|
|
if(p->state != RUNNABLE)
|
|
|
|
continue;
|
2006-09-06 17:27:19 +00:00
|
|
|
|
|
|
|
// Switch to chosen process. It is the process's job
|
2006-07-16 01:15:28 +00:00
|
|
|
// to release proc_table_lock and then reacquire it
|
|
|
|
// before jumping back to us.
|
2009-05-31 00:28:45 +00:00
|
|
|
cp = p;
|
|
|
|
usegment();
|
2006-07-16 01:15:28 +00:00
|
|
|
p->state = RUNNING;
|
kernel SMP interruptibility fixes.
Last year, right before I sent xv6 to the printer, I changed the
SETGATE calls so that interrupts would be disabled on entry to
interrupt handlers, and I added the nlock++ / nlock-- in trap()
so that interrupts would stay disabled while the hw handlers
(but not the syscall handler) did their work. I did this because
the kernel was otherwise causing Bochs to triple-fault in SMP
mode, and time was short.
Robert observed yesterday that something was keeping the SMP
preemption user test from working. It turned out that when I
simplified the lapic code I swapped the order of two register
writes that I didn't realize were order dependent. I fixed that
and then since I had everything paged in kept going and tried
to figure out why you can't leave interrupts on during interrupt
handlers. There are a few issues.
First, there must be some way to keep interrupts from "stacking
up" and overflowing the stack. Keeping interrupts off the whole
time solves this problem -- even if the clock tick handler runs
long enough that the next clock tick is waiting when it finishes,
keeping interrupts off means that the handler runs all the way
through the "iret" before the next handler begins. This is not
really a problem unless you are putting too many prints in trap
-- if the OS is doing its job right, the handlers should run
quickly and not stack up.
Second, if xv6 had page faults, then it would be important to
keep interrupts disabled between the start of the interrupt and
the time that cr2 was read, to avoid a scenario like:
p1 page faults [cr2 set to faulting address]
p1 starts executing trapasm.S
clock interrupt, p1 preempted, p2 starts executing
p2 page faults [cr2 set to another faulting address]
p2 starts, finishes fault handler
p1 rescheduled, reads cr2, sees wrong fault address
Alternately p1 could be rescheduled on the other cpu, in which
case it would still see the wrong cr2. That said, I think cr2
is the only interrupt state that isn't pushed onto the interrupt
stack atomically at fault time, and xv6 doesn't care. (This isn't
entirely hypothetical -- I debugged this problem on Plan 9.)
Third, and this is the big one, it is not safe to call cpu()
unless interrupts are disabled. If interrupts are enabled then
there is no guarantee that, between the time cpu() looks up the
cpu id and the time that it the result gets used, the process
has not been rescheduled to the other cpu. For example, the
very commonly-used expression curproc[cpu()] (aka the macro cp)
can end up referring to the wrong proc: the code stores the
result of cpu() in %eax, gets rescheduled to the other cpu at
just the wrong instant, and then reads curproc[%eax].
We use curproc[cpu()] to get the current process a LOT. In that
particular case, if we arranged for the current curproc entry
to be addressed by %fs:0 and just use a different %fs on each
CPU, then we could safely get at curproc even with interrupts
disabled, since the read of %fs would be atomic with the read
of %fs:0. Alternately, we could have a curproc() function that
disables interrupts while computing curproc[cpu()]. I've done
that last one.
Even in the current kernel, with interrupts off on entry to trap,
interrupts are enabled inside release if there are no locks held.
Also, the scheduler's idle loop must be interruptible at times
so that the clock and disk interrupts (which might make processes
runnable) can be handled.
In addition to the rampant use of curproc[cpu()], this little
snippet from acquire is wrong on smp:
if(cpus[cpu()].nlock == 0)
cli();
cpus[cpu()].nlock++;
because if interrupts are off then we might call cpu(), get
rescheduled to a different cpu, look at cpus[oldcpu].nlock, and
wrongly decide not to disable interrupts on the new cpu. The
fix is to always call cli(). But this is wrong too:
if(holding(lock))
panic("acquire");
cli();
cpus[cpu()].nlock++;
because holding looks at cpu(). The fix is:
cli();
if(holding(lock))
panic("acquire");
cpus[cpu()].nlock++;
I've done that, and I changed cpu() to complain the first time
it gets called with interrupts disabled. (It gets called too
much to complain every time.)
I added new functions splhi and spllo that are like acquire and
release but without the locking:
void
splhi(void)
{
cli();
cpus[cpu()].nsplhi++;
}
void
spllo(void)
{
if(--cpus[cpu()].nsplhi == 0)
sti();
}
and I've used those to protect other sections of code that refer
to cpu() when interrupts would otherwise be disabled (basically
just curproc and setupsegs). I also use them in acquire/release
and got rid of nlock.
I'm not thrilled with the names, but I think the concept -- a
counted cli/sti -- is sound. Having them also replaces the
nlock++/nlock-- in trap.c and main.c, which is nice.
Final note: it's still not safe to enable interrupts in
the middle of trap() between lapic_eoi and returning
to user space. I don't understand why, but we get a
fault on pop %es because 0x10 is a bad segment
descriptor (!) and then the fault faults trying to go into
a new interrupt because 0x8 is a bad segment descriptor too!
Triple fault. I haven't debugged this yet.
2007-09-27 12:58:42 +00:00
|
|
|
swtch(&c->context, &p->context);
|
2006-09-06 17:27:19 +00:00
|
|
|
|
|
|
|
// Process is done running for now.
|
2006-07-16 01:15:28 +00:00
|
|
|
// It should have changed its p->state before coming back.
|
2009-05-31 00:28:45 +00:00
|
|
|
cp = 0;
|
|
|
|
usegment();
|
2006-07-12 01:48:35 +00:00
|
|
|
}
|
2006-07-16 01:15:28 +00:00
|
|
|
release(&proc_table_lock);
|
2007-09-27 21:25:37 +00:00
|
|
|
|
2006-06-12 15:22:12 +00:00
|
|
|
}
|
2006-07-16 01:15:28 +00:00
|
|
|
}
|
2006-06-12 15:22:12 +00:00
|
|
|
|
2006-07-16 01:15:28 +00:00
|
|
|
// Enter scheduler. Must already hold proc_table_lock
|
2009-05-31 00:28:45 +00:00
|
|
|
// and have changed cp->state.
|
2006-07-16 01:15:28 +00:00
|
|
|
void
|
|
|
|
sched(void)
|
|
|
|
{
|
2008-10-15 05:01:39 +00:00
|
|
|
int intena;
|
|
|
|
|
2009-03-08 22:07:13 +00:00
|
|
|
if(readeflags()&FL_IF)
|
kernel SMP interruptibility fixes.
Last year, right before I sent xv6 to the printer, I changed the
SETGATE calls so that interrupts would be disabled on entry to
interrupt handlers, and I added the nlock++ / nlock-- in trap()
so that interrupts would stay disabled while the hw handlers
(but not the syscall handler) did their work. I did this because
the kernel was otherwise causing Bochs to triple-fault in SMP
mode, and time was short.
Robert observed yesterday that something was keeping the SMP
preemption user test from working. It turned out that when I
simplified the lapic code I swapped the order of two register
writes that I didn't realize were order dependent. I fixed that
and then since I had everything paged in kept going and tried
to figure out why you can't leave interrupts on during interrupt
handlers. There are a few issues.
First, there must be some way to keep interrupts from "stacking
up" and overflowing the stack. Keeping interrupts off the whole
time solves this problem -- even if the clock tick handler runs
long enough that the next clock tick is waiting when it finishes,
keeping interrupts off means that the handler runs all the way
through the "iret" before the next handler begins. This is not
really a problem unless you are putting too many prints in trap
-- if the OS is doing its job right, the handlers should run
quickly and not stack up.
Second, if xv6 had page faults, then it would be important to
keep interrupts disabled between the start of the interrupt and
the time that cr2 was read, to avoid a scenario like:
p1 page faults [cr2 set to faulting address]
p1 starts executing trapasm.S
clock interrupt, p1 preempted, p2 starts executing
p2 page faults [cr2 set to another faulting address]
p2 starts, finishes fault handler
p1 rescheduled, reads cr2, sees wrong fault address
Alternately p1 could be rescheduled on the other cpu, in which
case it would still see the wrong cr2. That said, I think cr2
is the only interrupt state that isn't pushed onto the interrupt
stack atomically at fault time, and xv6 doesn't care. (This isn't
entirely hypothetical -- I debugged this problem on Plan 9.)
Third, and this is the big one, it is not safe to call cpu()
unless interrupts are disabled. If interrupts are enabled then
there is no guarantee that, between the time cpu() looks up the
cpu id and the time that it the result gets used, the process
has not been rescheduled to the other cpu. For example, the
very commonly-used expression curproc[cpu()] (aka the macro cp)
can end up referring to the wrong proc: the code stores the
result of cpu() in %eax, gets rescheduled to the other cpu at
just the wrong instant, and then reads curproc[%eax].
We use curproc[cpu()] to get the current process a LOT. In that
particular case, if we arranged for the current curproc entry
to be addressed by %fs:0 and just use a different %fs on each
CPU, then we could safely get at curproc even with interrupts
disabled, since the read of %fs would be atomic with the read
of %fs:0. Alternately, we could have a curproc() function that
disables interrupts while computing curproc[cpu()]. I've done
that last one.
Even in the current kernel, with interrupts off on entry to trap,
interrupts are enabled inside release if there are no locks held.
Also, the scheduler's idle loop must be interruptible at times
so that the clock and disk interrupts (which might make processes
runnable) can be handled.
In addition to the rampant use of curproc[cpu()], this little
snippet from acquire is wrong on smp:
if(cpus[cpu()].nlock == 0)
cli();
cpus[cpu()].nlock++;
because if interrupts are off then we might call cpu(), get
rescheduled to a different cpu, look at cpus[oldcpu].nlock, and
wrongly decide not to disable interrupts on the new cpu. The
fix is to always call cli(). But this is wrong too:
if(holding(lock))
panic("acquire");
cli();
cpus[cpu()].nlock++;
because holding looks at cpu(). The fix is:
cli();
if(holding(lock))
panic("acquire");
cpus[cpu()].nlock++;
I've done that, and I changed cpu() to complain the first time
it gets called with interrupts disabled. (It gets called too
much to complain every time.)
I added new functions splhi and spllo that are like acquire and
release but without the locking:
void
splhi(void)
{
cli();
cpus[cpu()].nsplhi++;
}
void
spllo(void)
{
if(--cpus[cpu()].nsplhi == 0)
sti();
}
and I've used those to protect other sections of code that refer
to cpu() when interrupts would otherwise be disabled (basically
just curproc and setupsegs). I also use them in acquire/release
and got rid of nlock.
I'm not thrilled with the names, but I think the concept -- a
counted cli/sti -- is sound. Having them also replaces the
nlock++/nlock-- in trap.c and main.c, which is nice.
Final note: it's still not safe to enable interrupts in
the middle of trap() between lapic_eoi and returning
to user space. I don't understand why, but we get a
fault on pop %es because 0x10 is a bad segment
descriptor (!) and then the fault faults trying to go into
a new interrupt because 0x8 is a bad segment descriptor too!
Triple fault. I haven't debugged this yet.
2007-09-27 12:58:42 +00:00
|
|
|
panic("sched interruptible");
|
2007-08-09 17:32:40 +00:00
|
|
|
if(cp->state == RUNNING)
|
2007-08-08 08:57:37 +00:00
|
|
|
panic("sched running");
|
2006-09-07 16:54:00 +00:00
|
|
|
if(!holding(&proc_table_lock))
|
2007-08-08 08:57:37 +00:00
|
|
|
panic("sched proc_table_lock");
|
2009-05-31 00:28:45 +00:00
|
|
|
if(c->ncli != 1)
|
2006-09-07 16:54:00 +00:00
|
|
|
panic("sched locks");
|
|
|
|
|
2009-05-31 00:28:45 +00:00
|
|
|
intena = c->intena;
|
|
|
|
swtch(&cp->context, &c->context);
|
|
|
|
c->intena = intena;
|
Changes to allow use of native x86 ELF compilers, which on my
Linux 2.4 box using gcc 3.4.6 don't seem to follow the same
conventions as the i386-jos-elf-gcc compilers.
Can run make 'TOOLPREFIX=' or edit the Makefile.
curproc[cpu()] can now be NULL, indicating that no proc is running.
This seemed safer to me than having curproc[0] and curproc[1]
both pointing at proc[0] potentially.
The old implementation of swtch depended on the stack frame layout
used inside swtch being okay to return from on the other stack
(exactly the V6 you are not expected to understand this).
It also could be called in two contexts: at boot time, to schedule
the very first process, and later, on behalf of a process, to sleep
or schedule some other process.
I split this into two functions: scheduler and swtch.
The scheduler is now a separate never-returning function, invoked
by each cpu once set up. The scheduler looks like:
scheduler() {
setjmp(cpu.context);
pick proc to schedule
blah blah blah
longjmp(proc.context)
}
The new swtch is intended to be called only when curproc[cpu()] is not NULL,
that is, only on behalf of a user proc. It does:
swtch() {
if(setjmp(proc.context) == 0)
longjmp(cpu.context)
}
to save the current proc context and then jump over to the scheduler,
running on the cpu stack.
Similarly the system call stubs are now in assembly in usys.S to avoid
needing to know the details of stack frame layout used by the compiler.
Also various changes in the debugging prints.
2006-07-11 01:07:40 +00:00
|
|
|
}
|
|
|
|
|
2006-07-16 01:15:28 +00:00
|
|
|
// Give up the CPU for one scheduling round.
|
Changes to allow use of native x86 ELF compilers, which on my
Linux 2.4 box using gcc 3.4.6 don't seem to follow the same
conventions as the i386-jos-elf-gcc compilers.
Can run make 'TOOLPREFIX=' or edit the Makefile.
curproc[cpu()] can now be NULL, indicating that no proc is running.
This seemed safer to me than having curproc[0] and curproc[1]
both pointing at proc[0] potentially.
The old implementation of swtch depended on the stack frame layout
used inside swtch being okay to return from on the other stack
(exactly the V6 you are not expected to understand this).
It also could be called in two contexts: at boot time, to schedule
the very first process, and later, on behalf of a process, to sleep
or schedule some other process.
I split this into two functions: scheduler and swtch.
The scheduler is now a separate never-returning function, invoked
by each cpu once set up. The scheduler looks like:
scheduler() {
setjmp(cpu.context);
pick proc to schedule
blah blah blah
longjmp(proc.context)
}
The new swtch is intended to be called only when curproc[cpu()] is not NULL,
that is, only on behalf of a user proc. It does:
swtch() {
if(setjmp(proc.context) == 0)
longjmp(cpu.context)
}
to save the current proc context and then jump over to the scheduler,
running on the cpu stack.
Similarly the system call stubs are now in assembly in usys.S to avoid
needing to know the details of stack frame layout used by the compiler.
Also various changes in the debugging prints.
2006-07-11 01:07:40 +00:00
|
|
|
void
|
2006-07-16 01:47:40 +00:00
|
|
|
yield(void)
|
Changes to allow use of native x86 ELF compilers, which on my
Linux 2.4 box using gcc 3.4.6 don't seem to follow the same
conventions as the i386-jos-elf-gcc compilers.
Can run make 'TOOLPREFIX=' or edit the Makefile.
curproc[cpu()] can now be NULL, indicating that no proc is running.
This seemed safer to me than having curproc[0] and curproc[1]
both pointing at proc[0] potentially.
The old implementation of swtch depended on the stack frame layout
used inside swtch being okay to return from on the other stack
(exactly the V6 you are not expected to understand this).
It also could be called in two contexts: at boot time, to schedule
the very first process, and later, on behalf of a process, to sleep
or schedule some other process.
I split this into two functions: scheduler and swtch.
The scheduler is now a separate never-returning function, invoked
by each cpu once set up. The scheduler looks like:
scheduler() {
setjmp(cpu.context);
pick proc to schedule
blah blah blah
longjmp(proc.context)
}
The new swtch is intended to be called only when curproc[cpu()] is not NULL,
that is, only on behalf of a user proc. It does:
swtch() {
if(setjmp(proc.context) == 0)
longjmp(cpu.context)
}
to save the current proc context and then jump over to the scheduler,
running on the cpu stack.
Similarly the system call stubs are now in assembly in usys.S to avoid
needing to know the details of stack frame layout used by the compiler.
Also various changes in the debugging prints.
2006-07-11 01:07:40 +00:00
|
|
|
{
|
2006-07-16 01:15:28 +00:00
|
|
|
acquire(&proc_table_lock);
|
2007-08-09 17:32:40 +00:00
|
|
|
cp->state = RUNNABLE;
|
2006-07-16 01:15:28 +00:00
|
|
|
sched();
|
|
|
|
release(&proc_table_lock);
|
2006-06-12 15:22:12 +00:00
|
|
|
}
|
2006-06-15 19:58:01 +00:00
|
|
|
|
2006-08-29 21:35:30 +00:00
|
|
|
// A fork child's very first scheduling by scheduler()
|
2007-08-30 17:39:56 +00:00
|
|
|
// will swtch here. "Return" to user space.
|
2006-07-16 01:47:40 +00:00
|
|
|
void
|
|
|
|
forkret(void)
|
|
|
|
{
|
|
|
|
// Still holding proc_table_lock from scheduler.
|
|
|
|
release(&proc_table_lock);
|
2006-09-06 17:27:19 +00:00
|
|
|
|
2006-07-16 01:47:40 +00:00
|
|
|
// Jump into assembly, never to return.
|
2007-08-10 16:37:27 +00:00
|
|
|
forkret1(cp->tf);
|
2006-07-16 01:47:40 +00:00
|
|
|
}
|
|
|
|
|
2006-07-16 01:15:28 +00:00
|
|
|
// Atomically release lock and sleep on chan.
|
|
|
|
// Reacquires lock when reawakened.
|
2006-06-15 19:58:01 +00:00
|
|
|
void
|
2006-07-16 01:15:28 +00:00
|
|
|
sleep(void *chan, struct spinlock *lk)
|
2006-06-15 19:58:01 +00:00
|
|
|
{
|
2007-08-09 17:32:40 +00:00
|
|
|
if(cp == 0)
|
Changes to allow use of native x86 ELF compilers, which on my
Linux 2.4 box using gcc 3.4.6 don't seem to follow the same
conventions as the i386-jos-elf-gcc compilers.
Can run make 'TOOLPREFIX=' or edit the Makefile.
curproc[cpu()] can now be NULL, indicating that no proc is running.
This seemed safer to me than having curproc[0] and curproc[1]
both pointing at proc[0] potentially.
The old implementation of swtch depended on the stack frame layout
used inside swtch being okay to return from on the other stack
(exactly the V6 you are not expected to understand this).
It also could be called in two contexts: at boot time, to schedule
the very first process, and later, on behalf of a process, to sleep
or schedule some other process.
I split this into two functions: scheduler and swtch.
The scheduler is now a separate never-returning function, invoked
by each cpu once set up. The scheduler looks like:
scheduler() {
setjmp(cpu.context);
pick proc to schedule
blah blah blah
longjmp(proc.context)
}
The new swtch is intended to be called only when curproc[cpu()] is not NULL,
that is, only on behalf of a user proc. It does:
swtch() {
if(setjmp(proc.context) == 0)
longjmp(cpu.context)
}
to save the current proc context and then jump over to the scheduler,
running on the cpu stack.
Similarly the system call stubs are now in assembly in usys.S to avoid
needing to know the details of stack frame layout used by the compiler.
Also various changes in the debugging prints.
2006-07-11 01:07:40 +00:00
|
|
|
panic("sleep");
|
2006-07-15 12:03:57 +00:00
|
|
|
|
2006-07-17 05:00:25 +00:00
|
|
|
if(lk == 0)
|
|
|
|
panic("sleep without lk");
|
|
|
|
|
2006-09-06 17:27:19 +00:00
|
|
|
// Must acquire proc_table_lock in order to
|
2006-07-16 01:15:28 +00:00
|
|
|
// change p->state and then call sched.
|
2006-09-06 17:27:19 +00:00
|
|
|
// Once we hold proc_table_lock, we can be
|
2006-07-16 01:15:28 +00:00
|
|
|
// guaranteed that we won't miss any wakeup
|
|
|
|
// (wakeup runs with proc_table_lock locked),
|
|
|
|
// so it's okay to release lk.
|
|
|
|
if(lk != &proc_table_lock){
|
|
|
|
acquire(&proc_table_lock);
|
|
|
|
release(lk);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Go to sleep.
|
2007-08-09 17:32:40 +00:00
|
|
|
cp->chan = chan;
|
|
|
|
cp->state = SLEEPING;
|
2006-07-16 01:15:28 +00:00
|
|
|
sched();
|
2006-07-15 12:03:57 +00:00
|
|
|
|
2006-07-16 01:15:28 +00:00
|
|
|
// Tidy up.
|
2007-08-09 17:32:40 +00:00
|
|
|
cp->chan = 0;
|
2006-07-16 01:15:28 +00:00
|
|
|
|
|
|
|
// Reacquire original lock.
|
|
|
|
if(lk != &proc_table_lock){
|
|
|
|
release(&proc_table_lock);
|
|
|
|
acquire(lk);
|
|
|
|
}
|
2006-06-15 19:58:01 +00:00
|
|
|
}
|
|
|
|
|
2007-08-22 06:01:32 +00:00
|
|
|
//PAGEBREAK!
|
2006-07-16 01:15:28 +00:00
|
|
|
// Wake up all processes sleeping on chan.
|
|
|
|
// Proc_table_lock must be held.
|
2007-08-24 20:22:55 +00:00
|
|
|
static void
|
2006-07-15 12:03:57 +00:00
|
|
|
wakeup1(void *chan)
|
2006-06-15 19:58:01 +00:00
|
|
|
{
|
|
|
|
struct proc *p;
|
|
|
|
|
2006-07-15 12:03:57 +00:00
|
|
|
for(p = proc; p < &proc[NPROC]; p++)
|
2006-07-16 01:15:28 +00:00
|
|
|
if(p->state == SLEEPING && p->chan == chan)
|
2006-06-15 19:58:01 +00:00
|
|
|
p->state = RUNNABLE;
|
2006-07-15 12:03:57 +00:00
|
|
|
}
|
|
|
|
|
2006-07-16 01:15:28 +00:00
|
|
|
// Wake up all processes sleeping on chan.
|
2006-07-15 12:03:57 +00:00
|
|
|
void
|
|
|
|
wakeup(void *chan)
|
|
|
|
{
|
|
|
|
acquire(&proc_table_lock);
|
|
|
|
wakeup1(chan);
|
2006-07-12 01:48:35 +00:00
|
|
|
release(&proc_table_lock);
|
2006-06-15 19:58:01 +00:00
|
|
|
}
|
2006-07-11 17:39:45 +00:00
|
|
|
|
2006-07-16 01:15:28 +00:00
|
|
|
// Kill the process with the given pid.
|
|
|
|
// Process won't actually exit until it returns
|
|
|
|
// to user space (see trap in trap.c).
|
|
|
|
int
|
2007-08-28 19:14:43 +00:00
|
|
|
kill(int pid)
|
2006-07-11 17:39:45 +00:00
|
|
|
{
|
2006-07-16 01:15:28 +00:00
|
|
|
struct proc *p;
|
|
|
|
|
|
|
|
acquire(&proc_table_lock);
|
|
|
|
for(p = proc; p < &proc[NPROC]; p++){
|
|
|
|
if(p->pid == pid){
|
|
|
|
p->killed = 1;
|
|
|
|
// Wake process from sleep if necessary.
|
|
|
|
if(p->state == SLEEPING)
|
|
|
|
p->state = RUNNABLE;
|
|
|
|
release(&proc_table_lock);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
release(&proc_table_lock);
|
|
|
|
return -1;
|
2006-07-11 17:39:45 +00:00
|
|
|
}
|
|
|
|
|
2006-07-16 01:15:28 +00:00
|
|
|
// Exit the current process. Does not return.
|
2006-09-06 17:27:19 +00:00
|
|
|
// Exited processes remain in the zombie state
|
2006-07-16 01:15:28 +00:00
|
|
|
// until their parent calls wait() to find out they exited.
|
2006-07-11 17:39:45 +00:00
|
|
|
void
|
2007-08-28 19:14:43 +00:00
|
|
|
exit(void)
|
2006-07-11 17:39:45 +00:00
|
|
|
{
|
|
|
|
struct proc *p;
|
|
|
|
int fd;
|
|
|
|
|
2007-08-23 14:35:28 +00:00
|
|
|
if(cp == initproc)
|
2007-08-08 08:57:37 +00:00
|
|
|
panic("init exiting");
|
|
|
|
|
2006-07-16 01:15:28 +00:00
|
|
|
// Close all open files.
|
2006-07-11 17:39:45 +00:00
|
|
|
for(fd = 0; fd < NOFILE; fd++){
|
2006-09-06 18:38:56 +00:00
|
|
|
if(cp->ofile[fd]){
|
2006-09-06 18:43:45 +00:00
|
|
|
fileclose(cp->ofile[fd]);
|
2006-09-06 18:38:56 +00:00
|
|
|
cp->ofile[fd] = 0;
|
2006-07-11 17:39:45 +00:00
|
|
|
}
|
|
|
|
}
|
2006-09-06 17:27:19 +00:00
|
|
|
|
2007-08-22 06:01:32 +00:00
|
|
|
iput(cp->cwd);
|
2006-08-29 19:59:52 +00:00
|
|
|
cp->cwd = 0;
|
2006-07-11 17:39:45 +00:00
|
|
|
|
2006-07-12 01:48:35 +00:00
|
|
|
acquire(&proc_table_lock);
|
2006-07-11 17:39:45 +00:00
|
|
|
|
2007-10-20 18:25:38 +00:00
|
|
|
// Parent might be sleeping in wait().
|
2007-08-23 14:40:30 +00:00
|
|
|
wakeup1(cp->parent);
|
2006-07-11 17:39:45 +00:00
|
|
|
|
2007-08-23 14:35:28 +00:00
|
|
|
// Pass abandoned children to init.
|
|
|
|
for(p = proc; p < &proc[NPROC]; p++){
|
2007-08-23 14:40:30 +00:00
|
|
|
if(p->parent == cp){
|
|
|
|
p->parent = initproc;
|
2007-08-23 14:35:28 +00:00
|
|
|
if(p->state == ZOMBIE)
|
|
|
|
wakeup1(initproc);
|
2007-08-08 08:57:37 +00:00
|
|
|
}
|
2007-08-23 14:35:28 +00:00
|
|
|
}
|
2006-09-06 17:27:19 +00:00
|
|
|
|
2006-07-16 01:15:28 +00:00
|
|
|
// Jump into the scheduler, never to return.
|
2006-08-29 19:59:52 +00:00
|
|
|
cp->killed = 0;
|
2006-07-16 01:15:28 +00:00
|
|
|
cp->state = ZOMBIE;
|
|
|
|
sched();
|
|
|
|
panic("zombie exit");
|
2006-07-11 17:39:45 +00:00
|
|
|
}
|
2006-07-12 11:15:38 +00:00
|
|
|
|
2006-07-16 01:15:28 +00:00
|
|
|
// Wait for a child process to exit and return its pid.
|
|
|
|
// Return -1 if this process has no children.
|
2006-07-15 17:24:54 +00:00
|
|
|
int
|
2007-08-28 19:14:43 +00:00
|
|
|
wait(void)
|
2006-07-15 17:24:54 +00:00
|
|
|
{
|
|
|
|
struct proc *p;
|
2006-07-16 01:15:28 +00:00
|
|
|
int i, havekids, pid;
|
2006-07-15 17:24:54 +00:00
|
|
|
|
|
|
|
acquire(&proc_table_lock);
|
2006-07-16 01:15:28 +00:00
|
|
|
for(;;){
|
2006-08-29 21:35:30 +00:00
|
|
|
// Scan through table looking for zombie children.
|
2006-07-16 01:15:28 +00:00
|
|
|
havekids = 0;
|
|
|
|
for(i = 0; i < NPROC; i++){
|
|
|
|
p = &proc[i];
|
2006-09-07 01:56:22 +00:00
|
|
|
if(p->state == UNUSED)
|
|
|
|
continue;
|
2007-08-23 14:40:30 +00:00
|
|
|
if(p->parent == cp){
|
2009-05-31 00:28:45 +00:00
|
|
|
havekids = 1;
|
2006-07-16 01:15:28 +00:00
|
|
|
if(p->state == ZOMBIE){
|
|
|
|
// Found one.
|
|
|
|
kfree(p->mem, p->sz);
|
|
|
|
kfree(p->kstack, KSTACKSIZE);
|
|
|
|
pid = p->pid;
|
|
|
|
p->state = UNUSED;
|
|
|
|
p->pid = 0;
|
2007-08-23 14:40:30 +00:00
|
|
|
p->parent = 0;
|
2007-08-08 08:38:38 +00:00
|
|
|
p->name[0] = 0;
|
2006-07-16 01:15:28 +00:00
|
|
|
release(&proc_table_lock);
|
|
|
|
return pid;
|
|
|
|
}
|
2006-07-15 17:24:54 +00:00
|
|
|
}
|
|
|
|
}
|
2006-09-06 15:32:21 +00:00
|
|
|
|
2006-07-16 01:15:28 +00:00
|
|
|
// No point waiting if we don't have any children.
|
2007-08-08 10:29:42 +00:00
|
|
|
if(!havekids || cp->killed){
|
2006-07-15 17:24:54 +00:00
|
|
|
release(&proc_table_lock);
|
|
|
|
return -1;
|
|
|
|
}
|
2006-09-06 17:27:19 +00:00
|
|
|
|
2006-07-16 01:15:28 +00:00
|
|
|
// Wait for children to exit. (See wakeup1 call in proc_exit.)
|
2006-07-15 17:24:54 +00:00
|
|
|
sleep(cp, &proc_table_lock);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2006-09-07 15:45:38 +00:00
|
|
|
// Print a process listing to console. For debugging.
|
|
|
|
// Runs when user types ^P on console.
|
|
|
|
// No lock to avoid wedging a stuck machine further.
|
|
|
|
void
|
|
|
|
procdump(void)
|
|
|
|
{
|
2007-08-08 08:38:38 +00:00
|
|
|
static char *states[] = {
|
2007-08-08 09:42:36 +00:00
|
|
|
[UNUSED] "unused",
|
|
|
|
[EMBRYO] "embryo",
|
|
|
|
[SLEEPING] "sleep ",
|
|
|
|
[RUNNABLE] "runble",
|
|
|
|
[RUNNING] "run ",
|
|
|
|
[ZOMBIE] "zombie"
|
2007-08-08 08:38:38 +00:00
|
|
|
};
|
2007-08-20 19:37:15 +00:00
|
|
|
int i, j;
|
2006-09-07 15:45:38 +00:00
|
|
|
struct proc *p;
|
2007-08-08 08:38:38 +00:00
|
|
|
char *state;
|
2007-08-20 19:37:15 +00:00
|
|
|
uint pc[10];
|
2006-09-07 15:45:38 +00:00
|
|
|
|
2007-08-28 18:32:08 +00:00
|
|
|
for(i = 0; i < NPROC; i++){
|
2006-09-07 15:45:38 +00:00
|
|
|
p = &proc[i];
|
|
|
|
if(p->state == UNUSED)
|
|
|
|
continue;
|
2007-08-08 09:43:07 +00:00
|
|
|
if(p->state >= 0 && p->state < NELEM(states) && states[p->state])
|
2007-08-08 08:38:38 +00:00
|
|
|
state = states[p->state];
|
2007-08-08 09:42:36 +00:00
|
|
|
else
|
|
|
|
state = "???";
|
2007-08-20 19:37:15 +00:00
|
|
|
cprintf("%d %s %s", p->pid, state, p->name);
|
2007-08-28 18:32:08 +00:00
|
|
|
if(p->state == SLEEPING){
|
2008-10-15 05:14:10 +00:00
|
|
|
getcallerpcs((uint*)p->context->ebp+2, pc);
|
2007-08-20 19:37:15 +00:00
|
|
|
for(j=0; j<10 && pc[j] != 0; j++)
|
|
|
|
cprintf(" %p", pc[j]);
|
|
|
|
}
|
|
|
|
cprintf("\n");
|
2006-09-07 15:45:38 +00:00
|
|
|
}
|
|
|
|
}
|
2006-09-08 14:26:51 +00:00
|
|
|
|