| 1 | // Mutual exclusion spin locks. |
| 2 | |
| 3 | #include "types.h" |
| 4 | #include "defs.h" |
| 5 | #include "param.h" |
| 6 | #include "x86.h" |
| 7 | #include "memlayout.h" |
| 8 | #include "mmu.h" |
| 9 | #include "proc.h" |
| 10 | #include "spinlock.h" |
| 11 | |
| 12 | void |
| 13 | initlock(struct spinlock *lk, char *name) |
| 14 | { |
| 15 | lk->name = name; |
| 16 | lk->locked = 0; |
| 17 | lk->cpu = 0; |
| 18 | } |
| 19 | |
| 20 | // Acquire the lock. |
| 21 | // Loops (spins) until the lock is acquired. |
| 22 | // Holding a lock for a long time may cause |
| 23 | // other CPUs to waste time spinning to acquire it. |
| 24 | void |
| 25 | acquire(struct spinlock *lk) |
| 26 | { |
| 27 | pushcli(); // disable interrupts to avoid deadlock. |
| 28 | if(holding(lk)) |
| 29 | panic("acquire" ); |
| 30 | |
| 31 | // The xchg is atomic. |
| 32 | while(xchg(&lk->locked, 1) != 0) |
| 33 | ; |
| 34 | |
| 35 | // Tell the C compiler and the processor to not move loads or stores |
| 36 | // past this point, to ensure that the critical section's memory |
| 37 | // references happen after the lock is acquired. |
| 38 | __sync_synchronize(); |
| 39 | |
| 40 | // Record info about lock acquisition for debugging. |
| 41 | lk->cpu = mycpu(); |
| 42 | getcallerpcs(&lk, lk->pcs); |
| 43 | } |
| 44 | |
| 45 | // Release the lock. |
| 46 | void |
| 47 | release(struct spinlock *lk) |
| 48 | { |
| 49 | if(!holding(lk)) |
| 50 | panic("release" ); |
| 51 | |
| 52 | lk->pcs[0] = 0; |
| 53 | lk->cpu = 0; |
| 54 | |
| 55 | // Tell the C compiler and the processor to not move loads or stores |
| 56 | // past this point, to ensure that all the stores in the critical |
| 57 | // section are visible to other cores before the lock is released. |
| 58 | // Both the C compiler and the hardware may re-order loads and |
| 59 | // stores; __sync_synchronize() tells them both not to. |
| 60 | __sync_synchronize(); |
| 61 | |
| 62 | // Release the lock, equivalent to lk->locked = 0. |
| 63 | // This code can't use a C assignment, since it might |
| 64 | // not be atomic. A real OS would use C atomics here. |
| 65 | asm volatile("movl $0, %0" : "+m" (lk->locked) : ); |
| 66 | |
| 67 | popcli(); |
| 68 | } |
| 69 | |
| 70 | // Record the current call stack in pcs[] by following the %ebp chain. |
| 71 | void |
| 72 | getcallerpcs(void *v, uint pcs[]) |
| 73 | { |
| 74 | uint *ebp; |
| 75 | int i; |
| 76 | |
| 77 | ebp = (uint*)v - 2; |
| 78 | for(i = 0; i < 10; i++){ |
| 79 | if(ebp == 0 || ebp < (uint*)KERNBASE || ebp == (uint*)0xffffffff) |
| 80 | break; |
| 81 | pcs[i] = ebp[1]; // saved %eip |
| 82 | ebp = (uint*)ebp[0]; // saved %ebp |
| 83 | } |
| 84 | for(; i < 10; i++) |
| 85 | pcs[i] = 0; |
| 86 | } |
| 87 | |
| 88 | // Check whether this cpu is holding the lock. |
| 89 | int |
| 90 | holding(struct spinlock *lock) |
| 91 | { |
| 92 | int r; |
| 93 | pushcli(); |
| 94 | r = lock->locked && lock->cpu == mycpu(); |
| 95 | popcli(); |
| 96 | return r; |
| 97 | } |
| 98 | |
| 99 | |
| 100 | // Pushcli/popcli are like cli/sti except that they are matched: |
| 101 | // it takes two popcli to undo two pushcli. Also, if interrupts |
| 102 | // are off, then pushcli, popcli leaves them off. |
| 103 | |
| 104 | void |
| 105 | pushcli(void) |
| 106 | { |
| 107 | int eflags; |
| 108 | |
| 109 | eflags = readeflags(); |
| 110 | cli(); |
| 111 | if(mycpu()->ncli == 0) |
| 112 | mycpu()->intena = eflags & FL_IF; |
| 113 | mycpu()->ncli += 1; |
| 114 | } |
| 115 | |
| 116 | void |
| 117 | popcli(void) |
| 118 | { |
| 119 | if(readeflags()&FL_IF) |
| 120 | panic("popcli - interruptible" ); |
| 121 | if(--mycpu()->ncli < 0) |
| 122 | panic("popcli" ); |
| 123 | if(mycpu()->ncli == 0 && mycpu()->intena) |
| 124 | sti(); |
| 125 | } |
| 126 | |
| 127 | |