patch-1.3.93 linux/include/asm-sparc/asmmacro.h
Next file: linux/include/asm-sparc/atomic.h
Previous file: linux/include/asm-sparc/asi.h
Back to the patch index
Back to the overall index
- Lines: 208
- Date:
Sun Apr 21 12:30:33 1996
- Orig file:
v1.3.92/linux/include/asm-sparc/asmmacro.h
- Orig date:
Thu Jan 1 02:00:00 1970
diff -u --recursive --new-file v1.3.92/linux/include/asm-sparc/asmmacro.h linux/include/asm-sparc/asmmacro.h
@@ -0,0 +1,207 @@
+/* asmmacro.h: Assembler macros.
+ *
+ * Copyright (C) 1996 David S. Miller (davem@caipfs.rutgers.edu)
+ */
+
+#ifndef _SPARC_ASMMACRO_H
+#define _SPARC_ASMMACRO_H
+
+/* #define SMP_DEBUG */
+
+#define GET_PROCESSOR_ID(reg) \
+ rd %tbr, %reg; \
+ srl %reg, 12, %reg; \
+ and %reg, 3, %reg;
+
+#define GET_PROCESSOR_MID(reg, tmp) \
+ GET_PROCESSOR_ID(reg) \
+ set C_LABEL(mid_xlate), %tmp; \
+ ldub [%tmp + %reg], %reg;
+
+#define GET_PROCESSOR_OFFSET(reg) \
+ rd %tbr, %reg; \
+ srl %reg, 10, %reg; \
+ and %reg, 0xc, %reg;
+
+#define PROCESSOR_OFFSET_TO_ID(reg) \
+ srl %reg, 2, %reg;
+
+#define PROCESSOR_ID_TO_OFFSET(reg) \
+ sll %reg, 2, %reg;
+
+/* All trap entry points _must_ begin with this macro or else you
+ * lose. It makes sure the kernel has a proper window so that
+ * c-code can be called.
+ */
+#ifndef SMP_DEBUG
+#define SAVE_ALL \
+ sethi %hi(trap_setup), %l4; \
+ jmpl %l4 + %lo(trap_setup), %l6; \
+ nop;
+#else
+#define SAVE_ALL \
+ GET_PROCESSOR_ID(l4); \
+ set C_LABEL(trap_log), %l5; \
+ sll %l4, 11, %l6; \
+ add %l5, %l6, %l5; \
+ set C_LABEL(trap_log_ent), %l6; \
+ sll %l4, 2, %l4; \
+ add %l6, %l4, %l6; \
+ ld [%l6], %l6; \
+ sll %l6, 3, %l6; \
+ st %l1, [%l5 + %l6]; \
+ add %l5, 4, %l5; \
+ st %l0, [%l5 + %l6]; \
+ set C_LABEL(trap_log_ent), %l5; \
+ add %l5, %l4, %l5; \
+ srl %l6, 3, %l6; \
+ add %l6, 1, %l6; \
+ and %l6, 255, %l6; \
+ st %l6, [%l5]; \
+ sethi %hi(trap_setup), %l4; \
+ jmpl %l4 + %lo(trap_setup), %l6; \
+ nop;
+#endif
+
+/* All traps low-level code here must end with this macro.
+ * For SMP configurations the ret_trap_entry routine will
+ * have to appropriate code to actually release the kernel
+ * entry lock.
+ */
+#define RESTORE_ALL \
+ b ret_trap_entry; \
+ nop;
+
+#ifndef __SMP__
+
+#define ENTER_SYSCALL
+#define LEAVE_SYSCALL
+#define ENTER_IRQ
+#define LEAVE_IRQ
+
+#else
+
+#define INCREMENT_COUNTER(symbol, tmp1, tmp2) \
+ set C_LABEL(symbol), %tmp1; \
+ ld [%tmp1], %tmp2; \
+ add %tmp2, 1, %tmp2; \
+ st %tmp2, [%tmp1];
+
+#define DECREMENT_COUNTER(symbol, tmp1, tmp2) \
+ set C_LABEL(symbol), %tmp1; \
+ ld [%tmp1], %tmp2; \
+ sub %tmp2, 1, %tmp2; \
+ st %tmp2, [%tmp1];
+
+ /* This is so complicated I suggest you don't look at it. */
+#define ENTER_MASK(mask) \
+ GET_PROCESSOR_OFFSET(l4) \
+ set C_LABEL(smp_spinning), %l6; \
+ add %l6, %l4, %l6; \
+ mov 1, %l5; \
+ st %l5, [%l6]; \
+ set C_LABEL(smp_proc_in_lock), %l5; \
+ ld [%l5 + %l4], %l6; \
+ or %l6, mask, %l6; \
+ st %l6, [%l5 + %l4]; \
+1: \
+ set C_LABEL(kernel_flag), %l5; \
+ ldstub [%l5], %l6; \
+ cmp %l6, 0; \
+ be 3f; \
+ nop; \
+ set C_LABEL(active_kernel_processor), %l5; \
+ GET_PROCESSOR_ID(l4) \
+ ldub [%l5], %l6; \
+ cmp %l6, %l4; \
+ be 4f; \
+ nop; \
+2: \
+ GET_PROCESSOR_MID(l4, l5) \
+ set C_LABEL(sun4m_interrupts), %l5; \
+ ld [%l5], %l5; \
+ sll %l4, 12, %l4; \
+ add %l5, %l4, %l5; \
+ ld [%l5], %l4; \
+ sethi %hi(0x80000000), %l6; \
+ andcc %l6, %l4, %g0; \
+ be 5f; \
+ nop; \
+ st %l6, [%l5 + 4]; \
+ nop; nop; nop; \
+ ld [%l5], %g0; \
+ nop; nop; nop; \
+ or %l0, PSR_PIL, %l4; \
+ wr %l4, 0x0, %psr; \
+ nop; nop; nop; \
+ wr %l4, PSR_ET, %psr; \
+ nop; nop; nop; \
+ call C_LABEL(smp_message_irq); \
+ nop; \
+ wr %l0, 0x0, %psr; \
+ nop; nop; nop; \
+5: \
+ set C_LABEL(kernel_flag), %l5; \
+ ldub [%l5], %l6; \
+ cmp %l6, 0; \
+ bne 2b; \
+ nop; \
+ b 1b; \
+ nop; \
+3: \
+ GET_PROCESSOR_ID(l4) \
+ set C_LABEL(active_kernel_processor), %l5; \
+ stb %l4, [%l5]; \
+ GET_PROCESSOR_MID(l4, l5) \
+ set C_LABEL(irq_rcvreg), %l5; \
+ ld [%l5], %l5; \
+ st %l4, [%l5]; \
+4: \
+ GET_PROCESSOR_OFFSET(l4) \
+ set C_LABEL(smp_spinning), %l6; \
+ st %g0, [%l6 + %l4];
+
+#define ENTER_SYSCALL \
+ ENTER_MASK(SMP_FROM_SYSCALL) \
+ INCREMENT_COUNTER(kernel_counter, l6, l5) \
+ INCREMENT_COUNTER(syscall_count, l6, l5)
+
+#define ENTER_IRQ \
+ ENTER_MASK(SMP_FROM_INT) \
+ INCREMENT_COUNTER(kernel_counter, l6, l5)
+
+#define LEAVE_MASK(mask) \
+ GET_PROCESSOR_OFFSET(l4) \
+ set C_LABEL(smp_proc_in_lock), %l5; \
+ ld [%l5 + %l4], %l6; \
+ andn %l6, mask, %l6; \
+ st %l6, [%l5 + %l4];
+
+#define LEAVE_SYSCALL \
+ LEAVE_MASK(SMP_FROM_SYSCALL) \
+ DECREMENT_COUNTER(syscall_count, l6, l5) \
+ set C_LABEL(kernel_counter), %l6; \
+ ld [%l6], %l5; \
+ subcc %l5, 1, %l5; \
+ st %l5, [%l6]; \
+ bne 1f; \
+ nop; \
+ set C_LABEL(active_kernel_processor), %l6; \
+ mov NO_PROC_ID, %l5; \
+ stb %l5, [%l6]; \
+ set C_LABEL(kernel_flag), %l6; \
+ stb %g0, [%l6]; \
+1:
+
+#define LEAVE_IRQ \
+ LEAVE_MASK(SMP_FROM_INT) \
+ INCREMENT_COUNTER(syscall_count, l6, l5)
+
+
+#define RESTORE_ALL_FASTIRQ \
+ b ret_irq_entry; \
+ nop;
+
+#endif /* !(__SMP__) */
+
+#endif /* !(_SPARC_ASMMACRO_H) */
FUNET's LINUX-ADM group, linux-adm@nic.funet.fi
TCL-scripts by Sam Shen, slshen@lbl.gov
with Sam's (original) version of this