patch-2.2.14 linux/include/asm-s390/softirq.h

Next file: linux/include/asm-s390/spinlock.h
Previous file: linux/include/asm-s390/sockios.h
Back to the patch index
Back to the overall index

diff -u --recursive --new-file v2.2.13/linux/include/asm-s390/softirq.h linux/include/asm-s390/softirq.h
@@ -0,0 +1,149 @@
+/*
+ *  include/asm-s390/softirq.h
+ *
+ *  S390 version
+ *
+ *  Derived from "include/asm-i386/softirq.h"
+ */
+
+#ifndef __ASM_SOFTIRQ_H
+#define __ASM_SOFTIRQ_H
+
+#ifndef __LINUX_SMP_H
+#include <linux/smp.h>
+#endif
+
+#include <asm/atomic.h>
+#include <asm/hardirq.h>
+#include <asm/lowcore.h>
+
+#define get_active_bhs()        (bh_mask & bh_active)
+#define clear_active_bhs(x)     atomic_clear_mask((x),(atomic_t*) &bh_active)
+
+#ifdef __SMP__
+
+/*
+ * The locking mechanism for base handlers, to prevent re-entrancy,
+ * is entirely private to an implementation, it should not be
+ * referenced at all outside of this file.
+ */
+extern atomic_t global_bh_lock;
+extern atomic_t global_bh_count;
+extern spinlock_t s390_bh_lock;
+
+extern void synchronize_bh(void);
+
+static inline void start_bh_atomic(void)
+{
+	atomic_inc(&global_bh_lock);
+	synchronize_bh();
+}
+
+static inline void end_bh_atomic(void)
+{
+	atomic_dec(&global_bh_lock);
+}
+
+/* These are for the IRQs testing the lock */
+static inline int softirq_trylock(int cpu)
+{
+	if (!test_and_set_bit(0,&global_bh_count)) {
+		if (atomic_read(&global_bh_lock) == 0) {
+                        atomic_inc(&safe_get_cpu_lowcore(cpu).local_bh_count);
+			return 1;
+		}
+		clear_bit(0,&global_bh_count);
+	}
+	return 0;
+}
+
+static inline void softirq_endlock(int cpu)
+{
+        atomic_dec(&safe_get_cpu_lowcore(cpu).local_bh_count);
+	clear_bit(0,&global_bh_count);
+}
+
+#else
+
+extern inline void start_bh_atomic(void)
+{
+        atomic_inc(&S390_lowcore.local_bh_count);
+	barrier();
+}
+
+extern inline void end_bh_atomic(void)
+{
+	barrier();
+        atomic_dec(&S390_lowcore.local_bh_count);
+}
+
+/* These are for the irq's testing the lock */
+#define softirq_trylock(cpu)	(atomic_compare_and_swap(0,1,&S390_lowcore.local_bh_count) == 0)
+
+#define softirq_endlock(cpu)   atomic_set(&S390_lowcore.local_bh_count,0);
+#define synchronize_bh()       barrier()
+
+#endif	/* SMP */
+
+extern inline void init_bh(int nr, void (*routine)(void))
+{
+        unsigned long flags;
+
+        bh_base[nr] = routine;
+        atomic_set(&bh_mask_count[nr], 0);
+
+        spin_lock_irqsave(&s390_bh_lock, flags);
+        bh_mask |= 1 << nr;
+        spin_unlock_irqrestore(&s390_bh_lock, flags);
+}
+
+extern inline void remove_bh(int nr)
+{
+        unsigned long flags;
+
+        spin_lock_irqsave(&s390_bh_lock, flags);
+        bh_mask &= ~(1 << nr);
+        spin_unlock_irqrestore(&s390_bh_lock, flags);
+
+        synchronize_bh();
+        bh_base[nr] = NULL;
+}
+
+extern inline void mark_bh(int nr)
+{
+        set_bit(nr, &bh_active);
+}
+
+/*
+ * These use a mask count to correctly handle
+ * nested disable/enable calls
+ */
+extern inline void disable_bh(int nr)
+{
+        unsigned long flags;
+
+        spin_lock_irqsave(&s390_bh_lock, flags);
+        bh_mask &= ~(1 << nr);
+        atomic_inc(&bh_mask_count[nr]);
+        spin_unlock_irqrestore(&s390_bh_lock, flags);
+        synchronize_bh();
+}
+
+extern inline void enable_bh(int nr)
+{
+        unsigned long flags;
+
+        spin_lock_irqsave(&s390_bh_lock, flags);
+        if (atomic_dec_and_test(&bh_mask_count[nr]))
+                bh_mask |= 1 << nr;
+        spin_unlock_irqrestore(&s390_bh_lock, flags);
+}
+
+#endif	/* __ASM_SOFTIRQ_H */
+
+
+
+
+
+
+

FUNET's LINUX-ADM group, linux-adm@nic.funet.fi
TCL-scripts by Sam Shen (who was at: slshen@lbl.gov)