Remove unused define
[fio.git] / arch / arch-x86_64.h
index 972c623a0bfcd90c043ec776158606e10b93cf60..c8c3d18251caf721fe25bef5c85b6c765d675ee3 100644 (file)
 #define FIO_HAVE_SYSLET
 
 #define nop            __asm__ __volatile__("rep;nop": : :"memory")
-#define read_barrier() asm volatile("lfence":::"memory")
+#define read_barrier() __asm__ __volatile__("lfence":::"memory")
+#define write_barrier()        __asm__ __volatile__("sfence":::"memory")
 
-static inline unsigned long arch_ffz(unsigned long bitmask)
+static inline unsigned int arch_ffz(unsigned int bitmask)
 {
-       __asm__("bsfq %1,%0" :"=r" (bitmask) :"r" (~bitmask));
+       __asm__("bsfl %1,%0" :"=r" (bitmask) :"r" (~bitmask));
        return bitmask;
 }
 #define ARCH_HAVE_FFZ
 
+typedef struct {
+       unsigned int lock;
+} spinlock_t;
+
+static inline void spin_lock(spinlock_t *lock)
+{
+       short inc = 0x0100;
+
+       __asm__ __volatile__("xaddw %w0, %1\n"
+                       "1:\t"
+                       "cmpb %h0, %b0\n\t"
+                       "je 2f\n\t"
+                       "rep ; nop\n\t"
+                       "movb %1, %b0\n\t"
+                       "jmp 1b\n"
+                       "2:"
+                       : "+Q" (inc), "+m" (lock->lock)
+                       :
+                       : "memory", "cc");
+}
+
+static inline void spin_unlock(spinlock_t *lock)
+{
+       __asm__ __volatile__("incb %0"
+                       : "+m" (lock->lock)
+                       :
+                       : "memory", "cc");
+}
+
 #endif