#define read_barrier() asm volatile("bcr 15,0" : : : "memory")
#define write_barrier() asm volatile("bcr 15,0" : : : "memory")
+/*
+ * Fio needs monotonic (never lower), but not strict monotonic (never the same)
+ * so store clock fast is enough
+ */
static inline unsigned long long get_cpu_clock(void)
{
unsigned long long clk;
- __asm__ __volatile__("stck %0" : "=Q" (clk) : : "cc");
- return clk;
+ __asm__ __volatile__("stckf %0" : "=Q" (clk) : : "cc");
+ return clk>>12;
}
+#define ARCH_CPU_CLOCK_CYCLES_PER_USEC 1
+#define ARCH_HAVE_CPU_CLOCK
+
#define ARCH_HAVE_INIT
extern int tsc_reliable;
static inline int arch_init(char *envp[])
return 0;
}
-#define ARCH_HAVE_CPU_CLOCK
-
#endif
#include "hash.h"
#include "os/os.h"
-#ifdef ARCH_HAVE_CPU_CLOCK
+#if defined(ARCH_HAVE_CPU_CLOCK) && !defined(ARCH_CPU_CLOCK_CYCLES_PER_USEC)
static unsigned long cycles_per_usec;
static unsigned long inv_cycles_per_usec;
#endif
} else if (tv)
tv->last_cycles = t;
+#ifdef ARCH_CPU_CLOCK_CYCLES_PER_USEC
+ usecs = t / ARCH_CPU_CLOCK_CYCLES_PER_USEC;
+#else
usecs = (t * inv_cycles_per_usec) / 16777216UL;
+#endif
tp->tv_sec = usecs / 1000000;
tp->tv_usec = usecs % 1000000;
break;
}
}
-#ifdef ARCH_HAVE_CPU_CLOCK
+#if defined(ARCH_HAVE_CPU_CLOCK) && !defined(ARCH_CPU_CLOCK_CYCLES_PER_USEC)
static unsigned long get_cycles_per_usec(void)
{
struct timeval s, e;
#else
static int calibrate_cpu_clock(void)
{
+#ifdef ARCH_CPU_CLOCK_CYCLES_PER_USEC
+ return 0;
+#else
return 1;
-}
#endif
+}
+#endif // ARCH_HAVE_CPU_CLOCK
#ifndef CONFIG_TLS_THREAD
void fio_local_clock_init(int is_thread)