#define __NR_fadvise64 413
#endif
-#define nop do { } while (0)
-#define ffz(v) generic_ffz((v))
+#define nop do { } while (0)
+#define fio_ffz(v) generic_ffz((v))
#endif
ia64_intri_res; \
})
-static inline unsigned long ffz(unsigned long bitmask)
+static inline unsigned long fio_ffz(unsigned long bitmask)
{
return ia64_popcnt(bitmask & (~bitmask - 1));
}
return 31 - lz;
}
-static inline int ffz(unsigned long bitmask)
+static inline int fio_ffz(unsigned long bitmask)
{
if ((bitmask = ~bitmask) == 0)
return 32;
#define __NR_fadvise64 253
#endif
-#define nop asm volatile ("diag 0,0,68" : : : "memory")
-#define ffz(v) generic_ffz((v))
+#define nop asm volatile ("diag 0,0,68" : : : "memory")
+#define fio_ffz(v) generic_ffz((v))
#endif
#define nop __asm__ __volatile__("rep;nop": : :"memory")
-static inline unsigned long ffz(unsigned long bitmask)
+static inline unsigned long fio_ffz(unsigned long bitmask)
{
__asm__("bsfl %1,%0" :"=r" (bitmask) :"r" (~bitmask));
return bitmask;
#define nop __asm__ __volatile__("rep;nop": : :"memory")
-static inline unsigned long ffz(unsigned long bitmask)
+static inline unsigned long fio_ffz(unsigned long bitmask)
{
__asm__("bsfq %1,%0" :"=r" (bitmask) :"r" (~bitmask));
return bitmask;
*b = (i * BLOCKS_PER_MAP);
while ((*b) * td->o.rw_min_bs < f->real_file_size) {
if (f->file_map[i] != -1UL) {
- *b += ffz(f->file_map[i]);
+ *b += fio_ffz(f->file_map[i]);
f->last_free_lookup = i;
return 0;
}