Merge tag 'xfs-for-linus-4.7-rc1' of git://git.kernel.org/pub/scm/linux/kernel/git...
[linux-2.6-block.git] / include / linux / pmem.h
CommitLineData
61031952
RZ
1/*
2 * Copyright(c) 2015 Intel Corporation. All rights reserved.
3 *
4 * This program is free software; you can redistribute it and/or modify
5 * it under the terms of version 2 of the GNU General Public License as
6 * published by the Free Software Foundation.
7 *
8 * This program is distributed in the hope that it will be useful, but
9 * WITHOUT ANY WARRANTY; without even the implied warranty of
10 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
11 * General Public License for more details.
12 */
13#ifndef __PMEM_H__
14#define __PMEM_H__
15
16#include <linux/io.h>
5de490da 17#include <linux/uio.h>
61031952
RZ
18
19#ifdef CONFIG_ARCH_HAS_PMEM_API
96601adb 20#define ARCH_MEMREMAP_PMEM MEMREMAP_WB
40603526 21#include <asm/pmem.h>
61031952 22#else
96601adb
DW
23#define ARCH_MEMREMAP_PMEM MEMREMAP_WT
24/*
25 * These are simply here to enable compilation, all call sites gate
26 * calling these symbols with arch_has_pmem_api() and redirect to the
27 * implementation in asm/pmem.h.
28 */
29static inline bool __arch_has_wmb_pmem(void)
61031952 30{
96601adb 31 return false;
61031952
RZ
32}
33
96601adb 34static inline void arch_wmb_pmem(void)
61031952 35{
96601adb 36 BUG();
61031952
RZ
37}
38
61031952
RZ
39static inline void arch_memcpy_to_pmem(void __pmem *dst, const void *src,
40 size_t n)
41{
42 BUG();
43}
5de490da 44
fc0c2028
DW
45static inline int arch_memcpy_from_pmem(void *dst, const void __pmem *src,
46 size_t n)
47{
48 BUG();
49 return -EFAULT;
50}
51
5de490da
RZ
52static inline size_t arch_copy_from_iter_pmem(void __pmem *addr, size_t bytes,
53 struct iov_iter *i)
54{
55 BUG();
56 return 0;
57}
58
59static inline void arch_clear_pmem(void __pmem *addr, size_t size)
60{
61 BUG();
62}
3f4a2670
RZ
63
64static inline void arch_wb_cache_pmem(void __pmem *addr, size_t size)
65{
66 BUG();
67}
59e64739
DW
68
69static inline void arch_invalidate_pmem(void __pmem *addr, size_t size)
70{
71 BUG();
72}
61031952
RZ
73#endif
74
cba2e47a
TK
75static inline bool arch_has_pmem_api(void)
76{
77 return IS_ENABLED(CONFIG_ARCH_HAS_PMEM_API);
78}
79
80static inline int default_memcpy_from_pmem(void *dst, void __pmem const *src,
81 size_t size)
82{
83 memcpy(dst, (void __force *) src, size);
84 return 0;
85}
86
61031952 87/*
fc0c2028
DW
88 * memcpy_from_pmem - read from persistent memory with error handling
89 * @dst: destination buffer
90 * @src: source buffer
91 * @size: transfer length
92 *
93 * Returns 0 on success negative error code on failure.
61031952 94 */
fc0c2028
DW
95static inline int memcpy_from_pmem(void *dst, void __pmem const *src,
96 size_t size)
61031952 97{
cba2e47a
TK
98 if (arch_has_pmem_api())
99 return arch_memcpy_from_pmem(dst, src, size);
100 else
101 return default_memcpy_from_pmem(dst, src, size);
96601adb
DW
102}
103
61031952 104/**
96601adb 105 * arch_has_wmb_pmem - true if wmb_pmem() ensures durability
61031952
RZ
106 *
107 * For a given cpu implementation within an architecture it is possible
108 * that wmb_pmem() resolves to a nop. In the case this returns
109 * false, pmem api users are unable to ensure durability and may want to
110 * fall back to a different data consistency model, or otherwise notify
111 * the user.
112 */
96601adb 113static inline bool arch_has_wmb_pmem(void)
61031952 114{
96601adb 115 return arch_has_pmem_api() && __arch_has_wmb_pmem();
61031952
RZ
116}
117
118/*
119 * These defaults seek to offer decent performance and minimize the
120 * window between i/o completion and writes being durable on media.
121 * However, it is undefined / architecture specific whether
a639315d 122 * ARCH_MEMREMAP_PMEM + default_memcpy_to_pmem is sufficient for
61031952
RZ
123 * making data durable relative to i/o completion.
124 */
e836a256 125static inline void default_memcpy_to_pmem(void __pmem *dst, const void *src,
61031952
RZ
126 size_t size)
127{
128 memcpy((void __force *) dst, src, size);
129}
130
5de490da
RZ
131static inline size_t default_copy_from_iter_pmem(void __pmem *addr,
132 size_t bytes, struct iov_iter *i)
133{
134 return copy_from_iter_nocache((void __force *)addr, bytes, i);
135}
136
137static inline void default_clear_pmem(void __pmem *addr, size_t size)
138{
139 if (size == PAGE_SIZE && ((unsigned long)addr & ~PAGE_MASK) == 0)
140 clear_page((void __force *)addr);
141 else
142 memset((void __force *)addr, 0, size);
143}
144
61031952
RZ
145/**
146 * memcpy_to_pmem - copy data to persistent memory
147 * @dst: destination buffer for the copy
148 * @src: source buffer for the copy
149 * @n: length of the copy in bytes
150 *
151 * Perform a memory copy that results in the destination of the copy
152 * being effectively evicted from, or never written to, the processor
153 * cache hierarchy after the copy completes. After memcpy_to_pmem()
154 * data may still reside in cpu or platform buffers, so this operation
155 * must be followed by a wmb_pmem().
156 */
157static inline void memcpy_to_pmem(void __pmem *dst, const void *src, size_t n)
158{
159 if (arch_has_pmem_api())
160 arch_memcpy_to_pmem(dst, src, n);
161 else
162 default_memcpy_to_pmem(dst, src, n);
163}
164
165/**
166 * wmb_pmem - synchronize writes to persistent memory
167 *
168 * After a series of memcpy_to_pmem() operations this drains data from
169 * cpu write buffers and any platform (memory controller) buffers to
170 * ensure that written data is durable on persistent memory media.
171 */
172static inline void wmb_pmem(void)
173{
96601adb 174 if (arch_has_wmb_pmem())
61031952 175 arch_wmb_pmem();
96601adb
DW
176 else
177 wmb();
61031952 178}
5de490da
RZ
179
180/**
181 * copy_from_iter_pmem - copy data from an iterator to PMEM
182 * @addr: PMEM destination address
183 * @bytes: number of bytes to copy
184 * @i: iterator with source data
185 *
186 * Copy data from the iterator 'i' to the PMEM buffer starting at 'addr'.
187 * This function requires explicit ordering with a wmb_pmem() call.
188 */
189static inline size_t copy_from_iter_pmem(void __pmem *addr, size_t bytes,
190 struct iov_iter *i)
191{
192 if (arch_has_pmem_api())
193 return arch_copy_from_iter_pmem(addr, bytes, i);
194 return default_copy_from_iter_pmem(addr, bytes, i);
195}
196
197/**
198 * clear_pmem - zero a PMEM memory range
199 * @addr: virtual start address
200 * @size: number of bytes to zero
201 *
202 * Write zeros into the memory range starting at 'addr' for 'size' bytes.
203 * This function requires explicit ordering with a wmb_pmem() call.
204 */
205static inline void clear_pmem(void __pmem *addr, size_t size)
206{
207 if (arch_has_pmem_api())
208 arch_clear_pmem(addr, size);
209 else
210 default_clear_pmem(addr, size);
211}
3f4a2670 212
59e64739
DW
213/**
214 * invalidate_pmem - flush a pmem range from the cache hierarchy
215 * @addr: virtual start address
216 * @size: bytes to invalidate (internally aligned to cache line size)
217 *
218 * For platforms that support clearing poison this flushes any poisoned
219 * ranges out of the cache
220 */
221static inline void invalidate_pmem(void __pmem *addr, size_t size)
222{
223 if (arch_has_pmem_api())
224 arch_invalidate_pmem(addr, size);
225}
226
3f4a2670
RZ
227/**
228 * wb_cache_pmem - write back processor cache for PMEM memory range
229 * @addr: virtual start address
230 * @size: number of bytes to write back
231 *
232 * Write back the processor cache range starting at 'addr' for 'size' bytes.
233 * This function requires explicit ordering with a wmb_pmem() call.
234 */
235static inline void wb_cache_pmem(void __pmem *addr, size_t size)
236{
237 if (arch_has_pmem_api())
238 arch_wb_cache_pmem(addr, size);
239}
61031952 240#endif /* __PMEM_H__ */