Commit | Line | Data |
---|---|---|
4d88a97a DW |
1 | /* |
2 | * Copyright(c) 2013-2015 Intel Corporation. All rights reserved. | |
3 | * | |
4 | * This program is free software; you can redistribute it and/or modify | |
5 | * it under the terms of version 2 of the GNU General Public License as | |
6 | * published by the Free Software Foundation. | |
7 | * | |
8 | * This program is distributed in the hope that it will be useful, but | |
9 | * WITHOUT ANY WARRANTY; without even the implied warranty of | |
10 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU | |
11 | * General Public License for more details. | |
12 | */ | |
13 | #ifndef __ND_H__ | |
14 | #define __ND_H__ | |
1f7df6f8 | 15 | #include <linux/libnvdimm.h> |
200c79da | 16 | #include <linux/badblocks.h> |
f0dc089c | 17 | #include <linux/blkdev.h> |
4d88a97a DW |
18 | #include <linux/device.h> |
19 | #include <linux/mutex.h> | |
20 | #include <linux/ndctl.h> | |
bf9bccc1 | 21 | #include <linux/types.h> |
71999466 | 22 | #include <linux/nd.h> |
4a826c83 | 23 | #include "label.h" |
4d88a97a | 24 | |
8c2f7e86 | 25 | enum { |
5212e11f VV |
26 | /* |
27 | * Limits the maximum number of block apertures a dimm can | |
28 | * support and is an input to the geometry/on-disk-format of a | |
29 | * BTT instance | |
30 | */ | |
31 | ND_MAX_LANES = 256, | |
fcae6957 | 32 | INT_LBASIZE_ALIGNMENT = 64, |
3ae3d67b | 33 | NVDIMM_IO_ATOMIC = 1, |
8c2f7e86 DW |
34 | }; |
35 | ||
4d88a97a DW |
36 | struct nvdimm_drvdata { |
37 | struct device *dev; | |
02881768 | 38 | int nslabel_size; |
4d88a97a DW |
39 | struct nd_cmd_get_config_size nsarea; |
40 | void *data; | |
4a826c83 DW |
41 | int ns_current, ns_next; |
42 | struct resource dpa; | |
bf9bccc1 | 43 | struct kref kref; |
4d88a97a DW |
44 | }; |
45 | ||
e5ae3b25 DW |
46 | struct nd_region_data { |
47 | int ns_count; | |
48 | int ns_active; | |
595c7307 DW |
49 | unsigned int hints_shift; |
50 | void __iomem *flush_wpq[0]; | |
3d88002e DW |
51 | }; |
52 | ||
595c7307 DW |
53 | static inline void __iomem *ndrd_get_flush_wpq(struct nd_region_data *ndrd, |
54 | int dimm, int hint) | |
55 | { | |
56 | unsigned int num = 1 << ndrd->hints_shift; | |
57 | unsigned int mask = num - 1; | |
58 | ||
59 | return ndrd->flush_wpq[dimm * num + (hint & mask)]; | |
60 | } | |
61 | ||
62 | static inline void ndrd_set_flush_wpq(struct nd_region_data *ndrd, int dimm, | |
63 | int hint, void __iomem *flush) | |
64 | { | |
65 | unsigned int num = 1 << ndrd->hints_shift; | |
66 | unsigned int mask = num - 1; | |
67 | ||
68 | ndrd->flush_wpq[dimm * num + (hint & mask)] = flush; | |
69 | } | |
70 | ||
4a826c83 DW |
71 | static inline struct nd_namespace_index *to_namespace_index( |
72 | struct nvdimm_drvdata *ndd, int i) | |
73 | { | |
74 | if (i < 0) | |
75 | return NULL; | |
76 | ||
77 | return ndd->data + sizeof_namespace_index(ndd) * i; | |
78 | } | |
79 | ||
80 | static inline struct nd_namespace_index *to_current_namespace_index( | |
81 | struct nvdimm_drvdata *ndd) | |
82 | { | |
83 | return to_namespace_index(ndd, ndd->ns_current); | |
84 | } | |
85 | ||
86 | static inline struct nd_namespace_index *to_next_namespace_index( | |
87 | struct nvdimm_drvdata *ndd) | |
88 | { | |
89 | return to_namespace_index(ndd, ndd->ns_next); | |
90 | } | |
91 | ||
564e871a DW |
92 | unsigned sizeof_namespace_label(struct nvdimm_drvdata *ndd); |
93 | ||
94 | #define namespace_label_has(ndd, field) \ | |
95 | (offsetof(struct nd_namespace_label, field) \ | |
96 | < sizeof_namespace_label(ndd)) | |
97 | ||
4a826c83 DW |
98 | #define nd_dbg_dpa(r, d, res, fmt, arg...) \ |
99 | dev_dbg((r) ? &(r)->dev : (d)->dev, "%s: %.13s: %#llx @ %#llx " fmt, \ | |
100 | (r) ? dev_name((d)->dev) : "", res ? res->name : "null", \ | |
101 | (unsigned long long) (res ? resource_size(res) : 0), \ | |
102 | (unsigned long long) (res ? res->start : 0), ##arg) | |
103 | ||
bf9bccc1 DW |
104 | #define for_each_dpa_resource(ndd, res) \ |
105 | for (res = (ndd)->dpa.child; res; res = res->sibling) | |
106 | ||
4a826c83 DW |
107 | #define for_each_dpa_resource_safe(ndd, res, next) \ |
108 | for (res = (ndd)->dpa.child, next = res ? res->sibling : NULL; \ | |
109 | res; res = next, next = next ? next->sibling : NULL) | |
110 | ||
5212e11f VV |
111 | struct nd_percpu_lane { |
112 | int count; | |
113 | spinlock_t lock; | |
114 | }; | |
115 | ||
c4703ce1 DW |
116 | enum nd_label_flags { |
117 | ND_LABEL_REAP, | |
118 | }; | |
ae8219f1 DW |
119 | struct nd_label_ent { |
120 | struct list_head list; | |
c4703ce1 | 121 | unsigned long flags; |
ae8219f1 DW |
122 | struct nd_namespace_label *label; |
123 | }; | |
124 | ||
125 | enum nd_mapping_lock_class { | |
126 | ND_MAPPING_CLASS0, | |
127 | ND_MAPPING_UUID_SCAN, | |
128 | }; | |
129 | ||
44c462eb DW |
130 | struct nd_mapping { |
131 | struct nvdimm *nvdimm; | |
44c462eb DW |
132 | u64 start; |
133 | u64 size; | |
401c0a19 | 134 | int position; |
ae8219f1 DW |
135 | struct list_head labels; |
136 | struct mutex lock; | |
44c462eb DW |
137 | /* |
138 | * @ndd is for private use at region enable / disable time for | |
139 | * get_ndd() + put_ndd(), all other nd_mapping to ndd | |
140 | * conversions use to_ndd() which respects enabled state of the | |
141 | * nvdimm. | |
142 | */ | |
143 | struct nvdimm_drvdata *ndd; | |
144 | }; | |
145 | ||
1f7df6f8 DW |
146 | struct nd_region { |
147 | struct device dev; | |
1b40e09a | 148 | struct ida ns_ida; |
8c2f7e86 | 149 | struct ida btt_ida; |
e1455744 | 150 | struct ida pfn_ida; |
cd03412a | 151 | struct ida dax_ida; |
004f1afb | 152 | unsigned long flags; |
bf9bccc1 | 153 | struct device *ns_seed; |
8c2f7e86 | 154 | struct device *btt_seed; |
e1455744 | 155 | struct device *pfn_seed; |
cd03412a | 156 | struct device *dax_seed; |
1f7df6f8 DW |
157 | u16 ndr_mappings; |
158 | u64 ndr_size; | |
159 | u64 ndr_start; | |
8fc5c735 | 160 | int id, num_lanes, ro, numa_node, target_node; |
1f7df6f8 | 161 | void *provider_data; |
975750a9 | 162 | struct kernfs_node *bb_state; |
6a6bef90 | 163 | struct badblocks bb; |
eaf96153 | 164 | struct nd_interleave_set *nd_set; |
5212e11f | 165 | struct nd_percpu_lane __percpu *lane; |
1f7df6f8 DW |
166 | struct nd_mapping mapping[0]; |
167 | }; | |
168 | ||
047fc8a1 RZ |
169 | struct nd_blk_region { |
170 | int (*enable)(struct nvdimm_bus *nvdimm_bus, struct device *dev); | |
047fc8a1 RZ |
171 | int (*do_io)(struct nd_blk_region *ndbr, resource_size_t dpa, |
172 | void *iobuf, u64 len, int rw); | |
173 | void *blk_provider_data; | |
174 | struct nd_region nd_region; | |
175 | }; | |
176 | ||
4a826c83 DW |
177 | /* |
178 | * Lookup next in the repeating sequence of 01, 10, and 11. | |
179 | */ | |
180 | static inline unsigned nd_inc_seq(unsigned seq) | |
181 | { | |
182 | static const unsigned next[] = { 0, 2, 3, 1 }; | |
183 | ||
184 | return next[seq & 3]; | |
185 | } | |
f524bf27 | 186 | |
5212e11f | 187 | struct btt; |
8c2f7e86 DW |
188 | struct nd_btt { |
189 | struct device dev; | |
190 | struct nd_namespace_common *ndns; | |
5212e11f | 191 | struct btt *btt; |
8c2f7e86 | 192 | unsigned long lbasize; |
abe8b4e3 | 193 | u64 size; |
8c2f7e86 DW |
194 | u8 *uuid; |
195 | int id; | |
14e49454 VV |
196 | int initial_offset; |
197 | u16 version_major; | |
198 | u16 version_minor; | |
8c2f7e86 DW |
199 | }; |
200 | ||
e1455744 DW |
201 | enum nd_pfn_mode { |
202 | PFN_MODE_NONE, | |
203 | PFN_MODE_RAM, | |
204 | PFN_MODE_PMEM, | |
205 | }; | |
206 | ||
207 | struct nd_pfn { | |
208 | int id; | |
209 | u8 *uuid; | |
210 | struct device dev; | |
315c5625 | 211 | unsigned long align; |
e1455744 DW |
212 | unsigned long npfns; |
213 | enum nd_pfn_mode mode; | |
214 | struct nd_pfn_sb *pfn_sb; | |
215 | struct nd_namespace_common *ndns; | |
216 | }; | |
217 | ||
cd03412a DW |
218 | struct nd_dax { |
219 | struct nd_pfn nd_pfn; | |
220 | }; | |
221 | ||
4d88a97a DW |
222 | enum nd_async_mode { |
223 | ND_SYNC, | |
224 | ND_ASYNC, | |
225 | }; | |
226 | ||
41cd8b70 | 227 | int nd_integrity_init(struct gendisk *disk, unsigned long meta_size); |
bf9bccc1 | 228 | void wait_nvdimm_bus_probe_idle(struct device *dev); |
4d88a97a DW |
229 | void nd_device_register(struct device *dev); |
230 | void nd_device_unregister(struct device *dev, enum nd_async_mode mode); | |
71999466 | 231 | void nd_device_notify(struct device *dev, enum nvdimm_event event); |
bf9bccc1 DW |
232 | int nd_uuid_store(struct device *dev, u8 **uuid_out, const char *buf, |
233 | size_t len); | |
b2c48f9f | 234 | ssize_t nd_size_select_show(unsigned long current_size, |
1b40e09a | 235 | const unsigned long *supported, char *buf); |
b2c48f9f DW |
236 | ssize_t nd_size_select_store(struct device *dev, const char *buf, |
237 | unsigned long *current_size, const unsigned long *supported); | |
4d88a97a | 238 | int __init nvdimm_init(void); |
3d88002e | 239 | int __init nd_region_init(void); |
b3fde74e | 240 | int __init nd_label_init(void); |
4d88a97a | 241 | void nvdimm_exit(void); |
3d88002e | 242 | void nd_region_exit(void); |
bf9bccc1 DW |
243 | struct nvdimm; |
244 | struct nvdimm_drvdata *to_ndd(struct nd_mapping *nd_mapping); | |
aee65987 | 245 | int nvdimm_check_config_data(struct device *dev); |
4d88a97a DW |
246 | int nvdimm_init_nsarea(struct nvdimm_drvdata *ndd); |
247 | int nvdimm_init_config_data(struct nvdimm_drvdata *ndd); | |
2d657d17 AD |
248 | int nvdimm_get_config_data(struct nvdimm_drvdata *ndd, void *buf, |
249 | size_t offset, size_t len); | |
f524bf27 DW |
250 | int nvdimm_set_config_data(struct nvdimm_drvdata *ndd, size_t offset, |
251 | void *buf, size_t len); | |
59e64739 DW |
252 | long nvdimm_clear_poison(struct device *dev, phys_addr_t phys, |
253 | unsigned int len); | |
42237e39 | 254 | void nvdimm_set_aliasing(struct device *dev); |
8f078b38 | 255 | void nvdimm_set_locked(struct device *dev); |
d34cb808 | 256 | void nvdimm_clear_locked(struct device *dev); |
1cd73865 | 257 | int nvdimm_security_setup_events(struct device *dev); |
4c6926a2 DJ |
258 | #if IS_ENABLED(CONFIG_NVDIMM_KEYS) |
259 | int nvdimm_security_unlock(struct device *dev); | |
260 | #else | |
261 | static inline int nvdimm_security_unlock(struct device *dev) | |
262 | { | |
263 | return 0; | |
264 | } | |
265 | #endif | |
8c2f7e86 | 266 | struct nd_btt *to_nd_btt(struct device *dev); |
e1455744 DW |
267 | |
268 | struct nd_gen_sb { | |
269 | char reserved[SZ_4K - 8]; | |
270 | __le64 checksum; | |
271 | }; | |
272 | ||
273 | u64 nd_sb_checksum(struct nd_gen_sb *sb); | |
8c2f7e86 | 274 | #if IS_ENABLED(CONFIG_BTT) |
200c79da | 275 | int nd_btt_probe(struct device *dev, struct nd_namespace_common *ndns); |
8c2f7e86 DW |
276 | bool is_nd_btt(struct device *dev); |
277 | struct device *nd_btt_create(struct nd_region *nd_region); | |
278 | #else | |
e32bc729 | 279 | static inline int nd_btt_probe(struct device *dev, |
200c79da | 280 | struct nd_namespace_common *ndns) |
8c2f7e86 DW |
281 | { |
282 | return -ENODEV; | |
283 | } | |
284 | ||
285 | static inline bool is_nd_btt(struct device *dev) | |
286 | { | |
287 | return false; | |
288 | } | |
289 | ||
290 | static inline struct device *nd_btt_create(struct nd_region *nd_region) | |
291 | { | |
292 | return NULL; | |
293 | } | |
e1455744 | 294 | #endif |
8c2f7e86 | 295 | |
e1455744 DW |
296 | struct nd_pfn *to_nd_pfn(struct device *dev); |
297 | #if IS_ENABLED(CONFIG_NVDIMM_PFN) | |
0dd69643 OH |
298 | |
299 | #ifdef CONFIG_TRANSPARENT_HUGEPAGE | |
300 | #define PFN_DEFAULT_ALIGNMENT HPAGE_PMD_SIZE | |
301 | #else | |
302 | #define PFN_DEFAULT_ALIGNMENT PAGE_SIZE | |
303 | #endif | |
304 | ||
200c79da | 305 | int nd_pfn_probe(struct device *dev, struct nd_namespace_common *ndns); |
e1455744 DW |
306 | bool is_nd_pfn(struct device *dev); |
307 | struct device *nd_pfn_create(struct nd_region *nd_region); | |
cd03412a DW |
308 | struct device *nd_pfn_devinit(struct nd_pfn *nd_pfn, |
309 | struct nd_namespace_common *ndns); | |
c5ed9268 | 310 | int nd_pfn_validate(struct nd_pfn *nd_pfn, const char *sig); |
cd03412a | 311 | extern struct attribute_group nd_pfn_attribute_group; |
e1455744 | 312 | #else |
200c79da DW |
313 | static inline int nd_pfn_probe(struct device *dev, |
314 | struct nd_namespace_common *ndns) | |
e1455744 DW |
315 | { |
316 | return -ENODEV; | |
317 | } | |
318 | ||
319 | static inline bool is_nd_pfn(struct device *dev) | |
320 | { | |
321 | return false; | |
322 | } | |
323 | ||
324 | static inline struct device *nd_pfn_create(struct nd_region *nd_region) | |
325 | { | |
326 | return NULL; | |
327 | } | |
32ab0a3f | 328 | |
c5ed9268 | 329 | static inline int nd_pfn_validate(struct nd_pfn *nd_pfn, const char *sig) |
32ab0a3f DW |
330 | { |
331 | return -ENODEV; | |
332 | } | |
8c2f7e86 | 333 | #endif |
e1455744 | 334 | |
cd03412a DW |
335 | struct nd_dax *to_nd_dax(struct device *dev); |
336 | #if IS_ENABLED(CONFIG_NVDIMM_DAX) | |
c5ed9268 | 337 | int nd_dax_probe(struct device *dev, struct nd_namespace_common *ndns); |
cd03412a DW |
338 | bool is_nd_dax(struct device *dev); |
339 | struct device *nd_dax_create(struct nd_region *nd_region); | |
340 | #else | |
c5ed9268 DW |
341 | static inline int nd_dax_probe(struct device *dev, |
342 | struct nd_namespace_common *ndns) | |
343 | { | |
344 | return -ENODEV; | |
345 | } | |
346 | ||
cd03412a DW |
347 | static inline bool is_nd_dax(struct device *dev) |
348 | { | |
349 | return false; | |
350 | } | |
351 | ||
352 | static inline struct device *nd_dax_create(struct nd_region *nd_region) | |
353 | { | |
354 | return NULL; | |
355 | } | |
356 | #endif | |
357 | ||
3d88002e DW |
358 | int nd_region_to_nstype(struct nd_region *nd_region); |
359 | int nd_region_register_namespaces(struct nd_region *nd_region, int *err); | |
c12c48ce DW |
360 | u64 nd_region_interleave_set_cookie(struct nd_region *nd_region, |
361 | struct nd_namespace_index *nsindex); | |
86ef58a4 | 362 | u64 nd_region_interleave_set_altcookie(struct nd_region *nd_region); |
3d88002e DW |
363 | void nvdimm_bus_lock(struct device *dev); |
364 | void nvdimm_bus_unlock(struct device *dev); | |
365 | bool is_nvdimm_bus_locked(struct device *dev); | |
58138820 | 366 | int nvdimm_revalidate_disk(struct gendisk *disk); |
bf9bccc1 DW |
367 | void nvdimm_drvdata_release(struct kref *kref); |
368 | void put_ndd(struct nvdimm_drvdata *ndd); | |
4a826c83 DW |
369 | int nd_label_reserve_dpa(struct nvdimm_drvdata *ndd); |
370 | void nvdimm_free_dpa(struct nvdimm_drvdata *ndd, struct resource *res); | |
371 | struct resource *nvdimm_allocate_dpa(struct nvdimm_drvdata *ndd, | |
372 | struct nd_label_id *label_id, resource_size_t start, | |
373 | resource_size_t n); | |
8c2f7e86 | 374 | resource_size_t nvdimm_namespace_capacity(struct nd_namespace_common *ndns); |
08e6b3c6 | 375 | bool nvdimm_namespace_locked(struct nd_namespace_common *ndns); |
8c2f7e86 | 376 | struct nd_namespace_common *nvdimm_namespace_common_probe(struct device *dev); |
5212e11f | 377 | int nvdimm_namespace_attach_btt(struct nd_namespace_common *ndns); |
298f2bc5 | 378 | int nvdimm_namespace_detach_btt(struct nd_btt *nd_btt); |
5212e11f VV |
379 | const char *nvdimm_namespace_disk_name(struct nd_namespace_common *ndns, |
380 | char *name); | |
f979b13c | 381 | unsigned int pmem_sector_size(struct nd_namespace_common *ndns); |
a3901802 DW |
382 | void nvdimm_badblocks_populate(struct nd_region *nd_region, |
383 | struct badblocks *bb, const struct resource *res); | |
200c79da | 384 | #if IS_ENABLED(CONFIG_ND_CLAIM) |
e8d51348 | 385 | int nvdimm_setup_pfn(struct nd_pfn *nd_pfn, struct dev_pagemap *pgmap); |
200c79da DW |
386 | int devm_nsio_enable(struct device *dev, struct nd_namespace_io *nsio); |
387 | void devm_nsio_disable(struct device *dev, struct nd_namespace_io *nsio); | |
388 | #else | |
e8d51348 CH |
389 | static inline int nvdimm_setup_pfn(struct nd_pfn *nd_pfn, |
390 | struct dev_pagemap *pgmap) | |
ac515c08 | 391 | { |
e8d51348 | 392 | return -ENXIO; |
ac515c08 | 393 | } |
200c79da DW |
394 | static inline int devm_nsio_enable(struct device *dev, |
395 | struct nd_namespace_io *nsio) | |
396 | { | |
397 | return -ENXIO; | |
398 | } | |
399 | static inline void devm_nsio_disable(struct device *dev, | |
400 | struct nd_namespace_io *nsio) | |
401 | { | |
402 | } | |
403 | #endif | |
047fc8a1 | 404 | int nd_blk_region_init(struct nd_region *nd_region); |
e5ae3b25 | 405 | int nd_region_activate(struct nd_region *nd_region); |
f0dc089c DW |
406 | void __nd_iostat_start(struct bio *bio, unsigned long *start); |
407 | static inline bool nd_iostat_start(struct bio *bio, unsigned long *start) | |
408 | { | |
74d46992 | 409 | struct gendisk *disk = bio->bi_disk; |
f0dc089c DW |
410 | |
411 | if (!blk_queue_io_stat(disk->queue)) | |
412 | return false; | |
413 | ||
8d7c22ac | 414 | *start = jiffies; |
ddcf35d3 MC |
415 | generic_start_io_acct(disk->queue, bio_op(bio), bio_sectors(bio), |
416 | &disk->part0); | |
f0dc089c DW |
417 | return true; |
418 | } | |
8d7c22ac TK |
419 | static inline void nd_iostat_end(struct bio *bio, unsigned long start) |
420 | { | |
74d46992 | 421 | struct gendisk *disk = bio->bi_disk; |
8d7c22ac | 422 | |
ddcf35d3 | 423 | generic_end_io_acct(disk->queue, bio_op(bio), &disk->part0, start); |
8d7c22ac | 424 | } |
200c79da DW |
425 | static inline bool is_bad_pmem(struct badblocks *bb, sector_t sector, |
426 | unsigned int len) | |
427 | { | |
428 | if (bb->count) { | |
429 | sector_t first_bad; | |
430 | int num_bad; | |
431 | ||
432 | return !!badblocks_check(bb, sector, len / 512, &first_bad, | |
433 | &num_bad); | |
434 | } | |
435 | ||
436 | return false; | |
437 | } | |
047fc8a1 | 438 | resource_size_t nd_namespace_blk_validate(struct nd_namespace_blk *nsblk); |
6ec68954 | 439 | const u8 *nd_dev_to_uuid(struct device *dev); |
004f1afb | 440 | bool pmem_should_map_pages(struct device *dev); |
4d88a97a | 441 | #endif /* __ND_H__ */ |