Commit | Line | Data |
---|---|---|
5b497af4 | 1 | /* SPDX-License-Identifier: GPL-2.0-only */ |
b94d5230 DW |
2 | /* |
3 | * Copyright(c) 2013-2015 Intel Corporation. All rights reserved. | |
b94d5230 DW |
4 | */ |
5 | #ifndef __ND_CORE_H__ | |
6 | #define __ND_CORE_H__ | |
7 | #include <linux/libnvdimm.h> | |
8 | #include <linux/device.h> | |
eaf96153 DW |
9 | #include <linux/sizes.h> |
10 | #include <linux/mutex.h> | |
1b40e09a | 11 | #include <linux/nd.h> |
87a30e1f | 12 | #include "nd.h" |
b94d5230 | 13 | |
e6dfb2de DW |
14 | extern struct list_head nvdimm_bus_list; |
15 | extern struct mutex nvdimm_bus_list_mutex; | |
62232e45 | 16 | extern int nvdimm_major; |
7d988097 | 17 | extern struct workqueue_struct *nvdimm_wq; |
e6dfb2de | 18 | |
b94d5230 DW |
19 | struct nvdimm_bus { |
20 | struct nvdimm_bus_descriptor *nd_desc; | |
b70d31d0 | 21 | wait_queue_head_t wait; |
45def22c | 22 | struct list_head list; |
b94d5230 | 23 | struct device dev; |
eaf96153 | 24 | int id, probe_active; |
b70d31d0 | 25 | atomic_t ioctl_active; |
29b9aa0a | 26 | struct list_head mapping_list; |
3d88002e | 27 | struct mutex reconfig_mutex; |
aa9ad44a | 28 | struct badrange badrange; |
b94d5230 | 29 | }; |
45def22c | 30 | |
e6dfb2de DW |
31 | struct nvdimm { |
32 | unsigned long flags; | |
33 | void *provider_data; | |
e3654eca | 34 | unsigned long cmd_mask; |
e6dfb2de | 35 | struct device dev; |
eaf96153 | 36 | atomic_t busy; |
e5ae3b25 DW |
37 | int id, num_flush; |
38 | struct resource *flush_wpq; | |
d6548ae4 | 39 | const char *dimm_id; |
f2989396 DJ |
40 | struct { |
41 | const struct nvdimm_security_ops *ops; | |
42 | enum nvdimm_security_state state; | |
89fa9d8e | 43 | enum nvdimm_security_state ext_state; |
7d988097 DJ |
44 | unsigned int overwrite_tmo; |
45 | struct kernfs_node *overwrite_state; | |
f2989396 | 46 | } sec; |
7d988097 | 47 | struct delayed_work dwork; |
e6dfb2de DW |
48 | }; |
49 | ||
f2989396 | 50 | static inline enum nvdimm_security_state nvdimm_security_state( |
faa8bd6e | 51 | struct nvdimm *nvdimm, enum nvdimm_passphrase_type ptype) |
f2989396 DJ |
52 | { |
53 | if (!nvdimm->sec.ops) | |
54 | return -ENXIO; | |
55 | ||
faa8bd6e | 56 | return nvdimm->sec.ops->state(nvdimm, ptype); |
f2989396 | 57 | } |
37833fb7 | 58 | int nvdimm_security_freeze(struct nvdimm *nvdimm); |
03b65b22 DJ |
59 | #if IS_ENABLED(CONFIG_NVDIMM_KEYS) |
60 | int nvdimm_security_disable(struct nvdimm *nvdimm, unsigned int keyid); | |
d2a4ac73 | 61 | int nvdimm_security_update(struct nvdimm *nvdimm, unsigned int keyid, |
89fa9d8e DJ |
62 | unsigned int new_keyid, |
63 | enum nvdimm_passphrase_type pass_type); | |
64 | int nvdimm_security_erase(struct nvdimm *nvdimm, unsigned int keyid, | |
65 | enum nvdimm_passphrase_type pass_type); | |
7d988097 DJ |
66 | int nvdimm_security_overwrite(struct nvdimm *nvdimm, unsigned int keyid); |
67 | void nvdimm_security_overwrite_query(struct work_struct *work); | |
03b65b22 DJ |
68 | #else |
69 | static inline int nvdimm_security_disable(struct nvdimm *nvdimm, | |
70 | unsigned int keyid) | |
71 | { | |
72 | return -EOPNOTSUPP; | |
73 | } | |
89fa9d8e DJ |
74 | static inline int nvdimm_security_update(struct nvdimm *nvdimm, |
75 | unsigned int keyid, | |
76 | unsigned int new_keyid, | |
77 | enum nvdimm_passphrase_type pass_type) | |
d2a4ac73 DJ |
78 | { |
79 | return -EOPNOTSUPP; | |
80 | } | |
89fa9d8e DJ |
81 | static inline int nvdimm_security_erase(struct nvdimm *nvdimm, |
82 | unsigned int keyid, | |
83 | enum nvdimm_passphrase_type pass_type) | |
64e77c8c DJ |
84 | { |
85 | return -EOPNOTSUPP; | |
86 | } | |
7d988097 DJ |
87 | static inline int nvdimm_security_overwrite(struct nvdimm *nvdimm, |
88 | unsigned int keyid) | |
89 | { | |
90 | return -EOPNOTSUPP; | |
91 | } | |
92 | static inline void nvdimm_security_overwrite_query(struct work_struct *work) | |
93 | { | |
94 | } | |
03b65b22 | 95 | #endif |
f2989396 | 96 | |
762d067d DW |
97 | /** |
98 | * struct blk_alloc_info - tracking info for BLK dpa scanning | |
99 | * @nd_mapping: blk region mapping boundaries | |
100 | * @available: decremented in alias_dpa_busy as aliased PMEM is scanned | |
101 | * @busy: decremented in blk_dpa_busy to account for ranges already | |
102 | * handled by alias_dpa_busy | |
103 | * @res: alias_dpa_busy interprets this a free space range that needs to | |
104 | * be truncated to the valid BLK allocation starting DPA, blk_dpa_busy | |
105 | * treats it as a busy range that needs the aliased PMEM ranges | |
106 | * truncated. | |
107 | */ | |
108 | struct blk_alloc_info { | |
109 | struct nd_mapping *nd_mapping; | |
110 | resource_size_t available, busy; | |
111 | struct resource *res; | |
112 | }; | |
113 | ||
3d88002e | 114 | bool is_nvdimm(struct device *dev); |
3d88002e | 115 | bool is_nd_pmem(struct device *dev); |
c9e582aa | 116 | bool is_nd_volatile(struct device *dev); |
047fc8a1 | 117 | bool is_nd_blk(struct device *dev); |
c9e582aa DW |
118 | static inline bool is_nd_region(struct device *dev) |
119 | { | |
120 | return is_nd_pmem(dev) || is_nd_blk(dev) || is_nd_volatile(dev); | |
121 | } | |
122 | static inline bool is_memory(struct device *dev) | |
123 | { | |
124 | return is_nd_pmem(dev) || is_nd_volatile(dev); | |
125 | } | |
e6dfb2de | 126 | struct nvdimm_bus *walk_to_nvdimm_bus(struct device *nd_dev); |
45def22c | 127 | int __init nvdimm_bus_init(void); |
4d88a97a | 128 | void nvdimm_bus_exit(void); |
b354aba0 DW |
129 | void nvdimm_devs_exit(void); |
130 | void nd_region_devs_exit(void); | |
eaf96153 | 131 | void nd_region_probe_success(struct nvdimm_bus *nvdimm_bus, struct device *dev); |
1b40e09a | 132 | struct nd_region; |
98a29c39 | 133 | void nd_region_create_ns_seed(struct nd_region *nd_region); |
8c2f7e86 | 134 | void nd_region_create_btt_seed(struct nd_region *nd_region); |
2dc43331 | 135 | void nd_region_create_pfn_seed(struct nd_region *nd_region); |
cd03412a | 136 | void nd_region_create_dax_seed(struct nd_region *nd_region); |
eaf96153 | 137 | void nd_region_disable(struct nvdimm_bus *nvdimm_bus, struct device *dev); |
45def22c DW |
138 | int nvdimm_bus_create_ndctl(struct nvdimm_bus *nvdimm_bus); |
139 | void nvdimm_bus_destroy_ndctl(struct nvdimm_bus *nvdimm_bus); | |
4d88a97a | 140 | void nd_synchronize(void); |
1f7df6f8 DW |
141 | int nvdimm_bus_register_dimms(struct nvdimm_bus *nvdimm_bus); |
142 | int nvdimm_bus_register_regions(struct nvdimm_bus *nvdimm_bus); | |
eaf96153 | 143 | int nvdimm_bus_init_interleave_sets(struct nvdimm_bus *nvdimm_bus); |
8c2f7e86 | 144 | void __nd_device_register(struct device *dev); |
1f7df6f8 | 145 | int nd_match_dimm(struct device *dev, void *data); |
bf9bccc1 DW |
146 | struct nd_label_id; |
147 | char *nd_label_gen_id(struct nd_label_id *label_id, u8 *uuid, u32 flags); | |
148 | bool nd_is_uuid_unique(struct device *dev, u8 *uuid); | |
149 | struct nd_region; | |
150 | struct nvdimm_drvdata; | |
151 | struct nd_mapping; | |
ae8219f1 | 152 | void nd_mapping_free_labels(struct nd_mapping *nd_mapping); |
12e3129e KB |
153 | |
154 | int __reserve_free_pmem(struct device *dev, void *data); | |
155 | void release_free_pmem(struct nvdimm_bus *nvdimm_bus, | |
156 | struct nd_mapping *nd_mapping); | |
157 | ||
158 | resource_size_t nd_pmem_max_contiguous_dpa(struct nd_region *nd_region, | |
159 | struct nd_mapping *nd_mapping); | |
160 | resource_size_t nd_region_allocatable_dpa(struct nd_region *nd_region); | |
bf9bccc1 DW |
161 | resource_size_t nd_pmem_available_dpa(struct nd_region *nd_region, |
162 | struct nd_mapping *nd_mapping, resource_size_t *overlap); | |
a1f3e4d6 | 163 | resource_size_t nd_blk_available_dpa(struct nd_region *nd_region); |
bf9bccc1 | 164 | resource_size_t nd_region_available_dpa(struct nd_region *nd_region); |
ae86cbfe DW |
165 | int nd_region_conflict(struct nd_region *nd_region, resource_size_t start, |
166 | resource_size_t size); | |
bf9bccc1 DW |
167 | resource_size_t nvdimm_allocated_dpa(struct nvdimm_drvdata *ndd, |
168 | struct nd_label_id *label_id); | |
762d067d | 169 | int alias_dpa_busy(struct device *dev, void *data); |
1b40e09a DW |
170 | struct resource *nsblk_add_resource(struct nd_region *nd_region, |
171 | struct nvdimm_drvdata *ndd, struct nd_namespace_blk *nsblk, | |
172 | resource_size_t start); | |
0ba1c634 | 173 | int nvdimm_num_label_slots(struct nvdimm_drvdata *ndd); |
bf9bccc1 | 174 | void get_ndd(struct nvdimm_drvdata *ndd); |
8c2f7e86 | 175 | resource_size_t __nvdimm_namespace_capacity(struct nd_namespace_common *ndns); |
e1455744 DW |
176 | void nd_detach_ndns(struct device *dev, struct nd_namespace_common **_ndns); |
177 | void __nd_detach_ndns(struct device *dev, struct nd_namespace_common **_ndns); | |
178 | bool nd_attach_ndns(struct device *dev, struct nd_namespace_common *attach, | |
179 | struct nd_namespace_common **_ndns); | |
180 | bool __nd_attach_ndns(struct device *dev, struct nd_namespace_common *attach, | |
181 | struct nd_namespace_common **_ndns); | |
182 | ssize_t nd_namespace_store(struct device *dev, | |
183 | struct nd_namespace_common **_ndns, const char *buf, | |
184 | size_t len); | |
03dca343 | 185 | struct nd_pfn *to_nd_pfn_safe(struct device *dev); |
87a30e1f DW |
186 | bool is_nvdimm_bus(struct device *dev); |
187 | ||
188 | #ifdef CONFIG_PROVE_LOCKING | |
189 | extern struct class *nd_class; | |
190 | ||
191 | enum { | |
192 | LOCK_BUS, | |
193 | LOCK_NDCTL, | |
194 | LOCK_REGION, | |
195 | LOCK_DIMM = LOCK_REGION, | |
196 | LOCK_NAMESPACE, | |
197 | LOCK_CLAIM, | |
198 | }; | |
199 | ||
200 | static inline void debug_nvdimm_lock(struct device *dev) | |
201 | { | |
202 | if (is_nd_region(dev)) | |
203 | mutex_lock_nested(&dev->lockdep_mutex, LOCK_REGION); | |
204 | else if (is_nvdimm(dev)) | |
205 | mutex_lock_nested(&dev->lockdep_mutex, LOCK_DIMM); | |
206 | else if (is_nd_btt(dev) || is_nd_pfn(dev) || is_nd_dax(dev)) | |
207 | mutex_lock_nested(&dev->lockdep_mutex, LOCK_CLAIM); | |
208 | else if (dev->parent && (is_nd_region(dev->parent))) | |
209 | mutex_lock_nested(&dev->lockdep_mutex, LOCK_NAMESPACE); | |
210 | else if (is_nvdimm_bus(dev)) | |
211 | mutex_lock_nested(&dev->lockdep_mutex, LOCK_BUS); | |
212 | else if (dev->class && dev->class == nd_class) | |
213 | mutex_lock_nested(&dev->lockdep_mutex, LOCK_NDCTL); | |
214 | else | |
215 | dev_WARN(dev, "unknown lock level\n"); | |
216 | } | |
217 | ||
218 | static inline void debug_nvdimm_unlock(struct device *dev) | |
219 | { | |
220 | mutex_unlock(&dev->lockdep_mutex); | |
221 | } | |
222 | ||
223 | static inline void nd_device_lock(struct device *dev) | |
224 | { | |
225 | device_lock(dev); | |
226 | debug_nvdimm_lock(dev); | |
227 | } | |
228 | ||
229 | static inline void nd_device_unlock(struct device *dev) | |
230 | { | |
231 | debug_nvdimm_unlock(dev); | |
232 | device_unlock(dev); | |
233 | } | |
234 | #else | |
235 | static inline void nd_device_lock(struct device *dev) | |
236 | { | |
237 | device_lock(dev); | |
238 | } | |
239 | ||
240 | static inline void nd_device_unlock(struct device *dev) | |
241 | { | |
242 | device_unlock(dev); | |
243 | } | |
244 | ||
245 | static inline void debug_nvdimm_lock(struct device *dev) | |
246 | { | |
247 | } | |
248 | ||
249 | static inline void debug_nvdimm_unlock(struct device *dev) | |
250 | { | |
251 | } | |
252 | #endif | |
b94d5230 | 253 | #endif /* __ND_CORE_H__ */ |