Commit | Line | Data |
---|---|---|
5b497af4 | 1 | /* SPDX-License-Identifier: GPL-2.0-only */ |
b94d5230 DW |
2 | /* |
3 | * Copyright(c) 2013-2015 Intel Corporation. All rights reserved. | |
b94d5230 DW |
4 | */ |
5 | #ifndef __ND_CORE_H__ | |
6 | #define __ND_CORE_H__ | |
7 | #include <linux/libnvdimm.h> | |
8 | #include <linux/device.h> | |
eaf96153 DW |
9 | #include <linux/sizes.h> |
10 | #include <linux/mutex.h> | |
1b40e09a | 11 | #include <linux/nd.h> |
87a30e1f | 12 | #include "nd.h" |
b94d5230 | 13 | |
e6dfb2de DW |
14 | extern struct list_head nvdimm_bus_list; |
15 | extern struct mutex nvdimm_bus_list_mutex; | |
62232e45 | 16 | extern int nvdimm_major; |
7d988097 | 17 | extern struct workqueue_struct *nvdimm_wq; |
e6dfb2de | 18 | |
b94d5230 DW |
19 | struct nvdimm_bus { |
20 | struct nvdimm_bus_descriptor *nd_desc; | |
b70d31d0 | 21 | wait_queue_head_t wait; |
45def22c | 22 | struct list_head list; |
b94d5230 | 23 | struct device dev; |
eaf96153 | 24 | int id, probe_active; |
b70d31d0 | 25 | atomic_t ioctl_active; |
29b9aa0a | 26 | struct list_head mapping_list; |
3d88002e | 27 | struct mutex reconfig_mutex; |
aa9ad44a | 28 | struct badrange badrange; |
b94d5230 | 29 | }; |
45def22c | 30 | |
e6dfb2de DW |
31 | struct nvdimm { |
32 | unsigned long flags; | |
33 | void *provider_data; | |
e3654eca | 34 | unsigned long cmd_mask; |
e6dfb2de | 35 | struct device dev; |
eaf96153 | 36 | atomic_t busy; |
e5ae3b25 DW |
37 | int id, num_flush; |
38 | struct resource *flush_wpq; | |
d6548ae4 | 39 | const char *dimm_id; |
f2989396 DJ |
40 | struct { |
41 | const struct nvdimm_security_ops *ops; | |
d78c620a DW |
42 | unsigned long flags; |
43 | unsigned long ext_flags; | |
7d988097 DJ |
44 | unsigned int overwrite_tmo; |
45 | struct kernfs_node *overwrite_state; | |
f2989396 | 46 | } sec; |
7d988097 | 47 | struct delayed_work dwork; |
48001ea5 | 48 | const struct nvdimm_fw_ops *fw_ops; |
e6dfb2de DW |
49 | }; |
50 | ||
d78c620a | 51 | static inline unsigned long nvdimm_security_flags( |
faa8bd6e | 52 | struct nvdimm *nvdimm, enum nvdimm_passphrase_type ptype) |
f2989396 | 53 | { |
d78c620a DW |
54 | u64 flags; |
55 | const u64 state_flags = 1UL << NVDIMM_SECURITY_DISABLED | |
56 | | 1UL << NVDIMM_SECURITY_LOCKED | |
57 | | 1UL << NVDIMM_SECURITY_UNLOCKED | |
58 | | 1UL << NVDIMM_SECURITY_OVERWRITE; | |
59 | ||
f2989396 | 60 | if (!nvdimm->sec.ops) |
d78c620a | 61 | return 0; |
f2989396 | 62 | |
d78c620a DW |
63 | flags = nvdimm->sec.ops->get_flags(nvdimm, ptype); |
64 | /* disabled, locked, unlocked, and overwrite are mutually exclusive */ | |
65 | dev_WARN_ONCE(&nvdimm->dev, hweight64(flags & state_flags) > 1, | |
66 | "reported invalid security state: %#llx\n", | |
67 | (unsigned long long) flags); | |
68 | return flags; | |
f2989396 | 69 | } |
37833fb7 | 70 | int nvdimm_security_freeze(struct nvdimm *nvdimm); |
03b65b22 | 71 | #if IS_ENABLED(CONFIG_NVDIMM_KEYS) |
7b60422c | 72 | ssize_t nvdimm_security_store(struct device *dev, const char *buf, size_t len); |
7d988097 | 73 | void nvdimm_security_overwrite_query(struct work_struct *work); |
03b65b22 | 74 | #else |
7b60422c DW |
75 | static inline ssize_t nvdimm_security_store(struct device *dev, |
76 | const char *buf, size_t len) | |
7d988097 DJ |
77 | { |
78 | return -EOPNOTSUPP; | |
79 | } | |
80 | static inline void nvdimm_security_overwrite_query(struct work_struct *work) | |
81 | { | |
82 | } | |
03b65b22 | 83 | #endif |
f2989396 | 84 | |
762d067d DW |
85 | /** |
86 | * struct blk_alloc_info - tracking info for BLK dpa scanning | |
87 | * @nd_mapping: blk region mapping boundaries | |
88 | * @available: decremented in alias_dpa_busy as aliased PMEM is scanned | |
89 | * @busy: decremented in blk_dpa_busy to account for ranges already | |
90 | * handled by alias_dpa_busy | |
91 | * @res: alias_dpa_busy interprets this a free space range that needs to | |
92 | * be truncated to the valid BLK allocation starting DPA, blk_dpa_busy | |
93 | * treats it as a busy range that needs the aliased PMEM ranges | |
94 | * truncated. | |
95 | */ | |
96 | struct blk_alloc_info { | |
97 | struct nd_mapping *nd_mapping; | |
98 | resource_size_t available, busy; | |
99 | struct resource *res; | |
100 | }; | |
101 | ||
3d88002e | 102 | bool is_nvdimm(struct device *dev); |
3d88002e | 103 | bool is_nd_pmem(struct device *dev); |
c9e582aa | 104 | bool is_nd_volatile(struct device *dev); |
047fc8a1 | 105 | bool is_nd_blk(struct device *dev); |
c9e582aa DW |
106 | static inline bool is_nd_region(struct device *dev) |
107 | { | |
108 | return is_nd_pmem(dev) || is_nd_blk(dev) || is_nd_volatile(dev); | |
109 | } | |
110 | static inline bool is_memory(struct device *dev) | |
111 | { | |
112 | return is_nd_pmem(dev) || is_nd_volatile(dev); | |
113 | } | |
e6dfb2de | 114 | struct nvdimm_bus *walk_to_nvdimm_bus(struct device *nd_dev); |
45def22c | 115 | int __init nvdimm_bus_init(void); |
4d88a97a | 116 | void nvdimm_bus_exit(void); |
b354aba0 | 117 | void nvdimm_devs_exit(void); |
1b40e09a | 118 | struct nd_region; |
a2d1c7a6 | 119 | void nd_region_advance_seeds(struct nd_region *nd_region, struct device *dev); |
98a29c39 | 120 | void nd_region_create_ns_seed(struct nd_region *nd_region); |
8c2f7e86 | 121 | void nd_region_create_btt_seed(struct nd_region *nd_region); |
2dc43331 | 122 | void nd_region_create_pfn_seed(struct nd_region *nd_region); |
cd03412a | 123 | void nd_region_create_dax_seed(struct nd_region *nd_region); |
45def22c DW |
124 | int nvdimm_bus_create_ndctl(struct nvdimm_bus *nvdimm_bus); |
125 | void nvdimm_bus_destroy_ndctl(struct nvdimm_bus *nvdimm_bus); | |
4d88a97a | 126 | void nd_synchronize(void); |
8c2f7e86 | 127 | void __nd_device_register(struct device *dev); |
bf9bccc1 DW |
128 | struct nd_label_id; |
129 | char *nd_label_gen_id(struct nd_label_id *label_id, u8 *uuid, u32 flags); | |
130 | bool nd_is_uuid_unique(struct device *dev, u8 *uuid); | |
131 | struct nd_region; | |
132 | struct nvdimm_drvdata; | |
133 | struct nd_mapping; | |
ae8219f1 | 134 | void nd_mapping_free_labels(struct nd_mapping *nd_mapping); |
12e3129e KB |
135 | |
136 | int __reserve_free_pmem(struct device *dev, void *data); | |
137 | void release_free_pmem(struct nvdimm_bus *nvdimm_bus, | |
138 | struct nd_mapping *nd_mapping); | |
139 | ||
140 | resource_size_t nd_pmem_max_contiguous_dpa(struct nd_region *nd_region, | |
141 | struct nd_mapping *nd_mapping); | |
142 | resource_size_t nd_region_allocatable_dpa(struct nd_region *nd_region); | |
bf9bccc1 DW |
143 | resource_size_t nd_pmem_available_dpa(struct nd_region *nd_region, |
144 | struct nd_mapping *nd_mapping, resource_size_t *overlap); | |
a1f3e4d6 | 145 | resource_size_t nd_blk_available_dpa(struct nd_region *nd_region); |
bf9bccc1 | 146 | resource_size_t nd_region_available_dpa(struct nd_region *nd_region); |
ae86cbfe DW |
147 | int nd_region_conflict(struct nd_region *nd_region, resource_size_t start, |
148 | resource_size_t size); | |
bf9bccc1 DW |
149 | resource_size_t nvdimm_allocated_dpa(struct nvdimm_drvdata *ndd, |
150 | struct nd_label_id *label_id); | |
762d067d | 151 | int alias_dpa_busy(struct device *dev, void *data); |
1b40e09a DW |
152 | struct resource *nsblk_add_resource(struct nd_region *nd_region, |
153 | struct nvdimm_drvdata *ndd, struct nd_namespace_blk *nsblk, | |
154 | resource_size_t start); | |
0ba1c634 | 155 | int nvdimm_num_label_slots(struct nvdimm_drvdata *ndd); |
bf9bccc1 | 156 | void get_ndd(struct nvdimm_drvdata *ndd); |
8c2f7e86 | 157 | resource_size_t __nvdimm_namespace_capacity(struct nd_namespace_common *ndns); |
e1455744 DW |
158 | void nd_detach_ndns(struct device *dev, struct nd_namespace_common **_ndns); |
159 | void __nd_detach_ndns(struct device *dev, struct nd_namespace_common **_ndns); | |
160 | bool nd_attach_ndns(struct device *dev, struct nd_namespace_common *attach, | |
161 | struct nd_namespace_common **_ndns); | |
162 | bool __nd_attach_ndns(struct device *dev, struct nd_namespace_common *attach, | |
163 | struct nd_namespace_common **_ndns); | |
164 | ssize_t nd_namespace_store(struct device *dev, | |
165 | struct nd_namespace_common **_ndns, const char *buf, | |
166 | size_t len); | |
03dca343 | 167 | struct nd_pfn *to_nd_pfn_safe(struct device *dev); |
87a30e1f DW |
168 | bool is_nvdimm_bus(struct device *dev); |
169 | ||
8f4b01fc AK |
170 | #if IS_ENABLED(CONFIG_ND_CLAIM) |
171 | int devm_nsio_enable(struct device *dev, struct nd_namespace_io *nsio, | |
172 | resource_size_t size); | |
173 | void devm_nsio_disable(struct device *dev, struct nd_namespace_io *nsio); | |
174 | #else | |
175 | static inline int devm_nsio_enable(struct device *dev, | |
176 | struct nd_namespace_io *nsio, resource_size_t size) | |
177 | { | |
178 | return -ENXIO; | |
179 | } | |
180 | ||
181 | static inline void devm_nsio_disable(struct device *dev, | |
182 | struct nd_namespace_io *nsio) | |
183 | { | |
184 | } | |
185 | #endif | |
186 | ||
87a30e1f DW |
187 | #ifdef CONFIG_PROVE_LOCKING |
188 | extern struct class *nd_class; | |
189 | ||
190 | enum { | |
191 | LOCK_BUS, | |
192 | LOCK_NDCTL, | |
193 | LOCK_REGION, | |
194 | LOCK_DIMM = LOCK_REGION, | |
195 | LOCK_NAMESPACE, | |
196 | LOCK_CLAIM, | |
197 | }; | |
198 | ||
199 | static inline void debug_nvdimm_lock(struct device *dev) | |
200 | { | |
201 | if (is_nd_region(dev)) | |
202 | mutex_lock_nested(&dev->lockdep_mutex, LOCK_REGION); | |
203 | else if (is_nvdimm(dev)) | |
204 | mutex_lock_nested(&dev->lockdep_mutex, LOCK_DIMM); | |
205 | else if (is_nd_btt(dev) || is_nd_pfn(dev) || is_nd_dax(dev)) | |
206 | mutex_lock_nested(&dev->lockdep_mutex, LOCK_CLAIM); | |
207 | else if (dev->parent && (is_nd_region(dev->parent))) | |
208 | mutex_lock_nested(&dev->lockdep_mutex, LOCK_NAMESPACE); | |
209 | else if (is_nvdimm_bus(dev)) | |
210 | mutex_lock_nested(&dev->lockdep_mutex, LOCK_BUS); | |
211 | else if (dev->class && dev->class == nd_class) | |
212 | mutex_lock_nested(&dev->lockdep_mutex, LOCK_NDCTL); | |
213 | else | |
214 | dev_WARN(dev, "unknown lock level\n"); | |
215 | } | |
216 | ||
217 | static inline void debug_nvdimm_unlock(struct device *dev) | |
218 | { | |
219 | mutex_unlock(&dev->lockdep_mutex); | |
220 | } | |
221 | ||
222 | static inline void nd_device_lock(struct device *dev) | |
223 | { | |
224 | device_lock(dev); | |
225 | debug_nvdimm_lock(dev); | |
226 | } | |
227 | ||
228 | static inline void nd_device_unlock(struct device *dev) | |
229 | { | |
230 | debug_nvdimm_unlock(dev); | |
231 | device_unlock(dev); | |
232 | } | |
233 | #else | |
234 | static inline void nd_device_lock(struct device *dev) | |
235 | { | |
236 | device_lock(dev); | |
237 | } | |
238 | ||
239 | static inline void nd_device_unlock(struct device *dev) | |
240 | { | |
241 | device_unlock(dev); | |
242 | } | |
243 | ||
244 | static inline void debug_nvdimm_lock(struct device *dev) | |
245 | { | |
246 | } | |
247 | ||
248 | static inline void debug_nvdimm_unlock(struct device *dev) | |
249 | { | |
250 | } | |
251 | #endif | |
b94d5230 | 252 | #endif /* __ND_CORE_H__ */ |