Commit | Line | Data |
---|---|---|
5b497af4 | 1 | /* SPDX-License-Identifier: GPL-2.0-only */ |
b94d5230 DW |
2 | /* |
3 | * Copyright(c) 2013-2015 Intel Corporation. All rights reserved. | |
b94d5230 DW |
4 | */ |
5 | #ifndef __ND_CORE_H__ | |
6 | #define __ND_CORE_H__ | |
7 | #include <linux/libnvdimm.h> | |
8 | #include <linux/device.h> | |
eaf96153 DW |
9 | #include <linux/sizes.h> |
10 | #include <linux/mutex.h> | |
1b40e09a | 11 | #include <linux/nd.h> |
87a30e1f | 12 | #include "nd.h" |
b94d5230 | 13 | |
e6dfb2de DW |
14 | extern struct list_head nvdimm_bus_list; |
15 | extern struct mutex nvdimm_bus_list_mutex; | |
62232e45 | 16 | extern int nvdimm_major; |
7d988097 | 17 | extern struct workqueue_struct *nvdimm_wq; |
e6dfb2de | 18 | |
b94d5230 DW |
19 | struct nvdimm_bus { |
20 | struct nvdimm_bus_descriptor *nd_desc; | |
b70d31d0 | 21 | wait_queue_head_t wait; |
45def22c | 22 | struct list_head list; |
b94d5230 | 23 | struct device dev; |
eaf96153 | 24 | int id, probe_active; |
b70d31d0 | 25 | atomic_t ioctl_active; |
29b9aa0a | 26 | struct list_head mapping_list; |
3d88002e | 27 | struct mutex reconfig_mutex; |
aa9ad44a | 28 | struct badrange badrange; |
b94d5230 | 29 | }; |
45def22c | 30 | |
e6dfb2de DW |
31 | struct nvdimm { |
32 | unsigned long flags; | |
33 | void *provider_data; | |
e3654eca | 34 | unsigned long cmd_mask; |
e6dfb2de | 35 | struct device dev; |
eaf96153 | 36 | atomic_t busy; |
e5ae3b25 DW |
37 | int id, num_flush; |
38 | struct resource *flush_wpq; | |
d6548ae4 | 39 | const char *dimm_id; |
f2989396 DJ |
40 | struct { |
41 | const struct nvdimm_security_ops *ops; | |
d78c620a DW |
42 | unsigned long flags; |
43 | unsigned long ext_flags; | |
7d988097 DJ |
44 | unsigned int overwrite_tmo; |
45 | struct kernfs_node *overwrite_state; | |
f2989396 | 46 | } sec; |
7d988097 | 47 | struct delayed_work dwork; |
e6dfb2de DW |
48 | }; |
49 | ||
d78c620a | 50 | static inline unsigned long nvdimm_security_flags( |
faa8bd6e | 51 | struct nvdimm *nvdimm, enum nvdimm_passphrase_type ptype) |
f2989396 | 52 | { |
d78c620a DW |
53 | u64 flags; |
54 | const u64 state_flags = 1UL << NVDIMM_SECURITY_DISABLED | |
55 | | 1UL << NVDIMM_SECURITY_LOCKED | |
56 | | 1UL << NVDIMM_SECURITY_UNLOCKED | |
57 | | 1UL << NVDIMM_SECURITY_OVERWRITE; | |
58 | ||
f2989396 | 59 | if (!nvdimm->sec.ops) |
d78c620a | 60 | return 0; |
f2989396 | 61 | |
d78c620a DW |
62 | flags = nvdimm->sec.ops->get_flags(nvdimm, ptype); |
63 | /* disabled, locked, unlocked, and overwrite are mutually exclusive */ | |
64 | dev_WARN_ONCE(&nvdimm->dev, hweight64(flags & state_flags) > 1, | |
65 | "reported invalid security state: %#llx\n", | |
66 | (unsigned long long) flags); | |
67 | return flags; | |
f2989396 | 68 | } |
37833fb7 | 69 | int nvdimm_security_freeze(struct nvdimm *nvdimm); |
03b65b22 | 70 | #if IS_ENABLED(CONFIG_NVDIMM_KEYS) |
7b60422c | 71 | ssize_t nvdimm_security_store(struct device *dev, const char *buf, size_t len); |
7d988097 | 72 | void nvdimm_security_overwrite_query(struct work_struct *work); |
03b65b22 | 73 | #else |
7b60422c DW |
74 | static inline ssize_t nvdimm_security_store(struct device *dev, |
75 | const char *buf, size_t len) | |
7d988097 DJ |
76 | { |
77 | return -EOPNOTSUPP; | |
78 | } | |
79 | static inline void nvdimm_security_overwrite_query(struct work_struct *work) | |
80 | { | |
81 | } | |
03b65b22 | 82 | #endif |
f2989396 | 83 | |
762d067d DW |
84 | /** |
85 | * struct blk_alloc_info - tracking info for BLK dpa scanning | |
86 | * @nd_mapping: blk region mapping boundaries | |
87 | * @available: decremented in alias_dpa_busy as aliased PMEM is scanned | |
88 | * @busy: decremented in blk_dpa_busy to account for ranges already | |
89 | * handled by alias_dpa_busy | |
90 | * @res: alias_dpa_busy interprets this a free space range that needs to | |
91 | * be truncated to the valid BLK allocation starting DPA, blk_dpa_busy | |
92 | * treats it as a busy range that needs the aliased PMEM ranges | |
93 | * truncated. | |
94 | */ | |
95 | struct blk_alloc_info { | |
96 | struct nd_mapping *nd_mapping; | |
97 | resource_size_t available, busy; | |
98 | struct resource *res; | |
99 | }; | |
100 | ||
3d88002e | 101 | bool is_nvdimm(struct device *dev); |
3d88002e | 102 | bool is_nd_pmem(struct device *dev); |
c9e582aa | 103 | bool is_nd_volatile(struct device *dev); |
047fc8a1 | 104 | bool is_nd_blk(struct device *dev); |
c9e582aa DW |
105 | static inline bool is_nd_region(struct device *dev) |
106 | { | |
107 | return is_nd_pmem(dev) || is_nd_blk(dev) || is_nd_volatile(dev); | |
108 | } | |
109 | static inline bool is_memory(struct device *dev) | |
110 | { | |
111 | return is_nd_pmem(dev) || is_nd_volatile(dev); | |
112 | } | |
e6dfb2de | 113 | struct nvdimm_bus *walk_to_nvdimm_bus(struct device *nd_dev); |
45def22c | 114 | int __init nvdimm_bus_init(void); |
4d88a97a | 115 | void nvdimm_bus_exit(void); |
b354aba0 DW |
116 | void nvdimm_devs_exit(void); |
117 | void nd_region_devs_exit(void); | |
1b40e09a | 118 | struct nd_region; |
a2d1c7a6 | 119 | void nd_region_advance_seeds(struct nd_region *nd_region, struct device *dev); |
98a29c39 | 120 | void nd_region_create_ns_seed(struct nd_region *nd_region); |
8c2f7e86 | 121 | void nd_region_create_btt_seed(struct nd_region *nd_region); |
2dc43331 | 122 | void nd_region_create_pfn_seed(struct nd_region *nd_region); |
cd03412a | 123 | void nd_region_create_dax_seed(struct nd_region *nd_region); |
45def22c DW |
124 | int nvdimm_bus_create_ndctl(struct nvdimm_bus *nvdimm_bus); |
125 | void nvdimm_bus_destroy_ndctl(struct nvdimm_bus *nvdimm_bus); | |
4d88a97a | 126 | void nd_synchronize(void); |
1f7df6f8 DW |
127 | int nvdimm_bus_register_dimms(struct nvdimm_bus *nvdimm_bus); |
128 | int nvdimm_bus_register_regions(struct nvdimm_bus *nvdimm_bus); | |
eaf96153 | 129 | int nvdimm_bus_init_interleave_sets(struct nvdimm_bus *nvdimm_bus); |
8c2f7e86 | 130 | void __nd_device_register(struct device *dev); |
1f7df6f8 | 131 | int nd_match_dimm(struct device *dev, void *data); |
bf9bccc1 DW |
132 | struct nd_label_id; |
133 | char *nd_label_gen_id(struct nd_label_id *label_id, u8 *uuid, u32 flags); | |
134 | bool nd_is_uuid_unique(struct device *dev, u8 *uuid); | |
135 | struct nd_region; | |
136 | struct nvdimm_drvdata; | |
137 | struct nd_mapping; | |
ae8219f1 | 138 | void nd_mapping_free_labels(struct nd_mapping *nd_mapping); |
12e3129e KB |
139 | |
140 | int __reserve_free_pmem(struct device *dev, void *data); | |
141 | void release_free_pmem(struct nvdimm_bus *nvdimm_bus, | |
142 | struct nd_mapping *nd_mapping); | |
143 | ||
144 | resource_size_t nd_pmem_max_contiguous_dpa(struct nd_region *nd_region, | |
145 | struct nd_mapping *nd_mapping); | |
146 | resource_size_t nd_region_allocatable_dpa(struct nd_region *nd_region); | |
bf9bccc1 DW |
147 | resource_size_t nd_pmem_available_dpa(struct nd_region *nd_region, |
148 | struct nd_mapping *nd_mapping, resource_size_t *overlap); | |
a1f3e4d6 | 149 | resource_size_t nd_blk_available_dpa(struct nd_region *nd_region); |
bf9bccc1 | 150 | resource_size_t nd_region_available_dpa(struct nd_region *nd_region); |
ae86cbfe DW |
151 | int nd_region_conflict(struct nd_region *nd_region, resource_size_t start, |
152 | resource_size_t size); | |
bf9bccc1 DW |
153 | resource_size_t nvdimm_allocated_dpa(struct nvdimm_drvdata *ndd, |
154 | struct nd_label_id *label_id); | |
762d067d | 155 | int alias_dpa_busy(struct device *dev, void *data); |
1b40e09a DW |
156 | struct resource *nsblk_add_resource(struct nd_region *nd_region, |
157 | struct nvdimm_drvdata *ndd, struct nd_namespace_blk *nsblk, | |
158 | resource_size_t start); | |
0ba1c634 | 159 | int nvdimm_num_label_slots(struct nvdimm_drvdata *ndd); |
bf9bccc1 | 160 | void get_ndd(struct nvdimm_drvdata *ndd); |
8c2f7e86 | 161 | resource_size_t __nvdimm_namespace_capacity(struct nd_namespace_common *ndns); |
e1455744 DW |
162 | void nd_detach_ndns(struct device *dev, struct nd_namespace_common **_ndns); |
163 | void __nd_detach_ndns(struct device *dev, struct nd_namespace_common **_ndns); | |
164 | bool nd_attach_ndns(struct device *dev, struct nd_namespace_common *attach, | |
165 | struct nd_namespace_common **_ndns); | |
166 | bool __nd_attach_ndns(struct device *dev, struct nd_namespace_common *attach, | |
167 | struct nd_namespace_common **_ndns); | |
168 | ssize_t nd_namespace_store(struct device *dev, | |
169 | struct nd_namespace_common **_ndns, const char *buf, | |
170 | size_t len); | |
03dca343 | 171 | struct nd_pfn *to_nd_pfn_safe(struct device *dev); |
87a30e1f DW |
172 | bool is_nvdimm_bus(struct device *dev); |
173 | ||
174 | #ifdef CONFIG_PROVE_LOCKING | |
175 | extern struct class *nd_class; | |
176 | ||
177 | enum { | |
178 | LOCK_BUS, | |
179 | LOCK_NDCTL, | |
180 | LOCK_REGION, | |
181 | LOCK_DIMM = LOCK_REGION, | |
182 | LOCK_NAMESPACE, | |
183 | LOCK_CLAIM, | |
184 | }; | |
185 | ||
186 | static inline void debug_nvdimm_lock(struct device *dev) | |
187 | { | |
188 | if (is_nd_region(dev)) | |
189 | mutex_lock_nested(&dev->lockdep_mutex, LOCK_REGION); | |
190 | else if (is_nvdimm(dev)) | |
191 | mutex_lock_nested(&dev->lockdep_mutex, LOCK_DIMM); | |
192 | else if (is_nd_btt(dev) || is_nd_pfn(dev) || is_nd_dax(dev)) | |
193 | mutex_lock_nested(&dev->lockdep_mutex, LOCK_CLAIM); | |
194 | else if (dev->parent && (is_nd_region(dev->parent))) | |
195 | mutex_lock_nested(&dev->lockdep_mutex, LOCK_NAMESPACE); | |
196 | else if (is_nvdimm_bus(dev)) | |
197 | mutex_lock_nested(&dev->lockdep_mutex, LOCK_BUS); | |
198 | else if (dev->class && dev->class == nd_class) | |
199 | mutex_lock_nested(&dev->lockdep_mutex, LOCK_NDCTL); | |
200 | else | |
201 | dev_WARN(dev, "unknown lock level\n"); | |
202 | } | |
203 | ||
204 | static inline void debug_nvdimm_unlock(struct device *dev) | |
205 | { | |
206 | mutex_unlock(&dev->lockdep_mutex); | |
207 | } | |
208 | ||
209 | static inline void nd_device_lock(struct device *dev) | |
210 | { | |
211 | device_lock(dev); | |
212 | debug_nvdimm_lock(dev); | |
213 | } | |
214 | ||
215 | static inline void nd_device_unlock(struct device *dev) | |
216 | { | |
217 | debug_nvdimm_unlock(dev); | |
218 | device_unlock(dev); | |
219 | } | |
220 | #else | |
221 | static inline void nd_device_lock(struct device *dev) | |
222 | { | |
223 | device_lock(dev); | |
224 | } | |
225 | ||
226 | static inline void nd_device_unlock(struct device *dev) | |
227 | { | |
228 | device_unlock(dev); | |
229 | } | |
230 | ||
231 | static inline void debug_nvdimm_lock(struct device *dev) | |
232 | { | |
233 | } | |
234 | ||
235 | static inline void debug_nvdimm_unlock(struct device *dev) | |
236 | { | |
237 | } | |
238 | #endif | |
b94d5230 | 239 | #endif /* __ND_CORE_H__ */ |