Commit | Line | Data |
---|---|---|
5b497af4 | 1 | /* SPDX-License-Identifier: GPL-2.0-only */ |
b94d5230 DW |
2 | /* |
3 | * Copyright(c) 2013-2015 Intel Corporation. All rights reserved. | |
b94d5230 DW |
4 | */ |
5 | #ifndef __ND_CORE_H__ | |
6 | #define __ND_CORE_H__ | |
7 | #include <linux/libnvdimm.h> | |
8 | #include <linux/device.h> | |
eaf96153 DW |
9 | #include <linux/sizes.h> |
10 | #include <linux/mutex.h> | |
1b40e09a | 11 | #include <linux/nd.h> |
87a30e1f | 12 | #include "nd.h" |
b94d5230 | 13 | |
e6dfb2de DW |
14 | extern struct list_head nvdimm_bus_list; |
15 | extern struct mutex nvdimm_bus_list_mutex; | |
62232e45 | 16 | extern int nvdimm_major; |
7d988097 | 17 | extern struct workqueue_struct *nvdimm_wq; |
e6dfb2de | 18 | |
b94d5230 DW |
19 | struct nvdimm_bus { |
20 | struct nvdimm_bus_descriptor *nd_desc; | |
b70d31d0 | 21 | wait_queue_head_t wait; |
45def22c | 22 | struct list_head list; |
b94d5230 | 23 | struct device dev; |
eaf96153 | 24 | int id, probe_active; |
b70d31d0 | 25 | atomic_t ioctl_active; |
29b9aa0a | 26 | struct list_head mapping_list; |
3d88002e | 27 | struct mutex reconfig_mutex; |
aa9ad44a | 28 | struct badrange badrange; |
b94d5230 | 29 | }; |
45def22c | 30 | |
e6dfb2de DW |
31 | struct nvdimm { |
32 | unsigned long flags; | |
33 | void *provider_data; | |
e3654eca | 34 | unsigned long cmd_mask; |
e6dfb2de | 35 | struct device dev; |
eaf96153 | 36 | atomic_t busy; |
e5ae3b25 DW |
37 | int id, num_flush; |
38 | struct resource *flush_wpq; | |
d6548ae4 | 39 | const char *dimm_id; |
f2989396 DJ |
40 | struct { |
41 | const struct nvdimm_security_ops *ops; | |
d78c620a DW |
42 | unsigned long flags; |
43 | unsigned long ext_flags; | |
7d988097 DJ |
44 | unsigned int overwrite_tmo; |
45 | struct kernfs_node *overwrite_state; | |
f2989396 | 46 | } sec; |
7d988097 | 47 | struct delayed_work dwork; |
48001ea5 | 48 | const struct nvdimm_fw_ops *fw_ops; |
e6dfb2de DW |
49 | }; |
50 | ||
d78c620a | 51 | static inline unsigned long nvdimm_security_flags( |
faa8bd6e | 52 | struct nvdimm *nvdimm, enum nvdimm_passphrase_type ptype) |
f2989396 | 53 | { |
d78c620a DW |
54 | u64 flags; |
55 | const u64 state_flags = 1UL << NVDIMM_SECURITY_DISABLED | |
56 | | 1UL << NVDIMM_SECURITY_LOCKED | |
57 | | 1UL << NVDIMM_SECURITY_UNLOCKED | |
58 | | 1UL << NVDIMM_SECURITY_OVERWRITE; | |
59 | ||
f2989396 | 60 | if (!nvdimm->sec.ops) |
d78c620a | 61 | return 0; |
f2989396 | 62 | |
d78c620a DW |
63 | flags = nvdimm->sec.ops->get_flags(nvdimm, ptype); |
64 | /* disabled, locked, unlocked, and overwrite are mutually exclusive */ | |
65 | dev_WARN_ONCE(&nvdimm->dev, hweight64(flags & state_flags) > 1, | |
66 | "reported invalid security state: %#llx\n", | |
67 | (unsigned long long) flags); | |
68 | return flags; | |
f2989396 | 69 | } |
37833fb7 | 70 | int nvdimm_security_freeze(struct nvdimm *nvdimm); |
03b65b22 | 71 | #if IS_ENABLED(CONFIG_NVDIMM_KEYS) |
7b60422c | 72 | ssize_t nvdimm_security_store(struct device *dev, const char *buf, size_t len); |
7d988097 | 73 | void nvdimm_security_overwrite_query(struct work_struct *work); |
03b65b22 | 74 | #else |
7b60422c DW |
75 | static inline ssize_t nvdimm_security_store(struct device *dev, |
76 | const char *buf, size_t len) | |
7d988097 DJ |
77 | { |
78 | return -EOPNOTSUPP; | |
79 | } | |
80 | static inline void nvdimm_security_overwrite_query(struct work_struct *work) | |
81 | { | |
82 | } | |
03b65b22 | 83 | #endif |
f2989396 | 84 | |
3d88002e | 85 | bool is_nvdimm(struct device *dev); |
3d88002e | 86 | bool is_nd_pmem(struct device *dev); |
c9e582aa | 87 | bool is_nd_volatile(struct device *dev); |
c9e582aa DW |
88 | static inline bool is_nd_region(struct device *dev) |
89 | { | |
3b6c6c03 | 90 | return is_nd_pmem(dev) || is_nd_volatile(dev); |
c9e582aa DW |
91 | } |
92 | static inline bool is_memory(struct device *dev) | |
93 | { | |
94 | return is_nd_pmem(dev) || is_nd_volatile(dev); | |
95 | } | |
e6dfb2de | 96 | struct nvdimm_bus *walk_to_nvdimm_bus(struct device *nd_dev); |
45def22c | 97 | int __init nvdimm_bus_init(void); |
4d88a97a | 98 | void nvdimm_bus_exit(void); |
b354aba0 | 99 | void nvdimm_devs_exit(void); |
1b40e09a | 100 | struct nd_region; |
a2d1c7a6 | 101 | void nd_region_advance_seeds(struct nd_region *nd_region, struct device *dev); |
98a29c39 | 102 | void nd_region_create_ns_seed(struct nd_region *nd_region); |
8c2f7e86 | 103 | void nd_region_create_btt_seed(struct nd_region *nd_region); |
2dc43331 | 104 | void nd_region_create_pfn_seed(struct nd_region *nd_region); |
cd03412a | 105 | void nd_region_create_dax_seed(struct nd_region *nd_region); |
45def22c DW |
106 | int nvdimm_bus_create_ndctl(struct nvdimm_bus *nvdimm_bus); |
107 | void nvdimm_bus_destroy_ndctl(struct nvdimm_bus *nvdimm_bus); | |
4d88a97a | 108 | void nd_synchronize(void); |
8c2f7e86 | 109 | void __nd_device_register(struct device *dev); |
bf9bccc1 | 110 | struct nd_label_id; |
d1c6e08e DW |
111 | char *nd_label_gen_id(struct nd_label_id *label_id, const uuid_t *uuid, |
112 | u32 flags); | |
113 | bool nd_is_uuid_unique(struct device *dev, uuid_t *uuid); | |
bf9bccc1 DW |
114 | struct nd_region; |
115 | struct nvdimm_drvdata; | |
116 | struct nd_mapping; | |
ae8219f1 | 117 | void nd_mapping_free_labels(struct nd_mapping *nd_mapping); |
12e3129e KB |
118 | |
119 | int __reserve_free_pmem(struct device *dev, void *data); | |
120 | void release_free_pmem(struct nvdimm_bus *nvdimm_bus, | |
121 | struct nd_mapping *nd_mapping); | |
122 | ||
123 | resource_size_t nd_pmem_max_contiguous_dpa(struct nd_region *nd_region, | |
124 | struct nd_mapping *nd_mapping); | |
125 | resource_size_t nd_region_allocatable_dpa(struct nd_region *nd_region); | |
bf9bccc1 | 126 | resource_size_t nd_pmem_available_dpa(struct nd_region *nd_region, |
3b6c6c03 | 127 | struct nd_mapping *nd_mapping); |
bf9bccc1 | 128 | resource_size_t nd_region_available_dpa(struct nd_region *nd_region); |
ae86cbfe DW |
129 | int nd_region_conflict(struct nd_region *nd_region, resource_size_t start, |
130 | resource_size_t size); | |
bf9bccc1 DW |
131 | resource_size_t nvdimm_allocated_dpa(struct nvdimm_drvdata *ndd, |
132 | struct nd_label_id *label_id); | |
0ba1c634 | 133 | int nvdimm_num_label_slots(struct nvdimm_drvdata *ndd); |
bf9bccc1 | 134 | void get_ndd(struct nvdimm_drvdata *ndd); |
8c2f7e86 | 135 | resource_size_t __nvdimm_namespace_capacity(struct nd_namespace_common *ndns); |
e1455744 DW |
136 | void nd_detach_ndns(struct device *dev, struct nd_namespace_common **_ndns); |
137 | void __nd_detach_ndns(struct device *dev, struct nd_namespace_common **_ndns); | |
138 | bool nd_attach_ndns(struct device *dev, struct nd_namespace_common *attach, | |
139 | struct nd_namespace_common **_ndns); | |
140 | bool __nd_attach_ndns(struct device *dev, struct nd_namespace_common *attach, | |
141 | struct nd_namespace_common **_ndns); | |
142 | ssize_t nd_namespace_store(struct device *dev, | |
143 | struct nd_namespace_common **_ndns, const char *buf, | |
144 | size_t len); | |
03dca343 | 145 | struct nd_pfn *to_nd_pfn_safe(struct device *dev); |
87a30e1f DW |
146 | bool is_nvdimm_bus(struct device *dev); |
147 | ||
8f4b01fc AK |
148 | #if IS_ENABLED(CONFIG_ND_CLAIM) |
149 | int devm_nsio_enable(struct device *dev, struct nd_namespace_io *nsio, | |
150 | resource_size_t size); | |
151 | void devm_nsio_disable(struct device *dev, struct nd_namespace_io *nsio); | |
152 | #else | |
153 | static inline int devm_nsio_enable(struct device *dev, | |
154 | struct nd_namespace_io *nsio, resource_size_t size) | |
155 | { | |
156 | return -ENXIO; | |
157 | } | |
158 | ||
159 | static inline void devm_nsio_disable(struct device *dev, | |
160 | struct nd_namespace_io *nsio) | |
161 | { | |
162 | } | |
163 | #endif | |
164 | ||
3c5b9039 | 165 | #ifdef CONFIG_PROVE_NVDIMM_LOCKING |
87a30e1f DW |
166 | extern struct class *nd_class; |
167 | ||
168 | enum { | |
169 | LOCK_BUS, | |
170 | LOCK_NDCTL, | |
171 | LOCK_REGION, | |
172 | LOCK_DIMM = LOCK_REGION, | |
173 | LOCK_NAMESPACE, | |
174 | LOCK_CLAIM, | |
175 | }; | |
176 | ||
177 | static inline void debug_nvdimm_lock(struct device *dev) | |
178 | { | |
179 | if (is_nd_region(dev)) | |
180 | mutex_lock_nested(&dev->lockdep_mutex, LOCK_REGION); | |
181 | else if (is_nvdimm(dev)) | |
182 | mutex_lock_nested(&dev->lockdep_mutex, LOCK_DIMM); | |
183 | else if (is_nd_btt(dev) || is_nd_pfn(dev) || is_nd_dax(dev)) | |
184 | mutex_lock_nested(&dev->lockdep_mutex, LOCK_CLAIM); | |
185 | else if (dev->parent && (is_nd_region(dev->parent))) | |
186 | mutex_lock_nested(&dev->lockdep_mutex, LOCK_NAMESPACE); | |
187 | else if (is_nvdimm_bus(dev)) | |
188 | mutex_lock_nested(&dev->lockdep_mutex, LOCK_BUS); | |
189 | else if (dev->class && dev->class == nd_class) | |
190 | mutex_lock_nested(&dev->lockdep_mutex, LOCK_NDCTL); | |
191 | else | |
192 | dev_WARN(dev, "unknown lock level\n"); | |
193 | } | |
194 | ||
195 | static inline void debug_nvdimm_unlock(struct device *dev) | |
196 | { | |
197 | mutex_unlock(&dev->lockdep_mutex); | |
198 | } | |
199 | ||
200 | static inline void nd_device_lock(struct device *dev) | |
201 | { | |
202 | device_lock(dev); | |
203 | debug_nvdimm_lock(dev); | |
204 | } | |
205 | ||
206 | static inline void nd_device_unlock(struct device *dev) | |
207 | { | |
208 | debug_nvdimm_unlock(dev); | |
209 | device_unlock(dev); | |
210 | } | |
211 | #else | |
212 | static inline void nd_device_lock(struct device *dev) | |
213 | { | |
214 | device_lock(dev); | |
215 | } | |
216 | ||
217 | static inline void nd_device_unlock(struct device *dev) | |
218 | { | |
219 | device_unlock(dev); | |
220 | } | |
221 | ||
222 | static inline void debug_nvdimm_lock(struct device *dev) | |
223 | { | |
224 | } | |
225 | ||
226 | static inline void debug_nvdimm_unlock(struct device *dev) | |
227 | { | |
228 | } | |
229 | #endif | |
b94d5230 | 230 | #endif /* __ND_CORE_H__ */ |