Commit | Line | Data |
---|---|---|
cd9e9808 MB |
1 | #ifndef NVM_H |
2 | #define NVM_H | |
3 | ||
4 | enum { | |
5 | NVM_IO_OK = 0, | |
6 | NVM_IO_REQUEUE = 1, | |
7 | NVM_IO_DONE = 2, | |
8 | NVM_IO_ERR = 3, | |
9 | ||
10 | NVM_IOTYPE_NONE = 0, | |
11 | NVM_IOTYPE_GC = 1, | |
12 | }; | |
13 | ||
14 | #ifdef CONFIG_NVM | |
15 | ||
16 | #include <linux/blkdev.h> | |
17 | #include <linux/types.h> | |
18 | #include <linux/file.h> | |
19 | #include <linux/dmapool.h> | |
20 | ||
21 | enum { | |
22 | /* HW Responsibilities */ | |
23 | NVM_RSP_L2P = 1 << 0, | |
24 | NVM_RSP_ECC = 1 << 1, | |
25 | ||
26 | /* Physical Adressing Mode */ | |
27 | NVM_ADDRMODE_LINEAR = 0, | |
28 | NVM_ADDRMODE_CHANNEL = 1, | |
29 | ||
30 | /* Plane programming mode for LUN */ | |
31 | NVM_PLANE_SINGLE = 0, | |
32 | NVM_PLANE_DOUBLE = 1, | |
33 | NVM_PLANE_QUAD = 2, | |
34 | ||
35 | /* Status codes */ | |
36 | NVM_RSP_SUCCESS = 0x0, | |
37 | NVM_RSP_NOT_CHANGEABLE = 0x1, | |
38 | NVM_RSP_ERR_FAILWRITE = 0x40ff, | |
39 | NVM_RSP_ERR_EMPTYPAGE = 0x42ff, | |
40 | ||
41 | /* Device opcodes */ | |
42 | NVM_OP_HBREAD = 0x02, | |
43 | NVM_OP_HBWRITE = 0x81, | |
44 | NVM_OP_PWRITE = 0x91, | |
45 | NVM_OP_PREAD = 0x92, | |
46 | NVM_OP_ERASE = 0x90, | |
47 | ||
48 | /* PPA Command Flags */ | |
49 | NVM_IO_SNGL_ACCESS = 0x0, | |
50 | NVM_IO_DUAL_ACCESS = 0x1, | |
51 | NVM_IO_QUAD_ACCESS = 0x2, | |
52 | ||
53 | NVM_IO_SUSPEND = 0x80, | |
54 | NVM_IO_SLC_MODE = 0x100, | |
55 | NVM_IO_SCRAMBLE_DISABLE = 0x200, | |
56 | }; | |
57 | ||
58 | struct nvm_id_group { | |
59 | u8 mtype; | |
60 | u8 fmtype; | |
cd9e9808 MB |
61 | u8 num_ch; |
62 | u8 num_lun; | |
63 | u8 num_pln; | |
64 | u16 num_blk; | |
65 | u16 num_pg; | |
66 | u16 fpg_sz; | |
67 | u16 csecs; | |
68 | u16 sos; | |
69 | u32 trdt; | |
70 | u32 trdm; | |
71 | u32 tprt; | |
72 | u32 tprm; | |
73 | u32 tbet; | |
74 | u32 tbem; | |
75 | u32 mpos; | |
12be5edf | 76 | u32 mccap; |
cd9e9808 | 77 | u16 cpar; |
73387e7b | 78 | }; |
cd9e9808 MB |
79 | |
80 | struct nvm_addr_format { | |
81 | u8 ch_offset; | |
82 | u8 ch_len; | |
83 | u8 lun_offset; | |
84 | u8 lun_len; | |
85 | u8 pln_offset; | |
86 | u8 pln_len; | |
87 | u8 blk_offset; | |
88 | u8 blk_len; | |
89 | u8 pg_offset; | |
90 | u8 pg_len; | |
91 | u8 sect_offset; | |
92 | u8 sect_len; | |
cd9e9808 MB |
93 | }; |
94 | ||
95 | struct nvm_id { | |
96 | u8 ver_id; | |
97 | u8 vmnt; | |
98 | u8 cgrps; | |
cd9e9808 MB |
99 | u32 cap; |
100 | u32 dom; | |
101 | struct nvm_addr_format ppaf; | |
cd9e9808 MB |
102 | struct nvm_id_group groups[4]; |
103 | } __packed; | |
104 | ||
105 | struct nvm_target { | |
106 | struct list_head list; | |
107 | struct nvm_tgt_type *type; | |
108 | struct gendisk *disk; | |
109 | }; | |
110 | ||
111 | struct nvm_tgt_instance { | |
112 | struct nvm_tgt_type *tt; | |
113 | }; | |
114 | ||
115 | #define ADDR_EMPTY (~0ULL) | |
116 | ||
117 | #define NVM_VERSION_MAJOR 1 | |
118 | #define NVM_VERSION_MINOR 0 | |
119 | #define NVM_VERSION_PATCH 0 | |
120 | ||
cd9e9808 | 121 | #define NVM_BLK_BITS (16) |
7386af27 MB |
122 | #define NVM_PG_BITS (16) |
123 | #define NVM_SEC_BITS (8) | |
124 | #define NVM_PL_BITS (8) | |
125 | #define NVM_LUN_BITS (8) | |
cd9e9808 MB |
126 | #define NVM_CH_BITS (8) |
127 | ||
128 | struct ppa_addr { | |
7386af27 | 129 | /* Generic structure for all addresses */ |
cd9e9808 | 130 | union { |
cd9e9808 | 131 | struct { |
7386af27 MB |
132 | u64 blk : NVM_BLK_BITS; |
133 | u64 pg : NVM_PG_BITS; | |
b7ceb7d5 MB |
134 | u64 sec : NVM_SEC_BITS; |
135 | u64 pl : NVM_PL_BITS; | |
b7ceb7d5 MB |
136 | u64 lun : NVM_LUN_BITS; |
137 | u64 ch : NVM_CH_BITS; | |
cd9e9808 MB |
138 | } g; |
139 | ||
b7ceb7d5 | 140 | u64 ppa; |
cd9e9808 | 141 | }; |
7386af27 | 142 | }; |
cd9e9808 MB |
143 | |
144 | struct nvm_rq { | |
145 | struct nvm_tgt_instance *ins; | |
146 | struct nvm_dev *dev; | |
147 | ||
148 | struct bio *bio; | |
149 | ||
150 | union { | |
151 | struct ppa_addr ppa_addr; | |
152 | dma_addr_t dma_ppa_list; | |
153 | }; | |
154 | ||
155 | struct ppa_addr *ppa_list; | |
156 | ||
157 | void *metadata; | |
158 | dma_addr_t dma_metadata; | |
159 | ||
160 | uint8_t opcode; | |
161 | uint16_t nr_pages; | |
162 | uint16_t flags; | |
163 | }; | |
164 | ||
165 | static inline struct nvm_rq *nvm_rq_from_pdu(void *pdu) | |
166 | { | |
167 | return pdu - sizeof(struct nvm_rq); | |
168 | } | |
169 | ||
170 | static inline void *nvm_rq_to_pdu(struct nvm_rq *rqdata) | |
171 | { | |
172 | return rqdata + 1; | |
173 | } | |
174 | ||
175 | struct nvm_block; | |
176 | ||
177 | typedef int (nvm_l2p_update_fn)(u64, u32, __le64 *, void *); | |
11450469 | 178 | typedef int (nvm_bb_update_fn)(struct ppa_addr, int, u8 *, void *); |
cd9e9808 MB |
179 | typedef int (nvm_id_fn)(struct request_queue *, struct nvm_id *); |
180 | typedef int (nvm_get_l2p_tbl_fn)(struct request_queue *, u64, u32, | |
181 | nvm_l2p_update_fn *, void *); | |
11450469 | 182 | typedef int (nvm_op_bb_tbl_fn)(struct request_queue *, struct ppa_addr, int, |
cd9e9808 MB |
183 | nvm_bb_update_fn *, void *); |
184 | typedef int (nvm_op_set_bb_fn)(struct request_queue *, struct nvm_rq *, int); | |
185 | typedef int (nvm_submit_io_fn)(struct request_queue *, struct nvm_rq *); | |
186 | typedef int (nvm_erase_blk_fn)(struct request_queue *, struct nvm_rq *); | |
187 | typedef void *(nvm_create_dma_pool_fn)(struct request_queue *, char *); | |
188 | typedef void (nvm_destroy_dma_pool_fn)(void *); | |
189 | typedef void *(nvm_dev_dma_alloc_fn)(struct request_queue *, void *, gfp_t, | |
190 | dma_addr_t *); | |
191 | typedef void (nvm_dev_dma_free_fn)(void *, void*, dma_addr_t); | |
192 | ||
193 | struct nvm_dev_ops { | |
194 | nvm_id_fn *identity; | |
195 | nvm_get_l2p_tbl_fn *get_l2p_tbl; | |
196 | nvm_op_bb_tbl_fn *get_bb_tbl; | |
11450469 | 197 | nvm_op_set_bb_fn *set_bb_tbl; |
cd9e9808 MB |
198 | |
199 | nvm_submit_io_fn *submit_io; | |
200 | nvm_erase_blk_fn *erase_block; | |
201 | ||
202 | nvm_create_dma_pool_fn *create_dma_pool; | |
203 | nvm_destroy_dma_pool_fn *destroy_dma_pool; | |
204 | nvm_dev_dma_alloc_fn *dev_dma_alloc; | |
205 | nvm_dev_dma_free_fn *dev_dma_free; | |
206 | ||
aedf17f4 | 207 | unsigned int max_phys_sect; |
cd9e9808 MB |
208 | }; |
209 | ||
210 | struct nvm_lun { | |
211 | int id; | |
212 | ||
213 | int lun_id; | |
214 | int chnl_id; | |
215 | ||
216 | unsigned int nr_free_blocks; /* Number of unused blocks */ | |
217 | struct nvm_block *blocks; | |
218 | ||
219 | spinlock_t lock; | |
220 | }; | |
221 | ||
222 | struct nvm_block { | |
223 | struct list_head list; | |
224 | struct nvm_lun *lun; | |
225 | unsigned long id; | |
226 | ||
227 | void *priv; | |
228 | int type; | |
229 | }; | |
230 | ||
231 | struct nvm_dev { | |
232 | struct nvm_dev_ops *ops; | |
233 | ||
234 | struct list_head devices; | |
235 | struct list_head online_targets; | |
236 | ||
237 | /* Media manager */ | |
238 | struct nvmm_type *mt; | |
239 | void *mp; | |
240 | ||
241 | /* Device information */ | |
242 | int nr_chnls; | |
243 | int nr_planes; | |
244 | int luns_per_chnl; | |
245 | int sec_per_pg; /* only sectors for a single page */ | |
246 | int pgs_per_blk; | |
247 | int blks_per_lun; | |
248 | int sec_size; | |
249 | int oob_size; | |
7386af27 | 250 | struct nvm_addr_format ppaf; |
cd9e9808 MB |
251 | |
252 | /* Calculated/Cached values. These do not reflect the actual usable | |
253 | * blocks at run-time. | |
254 | */ | |
255 | int max_rq_size; | |
256 | int plane_mode; /* drive device in single, double or quad mode */ | |
257 | ||
258 | int sec_per_pl; /* all sectors across planes */ | |
259 | int sec_per_blk; | |
260 | int sec_per_lun; | |
261 | ||
262 | unsigned long total_pages; | |
263 | unsigned long total_blocks; | |
264 | int nr_luns; | |
265 | unsigned max_pages_per_blk; | |
266 | ||
267 | void *ppalist_pool; | |
268 | ||
269 | struct nvm_id identity; | |
270 | ||
271 | /* Backend device */ | |
272 | struct request_queue *q; | |
273 | char name[DISK_NAME_LEN]; | |
274 | }; | |
275 | ||
7386af27 MB |
276 | static inline struct ppa_addr generic_to_dev_addr(struct nvm_dev *dev, |
277 | struct ppa_addr r) | |
cd9e9808 MB |
278 | { |
279 | struct ppa_addr l; | |
280 | ||
7386af27 MB |
281 | l.ppa = ((u64)r.g.blk) << dev->ppaf.blk_offset; |
282 | l.ppa |= ((u64)r.g.pg) << dev->ppaf.pg_offset; | |
283 | l.ppa |= ((u64)r.g.sec) << dev->ppaf.sect_offset; | |
284 | l.ppa |= ((u64)r.g.pl) << dev->ppaf.pln_offset; | |
285 | l.ppa |= ((u64)r.g.lun) << dev->ppaf.lun_offset; | |
286 | l.ppa |= ((u64)r.g.ch) << dev->ppaf.ch_offset; | |
cd9e9808 MB |
287 | |
288 | return l; | |
289 | } | |
290 | ||
7386af27 MB |
291 | static inline struct ppa_addr dev_to_generic_addr(struct nvm_dev *dev, |
292 | struct ppa_addr r) | |
cd9e9808 MB |
293 | { |
294 | struct ppa_addr l; | |
295 | ||
7386af27 MB |
296 | /* |
297 | * (r.ppa << X offset) & X len bitmask. X eq. blk, pg, etc. | |
298 | */ | |
299 | l.g.blk = (r.ppa >> dev->ppaf.blk_offset) & | |
300 | (((1 << dev->ppaf.blk_len) - 1)); | |
301 | l.g.pg |= (r.ppa >> dev->ppaf.pg_offset) & | |
302 | (((1 << dev->ppaf.pg_len) - 1)); | |
303 | l.g.sec |= (r.ppa >> dev->ppaf.sect_offset) & | |
304 | (((1 << dev->ppaf.sect_len) - 1)); | |
305 | l.g.pl |= (r.ppa >> dev->ppaf.pln_offset) & | |
306 | (((1 << dev->ppaf.pln_len) - 1)); | |
307 | l.g.lun |= (r.ppa >> dev->ppaf.lun_offset) & | |
308 | (((1 << dev->ppaf.lun_len) - 1)); | |
309 | l.g.ch |= (r.ppa >> dev->ppaf.ch_offset) & | |
310 | (((1 << dev->ppaf.ch_len) - 1)); | |
cd9e9808 MB |
311 | |
312 | return l; | |
313 | } | |
314 | ||
cd9e9808 MB |
315 | static inline int ppa_empty(struct ppa_addr ppa_addr) |
316 | { | |
317 | return (ppa_addr.ppa == ADDR_EMPTY); | |
318 | } | |
319 | ||
320 | static inline void ppa_set_empty(struct ppa_addr *ppa_addr) | |
321 | { | |
322 | ppa_addr->ppa = ADDR_EMPTY; | |
323 | } | |
324 | ||
325 | static inline struct ppa_addr block_to_ppa(struct nvm_dev *dev, | |
326 | struct nvm_block *blk) | |
327 | { | |
328 | struct ppa_addr ppa; | |
329 | struct nvm_lun *lun = blk->lun; | |
330 | ||
331 | ppa.ppa = 0; | |
332 | ppa.g.blk = blk->id % dev->blks_per_lun; | |
333 | ppa.g.lun = lun->lun_id; | |
334 | ppa.g.ch = lun->chnl_id; | |
335 | ||
336 | return ppa; | |
337 | } | |
338 | ||
dece1635 | 339 | typedef blk_qc_t (nvm_tgt_make_rq_fn)(struct request_queue *, struct bio *); |
cd9e9808 MB |
340 | typedef sector_t (nvm_tgt_capacity_fn)(void *); |
341 | typedef int (nvm_tgt_end_io_fn)(struct nvm_rq *, int); | |
342 | typedef void *(nvm_tgt_init_fn)(struct nvm_dev *, struct gendisk *, int, int); | |
343 | typedef void (nvm_tgt_exit_fn)(void *); | |
344 | ||
345 | struct nvm_tgt_type { | |
346 | const char *name; | |
347 | unsigned int version[3]; | |
348 | ||
349 | /* target entry points */ | |
350 | nvm_tgt_make_rq_fn *make_rq; | |
351 | nvm_tgt_capacity_fn *capacity; | |
352 | nvm_tgt_end_io_fn *end_io; | |
353 | ||
354 | /* module-specific init/teardown */ | |
355 | nvm_tgt_init_fn *init; | |
356 | nvm_tgt_exit_fn *exit; | |
357 | ||
358 | /* For internal use */ | |
359 | struct list_head list; | |
360 | }; | |
361 | ||
362 | extern int nvm_register_target(struct nvm_tgt_type *); | |
363 | extern void nvm_unregister_target(struct nvm_tgt_type *); | |
364 | ||
365 | extern void *nvm_dev_dma_alloc(struct nvm_dev *, gfp_t, dma_addr_t *); | |
366 | extern void nvm_dev_dma_free(struct nvm_dev *, void *, dma_addr_t); | |
367 | ||
368 | typedef int (nvmm_register_fn)(struct nvm_dev *); | |
369 | typedef void (nvmm_unregister_fn)(struct nvm_dev *); | |
370 | typedef struct nvm_block *(nvmm_get_blk_fn)(struct nvm_dev *, | |
371 | struct nvm_lun *, unsigned long); | |
372 | typedef void (nvmm_put_blk_fn)(struct nvm_dev *, struct nvm_block *); | |
373 | typedef int (nvmm_open_blk_fn)(struct nvm_dev *, struct nvm_block *); | |
374 | typedef int (nvmm_close_blk_fn)(struct nvm_dev *, struct nvm_block *); | |
375 | typedef void (nvmm_flush_blk_fn)(struct nvm_dev *, struct nvm_block *); | |
376 | typedef int (nvmm_submit_io_fn)(struct nvm_dev *, struct nvm_rq *); | |
377 | typedef int (nvmm_end_io_fn)(struct nvm_rq *, int); | |
378 | typedef int (nvmm_erase_blk_fn)(struct nvm_dev *, struct nvm_block *, | |
379 | unsigned long); | |
380 | typedef struct nvm_lun *(nvmm_get_lun_fn)(struct nvm_dev *, int); | |
381 | typedef void (nvmm_free_blocks_print_fn)(struct nvm_dev *); | |
382 | ||
383 | struct nvmm_type { | |
384 | const char *name; | |
385 | unsigned int version[3]; | |
386 | ||
387 | nvmm_register_fn *register_mgr; | |
388 | nvmm_unregister_fn *unregister_mgr; | |
389 | ||
390 | /* Block administration callbacks */ | |
391 | nvmm_get_blk_fn *get_blk; | |
392 | nvmm_put_blk_fn *put_blk; | |
393 | nvmm_open_blk_fn *open_blk; | |
394 | nvmm_close_blk_fn *close_blk; | |
395 | nvmm_flush_blk_fn *flush_blk; | |
396 | ||
397 | nvmm_submit_io_fn *submit_io; | |
398 | nvmm_end_io_fn *end_io; | |
399 | nvmm_erase_blk_fn *erase_blk; | |
400 | ||
401 | /* Configuration management */ | |
402 | nvmm_get_lun_fn *get_lun; | |
403 | ||
404 | /* Statistics */ | |
405 | nvmm_free_blocks_print_fn *free_blocks_print; | |
406 | struct list_head list; | |
407 | }; | |
408 | ||
409 | extern int nvm_register_mgr(struct nvmm_type *); | |
410 | extern void nvm_unregister_mgr(struct nvmm_type *); | |
411 | ||
412 | extern struct nvm_block *nvm_get_blk(struct nvm_dev *, struct nvm_lun *, | |
413 | unsigned long); | |
414 | extern void nvm_put_blk(struct nvm_dev *, struct nvm_block *); | |
415 | ||
416 | extern int nvm_register(struct request_queue *, char *, | |
417 | struct nvm_dev_ops *); | |
418 | extern void nvm_unregister(char *); | |
419 | ||
420 | extern int nvm_submit_io(struct nvm_dev *, struct nvm_rq *); | |
421 | extern int nvm_erase_blk(struct nvm_dev *, struct nvm_block *); | |
422 | #else /* CONFIG_NVM */ | |
423 | struct nvm_dev_ops; | |
424 | ||
425 | static inline int nvm_register(struct request_queue *q, char *disk_name, | |
426 | struct nvm_dev_ops *ops) | |
427 | { | |
428 | return -EINVAL; | |
429 | } | |
430 | static inline void nvm_unregister(char *disk_name) {} | |
431 | #endif /* CONFIG_NVM */ | |
432 | #endif /* LIGHTNVM.H */ |