1 /* 2 * Copyright(c) 2013-2015 Intel Corporation. All rights reserved. 3 * 4 * This program is free software; you can redistribute it and/or modify 5 * it under the terms of version 2 of the GNU General Public License as 6 * published by the Free Software Foundation. 7 * 8 * This program is distributed in the hope that it will be useful, but 9 * WITHOUT ANY WARRANTY; without even the implied warranty of 10 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 11 * General Public License for more details. 12 */ 13 #ifndef __ND_CORE_H__ 14 #define __ND_CORE_H__ 15 #include <linux/libnvdimm.h> 16 #include <linux/device.h> 17 #include <linux/sizes.h> 18 #include <linux/mutex.h> 19 #include <linux/nd.h> 20 21 extern struct list_head nvdimm_bus_list; 22 extern struct mutex nvdimm_bus_list_mutex; 23 extern int nvdimm_major; 24 extern struct workqueue_struct *nvdimm_wq; 25 26 struct nvdimm_bus { 27 struct nvdimm_bus_descriptor *nd_desc; 28 wait_queue_head_t probe_wait; 29 struct list_head list; 30 struct device dev; 31 int id, probe_active; 32 struct list_head mapping_list; 33 struct mutex reconfig_mutex; 34 struct badrange badrange; 35 }; 36 37 struct nvdimm { 38 unsigned long flags; 39 void *provider_data; 40 unsigned long cmd_mask; 41 struct device dev; 42 atomic_t busy; 43 int id, num_flush; 44 struct resource *flush_wpq; 45 const char *dimm_id; 46 struct { 47 const struct nvdimm_security_ops *ops; 48 enum nvdimm_security_state state; 49 enum nvdimm_security_state ext_state; 50 unsigned int overwrite_tmo; 51 struct kernfs_node *overwrite_state; 52 } sec; 53 struct delayed_work dwork; 54 }; 55 56 static inline enum nvdimm_security_state nvdimm_security_state( 57 struct nvdimm *nvdimm, enum nvdimm_passphrase_type ptype) 58 { 59 if (!nvdimm->sec.ops) 60 return -ENXIO; 61 62 return nvdimm->sec.ops->state(nvdimm, ptype); 63 } 64 int nvdimm_security_freeze(struct nvdimm *nvdimm); 65 #if IS_ENABLED(CONFIG_NVDIMM_KEYS) 66 int nvdimm_security_disable(struct nvdimm *nvdimm, unsigned int keyid); 67 int nvdimm_security_update(struct nvdimm *nvdimm, unsigned int keyid, 68 unsigned int new_keyid, 69 enum nvdimm_passphrase_type pass_type); 70 int nvdimm_security_erase(struct nvdimm *nvdimm, unsigned int keyid, 71 enum nvdimm_passphrase_type pass_type); 72 int nvdimm_security_overwrite(struct nvdimm *nvdimm, unsigned int keyid); 73 void nvdimm_security_overwrite_query(struct work_struct *work); 74 #else 75 static inline int nvdimm_security_disable(struct nvdimm *nvdimm, 76 unsigned int keyid) 77 { 78 return -EOPNOTSUPP; 79 } 80 static inline int nvdimm_security_update(struct nvdimm *nvdimm, 81 unsigned int keyid, 82 unsigned int new_keyid, 83 enum nvdimm_passphrase_type pass_type) 84 { 85 return -EOPNOTSUPP; 86 } 87 static inline int nvdimm_security_erase(struct nvdimm *nvdimm, 88 unsigned int keyid, 89 enum nvdimm_passphrase_type pass_type) 90 { 91 return -EOPNOTSUPP; 92 } 93 static inline int nvdimm_security_overwrite(struct nvdimm *nvdimm, 94 unsigned int keyid) 95 { 96 return -EOPNOTSUPP; 97 } 98 static inline void nvdimm_security_overwrite_query(struct work_struct *work) 99 { 100 } 101 #endif 102 103 /** 104 * struct blk_alloc_info - tracking info for BLK dpa scanning 105 * @nd_mapping: blk region mapping boundaries 106 * @available: decremented in alias_dpa_busy as aliased PMEM is scanned 107 * @busy: decremented in blk_dpa_busy to account for ranges already 108 * handled by alias_dpa_busy 109 * @res: alias_dpa_busy interprets this a free space range that needs to 110 * be truncated to the valid BLK allocation starting DPA, blk_dpa_busy 111 * treats it as a busy range that needs the aliased PMEM ranges 112 * truncated. 113 */ 114 struct blk_alloc_info { 115 struct nd_mapping *nd_mapping; 116 resource_size_t available, busy; 117 struct resource *res; 118 }; 119 120 bool is_nvdimm(struct device *dev); 121 bool is_nd_pmem(struct device *dev); 122 bool is_nd_volatile(struct device *dev); 123 bool is_nd_blk(struct device *dev); 124 static inline bool is_nd_region(struct device *dev) 125 { 126 return is_nd_pmem(dev) || is_nd_blk(dev) || is_nd_volatile(dev); 127 } 128 static inline bool is_memory(struct device *dev) 129 { 130 return is_nd_pmem(dev) || is_nd_volatile(dev); 131 } 132 struct nvdimm_bus *walk_to_nvdimm_bus(struct device *nd_dev); 133 int __init nvdimm_bus_init(void); 134 void nvdimm_bus_exit(void); 135 void nvdimm_devs_exit(void); 136 void nd_region_devs_exit(void); 137 void nd_region_probe_success(struct nvdimm_bus *nvdimm_bus, struct device *dev); 138 struct nd_region; 139 void nd_region_create_ns_seed(struct nd_region *nd_region); 140 void nd_region_create_btt_seed(struct nd_region *nd_region); 141 void nd_region_create_pfn_seed(struct nd_region *nd_region); 142 void nd_region_create_dax_seed(struct nd_region *nd_region); 143 void nd_region_disable(struct nvdimm_bus *nvdimm_bus, struct device *dev); 144 int nvdimm_bus_create_ndctl(struct nvdimm_bus *nvdimm_bus); 145 void nvdimm_bus_destroy_ndctl(struct nvdimm_bus *nvdimm_bus); 146 void nd_synchronize(void); 147 int nvdimm_bus_register_dimms(struct nvdimm_bus *nvdimm_bus); 148 int nvdimm_bus_register_regions(struct nvdimm_bus *nvdimm_bus); 149 int nvdimm_bus_init_interleave_sets(struct nvdimm_bus *nvdimm_bus); 150 void __nd_device_register(struct device *dev); 151 int nd_match_dimm(struct device *dev, void *data); 152 struct nd_label_id; 153 char *nd_label_gen_id(struct nd_label_id *label_id, u8 *uuid, u32 flags); 154 bool nd_is_uuid_unique(struct device *dev, u8 *uuid); 155 struct nd_region; 156 struct nvdimm_drvdata; 157 struct nd_mapping; 158 void nd_mapping_free_labels(struct nd_mapping *nd_mapping); 159 160 int __reserve_free_pmem(struct device *dev, void *data); 161 void release_free_pmem(struct nvdimm_bus *nvdimm_bus, 162 struct nd_mapping *nd_mapping); 163 164 resource_size_t nd_pmem_max_contiguous_dpa(struct nd_region *nd_region, 165 struct nd_mapping *nd_mapping); 166 resource_size_t nd_region_allocatable_dpa(struct nd_region *nd_region); 167 resource_size_t nd_pmem_available_dpa(struct nd_region *nd_region, 168 struct nd_mapping *nd_mapping, resource_size_t *overlap); 169 resource_size_t nd_blk_available_dpa(struct nd_region *nd_region); 170 resource_size_t nd_region_available_dpa(struct nd_region *nd_region); 171 int nd_region_conflict(struct nd_region *nd_region, resource_size_t start, 172 resource_size_t size); 173 resource_size_t nvdimm_allocated_dpa(struct nvdimm_drvdata *ndd, 174 struct nd_label_id *label_id); 175 int alias_dpa_busy(struct device *dev, void *data); 176 struct resource *nsblk_add_resource(struct nd_region *nd_region, 177 struct nvdimm_drvdata *ndd, struct nd_namespace_blk *nsblk, 178 resource_size_t start); 179 int nvdimm_num_label_slots(struct nvdimm_drvdata *ndd); 180 void get_ndd(struct nvdimm_drvdata *ndd); 181 resource_size_t __nvdimm_namespace_capacity(struct nd_namespace_common *ndns); 182 void nd_detach_ndns(struct device *dev, struct nd_namespace_common **_ndns); 183 void __nd_detach_ndns(struct device *dev, struct nd_namespace_common **_ndns); 184 bool nd_attach_ndns(struct device *dev, struct nd_namespace_common *attach, 185 struct nd_namespace_common **_ndns); 186 bool __nd_attach_ndns(struct device *dev, struct nd_namespace_common *attach, 187 struct nd_namespace_common **_ndns); 188 ssize_t nd_namespace_store(struct device *dev, 189 struct nd_namespace_common **_ndns, const char *buf, 190 size_t len); 191 struct nd_pfn *to_nd_pfn_safe(struct device *dev); 192 #endif /* __ND_CORE_H__ */ 193