1/* SPDX-License-Identifier: GPL-2.0 */
2/*
3 * Copyright (C) 2017 Western Digital Corporation or its affiliates.
4 *
5 * This file is released under the GPL.
6 */
7
8#ifndef DM_ZONED_H
9#define DM_ZONED_H
10
11#include <linux/types.h>
12#include <linux/blkdev.h>
13#include <linux/device-mapper.h>
14#include <linux/dm-kcopyd.h>
15#include <linux/list.h>
16#include <linux/spinlock.h>
17#include <linux/mutex.h>
18#include <linux/workqueue.h>
19#include <linux/rwsem.h>
20#include <linux/rbtree.h>
21#include <linux/radix-tree.h>
22#include <linux/shrinker.h>
23
24/*
25 * dm-zoned creates block devices with 4KB blocks, always.
26 */
27#define DMZ_BLOCK_SHIFT 12
28#define DMZ_BLOCK_SIZE (1 << DMZ_BLOCK_SHIFT)
29#define DMZ_BLOCK_MASK (DMZ_BLOCK_SIZE - 1)
30
31#define DMZ_BLOCK_SHIFT_BITS (DMZ_BLOCK_SHIFT + 3)
32#define DMZ_BLOCK_SIZE_BITS (1 << DMZ_BLOCK_SHIFT_BITS)
33#define DMZ_BLOCK_MASK_BITS (DMZ_BLOCK_SIZE_BITS - 1)
34
35#define DMZ_BLOCK_SECTORS_SHIFT (DMZ_BLOCK_SHIFT - SECTOR_SHIFT)
36#define DMZ_BLOCK_SECTORS (DMZ_BLOCK_SIZE >> SECTOR_SHIFT)
37#define DMZ_BLOCK_SECTORS_MASK (DMZ_BLOCK_SECTORS - 1)
38
39/*
40 * 4KB block <-> 512B sector conversion.
41 */
42#define dmz_blk2sect(b) ((sector_t)(b) << DMZ_BLOCK_SECTORS_SHIFT)
43#define dmz_sect2blk(s) ((sector_t)(s) >> DMZ_BLOCK_SECTORS_SHIFT)
44
45#define dmz_bio_block(bio) dmz_sect2blk((bio)->bi_iter.bi_sector)
46#define dmz_bio_blocks(bio) dmz_sect2blk(bio_sectors(bio))
47
48struct dmz_metadata;
49struct dmz_reclaim;
50
51/*
52 * Zoned block device information.
53 */
54struct dmz_dev {
55 struct block_device *bdev;
56 struct dmz_metadata *metadata;
57 struct dmz_reclaim *reclaim;
58
59 uuid_t uuid;
60
61 sector_t capacity;
62
63 unsigned int dev_idx;
64
65 unsigned int nr_zones;
66 unsigned int zone_offset;
67
68 unsigned int flags;
69
70 sector_t zone_nr_sectors;
71
72 unsigned int nr_rnd;
73 atomic_t unmap_nr_rnd;
74 struct list_head unmap_rnd_list;
75 struct list_head map_rnd_list;
76
77 unsigned int nr_seq;
78 atomic_t unmap_nr_seq;
79 struct list_head unmap_seq_list;
80 struct list_head map_seq_list;
81};
82
83#define dmz_bio_chunk(zmd, bio) ((bio)->bi_iter.bi_sector >> \
84 dmz_zone_nr_sectors_shift(zmd))
85#define dmz_chunk_block(zmd, b) ((b) & (dmz_zone_nr_blocks(zmd) - 1))
86
87/* Device flags. */
88#define DMZ_BDEV_DYING (1 << 0)
89#define DMZ_CHECK_BDEV (2 << 0)
90#define DMZ_BDEV_REGULAR (4 << 0)
91
92/*
93 * Zone descriptor.
94 */
95struct dm_zone {
96 /* For listing the zone depending on its state */
97 struct list_head link;
98
99 /* Device containing this zone */
100 struct dmz_dev *dev;
101
102 /* Zone type and state */
103 unsigned long flags;
104
105 /* Zone activation reference count */
106 atomic_t refcount;
107
108 /* Zone id */
109 unsigned int id;
110
111 /* Zone write pointer block (relative to the zone start block) */
112 unsigned int wp_block;
113
114 /* Zone weight (number of valid blocks in the zone) */
115 unsigned int weight;
116
117 /* The chunk that the zone maps */
118 unsigned int chunk;
119
120 /*
121 * For a sequential data zone, pointer to the random zone
122 * used as a buffer for processing unaligned writes.
123 * For a buffer zone, this points back to the data zone.
124 */
125 struct dm_zone *bzone;
126};
127
128/*
129 * Zone flags.
130 */
131enum {
132 /* Zone write type */
133 DMZ_CACHE,
134 DMZ_RND,
135 DMZ_SEQ,
136
137 /* Zone critical condition */
138 DMZ_OFFLINE,
139 DMZ_READ_ONLY,
140
141 /* How the zone is being used */
142 DMZ_META,
143 DMZ_DATA,
144 DMZ_BUF,
145 DMZ_RESERVED,
146
147 /* Zone internal state */
148 DMZ_RECLAIM,
149 DMZ_SEQ_WRITE_ERR,
150 DMZ_RECLAIM_TERMINATE,
151};
152
153/*
154 * Zone data accessors.
155 */
156#define dmz_is_cache(z) test_bit(DMZ_CACHE, &(z)->flags)
157#define dmz_is_rnd(z) test_bit(DMZ_RND, &(z)->flags)
158#define dmz_is_seq(z) test_bit(DMZ_SEQ, &(z)->flags)
159#define dmz_is_empty(z) ((z)->wp_block == 0)
160#define dmz_is_offline(z) test_bit(DMZ_OFFLINE, &(z)->flags)
161#define dmz_is_readonly(z) test_bit(DMZ_READ_ONLY, &(z)->flags)
162#define dmz_in_reclaim(z) test_bit(DMZ_RECLAIM, &(z)->flags)
163#define dmz_is_reserved(z) test_bit(DMZ_RESERVED, &(z)->flags)
164#define dmz_seq_write_err(z) test_bit(DMZ_SEQ_WRITE_ERR, &(z)->flags)
165#define dmz_reclaim_should_terminate(z) \
166 test_bit(DMZ_RECLAIM_TERMINATE, &(z)->flags)
167
168#define dmz_is_meta(z) test_bit(DMZ_META, &(z)->flags)
169#define dmz_is_buf(z) test_bit(DMZ_BUF, &(z)->flags)
170#define dmz_is_data(z) test_bit(DMZ_DATA, &(z)->flags)
171
172#define dmz_weight(z) ((z)->weight)
173
174/*
175 * Message functions.
176 */
177#define dmz_dev_info(dev, format, args...) \
178 DMINFO("(%pg): " format, (dev)->bdev, ## args)
179
180#define dmz_dev_err(dev, format, args...) \
181 DMERR("(%pg): " format, (dev)->bdev, ## args)
182
183#define dmz_dev_warn(dev, format, args...) \
184 DMWARN("(%pg): " format, (dev)->bdev, ## args)
185
186#define dmz_dev_debug(dev, format, args...) \
187 DMDEBUG("(%pg): " format, (dev)->bdev, ## args)
188
189/*
190 * Functions defined in dm-zoned-metadata.c
191 */
192int dmz_ctr_metadata(struct dmz_dev *dev, int num_dev,
193 struct dmz_metadata **zmd, const char *devname);
194void dmz_dtr_metadata(struct dmz_metadata *zmd);
195int dmz_resume_metadata(struct dmz_metadata *zmd);
196
197void dmz_lock_map(struct dmz_metadata *zmd);
198void dmz_unlock_map(struct dmz_metadata *zmd);
199void dmz_lock_metadata(struct dmz_metadata *zmd);
200void dmz_unlock_metadata(struct dmz_metadata *zmd);
201void dmz_lock_flush(struct dmz_metadata *zmd);
202void dmz_unlock_flush(struct dmz_metadata *zmd);
203int dmz_flush_metadata(struct dmz_metadata *zmd);
204const char *dmz_metadata_label(struct dmz_metadata *zmd);
205
206sector_t dmz_start_sect(struct dmz_metadata *zmd, struct dm_zone *zone);
207sector_t dmz_start_block(struct dmz_metadata *zmd, struct dm_zone *zone);
208unsigned int dmz_nr_chunks(struct dmz_metadata *zmd);
209
210bool dmz_check_dev(struct dmz_metadata *zmd);
211bool dmz_dev_is_dying(struct dmz_metadata *zmd);
212
213#define DMZ_ALLOC_RND 0x01
214#define DMZ_ALLOC_CACHE 0x02
215#define DMZ_ALLOC_SEQ 0x04
216#define DMZ_ALLOC_RECLAIM 0x10
217
218struct dm_zone *dmz_alloc_zone(struct dmz_metadata *zmd,
219 unsigned int dev_idx, unsigned long flags);
220void dmz_free_zone(struct dmz_metadata *zmd, struct dm_zone *zone);
221
222void dmz_map_zone(struct dmz_metadata *zmd, struct dm_zone *zone,
223 unsigned int chunk);
224void dmz_unmap_zone(struct dmz_metadata *zmd, struct dm_zone *zone);
225unsigned int dmz_nr_zones(struct dmz_metadata *zmd);
226unsigned int dmz_nr_cache_zones(struct dmz_metadata *zmd);
227unsigned int dmz_nr_unmap_cache_zones(struct dmz_metadata *zmd);
228unsigned int dmz_nr_rnd_zones(struct dmz_metadata *zmd, int idx);
229unsigned int dmz_nr_unmap_rnd_zones(struct dmz_metadata *zmd, int idx);
230unsigned int dmz_nr_seq_zones(struct dmz_metadata *zmd, int idx);
231unsigned int dmz_nr_unmap_seq_zones(struct dmz_metadata *zmd, int idx);
232unsigned int dmz_zone_nr_blocks(struct dmz_metadata *zmd);
233unsigned int dmz_zone_nr_blocks_shift(struct dmz_metadata *zmd);
234unsigned int dmz_zone_nr_sectors(struct dmz_metadata *zmd);
235unsigned int dmz_zone_nr_sectors_shift(struct dmz_metadata *zmd);
236
237/*
238 * Activate a zone (increment its reference count).
239 */
240static inline void dmz_activate_zone(struct dm_zone *zone)
241{
242 atomic_inc(v: &zone->refcount);
243}
244
245int dmz_lock_zone_reclaim(struct dm_zone *zone);
246void dmz_unlock_zone_reclaim(struct dm_zone *zone);
247struct dm_zone *dmz_get_zone_for_reclaim(struct dmz_metadata *zmd,
248 unsigned int dev_idx, bool idle);
249
250struct dm_zone *dmz_get_chunk_mapping(struct dmz_metadata *zmd,
251 unsigned int chunk, enum req_op op);
252void dmz_put_chunk_mapping(struct dmz_metadata *zmd, struct dm_zone *zone);
253struct dm_zone *dmz_get_chunk_buffer(struct dmz_metadata *zmd,
254 struct dm_zone *dzone);
255
256int dmz_validate_blocks(struct dmz_metadata *zmd, struct dm_zone *zone,
257 sector_t chunk_block, unsigned int nr_blocks);
258int dmz_invalidate_blocks(struct dmz_metadata *zmd, struct dm_zone *zone,
259 sector_t chunk_block, unsigned int nr_blocks);
260int dmz_block_valid(struct dmz_metadata *zmd, struct dm_zone *zone,
261 sector_t chunk_block);
262int dmz_first_valid_block(struct dmz_metadata *zmd, struct dm_zone *zone,
263 sector_t *chunk_block);
264int dmz_copy_valid_blocks(struct dmz_metadata *zmd, struct dm_zone *from_zone,
265 struct dm_zone *to_zone);
266int dmz_merge_valid_blocks(struct dmz_metadata *zmd, struct dm_zone *from_zone,
267 struct dm_zone *to_zone, sector_t chunk_block);
268
269/*
270 * Functions defined in dm-zoned-reclaim.c
271 */
272int dmz_ctr_reclaim(struct dmz_metadata *zmd, struct dmz_reclaim **zrc, int idx);
273void dmz_dtr_reclaim(struct dmz_reclaim *zrc);
274void dmz_suspend_reclaim(struct dmz_reclaim *zrc);
275void dmz_resume_reclaim(struct dmz_reclaim *zrc);
276void dmz_reclaim_bio_acc(struct dmz_reclaim *zrc);
277void dmz_schedule_reclaim(struct dmz_reclaim *zrc);
278
279/*
280 * Functions defined in dm-zoned-target.c
281 */
282bool dmz_bdev_is_dying(struct dmz_dev *dmz_dev);
283bool dmz_check_bdev(struct dmz_dev *dmz_dev);
284
285/*
286 * Deactivate a zone. This decrement the zone reference counter
287 * indicating that all BIOs to the zone have completed when the count is 0.
288 */
289static inline void dmz_deactivate_zone(struct dm_zone *zone)
290{
291 dmz_reclaim_bio_acc(zrc: zone->dev->reclaim);
292 atomic_dec(v: &zone->refcount);
293}
294
295/*
296 * Test if a zone is active, that is, has a refcount > 0.
297 */
298static inline bool dmz_is_active(struct dm_zone *zone)
299{
300 return atomic_read(v: &zone->refcount);
301}
302
303#endif /* DM_ZONED_H */
304

source code of linux/drivers/md/dm-zoned.h