1 // SPDX-License-Identifier: GPL-2.0
2 #include <linux/vmalloc.h>
5 #define CREATE_TRACE_POINTS
6 #include "null_blk_trace.h"
8 /* zone_size in MBs to sectors. */
9 #define ZONE_SIZE_SHIFT 11
11 static inline unsigned int null_zone_no(struct nullb_device
*dev
, sector_t sect
)
13 return sect
>> ilog2(dev
->zone_size_sects
);
16 int null_init_zoned_dev(struct nullb_device
*dev
, struct request_queue
*q
)
18 sector_t dev_size
= (sector_t
)dev
->size
* 1024 * 1024;
22 if (!is_power_of_2(dev
->zone_size
)) {
23 pr_err("zone_size must be power-of-two\n");
26 if (dev
->zone_size
> dev
->size
) {
27 pr_err("Zone size larger than device capacity\n");
31 dev
->zone_size_sects
= dev
->zone_size
<< ZONE_SIZE_SHIFT
;
32 dev
->nr_zones
= dev_size
>>
33 (SECTOR_SHIFT
+ ilog2(dev
->zone_size_sects
));
34 dev
->zones
= kvmalloc_array(dev
->nr_zones
, sizeof(struct blk_zone
),
35 GFP_KERNEL
| __GFP_ZERO
);
39 if (dev
->zone_nr_conv
>= dev
->nr_zones
) {
40 dev
->zone_nr_conv
= dev
->nr_zones
- 1;
41 pr_info("changed the number of conventional zones to %u",
45 for (i
= 0; i
< dev
->zone_nr_conv
; i
++) {
46 struct blk_zone
*zone
= &dev
->zones
[i
];
49 zone
->len
= dev
->zone_size_sects
;
50 zone
->wp
= zone
->start
+ zone
->len
;
51 zone
->type
= BLK_ZONE_TYPE_CONVENTIONAL
;
52 zone
->cond
= BLK_ZONE_COND_NOT_WP
;
54 sector
+= dev
->zone_size_sects
;
57 for (i
= dev
->zone_nr_conv
; i
< dev
->nr_zones
; i
++) {
58 struct blk_zone
*zone
= &dev
->zones
[i
];
60 zone
->start
= zone
->wp
= sector
;
61 zone
->len
= dev
->zone_size_sects
;
62 zone
->type
= BLK_ZONE_TYPE_SEQWRITE_REQ
;
63 zone
->cond
= BLK_ZONE_COND_EMPTY
;
65 sector
+= dev
->zone_size_sects
;
68 q
->limits
.zoned
= BLK_ZONED_HM
;
69 blk_queue_flag_set(QUEUE_FLAG_ZONE_RESETALL
, q
);
70 blk_queue_required_elevator_features(q
, ELEVATOR_F_ZBD_SEQ_WRITE
);
75 int null_register_zoned_dev(struct nullb
*nullb
)
77 struct request_queue
*q
= nullb
->q
;
80 return blk_revalidate_disk_zones(nullb
->disk
);
82 blk_queue_chunk_sectors(q
, nullb
->dev
->zone_size_sects
);
83 q
->nr_zones
= blkdev_nr_zones(nullb
->disk
);
88 void null_free_zoned_dev(struct nullb_device
*dev
)
93 int null_report_zones(struct gendisk
*disk
, sector_t sector
,
94 unsigned int nr_zones
, report_zones_cb cb
, void *data
)
96 struct nullb
*nullb
= disk
->private_data
;
97 struct nullb_device
*dev
= nullb
->dev
;
98 unsigned int first_zone
, i
;
102 first_zone
= null_zone_no(dev
, sector
);
103 if (first_zone
>= dev
->nr_zones
)
106 nr_zones
= min(nr_zones
, dev
->nr_zones
- first_zone
);
107 trace_nullb_report_zones(nullb
, nr_zones
);
109 for (i
= 0; i
< nr_zones
; i
++) {
111 * Stacked DM target drivers will remap the zone information by
112 * modifying the zone information passed to the report callback.
113 * So use a local copy to avoid corruption of the device zone
116 memcpy(&zone
, &dev
->zones
[first_zone
+ i
],
117 sizeof(struct blk_zone
));
118 error
= cb(&zone
, i
, data
);
126 size_t null_zone_valid_read_len(struct nullb
*nullb
,
127 sector_t sector
, unsigned int len
)
129 struct nullb_device
*dev
= nullb
->dev
;
130 struct blk_zone
*zone
= &dev
->zones
[null_zone_no(dev
, sector
)];
131 unsigned int nr_sectors
= len
>> SECTOR_SHIFT
;
133 /* Read must be below the write pointer position */
134 if (zone
->type
== BLK_ZONE_TYPE_CONVENTIONAL
||
135 sector
+ nr_sectors
<= zone
->wp
)
138 if (sector
> zone
->wp
)
141 return (zone
->wp
- sector
) << SECTOR_SHIFT
;
144 static blk_status_t
null_zone_write(struct nullb_cmd
*cmd
, sector_t sector
,
145 unsigned int nr_sectors
)
147 struct nullb_device
*dev
= cmd
->nq
->dev
;
148 unsigned int zno
= null_zone_no(dev
, sector
);
149 struct blk_zone
*zone
= &dev
->zones
[zno
];
152 trace_nullb_zone_op(cmd
, zno
, zone
->cond
);
154 if (zone
->type
== BLK_ZONE_TYPE_CONVENTIONAL
)
155 return null_process_cmd(cmd
, REQ_OP_WRITE
, sector
, nr_sectors
);
157 switch (zone
->cond
) {
158 case BLK_ZONE_COND_FULL
:
159 /* Cannot write to a full zone */
160 return BLK_STS_IOERR
;
161 case BLK_ZONE_COND_EMPTY
:
162 case BLK_ZONE_COND_IMP_OPEN
:
163 case BLK_ZONE_COND_EXP_OPEN
:
164 case BLK_ZONE_COND_CLOSED
:
165 /* Writes must be at the write pointer position */
166 if (sector
!= zone
->wp
)
167 return BLK_STS_IOERR
;
169 if (zone
->cond
!= BLK_ZONE_COND_EXP_OPEN
)
170 zone
->cond
= BLK_ZONE_COND_IMP_OPEN
;
172 ret
= null_process_cmd(cmd
, REQ_OP_WRITE
, sector
, nr_sectors
);
173 if (ret
!= BLK_STS_OK
)
176 zone
->wp
+= nr_sectors
;
177 if (zone
->wp
== zone
->start
+ zone
->len
)
178 zone
->cond
= BLK_ZONE_COND_FULL
;
181 /* Invalid zone condition */
182 return BLK_STS_IOERR
;
186 static blk_status_t
null_zone_mgmt(struct nullb_cmd
*cmd
, enum req_opf op
,
189 struct nullb_device
*dev
= cmd
->nq
->dev
;
190 unsigned int zone_no
= null_zone_no(dev
, sector
);
191 struct blk_zone
*zone
= &dev
->zones
[zone_no
];
195 case REQ_OP_ZONE_RESET_ALL
:
196 for (i
= 0; i
< dev
->nr_zones
; i
++) {
197 if (zone
[i
].type
== BLK_ZONE_TYPE_CONVENTIONAL
)
199 zone
[i
].cond
= BLK_ZONE_COND_EMPTY
;
200 zone
[i
].wp
= zone
[i
].start
;
203 case REQ_OP_ZONE_RESET
:
204 if (zone
->type
== BLK_ZONE_TYPE_CONVENTIONAL
)
205 return BLK_STS_IOERR
;
207 zone
->cond
= BLK_ZONE_COND_EMPTY
;
208 zone
->wp
= zone
->start
;
210 case REQ_OP_ZONE_OPEN
:
211 if (zone
->type
== BLK_ZONE_TYPE_CONVENTIONAL
)
212 return BLK_STS_IOERR
;
213 if (zone
->cond
== BLK_ZONE_COND_FULL
)
214 return BLK_STS_IOERR
;
216 zone
->cond
= BLK_ZONE_COND_EXP_OPEN
;
218 case REQ_OP_ZONE_CLOSE
:
219 if (zone
->type
== BLK_ZONE_TYPE_CONVENTIONAL
)
220 return BLK_STS_IOERR
;
221 if (zone
->cond
== BLK_ZONE_COND_FULL
)
222 return BLK_STS_IOERR
;
224 if (zone
->wp
== zone
->start
)
225 zone
->cond
= BLK_ZONE_COND_EMPTY
;
227 zone
->cond
= BLK_ZONE_COND_CLOSED
;
229 case REQ_OP_ZONE_FINISH
:
230 if (zone
->type
== BLK_ZONE_TYPE_CONVENTIONAL
)
231 return BLK_STS_IOERR
;
233 zone
->cond
= BLK_ZONE_COND_FULL
;
234 zone
->wp
= zone
->start
+ zone
->len
;
237 return BLK_STS_NOTSUPP
;
240 trace_nullb_zone_op(cmd
, zone_no
, zone
->cond
);
244 blk_status_t
null_process_zoned_cmd(struct nullb_cmd
*cmd
, enum req_opf op
,
245 sector_t sector
, sector_t nr_sectors
)
249 return null_zone_write(cmd
, sector
, nr_sectors
);
250 case REQ_OP_ZONE_RESET
:
251 case REQ_OP_ZONE_RESET_ALL
:
252 case REQ_OP_ZONE_OPEN
:
253 case REQ_OP_ZONE_CLOSE
:
254 case REQ_OP_ZONE_FINISH
:
255 return null_zone_mgmt(cmd
, op
, sector
);
257 return null_process_cmd(cmd
, op
, sector
, nr_sectors
);