* sectors can be read successfully.
*/
if (recovery_mode)
- brq->data.blocks = queue_physical_block_size(mq->queue) >> 9;
+ brq->data.blocks = queue_physical_block_size(mq->queue) >> SECTOR_SHIFT;
/*
* Some controllers have HW issues while operating
}
if (rq_data_dir(req) == READ && brq->data.blocks >
- queue_physical_block_size(mq->queue) >> 9) {
+ queue_physical_block_size(mq->queue) >> SECTOR_SHIFT) {
/* Read one (native) sector at a time */
mmc_blk_read_single(mq, req);
return;
*/
ret = mmc_blk_alloc_rpmb_part(card, md,
card->part[idx].part_cfg,
- card->part[idx].size >> 9,
+ card->part[idx].size >> SECTOR_SHIFT,
card->part[idx].name);
if (ret)
return ret;
} else if (card->part[idx].size) {
ret = mmc_blk_alloc_part(card, md,
card->part[idx].part_cfg,
- card->part[idx].size >> 9,
+ card->part[idx].size >> SECTOR_SHIFT,
card->part[idx].force_ro,
card->part[idx].name,
card->part[idx].area_type);
return;
data->error = data_errors[get_random_u32_below(ARRAY_SIZE(data_errors))];
- data->bytes_xfered = get_random_u32_below(data->bytes_xfered >> 9) << 9;
+ data->bytes_xfered = get_random_u32_below(data->bytes_xfered >> SECTOR_SHIFT)
+ << SECTOR_SHIFT;
}
#else /* CONFIG_FAIL_MMC_REQUEST */