Lines Matching refs:nlb
1234 static uint16_t nvme_map_data(NvmeCtrl *n, uint32_t nlb, NvmeRequest *req) in nvme_map_data() argument
1240 size_t len = nvme_l2b(ns, nlb); in nvme_map_data()
1247 len += nvme_m2b(ns, nlb); in nvme_map_data()
1264 static uint16_t nvme_map_mdata(NvmeCtrl *n, uint32_t nlb, NvmeRequest *req) in nvme_map_mdata() argument
1267 size_t len = nvme_m2b(ns, nlb); in nvme_map_mdata()
1273 len += nvme_l2b(ns, nlb); in nvme_map_mdata()
1695 uint32_t nlb) in nvme_check_bounds() argument
1699 if (unlikely(UINT64_MAX - slba < nlb || slba + nlb > nsze)) { in nvme_check_bounds()
1700 trace_pci_nvme_err_invalid_lba_range(slba, nlb, nsze); in nvme_check_bounds()
1708 uint32_t nlb, int flags) in nvme_block_status_all() argument
1712 int64_t pnum = 0, bytes = nvme_l2b(ns, nlb); in nvme_block_status_all()
1745 uint32_t nlb) in nvme_check_dulbe() argument
1750 ret = nvme_block_status_all(ns, slba, nlb, BDRV_BLOCK_DATA); in nvme_check_dulbe()
1850 uint64_t slba, uint32_t nlb) in nvme_check_zone_write() argument
1863 if (slba < zone->w_ptr || slba + nlb > ezrwa) { in nvme_check_zone_write()
1875 if (unlikely((slba + nlb) > zcap)) { in nvme_check_zone_write()
1876 trace_pci_nvme_err_zone_boundary(slba, nlb, zcap); in nvme_check_zone_write()
1904 uint32_t nlb) in nvme_check_zone_read() argument
1914 end = slba + nlb; in nvme_check_zone_read()
2103 uint32_t nlb) in nvme_advance_zone_wp() argument
2105 zone->d.wp += nlb; in nvme_advance_zone_wp()
2131 uint32_t nlb; in nvme_finalize_zoned_write() local
2134 nlb = le16_to_cpu(rw->nlb) + 1; in nvme_finalize_zoned_write()
2140 uint64_t elba = slba + nlb - 1; in nvme_finalize_zoned_write()
2149 nvme_advance_zone_wp(ns, zone, nlb); in nvme_finalize_zoned_write()
2214 uint32_t nlb = (uint32_t)le16_to_cpu(rw->nlb) + 1; in nvme_rw_cb() local
2218 size_t mlen = nvme_m2b(ns, nlb); in nvme_rw_cb()
2230 status = nvme_map_mdata(nvme_ctrl(req), nlb, req); in nvme_rw_cb()
2310 uint32_t nlb = le16_to_cpu(rw->nlb) + 1; in nvme_verify_mdata_in_cb() local
2311 size_t mlen = nvme_m2b(ns, nlb); in nvme_verify_mdata_in_cb()
2473 uint32_t nlb = le16_to_cpu(rw->nlb) + 1; in nvme_compare_data_cb() local
2474 size_t mlen = nvme_m2b(ns, nlb); in nvme_compare_data_cb()
2542 uint32_t nlb; in nvme_dsm_md_cb() local
2550 nlb = le32_to_cpu(range->nlb); in nvme_dsm_md_cb()
2557 ret = nvme_block_status_all(ns, slba, nlb, BDRV_BLOCK_ZERO); in nvme_dsm_md_cb()
2568 nvme_m2b(ns, nlb), BDRV_REQ_MAY_UNMAP, in nvme_dsm_md_cb()
2584 uint32_t nlb; in nvme_dsm_cb() local
2600 nlb = le32_to_cpu(range->nlb); in nvme_dsm_cb()
2602 trace_pci_nvme_dsm_deallocate(slba, nlb); in nvme_dsm_cb()
2604 if (nlb > n->dmrsl) { in nvme_dsm_cb()
2605 trace_pci_nvme_dsm_single_range_limit_exceeded(nlb, n->dmrsl); in nvme_dsm_cb()
2609 if (nvme_check_bounds(ns, slba, nlb)) { in nvme_dsm_cb()
2610 trace_pci_nvme_err_invalid_lba_range(slba, nlb, in nvme_dsm_cb()
2616 nvme_l2b(ns, nlb), in nvme_dsm_cb()
2671 uint32_t nlb = le16_to_cpu(rw->nlb) + 1; in nvme_verify() local
2672 size_t len = nvme_l2b(ns, nlb); in nvme_verify()
2680 trace_pci_nvme_verify(nvme_cid(req), nvme_nsid(ns), slba, nlb); in nvme_verify()
2694 data_len += nvme_m2b(ns, nlb); in nvme_verify()
2701 status = nvme_check_bounds(ns, slba, nlb); in nvme_verify()
2707 status = nvme_check_dulbe(ns, slba, nlb); in nvme_verify()
2802 uint32_t *nlb, in nvme_copy_source_range_parse_format0_2() argument
2818 if (nlb) { in nvme_copy_source_range_parse_format0_2()
2819 *nlb = le16_to_cpu(_ranges[idx].nlb) + 1; in nvme_copy_source_range_parse_format0_2()
2837 uint32_t *nlb, in nvme_copy_source_range_parse_format1_3() argument
2853 if (nlb) { in nvme_copy_source_range_parse_format1_3()
2854 *nlb = le16_to_cpu(_ranges[idx].nlb) + 1; in nvme_copy_source_range_parse_format1_3()
2878 uint64_t *slba, uint32_t *nlb, in nvme_copy_source_range_parse() argument
2885 nvme_copy_source_range_parse_format0_2(ranges, idx, slba, nlb, snsid, in nvme_copy_source_range_parse()
2891 nvme_copy_source_range_parse_format1_3(ranges, idx, slba, nlb, snsid, in nvme_copy_source_range_parse()
2906 uint32_t nlb; in nvme_check_copy_mcl() local
2908 &nlb, NULL, NULL, NULL, NULL); in nvme_check_copy_mcl()
2909 copy_len += nlb; in nvme_check_copy_mcl()
2924 uint32_t nlb; in nvme_copy_out_completed_cb() local
2927 &nlb, NULL, NULL, NULL, NULL); in nvme_copy_out_completed_cb()
2937 nvme_advance_zone_wp(dns, iocb->zone, nlb); in nvme_copy_out_completed_cb()
2941 iocb->slba += nlb; in nvme_copy_out_completed_cb()
2951 uint32_t nlb; in nvme_copy_out_cb() local
2960 &nlb, NULL, NULL, NULL, NULL); in nvme_copy_out_cb()
2962 mlen = nvme_m2b(dns, nlb); in nvme_copy_out_cb()
2963 mbounce = iocb->bounce + nvme_l2b(dns, nlb); in nvme_copy_out_cb()
2986 uint32_t nlb; in nvme_copy_in_completed_cb() local
3001 &nlb, NULL, &apptag, &appmask, &reftag); in nvme_copy_in_completed_cb()
3003 trace_pci_nvme_copy_out(iocb->slba, nlb); in nvme_copy_in_completed_cb()
3005 len = nvme_l2b(sns, nlb); in nvme_copy_in_completed_cb()
3012 mlen = nvme_m2b(sns, nlb); in nvme_copy_in_completed_cb()
3013 mbounce = iocb->bounce + nvme_l2b(sns, nlb); in nvme_copy_in_completed_cb()
3030 mlen = nvme_m2b(dns, nlb); in nvme_copy_in_completed_cb()
3031 mbounce = iocb->bounce + nvme_l2b(dns, nlb); in nvme_copy_in_completed_cb()
3054 status = nvme_check_bounds(dns, iocb->slba, nlb); in nvme_copy_in_completed_cb()
3060 status = nvme_check_zone_write(dns, iocb->zone, iocb->slba, nlb); in nvme_copy_in_completed_cb()
3066 iocb->zone->w_ptr += nlb; in nvme_copy_in_completed_cb()
3093 uint32_t nlb; in nvme_copy_in_cb() local
3100 &nlb, NULL, NULL, NULL, NULL); in nvme_copy_in_cb()
3103 qemu_iovec_add(&iocb->iov, iocb->bounce + nvme_l2b(sns, nlb), in nvme_copy_in_cb()
3104 nvme_m2b(sns, nlb)); in nvme_copy_in_cb()
3182 uint32_t nlb; in nvme_do_copy() local
3198 &slba, &nlb, &snsid, NULL, NULL, NULL); in nvme_do_copy()
3211 if (((slba + nlb) > iocb->slba) && in nvme_do_copy()
3212 ((slba + nlb) < (iocb->slba + iocb->tcl))) { in nvme_do_copy()
3219 &slba, &nlb, NULL, NULL, NULL, NULL); in nvme_do_copy()
3275 len = nvme_l2b(sns, nlb); in nvme_do_copy()
3277 trace_pci_nvme_copy_source_range(slba, nlb); in nvme_do_copy()
3279 if (nlb > le16_to_cpu(sns->id_ns.mssrl)) { in nvme_do_copy()
3284 status = nvme_check_bounds(sns, slba, nlb); in nvme_do_copy()
3290 status = nvme_check_dulbe(sns, slba, nlb); in nvme_do_copy()
3297 status = nvme_check_zone_read(sns, slba, nlb); in nvme_do_copy()
3420 uint32_t nlb = le16_to_cpu(rw->nlb) + 1; in nvme_compare() local
3422 size_t data_len = nvme_l2b(ns, nlb); in nvme_compare()
3428 trace_pci_nvme_compare(nvme_cid(req), nvme_nsid(ns), slba, nlb); in nvme_compare()
3435 len += nvme_m2b(ns, nlb); in nvme_compare()
3447 status = nvme_check_bounds(ns, slba, nlb); in nvme_compare()
3453 status = nvme_check_dulbe(ns, slba, nlb); in nvme_compare()
3611 uint32_t nlb = (uint32_t)le16_to_cpu(rw->nlb) + 1; in nvme_read() local
3613 uint64_t data_size = nvme_l2b(ns, nlb); in nvme_read()
3620 mapped_size += nvme_m2b(ns, nlb); in nvme_read()
3631 trace_pci_nvme_read(nvme_cid(req), nvme_nsid(ns), nlb, mapped_size, slba); in nvme_read()
3638 status = nvme_check_bounds(ns, slba, nlb); in nvme_read()
3644 status = nvme_check_zone_read(ns, slba, nlb); in nvme_read()
3646 trace_pci_nvme_err_zone_read_not_ok(slba, nlb, status); in nvme_read()
3652 status = nvme_check_dulbe(ns, slba, nlb); in nvme_read()
3662 status = nvme_map_data(n, nlb, req); in nvme_read()
3680 uint32_t nlb) in nvme_do_write_fdp() argument
3684 uint64_t data_size = nvme_l2b(ns, nlb); in nvme_do_write_fdp()
3703 while (nlb) { in nvme_do_write_fdp()
3704 if (nlb < ru->ruamw) { in nvme_do_write_fdp()
3705 ru->ruamw -= nlb; in nvme_do_write_fdp()
3709 nlb -= ru->ruamw; in nvme_do_write_fdp()
3720 uint32_t nlb = (uint32_t)le16_to_cpu(rw->nlb) + 1; in nvme_do_write() local
3723 uint64_t data_size = nvme_l2b(ns, nlb); in nvme_do_write()
3732 mapped_size += nvme_m2b(ns, nlb); in nvme_do_write()
3738 mapped_size -= nvme_m2b(ns, nlb); in nvme_do_write()
3744 nvme_nsid(ns), nlb, mapped_size, slba); in nvme_do_write()
3753 status = nvme_check_bounds(ns, slba, nlb); in nvme_do_write()
3810 status = nvme_check_zone_write(ns, zone, slba, nlb); in nvme_do_write()
3821 zone->w_ptr += nlb; in nvme_do_write()
3824 nvme_do_write_fdp(n, req, slba, nlb); in nvme_do_write()
3834 status = nvme_map_data(n, nlb, req); in nvme_do_write()
4193 uint32_t nlb = elba - wp + 1; in nvme_zone_mgmt_send_zrwa_flush() local
4209 if (nlb % ns->zns.zrwafg) { in nvme_zone_mgmt_send_zrwa_flush()
4218 zone->w_ptr += nlb; in nvme_zone_mgmt_send_zrwa_flush()
4220 nvme_advance_zone_wp(ns, zone, nlb); in nvme_zone_mgmt_send_zrwa_flush()
7324 uint32_t nlb = (uint32_t)le16_to_cpu(rw->nlb); in nvme_atomic_write_check() local
7325 uint64_t elba = slba + nlb; in nvme_atomic_write_check()
7330 ((rw->nlb + 1) > atomic->atomic_max_write_size))) { in nvme_atomic_write_check()
7360 req_nlb = (uint32_t)le16_to_cpu(req_rw->nlb); in nvme_atomic_write_check()