Loading fs/cifs/smb2pdu.c +12 −6 Original line number Diff line number Diff line Loading @@ -2720,8 +2720,8 @@ smb2_new_read_req(void **buf, unsigned int *total_len, rdata->mr = smbd_register_mr( server->smbd_conn, rdata->pages, rdata->nr_pages, rdata->tailsz, true, need_invalidate); rdata->nr_pages, rdata->page_offset, rdata->tailsz, true, need_invalidate); if (!rdata->mr) return -ENOBUFS; Loading Loading @@ -3107,16 +3107,22 @@ smb2_async_writev(struct cifs_writedata *wdata, wdata->mr = smbd_register_mr( server->smbd_conn, wdata->pages, wdata->nr_pages, wdata->tailsz, false, need_invalidate); wdata->nr_pages, wdata->page_offset, wdata->tailsz, false, need_invalidate); if (!wdata->mr) { rc = -ENOBUFS; goto async_writev_out; } req->Length = 0; req->DataOffset = 0; if (wdata->nr_pages > 1) req->RemainingBytes = cpu_to_le32((wdata->nr_pages-1)*PAGE_SIZE + wdata->tailsz); cpu_to_le32( (wdata->nr_pages - 1) * wdata->pagesz - wdata->page_offset + wdata->tailsz ); else req->RemainingBytes = cpu_to_le32(wdata->tailsz); req->Channel = SMB2_CHANNEL_RDMA_V1_INVALIDATE; if (need_invalidate) req->Channel = SMB2_CHANNEL_RDMA_V1; Loading fs/cifs/smbdirect.c +45 −31 Original line number Diff line number Diff line Loading @@ -2299,14 +2299,6 @@ static void smbd_mr_recovery_work(struct work_struct *work) if (smbdirect_mr->state == MR_INVALIDATED || smbdirect_mr->state == MR_ERROR) { if (smbdirect_mr->state == MR_INVALIDATED) { ib_dma_unmap_sg( info->id->device, smbdirect_mr->sgl, smbdirect_mr->sgl_count, smbdirect_mr->dir); smbdirect_mr->state = MR_READY; } else if (smbdirect_mr->state == MR_ERROR) { /* recover this MR entry */ rc = ib_dereg_mr(smbdirect_mr->mr); if (rc) { Loading @@ -2314,6 +2306,7 @@ static void smbd_mr_recovery_work(struct work_struct *work) "ib_dereg_mr failed rc=%x\n", rc); smbd_disconnect_rdma_connection(info); continue; } smbdirect_mr->mr = ib_alloc_mr( Loading @@ -2326,10 +2319,17 @@ static void smbd_mr_recovery_work(struct work_struct *work) info->mr_type, info->max_frmr_depth); smbd_disconnect_rdma_connection(info); continue; } if (smbdirect_mr->state == MR_INVALIDATED) ib_dma_unmap_sg( info->id->device, smbdirect_mr->sgl, smbdirect_mr->sgl_count, smbdirect_mr->dir); smbdirect_mr->state = MR_READY; } /* smbdirect_mr->state is updated by this function * and is read and updated by I/O issuing CPUs trying * to get a MR, the call to atomic_inc_return Loading Loading @@ -2475,7 +2475,7 @@ static struct smbd_mr *get_mr(struct smbd_connection *info) */ struct smbd_mr *smbd_register_mr( struct smbd_connection *info, struct page *pages[], int num_pages, int tailsz, bool writing, bool need_invalidate) int offset, int tailsz, bool writing, bool need_invalidate) { struct smbd_mr *smbdirect_mr; int rc, i; Loading @@ -2498,17 +2498,31 @@ struct smbd_mr *smbd_register_mr( smbdirect_mr->sgl_count = num_pages; sg_init_table(smbdirect_mr->sgl, num_pages); for (i = 0; i < num_pages - 1; i++) sg_set_page(&smbdirect_mr->sgl[i], pages[i], PAGE_SIZE, 0); log_rdma_mr(INFO, "num_pages=0x%x offset=0x%x tailsz=0x%x\n", num_pages, offset, tailsz); if (num_pages == 1) { sg_set_page(&smbdirect_mr->sgl[0], pages[0], tailsz, offset); goto skip_multiple_pages; } /* We have at least two pages to register */ sg_set_page( &smbdirect_mr->sgl[0], pages[0], PAGE_SIZE - offset, offset); i = 1; while (i < num_pages - 1) { sg_set_page(&smbdirect_mr->sgl[i], pages[i], PAGE_SIZE, 0); i++; } sg_set_page(&smbdirect_mr->sgl[i], pages[i], tailsz ? tailsz : PAGE_SIZE, 0); skip_multiple_pages: dir = writing ? DMA_FROM_DEVICE : DMA_TO_DEVICE; smbdirect_mr->dir = dir; rc = ib_dma_map_sg(info->id->device, smbdirect_mr->sgl, num_pages, dir); if (!rc) { log_rdma_mr(INFO, "ib_dma_map_sg num_pages=%x dir=%x rc=%x\n", log_rdma_mr(ERR, "ib_dma_map_sg num_pages=%x dir=%x rc=%x\n", num_pages, dir, rc); goto dma_map_error; } Loading @@ -2516,8 +2530,8 @@ struct smbd_mr *smbd_register_mr( rc = ib_map_mr_sg(smbdirect_mr->mr, smbdirect_mr->sgl, num_pages, NULL, PAGE_SIZE); if (rc != num_pages) { log_rdma_mr(INFO, "ib_map_mr_sg failed rc = %x num_pages = %x\n", log_rdma_mr(ERR, "ib_map_mr_sg failed rc = %d num_pages = %x\n", rc, num_pages); goto map_mr_error; } Loading fs/cifs/smbdirect.h +1 −1 Original line number Diff line number Diff line Loading @@ -321,7 +321,7 @@ struct smbd_mr { /* Interfaces to register and deregister MR for RDMA read/write */ struct smbd_mr *smbd_register_mr( struct smbd_connection *info, struct page *pages[], int num_pages, int tailsz, bool writing, bool need_invalidate); int offset, int tailsz, bool writing, bool need_invalidate); int smbd_deregister_mr(struct smbd_mr *mr); #else Loading Loading
fs/cifs/smb2pdu.c +12 −6 Original line number Diff line number Diff line Loading @@ -2720,8 +2720,8 @@ smb2_new_read_req(void **buf, unsigned int *total_len, rdata->mr = smbd_register_mr( server->smbd_conn, rdata->pages, rdata->nr_pages, rdata->tailsz, true, need_invalidate); rdata->nr_pages, rdata->page_offset, rdata->tailsz, true, need_invalidate); if (!rdata->mr) return -ENOBUFS; Loading Loading @@ -3107,16 +3107,22 @@ smb2_async_writev(struct cifs_writedata *wdata, wdata->mr = smbd_register_mr( server->smbd_conn, wdata->pages, wdata->nr_pages, wdata->tailsz, false, need_invalidate); wdata->nr_pages, wdata->page_offset, wdata->tailsz, false, need_invalidate); if (!wdata->mr) { rc = -ENOBUFS; goto async_writev_out; } req->Length = 0; req->DataOffset = 0; if (wdata->nr_pages > 1) req->RemainingBytes = cpu_to_le32((wdata->nr_pages-1)*PAGE_SIZE + wdata->tailsz); cpu_to_le32( (wdata->nr_pages - 1) * wdata->pagesz - wdata->page_offset + wdata->tailsz ); else req->RemainingBytes = cpu_to_le32(wdata->tailsz); req->Channel = SMB2_CHANNEL_RDMA_V1_INVALIDATE; if (need_invalidate) req->Channel = SMB2_CHANNEL_RDMA_V1; Loading
fs/cifs/smbdirect.c +45 −31 Original line number Diff line number Diff line Loading @@ -2299,14 +2299,6 @@ static void smbd_mr_recovery_work(struct work_struct *work) if (smbdirect_mr->state == MR_INVALIDATED || smbdirect_mr->state == MR_ERROR) { if (smbdirect_mr->state == MR_INVALIDATED) { ib_dma_unmap_sg( info->id->device, smbdirect_mr->sgl, smbdirect_mr->sgl_count, smbdirect_mr->dir); smbdirect_mr->state = MR_READY; } else if (smbdirect_mr->state == MR_ERROR) { /* recover this MR entry */ rc = ib_dereg_mr(smbdirect_mr->mr); if (rc) { Loading @@ -2314,6 +2306,7 @@ static void smbd_mr_recovery_work(struct work_struct *work) "ib_dereg_mr failed rc=%x\n", rc); smbd_disconnect_rdma_connection(info); continue; } smbdirect_mr->mr = ib_alloc_mr( Loading @@ -2326,10 +2319,17 @@ static void smbd_mr_recovery_work(struct work_struct *work) info->mr_type, info->max_frmr_depth); smbd_disconnect_rdma_connection(info); continue; } if (smbdirect_mr->state == MR_INVALIDATED) ib_dma_unmap_sg( info->id->device, smbdirect_mr->sgl, smbdirect_mr->sgl_count, smbdirect_mr->dir); smbdirect_mr->state = MR_READY; } /* smbdirect_mr->state is updated by this function * and is read and updated by I/O issuing CPUs trying * to get a MR, the call to atomic_inc_return Loading Loading @@ -2475,7 +2475,7 @@ static struct smbd_mr *get_mr(struct smbd_connection *info) */ struct smbd_mr *smbd_register_mr( struct smbd_connection *info, struct page *pages[], int num_pages, int tailsz, bool writing, bool need_invalidate) int offset, int tailsz, bool writing, bool need_invalidate) { struct smbd_mr *smbdirect_mr; int rc, i; Loading @@ -2498,17 +2498,31 @@ struct smbd_mr *smbd_register_mr( smbdirect_mr->sgl_count = num_pages; sg_init_table(smbdirect_mr->sgl, num_pages); for (i = 0; i < num_pages - 1; i++) sg_set_page(&smbdirect_mr->sgl[i], pages[i], PAGE_SIZE, 0); log_rdma_mr(INFO, "num_pages=0x%x offset=0x%x tailsz=0x%x\n", num_pages, offset, tailsz); if (num_pages == 1) { sg_set_page(&smbdirect_mr->sgl[0], pages[0], tailsz, offset); goto skip_multiple_pages; } /* We have at least two pages to register */ sg_set_page( &smbdirect_mr->sgl[0], pages[0], PAGE_SIZE - offset, offset); i = 1; while (i < num_pages - 1) { sg_set_page(&smbdirect_mr->sgl[i], pages[i], PAGE_SIZE, 0); i++; } sg_set_page(&smbdirect_mr->sgl[i], pages[i], tailsz ? tailsz : PAGE_SIZE, 0); skip_multiple_pages: dir = writing ? DMA_FROM_DEVICE : DMA_TO_DEVICE; smbdirect_mr->dir = dir; rc = ib_dma_map_sg(info->id->device, smbdirect_mr->sgl, num_pages, dir); if (!rc) { log_rdma_mr(INFO, "ib_dma_map_sg num_pages=%x dir=%x rc=%x\n", log_rdma_mr(ERR, "ib_dma_map_sg num_pages=%x dir=%x rc=%x\n", num_pages, dir, rc); goto dma_map_error; } Loading @@ -2516,8 +2530,8 @@ struct smbd_mr *smbd_register_mr( rc = ib_map_mr_sg(smbdirect_mr->mr, smbdirect_mr->sgl, num_pages, NULL, PAGE_SIZE); if (rc != num_pages) { log_rdma_mr(INFO, "ib_map_mr_sg failed rc = %x num_pages = %x\n", log_rdma_mr(ERR, "ib_map_mr_sg failed rc = %d num_pages = %x\n", rc, num_pages); goto map_mr_error; } Loading
fs/cifs/smbdirect.h +1 −1 Original line number Diff line number Diff line Loading @@ -321,7 +321,7 @@ struct smbd_mr { /* Interfaces to register and deregister MR for RDMA read/write */ struct smbd_mr *smbd_register_mr( struct smbd_connection *info, struct page *pages[], int num_pages, int tailsz, bool writing, bool need_invalidate); int offset, int tailsz, bool writing, bool need_invalidate); int smbd_deregister_mr(struct smbd_mr *mr); #else Loading