diff --git a/src/ucp/core/ucp_request.c b/src/ucp/core/ucp_request.c index 7c0ef1dfefb..d31fbb2fc8d 100644 --- a/src/ucp/core/ucp_request.c +++ b/src/ucp/core/ucp_request.c @@ -353,7 +353,7 @@ ucp_request_send_start(ucp_request_t *req, ssize_t max_short, return UCS_ERR_NO_PROGRESS; } -int ucp_request_mrail_reg(ucp_request_t *req) +int ucp_request_rndv_reg(ucp_request_t *req) { ucp_ep_t *ep = req->send.ep; ucp_dt_state_t *state = &req->send.state.dt; @@ -364,7 +364,7 @@ int ucp_request_mrail_reg(ucp_request_t *req) ucs_assert(UCP_DT_IS_CONTIG(req->send.datatype)); - ucp_dt_clear_rails(state); + ucp_dt_clear_rndv_lanes(state); for (i = 0; i < ucp_ep_rndv_num_lanes(ep); i++) { lane = ucp_ep_get_rndv_get_lane(ep, i); @@ -381,13 +381,13 @@ int ucp_request_mrail_reg(ucp_request_t *req) return cnt; } -void ucp_request_mrail_dereg(ucp_request_t *req) +void ucp_request_rndv_dereg(ucp_request_t *req) { ucp_dt_state_t *state = &req->send.state.dt; ucs_status_t status; int i; - for (i = 0; i < UCP_MAX_RAILS; i++) { + for (i = 0; i < UCP_MAX_RNDV_LANES; i++) { if (state->dt.contig[i].memh != UCT_MEM_HANDLE_NULL) { status = uct_md_mem_dereg(ucp_ep_md(req->send.ep, ucp_ep_get_rndv_get_lane(req->send.ep, i)), @@ -397,7 +397,7 @@ void ucp_request_mrail_dereg(ucp_request_t *req) } req->send.reg_rsc = UCP_NULL_RESOURCE; - ucp_dt_clear_rails(state); + ucp_dt_clear_rndv_lanes(state); } diff --git a/src/ucp/core/ucp_request.h b/src/ucp/core/ucp_request.h index 2971cf55b94..9e7c0361ed3 100644 --- a/src/ucp/core/ucp_request.h +++ b/src/ucp/core/ucp_request.h @@ -87,7 +87,7 @@ enum { typedef struct ucp_rndv_get_rkey { ucp_lane_index_t lane_num; /* number of rkeys obtained from peer */ ucp_lane_index_t lane_idx; - uct_rkey_bundle_t rkey_bundle[UCP_MAX_RAILS]; + uct_rkey_bundle_t rkey_bundle[UCP_MAX_RNDV_LANES]; } ucp_rndv_get_rkey_t; @@ -239,7 +239,7 @@ ucs_status_t ucp_request_send_start(ucp_request_t *req, ssize_t max_short, size_t zcopy_thresh, size_t multi_thresh, size_t rndv_thresh, const ucp_proto_t *proto); -int ucp_request_mrail_reg(ucp_request_t *req); -void ucp_request_mrail_dereg(ucp_request_t *req); +int ucp_request_rndv_reg(ucp_request_t *req); +void ucp_request_rndv_dereg(ucp_request_t *req); #endif diff --git a/src/ucp/core/ucp_request.inl b/src/ucp/core/ucp_request.inl index b6343601441..1050b3507e2 100644 --- a/src/ucp/core/ucp_request.inl +++ b/src/ucp/core/ucp_request.inl @@ -209,7 +209,7 @@ ucp_request_send_state_reset(ucp_request_t *req, /* Fall through */ case UCP_REQUEST_SEND_PROTO_RNDV_GET: if (UCP_DT_IS_CONTIG(req->send.datatype)) { - ucp_dt_clear_rails(&req->send.state.dt); + ucp_dt_clear_rndv_lanes(&req->send.state.dt); } /* Fall through */ case UCP_REQUEST_SEND_PROTO_ZCOPY_AM: @@ -340,7 +340,7 @@ ucp_request_rndv_get_create(ucp_request_t *req) req->send.rndv_get.rkey->lane_idx = 0; req->send.rndv_get.rkey->lane_num = 0; - for (i = 0; i < UCP_MAX_RAILS; i++) { + for (i = 0; i < UCP_MAX_RNDV_LANES; i++) { ucp_tag_rndv_rkey(req, i)->rkey = UCT_INVALID_RKEY; } } @@ -354,7 +354,7 @@ ucp_request_rndv_get_release(ucp_request_t *req) ucs_assert(req->send.rndv_get.rkey != NULL); - for (i = 0; i < UCP_MAX_RAILS; i++) { + for (i = 0; i < UCP_MAX_RNDV_LANES; i++) { if (ucp_tag_rndv_rkey(req, i)->rkey != UCT_INVALID_RKEY) { uct_rkey_release(ucp_tag_rndv_rkey(req, i)); } diff --git a/src/ucp/core/ucp_types.h b/src/ucp/core/ucp_types.h index 19cdd8fa1c9..d10e91e55d1 100644 --- a/src/ucp/core/ucp_types.h +++ b/src/ucp/core/ucp_types.h @@ -30,7 +30,7 @@ UCP_UINT_TYPE(UCP_MD_INDEX_BITS) ucp_md_map_t; /* Lanes */ #define UCP_MAX_LANES 8 -#define UCP_MAX_RAILS 4 +#define UCP_MAX_RNDV_LANES 4 #define UCP_NULL_LANE ((ucp_lane_index_t)-1) typedef uint8_t ucp_lane_index_t; UCP_UINT_TYPE(UCP_MAX_LANES) ucp_lane_map_t; diff --git a/src/ucp/core/ucp_worker.c b/src/ucp/core/ucp_worker.c index e5e143185e7..76a82a309f2 100644 --- a/src/ucp/core/ucp_worker.c +++ b/src/ucp/core/ucp_worker.c @@ -1109,7 +1109,7 @@ ucs_status_t ucp_worker_create(ucp_context_h context, /* Create epoll set which combines events from all transports */ status = ucp_worker_wakeup_init(worker, params); if (status != UCS_OK) { - goto err_mrail_mp_cleanup; + goto err_rndv_lanes_mp_cleanup; } if (params->field_mask & UCP_WORKER_PARAM_FIELD_CPU_MASK) { @@ -1142,7 +1142,7 @@ ucs_status_t ucp_worker_create(ucp_context_h context, err_close_ifaces: ucp_worker_close_ifaces(worker); ucp_worker_wakeup_cleanup(worker); -err_mrail_mp_cleanup: +err_rndv_lanes_mp_cleanup: ucs_mpool_cleanup(&worker->rndv_get_mp, 1); err_req_mp_cleanup: ucs_mpool_cleanup(&worker->req_mp, 1); diff --git a/src/ucp/dt/dt.h b/src/ucp/dt/dt.h index 57ed0925e79..7e45cf36562 100644 --- a/src/ucp/dt/dt.h +++ b/src/ucp/dt/dt.h @@ -27,7 +27,7 @@ typedef struct ucp_dt_state { union { struct { uct_mem_h memh; - } contig[UCP_MAX_RAILS]; + } contig[UCP_MAX_RNDV_LANES]; }; struct { size_t iov_offset; /* Offset in the IOV item */ @@ -123,24 +123,24 @@ ucp_dt_unpack(ucp_datatype_t datatype, void *buffer, size_t buffer_size, } static UCS_F_ALWAYS_INLINE void -ucp_dt_clear_rails(ucp_dt_state_t *state) +ucp_dt_clear_rndv_lanes(ucp_dt_state_t *state) { int i; - for (i = 0; i < UCP_MAX_RAILS; i++) { + for (i = 0; i < UCP_MAX_RNDV_LANES; i++) { state->dt.contig[i].memh = UCT_MEM_HANDLE_NULL; } } static UCS_F_ALWAYS_INLINE int -ucp_dt_is_empty_rail(ucp_dt_state_t *state, int rail) +ucp_dt_is_empty_rndv_lane(ucp_dt_state_t *state, int idx) { - return state->dt.contig[rail].memh == UCT_MEM_HANDLE_NULL; + return state->dt.contig[idx].memh == UCT_MEM_HANDLE_NULL; } static UCS_F_ALWAYS_INLINE int -ucp_dt_have_rails(ucp_dt_state_t *state) +ucp_dt_have_rndv_lanes(ucp_dt_state_t *state) { - return !ucp_dt_is_empty_rail(state, 0); + return !ucp_dt_is_empty_rndv_lane(state, 0); } #endif diff --git a/src/ucp/tag/rndv.c b/src/ucp/tag/rndv.c index 6ba69aa6eb4..c5061d2c31b 100644 --- a/src/ucp/tag/rndv.c +++ b/src/ucp/tag/rndv.c @@ -48,7 +48,7 @@ static void ucp_rndv_rma_request_send_buffer_dereg(ucp_request_t *sreq) * (state->dt.contig.memh != UCT_MEM_HANDLE_NULL) */ if (UCP_DT_IS_CONTIG(sreq->send.datatype)) { - ucp_request_mrail_dereg(sreq); + ucp_request_rndv_dereg(sreq); } } @@ -87,12 +87,12 @@ static size_t ucp_tag_rndv_pack_rkeys(ucp_request_t *sreq, void *rkey_buf, uint1 ucs_assert(UCP_DT_IS_CONTIG(sreq->send.datatype)); - cnt = ucp_request_mrail_reg(sreq); - ucs_assert_always(cnt <= UCP_MAX_RAILS); + cnt = ucp_request_rndv_reg(sreq); + ucs_assert_always(cnt <= UCP_MAX_RNDV_LANES); if (cnt) { for (i = 0; i < ucp_ep_rndv_num_lanes(ep); i++) { - ucs_assert_always(!ucp_dt_is_empty_rail(state, i)); + ucs_assert_always(!ucp_dt_is_empty_rndv_lane(state, i)); lane = ucp_ep_get_rndv_get_lane(ep, i); @@ -336,9 +336,9 @@ UCS_PROFILE_FUNC(ucs_status_t, ucp_proto_progress_rndv_get_zcopy, (self), } /* rndv_req is the internal request to perform the get operation */ - if (!ucp_dt_have_rails(&rndv_req->send.state.dt)) { + if (!ucp_dt_have_rndv_lanes(&rndv_req->send.state.dt)) { /* TODO Not all UCTs need registration on the recv side */ - ucp_request_mrail_reg(rndv_req); + ucp_request_rndv_reg(rndv_req); rndv_req->send.reg_rsc = UCP_NULL_RESOURCE; rndv_req->send.rndv_get.rkey->lane_idx = 0; } diff --git a/src/ucp/wireup/select.c b/src/ucp/wireup/select.c index c8c6df9c0b1..c291cb33a7c 100644 --- a/src/ucp/wireup/select.c +++ b/src/ucp/wireup/select.c @@ -822,7 +822,7 @@ static ucs_status_t ucp_wireup_add_rndv_lanes(ucp_ep_h ep, (params->err_mode == UCP_ERR_HANDLING_MODE_PEER)) { max_lanes = ucs_min(1, ep->worker->context->config.ext.max_rndv_lanes); } else { - max_lanes = ucs_min(UCP_MAX_RAILS, ep->worker->context->config.ext.max_rndv_lanes); + max_lanes = ucs_min(UCP_MAX_RNDV_LANES, ep->worker->context->config.ext.max_rndv_lanes); } while (rndv_lanes < max_lanes) {