/** * Copyright (C) Mellanox Technologies Ltd. 2001-2014. ALL RIGHTS RESERVED. * * See file LICENSE for terms. */ static UCS_F_ALWAYS_INLINE void uct_ud_ep_ctl_op_schedule(uct_ud_iface_t *iface, uct_ud_ep_t *ep) { ucs_arbiter_group_push_elem(&ep->tx.pending.group, &ep->tx.pending.elem); ucs_arbiter_group_schedule(&iface->tx.pending_q, &ep->tx.pending.group); } /** * schedule control operation. */ static UCS_F_ALWAYS_INLINE void uct_ud_ep_ctl_op_add(uct_ud_iface_t *iface, uct_ud_ep_t *ep, int op) { ep->tx.pending.ops |= op; uct_ud_ep_ctl_op_schedule(iface, ep); } static UCS_F_ALWAYS_INLINE void uct_ud_ep_tx_stop(uct_ud_ep_t *ep) { ep->tx.max_psn = ep->tx.psn; } /* * check iface resources:tx_queue and return * prefetched/cached skb * * NOTE: caller must not return skb to mpool until it is * removed from the cache * skb is removed from cache by * uct_ud_iface_complete_tx_inl() * uct_ud_iface_complete_tx_skb() * * In case of error flow caller must do nothing with the skb */ static UCS_F_ALWAYS_INLINE uct_ud_send_skb_t *uct_ud_iface_get_tx_skb(uct_ud_iface_t *iface, uct_ud_ep_t *ep) { uct_ud_send_skb_t *skb; if (ucs_unlikely(!uct_ud_iface_can_tx(iface))) { UCS_STATS_UPDATE_COUNTER(ep->super.stats, UCT_EP_STAT_NO_RES, 1); return NULL; } skb = iface->tx.skb; if (ucs_unlikely(skb == NULL)) { skb = ucs_mpool_get(&iface->tx.mp); if (skb == NULL) { ucs_trace_data("iface=%p out of tx skbs", iface); UCT_TL_IFACE_STAT_TX_NO_DESC(&iface->super.super); return NULL; } iface->tx.skb = skb; } VALGRIND_MAKE_MEM_DEFINED(skb, sizeof *skb); ucs_prefetch(skb->neth); return skb; } /* same as above but also check ep resources: window&connection state */ static UCS_F_ALWAYS_INLINE uct_ud_send_skb_t * uct_ud_ep_get_tx_skb(uct_ud_iface_t *iface, uct_ud_ep_t *ep) { if (ucs_unlikely(!uct_ud_ep_is_connected(ep) || uct_ud_ep_no_window(ep) || uct_ud_iface_has_pending_async_ev(iface))) { ucs_trace_poll("iface=%p ep=%p (%d->%d) no ep resources (psn=%u max_psn=%u)", iface, ep, ep->ep_id, ep->dest_ep_id, (unsigned)ep->tx.psn, (unsigned)ep->tx.max_psn); UCS_STATS_UPDATE_COUNTER(ep->super.stats, UCT_EP_STAT_NO_RES, 1); return NULL; } return uct_ud_iface_get_tx_skb(iface, ep); } static UCS_F_ALWAYS_INLINE void uct_ud_am_set_zcopy_desc(uct_ud_send_skb_t *skb, const uct_iov_t *iov, size_t iovcnt, uct_completion_t *comp) { uct_ud_zcopy_desc_t *zdesc; size_t iov_it_length; size_t iov_it; skb->flags |= UCT_UD_SEND_SKB_FLAG_ZCOPY; zdesc = uct_ud_zcopy_desc(skb); zdesc->iovcnt = iovcnt; for (iov_it = 0; iov_it < iovcnt; ++iov_it) { iov_it_length = uct_iov_get_length(iov + iov_it); ucs_assert(iov_it_length <= UINT16_MAX); zdesc->iov[iov_it].buffer = iov[iov_it].buffer; zdesc->iov[iov_it].length = iov_it_length; } if (comp != NULL) { skb->flags |= UCT_UD_SEND_SKB_FLAG_COMP; zdesc->super.comp = comp; } } static UCS_F_ALWAYS_INLINE void uct_ud_iface_complete_tx_inl(uct_ud_iface_t *iface, uct_ud_ep_t *ep, uct_ud_send_skb_t *skb, void *data, const void *buffer, unsigned length) { iface->tx.skb = ucs_mpool_get(&iface->tx.mp); ep->tx.psn++; skb->len += length; memcpy(data, buffer, length); ucs_queue_push(&ep->tx.window, &skb->queue); ep->tx.slow_tick = iface->async.slow_tick; ucs_wtimer_add(&iface->async.slow_timer, &ep->slow_timer, uct_ud_iface_get_async_time(iface) - ucs_twheel_get_time(&iface->async.slow_timer) + ep->tx.slow_tick); ep->tx.send_time = uct_ud_iface_get_async_time(iface); } static UCS_F_ALWAYS_INLINE void uct_ud_iface_complete_tx_skb(uct_ud_iface_t *iface, uct_ud_ep_t *ep, uct_ud_send_skb_t *skb) { iface->tx.skb = ucs_mpool_get(&iface->tx.mp); ep->tx.psn++; ucs_queue_push(&ep->tx.window, &skb->queue); ep->tx.slow_tick = iface->async.slow_tick; ucs_wtimer_add(&iface->async.slow_timer, &ep->slow_timer, uct_ud_iface_get_async_time(iface) - ucs_twheel_get_time(&iface->async.slow_timer) + ep->tx.slow_tick); ep->tx.send_time = uct_ud_iface_get_async_time(iface); } static UCS_F_ALWAYS_INLINE void uct_ud_am_set_neth(uct_ud_neth_t *neth, uct_ud_ep_t *ep, uint8_t id) { uct_ud_neth_init_data(ep, neth); uct_ud_neth_set_type_am(ep, neth, id); uct_ud_neth_ack_req(ep, neth); } static UCS_F_ALWAYS_INLINE ucs_status_t uct_ud_am_common(uct_ud_iface_t *iface, uct_ud_ep_t *ep, uint8_t id, uct_ud_send_skb_t **skb_p) { uct_ud_send_skb_t *skb; UCT_CHECK_AM_ID(id); skb = uct_ud_ep_get_tx_skb(iface, ep); if (!skb) { return UCS_ERR_NO_RESOURCE; } /* either we are executing pending operations, or there are no any pending * elements, or the only pending element is for sending control messages * (we don't care about reordering with respect to control messages) */ ucs_assertv((ep->flags & UCT_UD_EP_FLAG_IN_PENDING) || ucs_arbiter_group_is_empty(&ep->tx.pending.group) || ucs_arbiter_elem_is_only(&ep->tx.pending.group, &ep->tx.pending.elem), "out-of-order send detected for ep %p am %d ep_pending %d arbtail %p arbelem %p", ep, id, (ep->flags & UCT_UD_EP_FLAG_IN_PENDING), ep->tx.pending.group.tail, &ep->tx.pending.elem); uct_ud_am_set_neth(skb->neth, ep, id); *skb_p = skb; return UCS_OK; } static UCS_F_ALWAYS_INLINE size_t uct_ud_skb_bcopy(uct_ud_send_skb_t *skb, uct_pack_callback_t pack_cb, void *arg) { size_t payload_len; payload_len = pack_cb(skb->neth + 1, arg); skb->len = sizeof(skb->neth[0]) + payload_len; return payload_len; }