/**
* Copyright (C) Mellanox Technologies Ltd. 2001-2014. ALL RIGHTS RESERVED.
*
* See file LICENSE for terms.
*/
static UCS_F_ALWAYS_INLINE void
uct_ud_ep_ctl_op_schedule(uct_ud_iface_t *iface, uct_ud_ep_t *ep)
{
ucs_arbiter_group_push_elem(&ep->tx.pending.group,
&ep->tx.pending.elem);
ucs_arbiter_group_schedule(&iface->tx.pending_q, &ep->tx.pending.group);
}
/**
* schedule control operation.
*/
static UCS_F_ALWAYS_INLINE void
uct_ud_ep_ctl_op_add(uct_ud_iface_t *iface, uct_ud_ep_t *ep, int op)
{
ep->tx.pending.ops |= op;
uct_ud_ep_ctl_op_schedule(iface, ep);
}
static UCS_F_ALWAYS_INLINE void
uct_ud_ep_tx_stop(uct_ud_ep_t *ep)
{
ep->tx.max_psn = ep->tx.psn;
}
/*
* check iface resources:tx_queue and return
* prefetched/cached skb
*
* NOTE: caller must not return skb to mpool until it is
* removed from the cache
* skb is removed from cache by
* uct_ud_iface_complete_tx_inl()
* uct_ud_iface_complete_tx_skb()
*
* In case of error flow caller must do nothing with the skb
*/
static UCS_F_ALWAYS_INLINE
uct_ud_send_skb_t *uct_ud_iface_get_tx_skb(uct_ud_iface_t *iface,
uct_ud_ep_t *ep)
{
uct_ud_send_skb_t *skb;
if (ucs_unlikely(!uct_ud_iface_can_tx(iface))) {
UCS_STATS_UPDATE_COUNTER(ep->super.stats, UCT_EP_STAT_NO_RES, 1);
return NULL;
}
skb = iface->tx.skb;
if (ucs_unlikely(skb == NULL)) {
skb = ucs_mpool_get(&iface->tx.mp);
if (skb == NULL) {
ucs_trace_data("iface=%p out of tx skbs", iface);
UCT_TL_IFACE_STAT_TX_NO_DESC(&iface->super.super);
return NULL;
}
iface->tx.skb = skb;
}
VALGRIND_MAKE_MEM_DEFINED(skb, sizeof *skb);
ucs_prefetch(skb->neth);
return skb;
}
/* same as above but also check ep resources: window&connection state */
static UCS_F_ALWAYS_INLINE uct_ud_send_skb_t *
uct_ud_ep_get_tx_skb(uct_ud_iface_t *iface, uct_ud_ep_t *ep)
{
if (ucs_unlikely(!uct_ud_ep_is_connected(ep) ||
uct_ud_ep_no_window(ep) ||
uct_ud_iface_has_pending_async_ev(iface))) {
ucs_trace_poll("iface=%p ep=%p (%d->%d) no ep resources (psn=%u max_psn=%u)",
iface, ep, ep->ep_id, ep->dest_ep_id,
(unsigned)ep->tx.psn,
(unsigned)ep->tx.max_psn);
UCS_STATS_UPDATE_COUNTER(ep->super.stats, UCT_EP_STAT_NO_RES, 1);
return NULL;
}
return uct_ud_iface_get_tx_skb(iface, ep);
}
static UCS_F_ALWAYS_INLINE void
uct_ud_am_set_zcopy_desc(uct_ud_send_skb_t *skb, const uct_iov_t *iov, size_t iovcnt,
uct_completion_t *comp)
{
uct_ud_zcopy_desc_t *zdesc;
size_t iov_it_length;
size_t iov_it;
skb->flags |= UCT_UD_SEND_SKB_FLAG_ZCOPY;
zdesc = uct_ud_zcopy_desc(skb);
zdesc->iovcnt = iovcnt;
for (iov_it = 0; iov_it < iovcnt; ++iov_it) {
iov_it_length = uct_iov_get_length(iov + iov_it);
ucs_assert(iov_it_length <= UINT16_MAX);
zdesc->iov[iov_it].buffer = iov[iov_it].buffer;
zdesc->iov[iov_it].length = iov_it_length;
}
if (comp != NULL) {
skb->flags |= UCT_UD_SEND_SKB_FLAG_COMP;
zdesc->super.comp = comp;
}
}
static UCS_F_ALWAYS_INLINE void
uct_ud_iface_complete_tx_inl(uct_ud_iface_t *iface, uct_ud_ep_t *ep,
uct_ud_send_skb_t *skb, void *data,
const void *buffer, unsigned length)
{
iface->tx.skb = ucs_mpool_get(&iface->tx.mp);
ep->tx.psn++;
skb->len += length;
memcpy(data, buffer, length);
ucs_queue_push(&ep->tx.window, &skb->queue);
ep->tx.slow_tick = iface->async.slow_tick;
ucs_wtimer_add(&iface->async.slow_timer, &ep->slow_timer,
uct_ud_iface_get_async_time(iface) -
ucs_twheel_get_time(&iface->async.slow_timer) +
ep->tx.slow_tick);
ep->tx.send_time = uct_ud_iface_get_async_time(iface);
}
static UCS_F_ALWAYS_INLINE void
uct_ud_iface_complete_tx_skb(uct_ud_iface_t *iface, uct_ud_ep_t *ep,
uct_ud_send_skb_t *skb)
{
iface->tx.skb = ucs_mpool_get(&iface->tx.mp);
ep->tx.psn++;
ucs_queue_push(&ep->tx.window, &skb->queue);
ep->tx.slow_tick = iface->async.slow_tick;
ucs_wtimer_add(&iface->async.slow_timer, &ep->slow_timer,
uct_ud_iface_get_async_time(iface) -
ucs_twheel_get_time(&iface->async.slow_timer) +
ep->tx.slow_tick);
ep->tx.send_time = uct_ud_iface_get_async_time(iface);
}
static UCS_F_ALWAYS_INLINE void
uct_ud_am_set_neth(uct_ud_neth_t *neth, uct_ud_ep_t *ep, uint8_t id)
{
uct_ud_neth_init_data(ep, neth);
uct_ud_neth_set_type_am(ep, neth, id);
uct_ud_neth_ack_req(ep, neth);
}
static UCS_F_ALWAYS_INLINE ucs_status_t
uct_ud_am_common(uct_ud_iface_t *iface, uct_ud_ep_t *ep, uint8_t id,
uct_ud_send_skb_t **skb_p)
{
uct_ud_send_skb_t *skb;
UCT_CHECK_AM_ID(id);
skb = uct_ud_ep_get_tx_skb(iface, ep);
if (!skb) {
return UCS_ERR_NO_RESOURCE;
}
/* either we are executing pending operations, or there are no any pending
* elements, or the only pending element is for sending control messages
* (we don't care about reordering with respect to control messages)
*/
ucs_assertv((ep->flags & UCT_UD_EP_FLAG_IN_PENDING) ||
ucs_arbiter_group_is_empty(&ep->tx.pending.group) ||
ucs_arbiter_elem_is_only(&ep->tx.pending.group, &ep->tx.pending.elem),
"out-of-order send detected for ep %p am %d ep_pending %d arbtail %p arbelem %p",
ep, id, (ep->flags & UCT_UD_EP_FLAG_IN_PENDING),
ep->tx.pending.group.tail,
&ep->tx.pending.elem);
uct_ud_am_set_neth(skb->neth, ep, id);
*skb_p = skb;
return UCS_OK;
}
static UCS_F_ALWAYS_INLINE size_t
uct_ud_skb_bcopy(uct_ud_send_skb_t *skb, uct_pack_callback_t pack_cb, void *arg)
{
size_t payload_len;
payload_len = pack_cb(skb->neth + 1, arg);
skb->len = sizeof(skb->neth[0]) + payload_len;
return payload_len;
}