// SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB
/* Copyright (c) 2015 - 2021 Intel Corporation */
#include "main.h"
#include "trace.h"
static void irdma_cm_post_event(struct irdma_cm_event *event);
static void irdma_disconnect_worker(struct work_struct *work);
/**
* irdma_free_sqbuf - put back puda buffer if refcount is 0
* @vsi: The VSI structure of the device
* @bufp: puda buffer to free
*/
void irdma_free_sqbuf(struct irdma_sc_vsi *vsi, void *bufp)
{
struct irdma_puda_buf *buf = bufp;
struct irdma_puda_rsrc *ilq = vsi->ilq;
if (refcount_dec_and_test(&buf->refcount))
irdma_puda_ret_bufpool(ilq, buf);
}
/**
* irdma_record_ird_ord - Record IRD/ORD passed in
* @cm_node: connection's node
* @conn_ird: connection IRD
* @conn_ord: connection ORD
*/
static void irdma_record_ird_ord(struct irdma_cm_node *cm_node, u32 conn_ird,
u32 conn_ord)
{
if (conn_ird > cm_node->dev->hw_attrs.max_hw_ird)
conn_ird = cm_node->dev->hw_attrs.max_hw_ird;
if (conn_ord > cm_node->dev->hw_attrs.max_hw_ord)
conn_ord = cm_node->dev->hw_attrs.max_hw_ord;
else if (!conn_ord && cm_node->send_rdma0_op == SEND_RDMA_READ_ZERO)
conn_ord = 1;
cm_node->ird_size = conn_ird;
cm_node->ord_size = conn_ord;
}
/**
* irdma_copy_ip_ntohl - copy IP address from network to host
* @dst: IP address in host order
* @src: IP address in network order (big endian)
*/
void irdma_copy_ip_ntohl(u32 *dst, __be32 *src)
{
*dst++ = ntohl(*src++);
*dst++ = ntohl(*src++);
*dst++ = ntohl(*src++);
*dst = ntohl(*src);
}
/**
* irdma_copy_ip_htonl - copy IP address from host to network order
* @dst: IP address in network order (big endian)
* @src: IP address in host order
*/
void irdma_copy_ip_htonl(__be32 *dst, u32 *src)
{
*dst++ = htonl(*src++);
*dst++ = htonl(*src++);
*dst++ = htonl(*src++);
*dst = htonl(*src);
}
/**
* irdma_get_addr_info
* @cm_node: contains ip/tcp info
* @cm_info: to get a copy of the cm_node ip/tcp info
*/
static void irdma_get_addr_info(struct irdma_cm_node *cm_node,
struct irdma_cm_info *cm_info)
{
memset(cm_info, 0, sizeof (*cm_info));
cm_info->ipv4 = cm_node->ipv4;
cm_info->vlan_id = cm_node->vlan_id;
memcpy(cm_info->loc_addr, cm_node->loc_addr, sizeof (cm_info->loc_addr));
memcpy(cm_info->rem_addr, cm_node->rem_addr, sizeof (cm_info->rem_addr));
cm_info->loc_port = cm_node->loc_port;
cm_info->rem_port = cm_node->rem_port;
}
/**
* irdma_fill_sockaddr4 - fill in addr info for IPv4 connection
* @cm_node: connection's node
* @event: upper layer's cm event
*/
static inline void irdma_fill_sockaddr4(struct irdma_cm_node *cm_node,
struct iw_cm_event *event)
{
struct sockaddr_in *laddr = (struct sockaddr_in *)&event->local_addr;
struct sockaddr_in *raddr = (struct sockaddr_in *)&event->remote_addr;
laddr->sin_family = AF_INET;
raddr->sin_family = AF_INET;
laddr->sin_port = htons(cm_node->loc_port);
raddr->sin_port = htons(cm_node->rem_port);
laddr->sin_addr.s_addr = htonl(cm_node->loc_addr[0]);
raddr->sin_addr.s_addr = htonl(cm_node->rem_addr[0]);
}
/**
* irdma_fill_sockaddr6 - fill in addr info for IPv6 connection
* @cm_node: connection's node
* @event: upper layer's cm event
*/
static inline void irdma_fill_sockaddr6(struct irdma_cm_node *cm_node,
struct iw_cm_event *event)
{
struct sockaddr_in6 *laddr6 = (struct sockaddr_in6 *)&event->local_addr;
struct sockaddr_in6 *raddr6 = (struct sockaddr_in6 *)&event->remote_addr;
laddr6->sin6_family = AF_INET6;
raddr6->sin6_family = AF_INET6;
laddr6->sin6_port = htons(cm_node->loc_port);
raddr6->sin6_port = htons(cm_node->rem_port);
irdma_copy_ip_htonl(laddr6->sin6_addr.in6_u.u6_addr32,
cm_node->loc_addr);
irdma_copy_ip_htonl(raddr6->sin6_addr.in6_u.u6_addr32,
cm_node->rem_addr);
}
/**
* irdma_get_cmevent_info - for cm event upcall
* @cm_node: connection's node
* @cm_id: upper layers cm struct for the event
* @event: upper layer's cm event
*/
static inline void irdma_get_cmevent_info(struct irdma_cm_node *cm_node,
struct iw_cm_id *cm_id,
struct iw_cm_event *event)
{
memcpy(&event->local_addr, &cm_id->m_local_addr,
sizeof (event->local_addr));
memcpy(&event->remote_addr, &cm_id->m_remote_addr,
sizeof (event->remote_addr));
if (cm_node) {
event->private_data = cm_node->pdata_buf;
event->private_data_len = (u8)cm_node->pdata.size;
event->ird = cm_node->ird_size;
event->ord = cm_node->ord_size;
}
}
/**
* irdma_send_cm_event - upcall cm's event handler
* @cm_node: connection's node
* @cm_id: upper layer's cm info struct
* @type: Event type to indicate
* @status: status for the event type
*/
static int irdma_send_cm_event(struct irdma_cm_node *cm_node,
struct iw_cm_id *cm_id,
enum iw_cm_event_type type, int status)
{
struct iw_cm_event event = {};
event.event = type;
event.status = status;
trace_irdma_send_cm_event(cm_node, cm_id, type, status,
__builtin_return_address(0));
ibdev_dbg(&cm_node->iwdev->ibdev,
"CM: cm_node %p cm_id=%p state=%d accel=%d event_type=%d status=%d\n" ,
cm_node, cm_id, cm_node->accelerated, cm_node->state, type,
status);
switch (type) {
case IW_CM_EVENT_CONNECT_REQUEST:
if (cm_node->ipv4)
irdma_fill_sockaddr4(cm_node, &event);
else
irdma_fill_sockaddr6(cm_node, &event);
event.provider_data = cm_node;
event.private_data = cm_node->pdata_buf;
event.private_data_len = (u8)cm_node->pdata.size;
event.ird = cm_node->ird_size;
break ;
case IW_CM_EVENT_CONNECT_REPLY:
irdma_get_cmevent_info(cm_node, cm_id, &event);
break ;
case IW_CM_EVENT_ESTABLISHED:
event.ird = cm_node->ird_size;
event.ord = cm_node->ord_size;
break ;
case IW_CM_EVENT_DISCONNECT:
case IW_CM_EVENT_CLOSE:
/* Wait if we are in RTS but havent issued the iwcm event upcall */
if (!cm_node->accelerated)
wait_for_completion(&cm_node->establish_comp);
break ;
default :
return -EINVAL;
}
return cm_id->event_handler(cm_id, &event);
}
/**
* irdma_timer_list_prep - add connection nodes to a list to perform timer tasks
* @cm_core: cm's core
* @timer_list: a timer list to which cm_node will be selected
*/
static void irdma_timer_list_prep(struct irdma_cm_core *cm_core,
struct list_head *timer_list)
{
struct irdma_cm_node *cm_node;
int bkt;
hash_for_each_rcu(cm_core->cm_hash_tbl, bkt, cm_node, list) {
if ((cm_node->close_entry || cm_node->send_entry) &&
refcount_inc_not_zero(&cm_node->refcnt))
list_add(&cm_node->timer_entry, timer_list);
}
}
/**
* irdma_create_event - create cm event
* @cm_node: connection's node
* @type: Event type to generate
*/
static struct irdma_cm_event *irdma_create_event(struct irdma_cm_node *cm_node,
enum irdma_cm_event_type type)
{
struct irdma_cm_event *event;
if (!cm_node->cm_id)
return NULL;
event = kzalloc(sizeof (*event), GFP_ATOMIC);
if (!event)
return NULL;
event->type = type;
event->cm_node = cm_node;
memcpy(event->cm_info.rem_addr, cm_node->rem_addr,
sizeof (event->cm_info.rem_addr));
memcpy(event->cm_info.loc_addr, cm_node->loc_addr,
sizeof (event->cm_info.loc_addr));
event->cm_info.rem_port = cm_node->rem_port;
event->cm_info.loc_port = cm_node->loc_port;
event->cm_info.cm_id = cm_node->cm_id;
ibdev_dbg(&cm_node->iwdev->ibdev,
"CM: node=%p event=%p type=%u dst=%pI4 src=%pI4\n" , cm_node,
event, type, event->cm_info.loc_addr,
event->cm_info.rem_addr);
trace_irdma_create_event(cm_node, type, __builtin_return_address(0));
irdma_cm_post_event(event);
return event;
}
/**
* irdma_free_retrans_entry - free send entry
* @cm_node: connection's node
*/
static void irdma_free_retrans_entry(struct irdma_cm_node *cm_node)
{
struct irdma_device *iwdev = cm_node->iwdev;
struct irdma_timer_entry *send_entry;
send_entry = cm_node->send_entry;
if (!send_entry)
return ;
cm_node->send_entry = NULL;
irdma_free_sqbuf(&iwdev->vsi, send_entry->sqbuf);
kfree(send_entry);
refcount_dec(&cm_node->refcnt);
}
/**
* irdma_cleanup_retrans_entry - free send entry with lock
* @cm_node: connection's node
*/
static void irdma_cleanup_retrans_entry(struct irdma_cm_node *cm_node)
{
unsigned long flags;
spin_lock_irqsave(&cm_node->retrans_list_lock, flags);
irdma_free_retrans_entry(cm_node);
spin_unlock_irqrestore(&cm_node->retrans_list_lock, flags);
}
/**
* irdma_form_ah_cm_frame - get a free packet and build frame with address handle
* @cm_node: connection's node ionfo to use in frame
* @options: pointer to options info
* @hdr: pointer mpa header
* @pdata: pointer to private data
* @flags: indicates FIN or ACK
*/
static struct irdma_puda_buf *irdma_form_ah_cm_frame(struct irdma_cm_node *cm_node,
struct irdma_kmem_info *options,
struct irdma_kmem_info *hdr,
struct irdma_mpa_priv_info *pdata,
u8 flags)
{
struct irdma_puda_buf *sqbuf;
struct irdma_sc_vsi *vsi = &cm_node->iwdev->vsi;
u8 *buf;
struct tcphdr *tcph;
u16 pktsize;
u32 opts_len = 0;
u32 pd_len = 0;
u32 hdr_len = 0;
if (!cm_node->ah || !cm_node->ah->ah_info.ah_valid) {
ibdev_dbg(&cm_node->iwdev->ibdev, "CM: AH invalid\n" );
return NULL;
}
sqbuf = irdma_puda_get_bufpool(vsi->ilq);
if (!sqbuf) {
ibdev_dbg(&cm_node->iwdev->ibdev, "CM: SQ buf NULL\n" );
return NULL;
}
sqbuf->ah_id = cm_node->ah->ah_info.ah_idx;
buf = sqbuf->mem.va;
if (options)
opts_len = (u32)options->size;
if (hdr)
hdr_len = hdr->size;
if (pdata)
pd_len = pdata->size;
pktsize = sizeof (*tcph) + opts_len + hdr_len + pd_len;
memset(buf, 0, sizeof (*tcph));
sqbuf->totallen = pktsize;
sqbuf->tcphlen = sizeof (*tcph) + opts_len;
sqbuf->scratch = cm_node;
tcph = (struct tcphdr *)buf;
buf += sizeof (*tcph);
tcph->source = htons(cm_node->loc_port);
tcph->dest = htons(cm_node->rem_port);
tcph->seq = htonl(cm_node->tcp_cntxt.loc_seq_num);
if (flags & SET_ACK) {
cm_node->tcp_cntxt.loc_ack_num = cm_node->tcp_cntxt.rcv_nxt;
tcph->ack_seq = htonl(cm_node->tcp_cntxt.loc_ack_num);
tcph->ack = 1;
} else {
tcph->ack_seq = 0;
}
if (flags & SET_SYN) {
cm_node->tcp_cntxt.loc_seq_num++;
tcph->syn = 1;
} else {
cm_node->tcp_cntxt.loc_seq_num += hdr_len + pd_len;
}
if (flags & SET_FIN) {
cm_node->tcp_cntxt.loc_seq_num++;
tcph->fin = 1;
}
if (flags & SET_RST)
tcph->rst = 1;
tcph->doff = (u16)((sizeof (*tcph) + opts_len + 3) >> 2);
sqbuf->tcphlen = tcph->doff << 2;
tcph->window = htons(cm_node->tcp_cntxt.rcv_wnd);
tcph->urg_ptr = 0;
if (opts_len) {
memcpy(buf, options->addr, opts_len);
buf += opts_len;
}
if (hdr_len) {
memcpy(buf, hdr->addr, hdr_len);
buf += hdr_len;
}
if (pdata && pdata->addr)
memcpy(buf, pdata->addr, pdata->size);
refcount_set(&sqbuf->refcount, 1);
print_hex_dump_debug("ILQ: TRANSMIT ILQ BUFFER" , DUMP_PREFIX_OFFSET,
16, 8, sqbuf->mem.va, sqbuf->totallen, false );
return sqbuf;
}
/**
* irdma_form_uda_cm_frame - get a free packet and build frame full tcpip packet
* @cm_node: connection's node ionfo to use in frame
* @options: pointer to options info
* @hdr: pointer mpa header
* @pdata: pointer to private data
* @flags: indicates FIN or ACK
*/
static struct irdma_puda_buf *irdma_form_uda_cm_frame(struct irdma_cm_node *cm_node,
struct irdma_kmem_info *options,
struct irdma_kmem_info *hdr,
struct irdma_mpa_priv_info *pdata,
u8 flags)
{
struct irdma_puda_buf *sqbuf;
struct irdma_sc_vsi *vsi = &cm_node->iwdev->vsi;
u8 *buf;
struct tcphdr *tcph;
struct iphdr *iph;
struct ipv6hdr *ip6h;
struct ethhdr *ethh;
u16 pktsize;
u16 eth_hlen = ETH_HLEN;
u32 opts_len = 0;
u32 pd_len = 0;
u32 hdr_len = 0;
u16 vtag;
sqbuf = irdma_puda_get_bufpool(vsi->ilq);
if (!sqbuf)
return NULL;
buf = sqbuf->mem.va;
if (options)
opts_len = (u32)options->size;
if (hdr)
hdr_len = hdr->size;
if (pdata)
pd_len = pdata->size;
if (cm_node->vlan_id < VLAN_N_VID)
eth_hlen += 4;
if (cm_node->ipv4)
pktsize = sizeof (*iph) + sizeof (*tcph);
else
pktsize = sizeof (*ip6h) + sizeof (*tcph);
pktsize += opts_len + hdr_len + pd_len;
memset(buf, 0, eth_hlen + pktsize);
sqbuf->totallen = pktsize + eth_hlen;
sqbuf->maclen = eth_hlen;
sqbuf->tcphlen = sizeof (*tcph) + opts_len;
sqbuf->scratch = cm_node;
ethh = (struct ethhdr *)buf;
buf += eth_hlen;
if (cm_node->do_lpb)
sqbuf->do_lpb = true ;
if (cm_node->ipv4) {
sqbuf->ipv4 = true ;
iph = (struct iphdr *)buf;
buf += sizeof (*iph);
tcph = (struct tcphdr *)buf;
buf += sizeof (*tcph);
ether_addr_copy(ethh->h_dest, cm_node->rem_mac);
ether_addr_copy(ethh->h_source, cm_node->loc_mac);
if (cm_node->vlan_id < VLAN_N_VID) {
((struct vlan_ethhdr *)ethh)->h_vlan_proto =
htons(ETH_P_8021Q);
vtag = (cm_node->user_pri << VLAN_PRIO_SHIFT) |
cm_node->vlan_id;
((struct vlan_ethhdr *)ethh)->h_vlan_TCI = htons(vtag);
((struct vlan_ethhdr *)ethh)->h_vlan_encapsulated_proto =
htons(ETH_P_IP);
} else {
ethh->h_proto = htons(ETH_P_IP);
}
iph->version = IPVERSION;
iph->ihl = 5; /* 5 * 4Byte words, IP headr len */
iph->tos = cm_node->tos;
iph->tot_len = htons(pktsize);
iph->id = htons(++cm_node->tcp_cntxt.loc_id);
iph->frag_off = htons(0x4000);
iph->ttl = 0x40;
iph->protocol = IPPROTO_TCP;
iph->saddr = htonl(cm_node->loc_addr[0]);
iph->daddr = htonl(cm_node->rem_addr[0]);
} else {
sqbuf->ipv4 = false ;
ip6h = (struct ipv6hdr *)buf;
buf += sizeof (*ip6h);
tcph = (struct tcphdr *)buf;
buf += sizeof (*tcph);
ether_addr_copy(ethh->h_dest, cm_node->rem_mac);
ether_addr_copy(ethh->h_source, cm_node->loc_mac);
if (cm_node->vlan_id < VLAN_N_VID) {
((struct vlan_ethhdr *)ethh)->h_vlan_proto =
htons(ETH_P_8021Q);
vtag = (cm_node->user_pri << VLAN_PRIO_SHIFT) |
cm_node->vlan_id;
((struct vlan_ethhdr *)ethh)->h_vlan_TCI = htons(vtag);
((struct vlan_ethhdr *)ethh)->h_vlan_encapsulated_proto =
htons(ETH_P_IPV6);
} else {
ethh->h_proto = htons(ETH_P_IPV6);
}
ip6h->version = 6;
ip6h->priority = cm_node->tos >> 4;
ip6h->flow_lbl[0] = cm_node->tos << 4;
ip6h->flow_lbl[1] = 0;
ip6h->flow_lbl[2] = 0;
ip6h->payload_len = htons(pktsize - sizeof (*ip6h));
ip6h->nexthdr = 6;
ip6h->hop_limit = 128;
irdma_copy_ip_htonl(ip6h->saddr.in6_u.u6_addr32,
cm_node->loc_addr);
irdma_copy_ip_htonl(ip6h->daddr.in6_u.u6_addr32,
cm_node->rem_addr);
}
tcph->source = htons(cm_node->loc_port);
tcph->dest = htons(cm_node->rem_port);
tcph->seq = htonl(cm_node->tcp_cntxt.loc_seq_num);
if (flags & SET_ACK) {
cm_node->tcp_cntxt.loc_ack_num = cm_node->tcp_cntxt.rcv_nxt;
tcph->ack_seq = htonl(cm_node->tcp_cntxt.loc_ack_num);
tcph->ack = 1;
} else {
tcph->ack_seq = 0;
}
if (flags & SET_SYN) {
cm_node->tcp_cntxt.loc_seq_num++;
tcph->syn = 1;
} else {
cm_node->tcp_cntxt.loc_seq_num += hdr_len + pd_len;
}
if (flags & SET_FIN) {
cm_node->tcp_cntxt.loc_seq_num++;
tcph->fin = 1;
}
if (flags & SET_RST)
tcph->rst = 1;
tcph->doff = (u16)((sizeof (*tcph) + opts_len + 3) >> 2);
sqbuf->tcphlen = tcph->doff << 2;
tcph->window = htons(cm_node->tcp_cntxt.rcv_wnd);
tcph->urg_ptr = 0;
if (opts_len) {
memcpy(buf, options->addr, opts_len);
buf += opts_len;
}
if (hdr_len) {
memcpy(buf, hdr->addr, hdr_len);
buf += hdr_len;
}
if (pdata && pdata->addr)
memcpy(buf, pdata->addr, pdata->size);
refcount_set(&sqbuf->refcount, 1);
print_hex_dump_debug("ILQ: TRANSMIT ILQ BUFFER" , DUMP_PREFIX_OFFSET,
16, 8, sqbuf->mem.va, sqbuf->totallen, false );
return sqbuf;
}
/**
* irdma_send_reset - Send RST packet
* @cm_node: connection's node
*/
int irdma_send_reset(struct irdma_cm_node *cm_node)
{
struct irdma_puda_buf *sqbuf;
int flags = SET_RST | SET_ACK;
trace_irdma_send_reset(cm_node, 0, __builtin_return_address(0));
sqbuf = cm_node->cm_core->form_cm_frame(cm_node, NULL, NULL, NULL,
flags);
if (!sqbuf)
return -ENOMEM;
ibdev_dbg(&cm_node->iwdev->ibdev,
"CM: caller: %pS cm_node %p cm_id=%p accel=%d state=%d rem_port=0x%04x, loc_port=0x%04x rem_addr=%pI4 loc_addr=%pI4\n" ,
__builtin_return_address(0), cm_node, cm_node->cm_id,
cm_node->accelerated, cm_node->state, cm_node->rem_port,
cm_node->loc_port, cm_node->rem_addr, cm_node->loc_addr);
return irdma_schedule_cm_timer(cm_node, sqbuf, IRDMA_TIMER_TYPE_SEND, 0,
1);
}
/**
* irdma_active_open_err - send event for active side cm error
* @cm_node: connection's node
* @reset: Flag to send reset or not
*/
static void irdma_active_open_err(struct irdma_cm_node *cm_node, bool reset)
{
trace_irdma_active_open_err(cm_node, reset,
__builtin_return_address(0));
irdma_cleanup_retrans_entry(cm_node);
cm_node->cm_core->stats_connect_errs++;
if (reset) {
ibdev_dbg(&cm_node->iwdev->ibdev,
"CM: cm_node=%p state=%d\n" , cm_node,
cm_node->state);
refcount_inc(&cm_node->refcnt);
irdma_send_reset(cm_node);
}
cm_node->state = IRDMA_CM_STATE_CLOSED;
irdma_create_event(cm_node, IRDMA_CM_EVENT_ABORTED);
}
/**
* irdma_passive_open_err - handle passive side cm error
* @cm_node: connection's node
* @reset: send reset or just free cm_node
*/
static void irdma_passive_open_err(struct irdma_cm_node *cm_node, bool reset)
{
irdma_cleanup_retrans_entry(cm_node);
cm_node->cm_core->stats_passive_errs++;
cm_node->state = IRDMA_CM_STATE_CLOSED;
ibdev_dbg(&cm_node->iwdev->ibdev, "CM: cm_node=%p state =%d\n" ,
cm_node, cm_node->state);
trace_irdma_passive_open_err(cm_node, reset,
__builtin_return_address(0));
if (reset)
irdma_send_reset(cm_node);
else
irdma_rem_ref_cm_node(cm_node);
}
/**
* irdma_event_connect_error - to create connect error event
* @event: cm information for connect event
*/
static void irdma_event_connect_error(struct irdma_cm_event *event)
{
struct irdma_qp *iwqp;
struct iw_cm_id *cm_id;
cm_id = event->cm_node->cm_id;
if (!cm_id)
return ;
iwqp = cm_id->provider_data;
if (!iwqp || !iwqp->iwdev)
return ;
iwqp->cm_id = NULL;
cm_id->provider_data = NULL;
irdma_send_cm_event(event->cm_node, cm_id, IW_CM_EVENT_CONNECT_REPLY,
-ECONNRESET);
irdma_rem_ref_cm_node(event->cm_node);
}
/**
* irdma_process_options - process options from TCP header
* @cm_node: connection's node
* @optionsloc: point to start of options
* @optionsize: size of all options
* @syn_pkt: flag if syn packet
*/
static int irdma_process_options(struct irdma_cm_node *cm_node, u8 *optionsloc,
u32 optionsize, u32 syn_pkt)
{
u32 tmp;
u32 offset = 0;
union all_known_options *all_options;
char got_mss_option = 0;
while (offset < optionsize) {
all_options = (union all_known_options *)(optionsloc + offset);
switch (all_options->base.optionnum) {
case OPTION_NUM_EOL:
offset = optionsize;
break ;
case OPTION_NUM_NONE:
offset += 1;
continue ;
case OPTION_NUM_MSS:
ibdev_dbg(&cm_node->iwdev->ibdev,
"CM: MSS Length: %d Offset: %d Size: %d\n" ,
all_options->mss.len, offset, optionsize);
got_mss_option = 1;
if (all_options->mss.len != 4)
return -EINVAL;
tmp = ntohs(all_options->mss.mss);
if ((cm_node->ipv4 &&
(tmp + IRDMA_MTU_TO_MSS_IPV4) < IRDMA_MIN_MTU_IPV4) ||
(!cm_node->ipv4 &&
(tmp + IRDMA_MTU_TO_MSS_IPV6) < IRDMA_MIN_MTU_IPV6))
return -EINVAL;
if (tmp < cm_node->tcp_cntxt.mss)
cm_node->tcp_cntxt.mss = tmp;
break ;
case OPTION_NUM_WINDOW_SCALE:
cm_node->tcp_cntxt.snd_wscale =
all_options->windowscale.shiftcount;
break ;
default :
ibdev_dbg(&cm_node->iwdev->ibdev,
"CM: Unsupported TCP Option: %x\n" ,
all_options->base.optionnum);
break ;
}
offset += all_options->base.len;
}
if (!got_mss_option && syn_pkt)
cm_node->tcp_cntxt.mss = IRDMA_CM_DEFAULT_MSS;
return 0;
}
/**
* irdma_handle_tcp_options - setup TCP context info after parsing TCP options
* @cm_node: connection's node
* @tcph: pointer tcp header
* @optionsize: size of options rcvd
* @passive: active or passive flag
*/
static int irdma_handle_tcp_options(struct irdma_cm_node *cm_node,
struct tcphdr *tcph, int optionsize,
int passive)
{
u8 *optionsloc = (u8 *)&tcph[1];
int ret;
if (optionsize) {
ret = irdma_process_options(cm_node, optionsloc, optionsize,
(u32)tcph->syn);
if (ret) {
ibdev_dbg(&cm_node->iwdev->ibdev,
"CM: Node %p, Sending Reset\n" , cm_node);
if (passive)
irdma_passive_open_err(cm_node, true );
else
irdma_active_open_err(cm_node, true );
return ret;
}
}
cm_node->tcp_cntxt.snd_wnd = ntohs(tcph->window)
<< cm_node->tcp_cntxt.snd_wscale;
if (cm_node->tcp_cntxt.snd_wnd > cm_node->tcp_cntxt.max_snd_wnd)
cm_node->tcp_cntxt.max_snd_wnd = cm_node->tcp_cntxt.snd_wnd;
return 0;
}
/**
* irdma_build_mpa_v1 - build a MPA V1 frame
* @cm_node: connection's node
* @start_addr: address where to build frame
* @mpa_key: to do read0 or write0
*/
static void irdma_build_mpa_v1(struct irdma_cm_node *cm_node, void *start_addr,
u8 mpa_key)
{
struct ietf_mpa_v1 *mpa_frame = start_addr;
switch (mpa_key) {
case MPA_KEY_REQUEST:
memcpy(mpa_frame->key, IEFT_MPA_KEY_REQ, IETF_MPA_KEY_SIZE);
break ;
case MPA_KEY_REPLY:
memcpy(mpa_frame->key, IEFT_MPA_KEY_REP, IETF_MPA_KEY_SIZE);
break ;
default :
break ;
}
mpa_frame->flags = IETF_MPA_FLAGS_CRC;
mpa_frame->rev = cm_node->mpa_frame_rev;
mpa_frame->priv_data_len = htons(cm_node->pdata.size);
}
/**
* irdma_build_mpa_v2 - build a MPA V2 frame
* @cm_node: connection's node
* @start_addr: buffer start address
* @mpa_key: to do read0 or write0
*/
static void irdma_build_mpa_v2(struct irdma_cm_node *cm_node, void *start_addr,
u8 mpa_key)
{
struct ietf_mpa_v2 *mpa_frame = start_addr;
struct ietf_rtr_msg *rtr_msg = &mpa_frame->rtr_msg;
u16 ctrl_ird, ctrl_ord;
/* initialize the upper 5 bytes of the frame */
irdma_build_mpa_v1(cm_node, start_addr, mpa_key);
mpa_frame->flags |= IETF_MPA_V2_FLAG;
if (cm_node->iwdev->iw_ooo) {
mpa_frame->flags |= IETF_MPA_FLAGS_MARKERS;
cm_node->rcv_mark_en = true ;
}
mpa_frame->priv_data_len = cpu_to_be16(be16_to_cpu(mpa_frame->priv_data_len) +
IETF_RTR_MSG_SIZE);
/* initialize RTR msg */
if (cm_node->mpav2_ird_ord == IETF_NO_IRD_ORD) {
ctrl_ird = IETF_NO_IRD_ORD;
ctrl_ord = IETF_NO_IRD_ORD;
} else {
ctrl_ird = (cm_node->ird_size > IETF_NO_IRD_ORD) ?
IETF_NO_IRD_ORD :
cm_node->ird_size;
ctrl_ord = (cm_node->ord_size > IETF_NO_IRD_ORD) ?
IETF_NO_IRD_ORD :
cm_node->ord_size;
}
ctrl_ird |= IETF_PEER_TO_PEER;
switch (mpa_key) {
case MPA_KEY_REQUEST:
ctrl_ord |= IETF_RDMA0_WRITE;
ctrl_ord |= IETF_RDMA0_READ;
break ;
case MPA_KEY_REPLY:
switch (cm_node->send_rdma0_op) {
case SEND_RDMA_WRITE_ZERO:
ctrl_ord |= IETF_RDMA0_WRITE;
break ;
case SEND_RDMA_READ_ZERO:
ctrl_ord |= IETF_RDMA0_READ;
break ;
}
break ;
default :
break ;
}
rtr_msg->ctrl_ird = htons(ctrl_ird);
rtr_msg->ctrl_ord = htons(ctrl_ord);
}
/**
* irdma_cm_build_mpa_frame - build mpa frame for mpa version 1 or version 2
* @cm_node: connection's node
* @mpa: mpa: data buffer
* @mpa_key: to do read0 or write0
*/
static int irdma_cm_build_mpa_frame(struct irdma_cm_node *cm_node,
struct irdma_kmem_info *mpa, u8 mpa_key)
{
int hdr_len = 0;
switch (cm_node->mpa_frame_rev) {
case IETF_MPA_V1:
hdr_len = sizeof (struct ietf_mpa_v1);
irdma_build_mpa_v1(cm_node, mpa->addr, mpa_key);
break ;
case IETF_MPA_V2:
hdr_len = sizeof (struct ietf_mpa_v2);
irdma_build_mpa_v2(cm_node, mpa->addr, mpa_key);
break ;
default :
break ;
}
return hdr_len;
}
/**
* irdma_send_mpa_request - active node send mpa request to passive node
* @cm_node: connection's node
*/
static int irdma_send_mpa_request(struct irdma_cm_node *cm_node)
{
struct irdma_puda_buf *sqbuf;
cm_node->mpa_hdr.addr = &cm_node->mpa_v2_frame;
cm_node->mpa_hdr.size = irdma_cm_build_mpa_frame(cm_node,
&cm_node->mpa_hdr,
MPA_KEY_REQUEST);
if (!cm_node->mpa_hdr.size) {
ibdev_dbg(&cm_node->iwdev->ibdev,
"CM: mpa size = %d\n" , cm_node->mpa_hdr.size);
return -EINVAL;
}
sqbuf = cm_node->cm_core->form_cm_frame(cm_node, NULL,
&cm_node->mpa_hdr,
&cm_node->pdata, SET_ACK);
if (!sqbuf)
return -ENOMEM;
return irdma_schedule_cm_timer(cm_node, sqbuf, IRDMA_TIMER_TYPE_SEND, 1,
0);
}
/**
* irdma_send_mpa_reject -
* @cm_node: connection's node
* @pdata: reject data for connection
* @plen: length of reject data
*/
static int irdma_send_mpa_reject(struct irdma_cm_node *cm_node,
const void *pdata, u8 plen)
{
struct irdma_puda_buf *sqbuf;
struct irdma_mpa_priv_info priv_info;
cm_node->mpa_hdr.addr = &cm_node->mpa_v2_frame;
cm_node->mpa_hdr.size = irdma_cm_build_mpa_frame(cm_node,
&cm_node->mpa_hdr,
MPA_KEY_REPLY);
cm_node->mpa_frame.flags |= IETF_MPA_FLAGS_REJECT;
priv_info.addr = pdata;
priv_info.size = plen;
sqbuf = cm_node->cm_core->form_cm_frame(cm_node, NULL,
&cm_node->mpa_hdr, &priv_info,
SET_ACK | SET_FIN);
if (!sqbuf)
return -ENOMEM;
cm_node->state = IRDMA_CM_STATE_FIN_WAIT1;
return irdma_schedule_cm_timer(cm_node, sqbuf, IRDMA_TIMER_TYPE_SEND, 1,
0);
}
/**
* irdma_negotiate_mpa_v2_ird_ord - negotiate MPAv2 IRD/ORD
* @cm_node: connection's node
* @buf: Data pointer
*/
static int irdma_negotiate_mpa_v2_ird_ord(struct irdma_cm_node *cm_node,
u8 *buf)
{
struct ietf_mpa_v2 *mpa_v2_frame;
struct ietf_rtr_msg *rtr_msg;
u16 ird_size;
u16 ord_size;
u16 ctrl_ord;
u16 ctrl_ird;
mpa_v2_frame = (struct ietf_mpa_v2 *)buf;
rtr_msg = &mpa_v2_frame->rtr_msg;
/* parse rtr message */
ctrl_ord = ntohs(rtr_msg->ctrl_ord);
ctrl_ird = ntohs(rtr_msg->ctrl_ird);
ird_size = ctrl_ird & IETF_NO_IRD_ORD;
ord_size = ctrl_ord & IETF_NO_IRD_ORD;
if (!(ctrl_ird & IETF_PEER_TO_PEER))
return -EOPNOTSUPP;
if (ird_size == IETF_NO_IRD_ORD || ord_size == IETF_NO_IRD_ORD) {
cm_node->mpav2_ird_ord = IETF_NO_IRD_ORD;
goto negotiate_done;
}
if (cm_node->state != IRDMA_CM_STATE_MPAREQ_SENT) {
/* responder */
if (!ord_size && (ctrl_ord & IETF_RDMA0_READ))
cm_node->ird_size = 1;
if (cm_node->ord_size > ird_size)
cm_node->ord_size = ird_size;
} else {
/* initiator */
if (!ird_size && (ctrl_ord & IETF_RDMA0_READ))
/* Remote peer doesn't support RDMA0_READ */
return -EOPNOTSUPP;
if (cm_node->ord_size > ird_size)
cm_node->ord_size = ird_size;
if (cm_node->ird_size < ord_size)
/* no resources available */
return -EINVAL;
}
negotiate_done:
if (ctrl_ord & IETF_RDMA0_READ)
cm_node->send_rdma0_op = SEND_RDMA_READ_ZERO;
else if (ctrl_ord & IETF_RDMA0_WRITE)
cm_node->send_rdma0_op = SEND_RDMA_WRITE_ZERO;
else
/* Not supported RDMA0 operation */
return -EOPNOTSUPP;
ibdev_dbg(&cm_node->iwdev->ibdev,
"CM: MPAV2 Negotiated ORD: %d, IRD: %d\n" ,
cm_node->ord_size, cm_node->ird_size);
trace_irdma_negotiate_mpa_v2(cm_node);
return 0;
}
/**
* irdma_parse_mpa - process an IETF MPA frame
* @cm_node: connection's node
* @buf: Data pointer
* @type: to return accept or reject
* @len: Len of mpa buffer
*/
static int irdma_parse_mpa(struct irdma_cm_node *cm_node, u8 *buf, u32 *type,
u32 len)
{
struct ietf_mpa_v1 *mpa_frame;
int mpa_hdr_len, priv_data_len, ret;
*type = IRDMA_MPA_REQUEST_ACCEPT;
if (len < sizeof (struct ietf_mpa_v1)) {
ibdev_dbg(&cm_node->iwdev->ibdev,
"CM: ietf buffer small (%x)\n" , len);
return -EINVAL;
}
mpa_frame = (struct ietf_mpa_v1 *)buf;
mpa_hdr_len = sizeof (struct ietf_mpa_v1);
priv_data_len = ntohs(mpa_frame->priv_data_len);
if (priv_data_len > IETF_MAX_PRIV_DATA_LEN) {
ibdev_dbg(&cm_node->iwdev->ibdev,
"CM: private_data too big %d\n" , priv_data_len);
return -EOVERFLOW;
}
if (mpa_frame->rev != IETF_MPA_V1 && mpa_frame->rev != IETF_MPA_V2) {
ibdev_dbg(&cm_node->iwdev->ibdev,
"CM: unsupported mpa rev = %d\n" , mpa_frame->rev);
return -EINVAL;
}
if (mpa_frame->rev > cm_node->mpa_frame_rev) {
ibdev_dbg(&cm_node->iwdev->ibdev, "CM: rev %d\n" ,
mpa_frame->rev);
return -EINVAL;
}
cm_node->mpa_frame_rev = mpa_frame->rev;
if (cm_node->state != IRDMA_CM_STATE_MPAREQ_SENT) {
if (memcmp(mpa_frame->key, IEFT_MPA_KEY_REQ,
IETF_MPA_KEY_SIZE)) {
ibdev_dbg(&cm_node->iwdev->ibdev,
"CM: Unexpected MPA Key received\n" );
return -EINVAL;
}
} else {
if (memcmp(mpa_frame->key, IEFT_MPA_KEY_REP,
IETF_MPA_KEY_SIZE)) {
ibdev_dbg(&cm_node->iwdev->ibdev,
"CM: Unexpected MPA Key received\n" );
return -EINVAL;
}
}
if (priv_data_len + mpa_hdr_len > len) {
ibdev_dbg(&cm_node->iwdev->ibdev,
"CM: ietf buffer len(%x + %x != %x)\n" ,
priv_data_len, mpa_hdr_len, len);
return -EOVERFLOW;
}
if (len > IRDMA_MAX_CM_BUF) {
ibdev_dbg(&cm_node->iwdev->ibdev,
"CM: ietf buffer large len = %d\n" , len);
return -EOVERFLOW;
}
switch (mpa_frame->rev) {
case IETF_MPA_V2:
mpa_hdr_len += IETF_RTR_MSG_SIZE;
ret = irdma_negotiate_mpa_v2_ird_ord(cm_node, buf);
if (ret)
return ret;
break ;
case IETF_MPA_V1:
default :
break ;
}
memcpy(cm_node->pdata_buf, buf + mpa_hdr_len, priv_data_len);
cm_node->pdata.size = priv_data_len;
if (mpa_frame->flags & IETF_MPA_FLAGS_REJECT)
*type = IRDMA_MPA_REQUEST_REJECT;
if (mpa_frame->flags & IETF_MPA_FLAGS_MARKERS)
cm_node->snd_mark_en = true ;
return 0;
}
/**
* irdma_schedule_cm_timer
* @cm_node: connection's node
* @sqbuf: buffer to send
* @type: if it is send or close
* @send_retrans: if rexmits to be done
* @close_when_complete: is cm_node to be removed
*
* note - cm_node needs to be protected before calling this. Encase in:
* irdma_rem_ref_cm_node(cm_core, cm_node);
* irdma_schedule_cm_timer(...)
* refcount_inc(&cm_node->refcnt);
*/
int irdma_schedule_cm_timer(struct irdma_cm_node *cm_node,
struct irdma_puda_buf *sqbuf,
enum irdma_timer_type type, int send_retrans,
int close_when_complete)
{
struct irdma_sc_vsi *vsi = &cm_node->iwdev->vsi;
struct irdma_cm_core *cm_core = cm_node->cm_core;
struct irdma_timer_entry *new_send;
u32 was_timer_set;
unsigned long flags;
new_send = kzalloc(sizeof (*new_send), GFP_ATOMIC);
if (!new_send) {
if (type != IRDMA_TIMER_TYPE_CLOSE)
irdma_free_sqbuf(vsi, sqbuf);
return -ENOMEM;
}
new_send->retrycount = IRDMA_DEFAULT_RETRYS;
new_send->retranscount = IRDMA_DEFAULT_RETRANS;
new_send->sqbuf = sqbuf;
new_send->timetosend = jiffies;
new_send->type = type;
new_send->send_retrans = send_retrans;
new_send->close_when_complete = close_when_complete;
if (type == IRDMA_TIMER_TYPE_CLOSE) {
new_send->timetosend += (HZ / 10);
if (cm_node->close_entry) {
kfree(new_send);
ibdev_dbg(&cm_node->iwdev->ibdev,
"CM: already close entry\n" );
return -EINVAL;
}
cm_node->close_entry = new_send;
} else { /* type == IRDMA_TIMER_TYPE_SEND */
spin_lock_irqsave(&cm_node->retrans_list_lock, flags);
cm_node->send_entry = new_send;
refcount_inc(&cm_node->refcnt);
spin_unlock_irqrestore(&cm_node->retrans_list_lock, flags);
new_send->timetosend = jiffies + IRDMA_RETRY_TIMEOUT;
refcount_inc(&sqbuf->refcount);
irdma_puda_send_buf(vsi->ilq, sqbuf);
if (!send_retrans) {
irdma_cleanup_retrans_entry(cm_node);
if (close_when_complete)
irdma_rem_ref_cm_node(cm_node);
return 0;
}
}
spin_lock_irqsave(&cm_core->ht_lock, flags);
was_timer_set = timer_pending(&cm_core->tcp_timer);
if (!was_timer_set) {
cm_core->tcp_timer.expires = new_send->timetosend;
add_timer(&cm_core->tcp_timer);
}
spin_unlock_irqrestore(&cm_core->ht_lock, flags);
return 0;
}
/**
* irdma_retrans_expired - Could not rexmit the packet
* @cm_node: connection's node
*/
static void irdma_retrans_expired(struct irdma_cm_node *cm_node)
{
enum irdma_cm_node_state state = cm_node->state;
cm_node->state = IRDMA_CM_STATE_CLOSED;
switch (state) {
case IRDMA_CM_STATE_SYN_RCVD:
case IRDMA_CM_STATE_CLOSING:
irdma_rem_ref_cm_node(cm_node);
break ;
case IRDMA_CM_STATE_FIN_WAIT1:
case IRDMA_CM_STATE_LAST_ACK:
irdma_send_reset(cm_node);
break ;
default :
refcount_inc(&cm_node->refcnt);
irdma_send_reset(cm_node);
irdma_create_event(cm_node, IRDMA_CM_EVENT_ABORTED);
break ;
}
}
/**
* irdma_handle_close_entry - for handling retry/timeouts
* @cm_node: connection's node
* @rem_node: flag for remove cm_node
*/
static void irdma_handle_close_entry(struct irdma_cm_node *cm_node,
u32 rem_node)
{
struct irdma_timer_entry *close_entry = cm_node->close_entry;
struct irdma_qp *iwqp;
unsigned long flags;
if (!close_entry)
return ;
iwqp = (struct irdma_qp *)close_entry->sqbuf;
if (iwqp) {
spin_lock_irqsave(&iwqp->lock, flags);
if (iwqp->cm_id) {
iwqp->hw_tcp_state = IRDMA_TCP_STATE_CLOSED;
iwqp->hw_iwarp_state = IRDMA_QP_STATE_ERROR;
iwqp->last_aeq = IRDMA_AE_RESET_SENT;
iwqp->ibqp_state = IB_QPS_ERR;
spin_unlock_irqrestore(&iwqp->lock, flags);
irdma_cm_disconn(iwqp);
} else {
spin_unlock_irqrestore(&iwqp->lock, flags);
}
} else if (rem_node) {
/* TIME_WAIT state */
irdma_rem_ref_cm_node(cm_node);
}
kfree(close_entry);
cm_node->close_entry = NULL;
}
/**
* irdma_cm_timer_tick - system's timer expired callback
* @t: Pointer to timer_list
*/
static void irdma_cm_timer_tick(struct timer_list *t)
{
unsigned long nexttimeout = jiffies + IRDMA_LONG_TIME;
struct irdma_cm_node *cm_node;
struct irdma_timer_entry *send_entry, *close_entry;
struct list_head *list_core_temp;
struct list_head *list_node;
struct irdma_cm_core *cm_core = timer_container_of(cm_core, t,
tcp_timer);
struct irdma_sc_vsi *vsi;
u32 settimer = 0;
unsigned long timetosend;
unsigned long flags;
struct list_head timer_list;
INIT_LIST_HEAD(&timer_list);
rcu_read_lock();
irdma_timer_list_prep(cm_core, &timer_list);
rcu_read_unlock();
list_for_each_safe (list_node, list_core_temp, &timer_list) {
cm_node = container_of(list_node, struct irdma_cm_node,
timer_entry);
close_entry = cm_node->close_entry;
if (close_entry) {
if (time_after(close_entry->timetosend, jiffies)) {
if (nexttimeout > close_entry->timetosend ||
!settimer) {
nexttimeout = close_entry->timetosend;
settimer = 1;
}
} else {
irdma_handle_close_entry(cm_node, 1);
}
}
spin_lock_irqsave(&cm_node->retrans_list_lock, flags);
send_entry = cm_node->send_entry;
if (!send_entry)
goto done;
if (time_after(send_entry->timetosend, jiffies)) {
if (cm_node->state != IRDMA_CM_STATE_OFFLOADED) {
if (nexttimeout > send_entry->timetosend ||
!settimer) {
nexttimeout = send_entry->timetosend;
settimer = 1;
}
} else {
irdma_free_retrans_entry(cm_node);
}
goto done;
}
if (cm_node->state == IRDMA_CM_STATE_OFFLOADED ||
cm_node->state == IRDMA_CM_STATE_CLOSED) {
irdma_free_retrans_entry(cm_node);
goto done;
}
if (!send_entry->retranscount || !send_entry->retrycount) {
irdma_free_retrans_entry(cm_node);
spin_unlock_irqrestore(&cm_node->retrans_list_lock,
flags);
irdma_retrans_expired(cm_node);
cm_node->state = IRDMA_CM_STATE_CLOSED;
spin_lock_irqsave(&cm_node->retrans_list_lock, flags);
goto done;
}
spin_unlock_irqrestore(&cm_node->retrans_list_lock, flags);
vsi = &cm_node->iwdev->vsi;
if (!cm_node->ack_rcvd) {
refcount_inc(&send_entry->sqbuf->refcount);
irdma_puda_send_buf(vsi->ilq, send_entry->sqbuf);
cm_node->cm_core->stats_pkt_retrans++;
}
spin_lock_irqsave(&cm_node->retrans_list_lock, flags);
if (send_entry->send_retrans) {
send_entry->retranscount--;
timetosend = (IRDMA_RETRY_TIMEOUT <<
(IRDMA_DEFAULT_RETRANS -
send_entry->retranscount));
send_entry->timetosend = jiffies +
min(timetosend, IRDMA_MAX_TIMEOUT);
if (nexttimeout > send_entry->timetosend || !settimer) {
nexttimeout = send_entry->timetosend;
settimer = 1;
}
} else {
int close_when_complete;
close_when_complete = send_entry->close_when_complete;
irdma_free_retrans_entry(cm_node);
if (close_when_complete)
irdma_rem_ref_cm_node(cm_node);
}
done:
spin_unlock_irqrestore(&cm_node->retrans_list_lock, flags);
irdma_rem_ref_cm_node(cm_node);
}
if (settimer) {
spin_lock_irqsave(&cm_core->ht_lock, flags);
if (!timer_pending(&cm_core->tcp_timer)) {
cm_core->tcp_timer.expires = nexttimeout;
add_timer(&cm_core->tcp_timer);
}
spin_unlock_irqrestore(&cm_core->ht_lock, flags);
}
}
/**
* irdma_send_syn - send SYN packet
* @cm_node: connection's node
* @sendack: flag to set ACK bit or not
*/
int irdma_send_syn(struct irdma_cm_node *cm_node, u32 sendack)
{
struct irdma_puda_buf *sqbuf;
int flags = SET_SYN;
char optionsbuf[sizeof (struct option_mss) +
sizeof (struct option_windowscale) +
sizeof (struct option_base) + TCP_OPTIONS_PADDING];
struct irdma_kmem_info opts;
int optionssize = 0;
/* Sending MSS option */
union all_known_options *options;
opts.addr = optionsbuf;
if (!cm_node)
return -EINVAL;
options = (union all_known_options *)&optionsbuf[optionssize];
options->mss.optionnum = OPTION_NUM_MSS;
options->mss.len = sizeof (struct option_mss);
options->mss.mss = htons(cm_node->tcp_cntxt.mss);
optionssize += sizeof (struct option_mss);
options = (union all_known_options *)&optionsbuf[optionssize];
options->windowscale.optionnum = OPTION_NUM_WINDOW_SCALE;
options->windowscale.len = sizeof (struct option_windowscale);
options->windowscale.shiftcount = cm_node->tcp_cntxt.rcv_wscale;
optionssize += sizeof (struct option_windowscale);
options = (union all_known_options *)&optionsbuf[optionssize];
options->eol = OPTION_NUM_EOL;
optionssize += 1;
if (sendack)
flags |= SET_ACK;
opts.size = optionssize;
sqbuf = cm_node->cm_core->form_cm_frame(cm_node, &opts, NULL, NULL,
flags);
if (!sqbuf)
return -ENOMEM;
return irdma_schedule_cm_timer(cm_node, sqbuf, IRDMA_TIMER_TYPE_SEND, 1,
0);
}
/**
* irdma_send_ack - Send ACK packet
* @cm_node: connection's node
*/
void irdma_send_ack(struct irdma_cm_node *cm_node)
{
struct irdma_puda_buf *sqbuf;
struct irdma_sc_vsi *vsi = &cm_node->iwdev->vsi;
sqbuf = cm_node->cm_core->form_cm_frame(cm_node, NULL, NULL, NULL,
SET_ACK);
if (sqbuf)
irdma_puda_send_buf(vsi->ilq, sqbuf);
}
/**
* irdma_send_fin - Send FIN pkt
* @cm_node: connection's node
*/
static int irdma_send_fin(struct irdma_cm_node *cm_node)
{
struct irdma_puda_buf *sqbuf;
sqbuf = cm_node->cm_core->form_cm_frame(cm_node, NULL, NULL, NULL,
SET_ACK | SET_FIN);
if (!sqbuf)
return -ENOMEM;
return irdma_schedule_cm_timer(cm_node, sqbuf, IRDMA_TIMER_TYPE_SEND, 1,
0);
}
/**
* irdma_find_listener - find a cm node listening on this addr-port pair
* @cm_core: cm's core
* @dst_addr: listener ip addr
* @ipv4: flag indicating IPv4 when true
* @dst_port: listener tcp port num
* @vlan_id: virtual LAN ID
* @listener_state: state to match with listen node's
*/
static struct irdma_cm_listener *
irdma_find_listener(struct irdma_cm_core *cm_core, u32 *dst_addr, bool ipv4,
u16 dst_port, u16 vlan_id,
enum irdma_cm_listener_state listener_state)
{
struct irdma_cm_listener *listen_node;
static const u32 ip_zero[4] = { 0, 0, 0, 0 };
u32 listen_addr[4];
u16 listen_port;
unsigned long flags;
/* walk list and find cm_node associated with this session ID */
spin_lock_irqsave(&cm_core->listen_list_lock, flags);
list_for_each_entry (listen_node, &cm_core->listen_list, list) {
memcpy(listen_addr, listen_node->loc_addr, sizeof (listen_addr));
listen_port = listen_node->loc_port;
if (listen_node->ipv4 != ipv4 || listen_port != dst_port ||
!(listener_state & listen_node->listener_state))
continue ;
/* compare node pair, return node handle if a match */
if (!memcmp(listen_addr, ip_zero, sizeof (listen_addr)) ||
(!memcmp(listen_addr, dst_addr, sizeof (listen_addr)) &&
vlan_id == listen_node->vlan_id)) {
refcount_inc(&listen_node->refcnt);
spin_unlock_irqrestore(&cm_core->listen_list_lock,
flags);
trace_irdma_find_listener(listen_node);
return listen_node;
}
}
spin_unlock_irqrestore(&cm_core->listen_list_lock, flags);
return NULL;
}
/**
* irdma_del_multiple_qhash - Remove qhash and child listens
* @iwdev: iWarp device
* @cm_info: CM info for parent listen node
* @cm_parent_listen_node: The parent listen node
*/
static int irdma_del_multiple_qhash(struct irdma_device *iwdev,
struct irdma_cm_info *cm_info,
struct irdma_cm_listener *cm_parent_listen_node)
{
struct irdma_cm_listener *child_listen_node;
struct list_head *pos, *tpos;
unsigned long flags;
int ret = -EINVAL;
spin_lock_irqsave(&iwdev->cm_core.listen_list_lock, flags);
list_for_each_safe (pos, tpos,
&cm_parent_listen_node->child_listen_list) {
child_listen_node = list_entry(pos, struct irdma_cm_listener,
child_listen_list);
if (child_listen_node->ipv4)
ibdev_dbg(&iwdev->ibdev,
"CM: removing child listen for IP=%pI4, port=%d, vlan=%d\n" ,
child_listen_node->loc_addr,
child_listen_node->loc_port,
child_listen_node->vlan_id);
else
ibdev_dbg(&iwdev->ibdev,
"CM: removing child listen for IP=%pI6, port=%d, vlan=%d\n" ,
child_listen_node->loc_addr,
child_listen_node->loc_port,
child_listen_node->vlan_id);
trace_irdma_del_multiple_qhash(child_listen_node);
list_del(pos);
memcpy(cm_info->loc_addr, child_listen_node->loc_addr,
sizeof (cm_info->loc_addr));
cm_info->vlan_id = child_listen_node->vlan_id;
if (child_listen_node->qhash_set) {
ret = irdma_manage_qhash(iwdev, cm_info,
IRDMA_QHASH_TYPE_TCP_SYN,
IRDMA_QHASH_MANAGE_TYPE_DELETE,
NULL, false );
child_listen_node->qhash_set = false ;
} else {
ret = 0;
}
ibdev_dbg(&iwdev->ibdev,
"CM: Child listen node freed = %p\n" ,
child_listen_node);
kfree(child_listen_node);
cm_parent_listen_node->cm_core->stats_listen_nodes_destroyed++;
}
spin_unlock_irqrestore(&iwdev->cm_core.listen_list_lock, flags);
return ret;
}
static u8 irdma_iw_get_vlan_prio(u32 *loc_addr, u8 prio, bool ipv4)
{
struct net_device *ndev = NULL;
rcu_read_lock();
if (ipv4) {
ndev = ip_dev_find(&init_net, htonl(loc_addr[0]));
} else if (IS_ENABLED(CONFIG_IPV6)) {
struct net_device *ip_dev;
struct in6_addr laddr6;
irdma_copy_ip_htonl(laddr6.in6_u.u6_addr32, loc_addr);
for_each_netdev_rcu (&init_net, ip_dev) {
if (ipv6_chk_addr(&init_net, &laddr6, ip_dev, 1)) {
ndev = ip_dev;
break ;
}
}
}
if (!ndev)
goto done;
if (is_vlan_dev(ndev))
prio = (vlan_dev_get_egress_qos_mask(ndev, prio) & VLAN_PRIO_MASK)
>> VLAN_PRIO_SHIFT;
if (ipv4)
dev_put(ndev);
done:
rcu_read_unlock();
return prio;
}
/**
* irdma_get_vlan_mac_ipv6 - Gets the vlan and mac
* @addr: local IPv6 address
* @vlan_id: vlan id for the given IPv6 address
* @mac: mac address for the given IPv6 address
*
* Returns the vlan id and mac for an IPv6 address.
*/
void irdma_get_vlan_mac_ipv6(u32 *addr, u16 *vlan_id, u8 *mac)
{
struct net_device *ip_dev = NULL;
struct in6_addr laddr6;
if (!IS_ENABLED(CONFIG_IPV6))
return ;
irdma_copy_ip_htonl(laddr6.in6_u.u6_addr32, addr);
if (vlan_id)
*vlan_id = 0xFFFF; /* Match rdma_vlan_dev_vlan_id() */
if (mac)
eth_zero_addr(mac);
rcu_read_lock();
for_each_netdev_rcu (&init_net, ip_dev) {
if (ipv6_chk_addr(&init_net, &laddr6, ip_dev, 1)) {
if (vlan_id)
*vlan_id = rdma_vlan_dev_vlan_id(ip_dev);
if (ip_dev->dev_addr && mac)
ether_addr_copy(mac, ip_dev->dev_addr);
break ;
}
}
rcu_read_unlock();
}
/**
* irdma_get_vlan_ipv4 - Returns the vlan_id for IPv4 address
* @addr: local IPv4 address
*/
u16 irdma_get_vlan_ipv4(u32 *addr)
{
struct net_device *netdev;
u16 vlan_id = 0xFFFF;
netdev = ip_dev_find(&init_net, htonl(addr[0]));
if (netdev) {
vlan_id = rdma_vlan_dev_vlan_id(netdev);
dev_put(netdev);
}
return vlan_id;
}
/**
* irdma_add_mqh_6 - Adds multiple qhashes for IPv6
* @iwdev: iWarp device
* @cm_info: CM info for parent listen node
* @cm_parent_listen_node: The parent listen node
*
* Adds a qhash and a child listen node for every IPv6 address
* on the adapter and adds the associated qhash filter
*/
static int irdma_add_mqh_6(struct irdma_device *iwdev,
struct irdma_cm_info *cm_info,
struct irdma_cm_listener *cm_parent_listen_node)
{
struct net_device *ip_dev;
struct inet6_dev *idev;
struct inet6_ifaddr *ifp, *tmp;
struct irdma_cm_listener *child_listen_node;
unsigned long flags;
int ret = 0;
rtnl_lock();
for_each_netdev(&init_net, ip_dev) {
if (!(ip_dev->flags & IFF_UP))
continue ;
if (((rdma_vlan_dev_vlan_id(ip_dev) >= VLAN_N_VID) ||
(rdma_vlan_dev_real_dev(ip_dev) != iwdev->netdev)) &&
ip_dev != iwdev->netdev)
continue ;
idev = __in6_dev_get(ip_dev);
if (!idev) {
ibdev_dbg(&iwdev->ibdev, "CM: idev == NULL\n" );
break ;
}
list_for_each_entry_safe (ifp, tmp, &idev->addr_list, if_list) {
ibdev_dbg(&iwdev->ibdev, "CM: IP=%pI6, vlan_id=%d, MAC=%pM\n" ,
&ifp->addr, rdma_vlan_dev_vlan_id(ip_dev),
ip_dev->dev_addr);
child_listen_node = kzalloc(sizeof (*child_listen_node), GFP_KERNEL);
ibdev_dbg(&iwdev->ibdev, "CM: Allocating child listener %p\n" ,
child_listen_node);
if (!child_listen_node) {
ibdev_dbg(&iwdev->ibdev, "CM: listener memory allocation\n" );
ret = -ENOMEM;
goto exit ;
}
cm_info->vlan_id = rdma_vlan_dev_vlan_id(ip_dev);
cm_parent_listen_node->vlan_id = cm_info->vlan_id;
memcpy(child_listen_node, cm_parent_listen_node,
sizeof (*child_listen_node));
irdma_copy_ip_ntohl(child_listen_node->loc_addr,
ifp->addr.in6_u.u6_addr32);
memcpy(cm_info->loc_addr, child_listen_node->loc_addr,
sizeof (cm_info->loc_addr));
if (!iwdev->vsi.dscp_mode)
cm_info->user_pri =
irdma_iw_get_vlan_prio(child_listen_node->loc_addr,
cm_info->user_pri,
false );
ret = irdma_manage_qhash(iwdev, cm_info,
IRDMA_QHASH_TYPE_TCP_SYN,
IRDMA_QHASH_MANAGE_TYPE_ADD,
NULL, true );
if (ret) {
kfree(child_listen_node);
continue ;
}
trace_irdma_add_mqh_6(iwdev, child_listen_node,
ip_dev->dev_addr);
child_listen_node->qhash_set = true ;
spin_lock_irqsave(&iwdev->cm_core.listen_list_lock, flags);
list_add(&child_listen_node->child_listen_list,
&cm_parent_listen_node->child_listen_list);
spin_unlock_irqrestore(&iwdev->cm_core.listen_list_lock, flags);
cm_parent_listen_node->cm_core->stats_listen_nodes_created++;
}
}
exit :
rtnl_unlock();
return ret;
}
/**
* irdma_add_mqh_4 - Adds multiple qhashes for IPv4
* @iwdev: iWarp device
* @cm_info: CM info for parent listen node
* @cm_parent_listen_node: The parent listen node
*
* Adds a qhash and a child listen node for every IPv4 address
* on the adapter and adds the associated qhash filter
*/
static int irdma_add_mqh_4(struct irdma_device *iwdev,
struct irdma_cm_info *cm_info,
struct irdma_cm_listener *cm_parent_listen_node)
{
struct net_device *ip_dev;
struct in_device *idev;
struct irdma_cm_listener *child_listen_node;
unsigned long flags;
const struct in_ifaddr *ifa;
int ret = 0;
rtnl_lock();
for_each_netdev(&init_net, ip_dev) {
if (!(ip_dev->flags & IFF_UP))
continue ;
if (((rdma_vlan_dev_vlan_id(ip_dev) >= VLAN_N_VID) ||
(rdma_vlan_dev_real_dev(ip_dev) != iwdev->netdev)) &&
ip_dev != iwdev->netdev)
continue ;
idev = in_dev_get(ip_dev);
if (!idev)
continue ;
in_dev_for_each_ifa_rtnl(ifa, idev) {
ibdev_dbg(&iwdev->ibdev,
"CM: Allocating child CM Listener forIP=%pI4, vlan_id=%d, MAC=%pM\n" ,
&ifa->ifa_address, rdma_vlan_dev_vlan_id(ip_dev),
ip_dev->dev_addr);
child_listen_node = kzalloc(sizeof (*child_listen_node), GFP_KERNEL);
cm_parent_listen_node->cm_core->stats_listen_nodes_created++;
ibdev_dbg(&iwdev->ibdev, "CM: Allocating child listener %p\n" ,
child_listen_node);
if (!child_listen_node) {
ibdev_dbg(&iwdev->ibdev, "CM: listener memory allocation\n" );
in_dev_put(idev);
ret = -ENOMEM;
goto exit ;
}
cm_info->vlan_id = rdma_vlan_dev_vlan_id(ip_dev);
cm_parent_listen_node->vlan_id = cm_info->vlan_id;
memcpy(child_listen_node, cm_parent_listen_node,
sizeof (*child_listen_node));
child_listen_node->loc_addr[0] =
ntohl(ifa->ifa_address);
memcpy(cm_info->loc_addr, child_listen_node->loc_addr,
sizeof (cm_info->loc_addr));
if (!iwdev->vsi.dscp_mode)
cm_info->user_pri =
irdma_iw_get_vlan_prio(child_listen_node->loc_addr,
cm_info->user_pri,
true );
ret = irdma_manage_qhash(iwdev, cm_info,
IRDMA_QHASH_TYPE_TCP_SYN,
IRDMA_QHASH_MANAGE_TYPE_ADD,
NULL, true );
if (ret) {
kfree(child_listen_node);
cm_parent_listen_node->cm_core
->stats_listen_nodes_created--;
continue ;
}
trace_irdma_add_mqh_4(iwdev, child_listen_node,
ip_dev->dev_addr);
child_listen_node->qhash_set = true ;
spin_lock_irqsave(&iwdev->cm_core.listen_list_lock,
flags);
list_add(&child_listen_node->child_listen_list,
&cm_parent_listen_node->child_listen_list);
spin_unlock_irqrestore(&iwdev->cm_core.listen_list_lock, flags);
}
in_dev_put(idev);
}
exit :
rtnl_unlock();
return ret;
}
/**
* irdma_add_mqh - Adds multiple qhashes
* @iwdev: iWarp device
* @cm_info: CM info for parent listen node
* @cm_listen_node: The parent listen node
*/
static int irdma_add_mqh(struct irdma_device *iwdev,
struct irdma_cm_info *cm_info,
struct irdma_cm_listener *cm_listen_node)
{
if (cm_info->ipv4)
return irdma_add_mqh_4(iwdev, cm_info, cm_listen_node);
else
return irdma_add_mqh_6(iwdev, cm_info, cm_listen_node);
}
/**
* irdma_reset_list_prep - add connection nodes slated for reset to list
* @cm_core: cm's core
* @listener: pointer to listener node
* @reset_list: a list to which cm_node will be selected
*/
static void irdma_reset_list_prep(struct irdma_cm_core *cm_core,
struct irdma_cm_listener *listener,
struct list_head *reset_list)
{
struct irdma_cm_node *cm_node;
int bkt;
hash_for_each_rcu(cm_core->cm_hash_tbl, bkt, cm_node, list) {
if (cm_node->listener == listener &&
!cm_node->accelerated &&
refcount_inc_not_zero(&cm_node->refcnt))
list_add(&cm_node->reset_entry, reset_list);
}
}
/**
* irdma_dec_refcnt_listen - delete listener and associated cm nodes
* @cm_core: cm's core
* @listener: pointer to listener node
* @free_hanging_nodes: to free associated cm_nodes
* @apbvt_del: flag to delete the apbvt
*/
static int irdma_dec_refcnt_listen(struct irdma_cm_core *cm_core,
struct irdma_cm_listener *listener,
int free_hanging_nodes, bool apbvt_del)
{
int err;
struct list_head *list_pos;
struct list_head *list_temp;
struct irdma_cm_node *cm_node;
struct list_head reset_list;
struct irdma_cm_info nfo;
enum irdma_cm_node_state old_state;
unsigned long flags;
trace_irdma_dec_refcnt_listen(listener, __builtin_return_address(0));
/* free non-accelerated child nodes for this listener */
INIT_LIST_HEAD(&reset_list);
if (free_hanging_nodes) {
rcu_read_lock();
irdma_reset_list_prep(cm_core, listener, &reset_list);
rcu_read_unlock();
}
list_for_each_safe (list_pos, list_temp, &reset_list) {
cm_node = container_of(list_pos, struct irdma_cm_node,
reset_entry);
if (cm_node->state >= IRDMA_CM_STATE_FIN_WAIT1) {
irdma_rem_ref_cm_node(cm_node);
continue ;
}
irdma_cleanup_retrans_entry(cm_node);
err = irdma_send_reset(cm_node);
if (err) {
cm_node->state = IRDMA_CM_STATE_CLOSED;
ibdev_dbg(&cm_node->iwdev->ibdev,
"CM: send reset failed\n" );
} else {
old_state = cm_node->state;
cm_node->state = IRDMA_CM_STATE_LISTENER_DESTROYED;
if (old_state != IRDMA_CM_STATE_MPAREQ_RCVD)
irdma_rem_ref_cm_node(cm_node);
}
}
if (refcount_dec_and_test(&listener->refcnt)) {
spin_lock_irqsave(&cm_core->listen_list_lock, flags);
list_del(&listener->list);
spin_unlock_irqrestore(&cm_core->listen_list_lock, flags);
if (apbvt_del)
irdma_del_apbvt(listener->iwdev,
listener->apbvt_entry);
memcpy(nfo.loc_addr, listener->loc_addr, sizeof (nfo.loc_addr));
nfo.loc_port = listener->loc_port;
nfo.ipv4 = listener->ipv4;
nfo.vlan_id = listener->vlan_id;
nfo.user_pri = listener->user_pri;
nfo.qh_qpid = listener->iwdev->vsi.ilq->qp_id;
if (!list_empty(&listener->child_listen_list)) {
irdma_del_multiple_qhash(listener->iwdev, &nfo,
listener);
} else {
if (listener->qhash_set)
irdma_manage_qhash(listener->iwdev,
&nfo,
IRDMA_QHASH_TYPE_TCP_SYN,
IRDMA_QHASH_MANAGE_TYPE_DELETE,
NULL, false );
}
cm_core->stats_listen_destroyed++;
cm_core->stats_listen_nodes_destroyed++;
ibdev_dbg(&listener->iwdev->ibdev,
"CM: loc_port=0x%04x loc_addr=%pI4 cm_listen_node=%p cm_id=%p qhash_set=%d vlan_id=%d apbvt_del=%d\n" ,
listener->loc_port, listener->loc_addr, listener,
listener->cm_id, listener->qhash_set,
listener->vlan_id, apbvt_del);
kfree(listener);
listener = NULL;
return 0;
}
return -EINVAL;
}
/**
* irdma_cm_del_listen - delete a listener
* @cm_core: cm's core
* @listener: passive connection's listener
* @apbvt_del: flag to delete apbvt
*/
static int irdma_cm_del_listen(struct irdma_cm_core *cm_core,
struct irdma_cm_listener *listener,
bool apbvt_del)
{
listener->listener_state = IRDMA_CM_LISTENER_PASSIVE_STATE;
listener->cm_id = NULL;
return irdma_dec_refcnt_listen(cm_core, listener, 1, apbvt_del);
}
/**
* irdma_addr_resolve_neigh - resolve neighbor address
* @iwdev: iwarp device structure
* @src_ip: local ip address
* @dst_ip: remote ip address
* @arpindex: if there is an arp entry
*/
static int irdma_addr_resolve_neigh(struct irdma_device *iwdev, u32 src_ip,
u32 dst_ip, int arpindex)
{
struct rtable *rt;
struct neighbour *neigh;
int rc = arpindex;
__be32 dst_ipaddr = htonl(dst_ip);
__be32 src_ipaddr = htonl(src_ip);
rt = ip_route_output(&init_net, dst_ipaddr, src_ipaddr, 0, 0,
RT_SCOPE_UNIVERSE);
if (IS_ERR(rt)) {
ibdev_dbg(&iwdev->ibdev, "CM: ip_route_output fail\n" );
return -EINVAL;
}
neigh = dst_neigh_lookup(&rt->dst, &dst_ipaddr);
if (!neigh)
goto exit ;
if (neigh->nud_state & NUD_VALID)
rc = irdma_add_arp(iwdev->rf, &dst_ip, true , neigh->ha);
else
neigh_event_send(neigh, NULL);
if (neigh)
neigh_release(neigh);
exit :
ip_rt_put(rt);
return rc;
}
/**
* irdma_get_dst_ipv6 - get destination cache entry via ipv6 lookup
* @src_addr: local ipv6 sock address
* @dst_addr: destination ipv6 sock address
*/
static struct dst_entry *irdma_get_dst_ipv6(struct sockaddr_in6 *src_addr,
struct sockaddr_in6 *dst_addr)
{
struct dst_entry *dst = NULL;
if ((IS_ENABLED(CONFIG_IPV6))) {
struct flowi6 fl6 = {};
fl6.daddr = dst_addr->sin6_addr;
fl6.saddr = src_addr->sin6_addr;
if (ipv6_addr_type(&fl6.daddr) & IPV6_ADDR_LINKLOCAL)
fl6.flowi6_oif = dst_addr->sin6_scope_id;
dst = ip6_route_output(&init_net, NULL, &fl6);
}
return dst;
}
/**
* irdma_addr_resolve_neigh_ipv6 - resolve neighbor ipv6 address
* @iwdev: iwarp device structure
* @src: local ip address
* @dest: remote ip address
* @arpindex: if there is an arp entry
*/
static int irdma_addr_resolve_neigh_ipv6(struct irdma_device *iwdev, u32 *src,
u32 *dest, int arpindex)
{
struct neighbour *neigh;
int rc = arpindex;
struct dst_entry *dst;
struct sockaddr_in6 dst_addr = {};
struct sockaddr_in6 src_addr = {};
dst_addr.sin6_family = AF_INET6;
irdma_copy_ip_htonl(dst_addr.sin6_addr.in6_u.u6_addr32, dest);
src_addr.sin6_family = AF_INET6;
irdma_copy_ip_htonl(src_addr.sin6_addr.in6_u.u6_addr32, src);
dst = irdma_get_dst_ipv6(&src_addr, &dst_addr);
if (!dst || dst->error) {
if (dst) {
dst_release(dst);
ibdev_dbg(&iwdev->ibdev,
"CM: ip6_route_output returned dst->error = %d\n" ,
dst->error);
}
return -EINVAL;
}
neigh = dst_neigh_lookup(dst, dst_addr.sin6_addr.in6_u.u6_addr32);
if (!neigh)
goto exit ;
ibdev_dbg(&iwdev->ibdev, "CM: dst_neigh_lookup MAC=%pM\n" ,
neigh->ha);
trace_irdma_addr_resolve(iwdev, neigh->ha);
if (neigh->nud_state & NUD_VALID)
rc = irdma_add_arp(iwdev->rf, dest, false , neigh->ha);
else
neigh_event_send(neigh, NULL);
if (neigh)
neigh_release(neigh);
exit :
dst_release(dst);
return rc;
}
/**
* irdma_find_node - find a cm node that matches the reference cm node
* @cm_core: cm's core
* @rem_port: remote tcp port num
* @rem_addr: remote ip addr
* @loc_port: local tcp port num
* @loc_addr: local ip addr
* @vlan_id: local VLAN ID
*/
struct irdma_cm_node *irdma_find_node(struct irdma_cm_core *cm_core,
u16 rem_port, u32 *rem_addr, u16 loc_port,
u32 *loc_addr, u16 vlan_id)
{
struct irdma_cm_node *cm_node;
u32 key = (rem_port << 16) | loc_port;
rcu_read_lock();
hash_for_each_possible_rcu(cm_core->cm_hash_tbl, cm_node, list, key) {
if (cm_node->vlan_id == vlan_id &&
cm_node->loc_port == loc_port && cm_node->rem_port == rem_port &&
!memcmp(cm_node->loc_addr, loc_addr, sizeof (cm_node->loc_addr)) &&
!memcmp(cm_node->rem_addr, rem_addr, sizeof (cm_node->rem_addr))) {
if (!refcount_inc_not_zero(&cm_node->refcnt))
goto exit ;
rcu_read_unlock();
trace_irdma_find_node(cm_node, 0, NULL);
return cm_node;
}
}
exit :
rcu_read_unlock();
/* no owner node */
return NULL;
}
/**
* irdma_add_hte_node - add a cm node to the hash table
* @cm_core: cm's core
* @cm_node: connection's node
*/
static void irdma_add_hte_node(struct irdma_cm_core *cm_core,
struct irdma_cm_node *cm_node)
{
unsigned long flags;
u32 key = (cm_node->rem_port << 16) | cm_node->loc_port;
spin_lock_irqsave(&cm_core->ht_lock, flags);
hash_add_rcu(cm_core->cm_hash_tbl, &cm_node->list, key);
spin_unlock_irqrestore(&cm_core->ht_lock, flags);
}
/**
* irdma_ipv4_is_lpb - check if loopback
* @loc_addr: local addr to compare
* @rem_addr: remote address
*/
bool irdma_ipv4_is_lpb(u32 loc_addr, u32 rem_addr)
{
return ipv4_is_loopback(htonl(rem_addr)) || (loc_addr == rem_addr);
}
/**
* irdma_ipv6_is_lpb - check if loopback
* @loc_addr: local addr to compare
* @rem_addr: remote address
*/
bool irdma_ipv6_is_lpb(u32 *loc_addr, u32 *rem_addr)
{
struct in6_addr raddr6;
irdma_copy_ip_htonl(raddr6.in6_u.u6_addr32, rem_addr);
return !memcmp(loc_addr, rem_addr, 16) || ipv6_addr_loopback(&raddr6);
}
/**
* irdma_cm_create_ah - create a cm address handle
* @cm_node: The connection manager node to create AH for
* @wait: Provides option to wait for ah creation or not
*/
static int irdma_cm_create_ah(struct irdma_cm_node *cm_node, bool wait)
{
struct irdma_ah_info ah_info = {};
struct irdma_device *iwdev = cm_node->iwdev;
ether_addr_copy(ah_info.mac_addr, iwdev->netdev->dev_addr);
ah_info.hop_ttl = 0x40;
ah_info.tc_tos = cm_node->tos;
ah_info.vsi = &iwdev->vsi;
if (cm_node->ipv4) {
ah_info.ipv4_valid = true ;
ah_info.dest_ip_addr[0] = cm_node->rem_addr[0];
ah_info.src_ip_addr[0] = cm_node->loc_addr[0];
ah_info.do_lpbk = irdma_ipv4_is_lpb(ah_info.src_ip_addr[0],
ah_info.dest_ip_addr[0]);
} else {
memcpy(ah_info.dest_ip_addr, cm_node->rem_addr,
sizeof (ah_info.dest_ip_addr));
memcpy(ah_info.src_ip_addr, cm_node->loc_addr,
sizeof (ah_info.src_ip_addr));
ah_info.do_lpbk = irdma_ipv6_is_lpb(ah_info.src_ip_addr,
ah_info.dest_ip_addr);
}
ah_info.vlan_tag = cm_node->vlan_id;
if (cm_node->vlan_id < VLAN_N_VID) {
ah_info.insert_vlan_tag = 1;
ah_info.vlan_tag |= cm_node->user_pri << VLAN_PRIO_SHIFT;
}
ah_info.dst_arpindex =
irdma_arp_table(iwdev->rf, ah_info.dest_ip_addr,
ah_info.ipv4_valid, NULL, IRDMA_ARP_RESOLVE);
if (irdma_puda_create_ah(&iwdev->rf->sc_dev, &ah_info, wait,
IRDMA_PUDA_RSRC_TYPE_ILQ, cm_node,
&cm_node->ah))
return -ENOMEM;
trace_irdma_create_ah(cm_node);
return 0;
}
/**
* irdma_cm_free_ah - free a cm address handle
* @cm_node: The connection manager node to create AH for
*/
static void irdma_cm_free_ah(struct irdma_cm_node *cm_node)
{
struct irdma_device *iwdev = cm_node->iwdev;
trace_irdma_cm_free_ah(cm_node);
irdma_puda_free_ah(&iwdev->rf->sc_dev, cm_node->ah);
cm_node->ah = NULL;
}
/**
* irdma_make_cm_node - create a new instance of a cm node
* @cm_core: cm's core
* @iwdev: iwarp device structure
* @cm_info: quad info for connection
* @listener: passive connection's listener
*/
static struct irdma_cm_node *
irdma_make_cm_node(struct irdma_cm_core *cm_core, struct irdma_device *iwdev,
struct irdma_cm_info *cm_info,
struct irdma_cm_listener *listener)
{
struct irdma_cm_node *cm_node;
int oldarpindex;
int arpindex;
struct net_device *netdev = iwdev->netdev;
/* create an hte and cm_node for this instance */
cm_node = kzalloc(sizeof (*cm_node), GFP_ATOMIC);
if (!cm_node)
return NULL;
/* set our node specific transport info */
cm_node->ipv4 = cm_info->ipv4;
cm_node->vlan_id = cm_info->vlan_id;
if (cm_node->vlan_id >= VLAN_N_VID && iwdev->dcb_vlan_mode)
cm_node->vlan_id = 0;
cm_node->tos = cm_info->tos;
cm_node->user_pri = cm_info->user_pri;
if (listener) {
if (listener->tos != cm_info->tos)
ibdev_warn(&iwdev->ibdev,
"application TOS[%d] and remote client TOS[%d] mismatch\n" ,
listener->tos, cm_info->tos);
if (iwdev->vsi.dscp_mode) {
cm_node->user_pri = listener->user_pri;
} else {
cm_node->tos = max(listener->tos, cm_info->tos);
cm_node->user_pri = rt_tos2priority(cm_node->tos);
cm_node->user_pri =
irdma_iw_get_vlan_prio(cm_info->loc_addr,
cm_node->user_pri,
cm_info->ipv4);
}
ibdev_dbg(&iwdev->ibdev,
"DCB: listener: TOS:[%d] UP:[%d]\n" , cm_node->tos,
cm_node->user_pri);
trace_irdma_listener_tos(iwdev, cm_node->tos,
cm_node->user_pri);
}
memcpy(cm_node->loc_addr, cm_info->loc_addr, sizeof (cm_node->loc_addr));
memcpy(cm_node->rem_addr, cm_info->rem_addr, sizeof (cm_node->rem_addr));
cm_node->loc_port = cm_info->loc_port;
cm_node->rem_port = cm_info->rem_port;
cm_node->mpa_frame_rev = IRDMA_CM_DEFAULT_MPA_VER;
cm_node->send_rdma0_op = SEND_RDMA_READ_ZERO;
cm_node->iwdev = iwdev;
cm_node->dev = &iwdev->rf->sc_dev;
cm_node->ird_size = cm_node->dev->hw_attrs.max_hw_ird;
cm_node->ord_size = cm_node->dev->hw_attrs.max_hw_ord;
cm_node->listener = listener;
cm_node->cm_id = cm_info->cm_id;
ether_addr_copy(cm_node->loc_mac, netdev->dev_addr);
spin_lock_init(&cm_node->retrans_list_lock);
cm_node->ack_rcvd = false ;
init_completion(&cm_node->establish_comp);
refcount_set(&cm_node->refcnt, 1);
/* associate our parent CM core */
cm_node->cm_core = cm_core;
cm_node->tcp_cntxt.loc_id = IRDMA_CM_DEFAULT_LOCAL_ID;
cm_node->tcp_cntxt.rcv_wscale = iwdev->rcv_wscale;
cm_node->tcp_cntxt.rcv_wnd = iwdev->rcv_wnd >> cm_node->tcp_cntxt.rcv_wscale;
if (cm_node->ipv4) {
cm_node->tcp_cntxt.loc_seq_num = secure_tcp_seq(htonl(cm_node->loc_addr[0]),
htonl(cm_node->rem_addr[0]),
htons(cm_node->loc_port),
htons(cm_node->rem_port));
cm_node->tcp_cntxt.mss = iwdev->vsi.mtu - IRDMA_MTU_TO_MSS_IPV4;
} else if (IS_ENABLED(CONFIG_IPV6)) {
__be32 loc[4] = {
htonl(cm_node->loc_addr[0]), htonl(cm_node->loc_addr[1]),
htonl(cm_node->loc_addr[2]), htonl(cm_node->loc_addr[3])
};
__be32 rem[4] = {
htonl(cm_node->rem_addr[0]), htonl(cm_node->rem_addr[1]),
htonl(cm_node->rem_addr[2]), htonl(cm_node->rem_addr[3])
};
cm_node->tcp_cntxt.loc_seq_num = secure_tcpv6_seq(loc, rem,
htons(cm_node->loc_port),
htons(cm_node->rem_port));
cm_node->tcp_cntxt.mss = iwdev->vsi.mtu - IRDMA_MTU_TO_MSS_IPV6;
}
if ((cm_node->ipv4 &&
irdma_ipv4_is_lpb(cm_node->loc_addr[0], cm_node->rem_addr[0])) ||
(!cm_node->ipv4 &&
irdma_ipv6_is_lpb(cm_node->loc_addr, cm_node->rem_addr))) {
cm_node->do_lpb = true ;
arpindex = irdma_arp_table(iwdev->rf, cm_node->rem_addr,
cm_node->ipv4, NULL,
IRDMA_ARP_RESOLVE);
} else {
oldarpindex = irdma_arp_table(iwdev->rf, cm_node->rem_addr,
cm_node->ipv4, NULL,
IRDMA_ARP_RESOLVE);
if (cm_node->ipv4)
arpindex = irdma_addr_resolve_neigh(iwdev,
cm_info->loc_addr[0],
cm_info->rem_addr[0],
oldarpindex);
else if (IS_ENABLED(CONFIG_IPV6))
arpindex = irdma_addr_resolve_neigh_ipv6(iwdev,
cm_info->loc_addr,
cm_info->rem_addr,
oldarpindex);
else
arpindex = -EINVAL;
}
if (arpindex < 0)
goto err;
ether_addr_copy(cm_node->rem_mac,
iwdev->rf->arp_table[arpindex].mac_addr);
irdma_add_hte_node(cm_core, cm_node);
cm_core->stats_nodes_created++;
return cm_node;
err:
kfree(cm_node);
return NULL;
}
static void irdma_destroy_connection(struct irdma_cm_node *cm_node)
{
struct irdma_cm_core *cm_core = cm_node->cm_core;
struct irdma_qp *iwqp;
struct irdma_cm_info nfo;
/* if the node is destroyed before connection was accelerated */
if (!cm_node->accelerated && cm_node->accept_pend) {
--> --------------------
--> maximum size reached
--> --------------------
Messung V0.5 C=97 H=98 G=97
¤ Dauer der Verarbeitung: 0.14 Sekunden
(vorverarbeitet)
¤
*© Formatika GbR, Deutschland