Lines Matching refs:skb_out
1194 struct sk_buff *skb_out; in cdc_ncm_fill_tx_frame() local
1220 skb_out = ctx->tx_curr_skb; in cdc_ncm_fill_tx_frame()
1223 if (!skb_out) { in cdc_ncm_fill_tx_frame()
1226 skb_out = alloc_skb(ctx->tx_curr_size, GFP_ATOMIC); in cdc_ncm_fill_tx_frame()
1232 if (skb_out == NULL) { in cdc_ncm_fill_tx_frame()
1238 if (skb_out == NULL) { in cdc_ncm_fill_tx_frame()
1248 skb_out = alloc_skb(ctx->tx_curr_size, GFP_ATOMIC); in cdc_ncm_fill_tx_frame()
1251 if (skb_out == NULL) { in cdc_ncm_fill_tx_frame()
1262 nth.nth16 = skb_put_zero(skb_out, sizeof(struct usb_cdc_ncm_nth16)); in cdc_ncm_fill_tx_frame()
1268 nth.nth32 = skb_put_zero(skb_out, sizeof(struct usb_cdc_ncm_nth32)); in cdc_ncm_fill_tx_frame()
1295 ndp.ndp16 = cdc_ncm_ndp16(ctx, skb_out, sign, skb->len + ctx->tx_modulus + ctx->tx_remainder); in cdc_ncm_fill_tx_frame()
1297 ndp.ndp32 = cdc_ncm_ndp32(ctx, skb_out, sign, skb->len + ctx->tx_modulus + ctx->tx_remainder); in cdc_ncm_fill_tx_frame()
1300 cdc_ncm_align_tail(skb_out, ctx->tx_modulus, ctx->tx_remainder, ctx->tx_curr_size); in cdc_ncm_fill_tx_frame()
1304 skb_out->len + skb->len + delayed_ndp_size > ctx->tx_curr_size) { in cdc_ncm_fill_tx_frame()
1332 ndp.ndp16->dpe16[index].wDatagramIndex = cpu_to_le16(skb_out->len); in cdc_ncm_fill_tx_frame()
1339 ndp.ndp32->dpe32[index].dwDatagramIndex = cpu_to_le32(skb_out->len); in cdc_ncm_fill_tx_frame()
1342 skb_put_data(skb_out, skb->data, skb->len); in cdc_ncm_fill_tx_frame()
1367 ctx->tx_curr_skb = skb_out; in cdc_ncm_fill_tx_frame()
1373 ctx->tx_curr_skb = skb_out; in cdc_ncm_fill_tx_frame()
1389 nth.nth16 = (struct usb_cdc_ncm_nth16 *)skb_out->data; in cdc_ncm_fill_tx_frame()
1390 cdc_ncm_align_tail(skb_out, ctx->tx_ndp_modulus, 0, ctx->tx_curr_size - ctx->max_ndp_size); in cdc_ncm_fill_tx_frame()
1391 nth.nth16->wNdpIndex = cpu_to_le16(skb_out->len); in cdc_ncm_fill_tx_frame()
1392 skb_put_data(skb_out, ctx->delayed_ndp16, ctx->max_ndp_size); in cdc_ncm_fill_tx_frame()
1397 nth.nth32 = (struct usb_cdc_ncm_nth32 *)skb_out->data; in cdc_ncm_fill_tx_frame()
1398 cdc_ncm_align_tail(skb_out, ctx->tx_ndp_modulus, 0, ctx->tx_curr_size - ctx->max_ndp_size); in cdc_ncm_fill_tx_frame()
1399 nth.nth32->dwNdpIndex = cpu_to_le32(skb_out->len); in cdc_ncm_fill_tx_frame()
1400 skb_put_data(skb_out, ctx->delayed_ndp32, ctx->max_ndp_size); in cdc_ncm_fill_tx_frame()
1416 skb_out->len > ctx->min_tx_pkt) { in cdc_ncm_fill_tx_frame()
1417 padding_count = ctx->tx_curr_size - skb_out->len; in cdc_ncm_fill_tx_frame()
1419 skb_put_zero(skb_out, padding_count); in cdc_ncm_fill_tx_frame()
1420 } else if (skb_out->len < ctx->tx_curr_size && in cdc_ncm_fill_tx_frame()
1421 (skb_out->len % dev->maxpacket) == 0) { in cdc_ncm_fill_tx_frame()
1422 skb_put_u8(skb_out, 0); /* force short packet */ in cdc_ncm_fill_tx_frame()
1427 nth.nth16 = (struct usb_cdc_ncm_nth16 *)skb_out->data; in cdc_ncm_fill_tx_frame()
1428 nth.nth16->wBlockLength = cpu_to_le16(skb_out->len); in cdc_ncm_fill_tx_frame()
1430 nth.nth32 = (struct usb_cdc_ncm_nth32 *)skb_out->data; in cdc_ncm_fill_tx_frame()
1431 nth.nth32->dwBlockLength = cpu_to_le32(skb_out->len); in cdc_ncm_fill_tx_frame()
1438 ctx->tx_overhead += skb_out->len - ctx->tx_curr_frame_payload; in cdc_ncm_fill_tx_frame()
1445 usbnet_set_skb_tx_stats(skb_out, n, in cdc_ncm_fill_tx_frame()
1446 (long)ctx->tx_curr_frame_payload - skb_out->len); in cdc_ncm_fill_tx_frame()
1448 return skb_out; in cdc_ncm_fill_tx_frame()
1501 struct sk_buff *skb_out; in cdc_ncm_tx_fixup() local
1517 skb_out = cdc_ncm_fill_tx_frame(dev, skb, cpu_to_le32(USB_CDC_NCM_NDP16_NOCRC_SIGN)); in cdc_ncm_tx_fixup()
1519 skb_out = cdc_ncm_fill_tx_frame(dev, skb, cpu_to_le32(USB_CDC_NCM_NDP32_NOCRC_SIGN)); in cdc_ncm_tx_fixup()
1522 return skb_out; in cdc_ncm_tx_fixup()