tcp_output.c revision 12016:0248e987199b
1/*
2 * CDDL HEADER START
3 *
4 * The contents of this file are subject to the terms of the
5 * Common Development and Distribution License (the "License").
6 * You may not use this file except in compliance with the License.
7 *
8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9 * or http://www.opensolaris.org/os/licensing.
10 * See the License for the specific language governing permissions
11 * and limitations under the License.
12 *
13 * When distributing Covered Code, include this CDDL HEADER in each
14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15 * If applicable, add the following below this CDDL HEADER, with the
16 * fields enclosed by brackets "[]" replaced with your own identifying
17 * information: Portions Copyright [yyyy] [name of copyright owner]
18 *
19 * CDDL HEADER END
20 */
21
22/*
23 * Copyright 2010 Sun Microsystems, Inc.  All rights reserved.
24 * Use is subject to license terms.
25 */
26
27/* This file contains all TCP output processing functions. */
28
29#include <sys/types.h>
30#include <sys/stream.h>
31#include <sys/strsun.h>
32#include <sys/strsubr.h>
33#include <sys/stropts.h>
34#include <sys/strlog.h>
35#define	_SUN_TPI_VERSION 2
36#include <sys/tihdr.h>
37#include <sys/suntpi.h>
38#include <sys/xti_inet.h>
39#include <sys/timod.h>
40#include <sys/pattr.h>
41#include <sys/squeue_impl.h>
42#include <sys/squeue.h>
43#include <sys/sockio.h>
44#include <sys/tsol/tnet.h>
45
46#include <inet/common.h>
47#include <inet/ip.h>
48#include <inet/tcp.h>
49#include <inet/tcp_impl.h>
50#include <inet/snmpcom.h>
51#include <inet/proto_set.h>
52#include <inet/ipsec_impl.h>
53#include <inet/ip_ndp.h>
54
55static mblk_t	*tcp_get_seg_mp(tcp_t *, uint32_t, int32_t *);
56static void	tcp_wput_cmdblk(queue_t *, mblk_t *);
57static void	tcp_wput_flush(tcp_t *, mblk_t *);
58static void	tcp_wput_iocdata(tcp_t *tcp, mblk_t *mp);
59static int	tcp_xmit_end(tcp_t *);
60static int	tcp_send(tcp_t *, const int, const int, const int,
61		    const int, int *, uint_t *, int *, mblk_t **, mblk_t *);
62static void	tcp_xmit_early_reset(char *, mblk_t *, uint32_t, uint32_t,
63		    int, ip_recv_attr_t *, ip_stack_t *, conn_t *);
64static boolean_t	tcp_send_rst_chk(tcp_stack_t *);
65static void	tcp_process_shrunk_swnd(tcp_t *, uint32_t);
66static void	tcp_fill_header(tcp_t *, uchar_t *, clock_t, int);
67
68/*
69 * Functions called directly via squeue having a prototype of edesc_t.
70 */
71static void	tcp_wput_nondata(void *, mblk_t *, void *, ip_recv_attr_t *);
72static void	tcp_wput_ioctl(void *, mblk_t *, void *, ip_recv_attr_t *);
73static void	tcp_wput_proto(void *, mblk_t *, void *, ip_recv_attr_t *);
74
75/*
76 * This controls how tiny a write must be before we try to copy it
77 * into the mblk on the tail of the transmit queue.  Not much
78 * speedup is observed for values larger than sixteen.  Zero will
79 * disable the optimisation.
80 */
81static int tcp_tx_pull_len = 16;
82
83void
84tcp_wput(queue_t *q, mblk_t *mp)
85{
86	conn_t	*connp = Q_TO_CONN(q);
87	tcp_t	*tcp;
88	void (*output_proc)();
89	t_scalar_t type;
90	uchar_t *rptr;
91	struct iocblk	*iocp;
92	size_t size;
93
94	ASSERT(connp->conn_ref >= 2);
95
96	switch (DB_TYPE(mp)) {
97	case M_DATA:
98		tcp = connp->conn_tcp;
99		ASSERT(tcp != NULL);
100
101		size = msgdsize(mp);
102
103		mutex_enter(&tcp->tcp_non_sq_lock);
104		tcp->tcp_squeue_bytes += size;
105		if (TCP_UNSENT_BYTES(tcp) > connp->conn_sndbuf) {
106			tcp_setqfull(tcp);
107		}
108		mutex_exit(&tcp->tcp_non_sq_lock);
109
110		CONN_INC_REF(connp);
111		SQUEUE_ENTER_ONE(connp->conn_sqp, mp, tcp_output, connp,
112		    NULL, tcp_squeue_flag, SQTAG_TCP_OUTPUT);
113		return;
114
115	case M_CMD:
116		tcp_wput_cmdblk(q, mp);
117		return;
118
119	case M_PROTO:
120	case M_PCPROTO:
121		/*
122		 * if it is a snmp message, don't get behind the squeue
123		 */
124		tcp = connp->conn_tcp;
125		rptr = mp->b_rptr;
126		if ((mp->b_wptr - rptr) >= sizeof (t_scalar_t)) {
127			type = ((union T_primitives *)rptr)->type;
128		} else {
129			if (connp->conn_debug) {
130				(void) strlog(TCP_MOD_ID, 0, 1,
131				    SL_ERROR|SL_TRACE,
132				    "tcp_wput_proto, dropping one...");
133			}
134			freemsg(mp);
135			return;
136		}
137		if (type == T_SVR4_OPTMGMT_REQ) {
138			/*
139			 * All Solaris components should pass a db_credp
140			 * for this TPI message, hence we ASSERT.
141			 * But in case there is some other M_PROTO that looks
142			 * like a TPI message sent by some other kernel
143			 * component, we check and return an error.
144			 */
145			cred_t	*cr = msg_getcred(mp, NULL);
146
147			ASSERT(cr != NULL);
148			if (cr == NULL) {
149				tcp_err_ack(tcp, mp, TSYSERR, EINVAL);
150				return;
151			}
152			if (snmpcom_req(q, mp, tcp_snmp_set, ip_snmp_get,
153			    cr)) {
154				/*
155				 * This was a SNMP request
156				 */
157				return;
158			} else {
159				output_proc = tcp_wput_proto;
160			}
161		} else {
162			output_proc = tcp_wput_proto;
163		}
164		break;
165	case M_IOCTL:
166		/*
167		 * Most ioctls can be processed right away without going via
168		 * squeues - process them right here. Those that do require
169		 * squeue (currently _SIOCSOCKFALLBACK)
170		 * are processed by tcp_wput_ioctl().
171		 */
172		iocp = (struct iocblk *)mp->b_rptr;
173		tcp = connp->conn_tcp;
174
175		switch (iocp->ioc_cmd) {
176		case TCP_IOC_ABORT_CONN:
177			tcp_ioctl_abort_conn(q, mp);
178			return;
179		case TI_GETPEERNAME:
180		case TI_GETMYNAME:
181			mi_copyin(q, mp, NULL,
182			    SIZEOF_STRUCT(strbuf, iocp->ioc_flag));
183			return;
184
185		default:
186			output_proc = tcp_wput_ioctl;
187			break;
188		}
189		break;
190	default:
191		output_proc = tcp_wput_nondata;
192		break;
193	}
194
195	CONN_INC_REF(connp);
196	SQUEUE_ENTER_ONE(connp->conn_sqp, mp, output_proc, connp,
197	    NULL, tcp_squeue_flag, SQTAG_TCP_WPUT_OTHER);
198}
199
200/*
201 * The TCP normal data output path.
202 * NOTE: the logic of the fast path is duplicated from this function.
203 */
204void
205tcp_wput_data(tcp_t *tcp, mblk_t *mp, boolean_t urgent)
206{
207	int		len;
208	mblk_t		*local_time;
209	mblk_t		*mp1;
210	uint32_t	snxt;
211	int		tail_unsent;
212	int		tcpstate;
213	int		usable = 0;
214	mblk_t		*xmit_tail;
215	int32_t		mss;
216	int32_t		num_sack_blk = 0;
217	int32_t		total_hdr_len;
218	int32_t		tcp_hdr_len;
219	int		rc;
220	tcp_stack_t	*tcps = tcp->tcp_tcps;
221	conn_t		*connp = tcp->tcp_connp;
222	clock_t		now = LBOLT_FASTPATH;
223
224	tcpstate = tcp->tcp_state;
225	if (mp == NULL) {
226		/*
227		 * tcp_wput_data() with NULL mp should only be called when
228		 * there is unsent data.
229		 */
230		ASSERT(tcp->tcp_unsent > 0);
231		/* Really tacky... but we need this for detached closes. */
232		len = tcp->tcp_unsent;
233		goto data_null;
234	}
235
236	ASSERT(mp->b_datap->db_type == M_DATA);
237	/*
238	 * Don't allow data after T_ORDREL_REQ or T_DISCON_REQ,
239	 * or before a connection attempt has begun.
240	 */
241	if (tcpstate < TCPS_SYN_SENT || tcpstate > TCPS_CLOSE_WAIT ||
242	    (tcp->tcp_valid_bits & TCP_FSS_VALID) != 0) {
243		if ((tcp->tcp_valid_bits & TCP_FSS_VALID) != 0) {
244#ifdef DEBUG
245			cmn_err(CE_WARN,
246			    "tcp_wput_data: data after ordrel, %s",
247			    tcp_display(tcp, NULL,
248			    DISP_ADDR_AND_PORT));
249#else
250			if (connp->conn_debug) {
251				(void) strlog(TCP_MOD_ID, 0, 1,
252				    SL_TRACE|SL_ERROR,
253				    "tcp_wput_data: data after ordrel, %s\n",
254				    tcp_display(tcp, NULL,
255				    DISP_ADDR_AND_PORT));
256			}
257#endif /* DEBUG */
258		}
259		if (tcp->tcp_snd_zcopy_aware &&
260		    (mp->b_datap->db_struioflag & STRUIO_ZCNOTIFY))
261			tcp_zcopy_notify(tcp);
262		freemsg(mp);
263		mutex_enter(&tcp->tcp_non_sq_lock);
264		if (tcp->tcp_flow_stopped &&
265		    TCP_UNSENT_BYTES(tcp) <= connp->conn_sndlowat) {
266			tcp_clrqfull(tcp);
267		}
268		mutex_exit(&tcp->tcp_non_sq_lock);
269		return;
270	}
271
272	/* Strip empties */
273	for (;;) {
274		ASSERT((uintptr_t)(mp->b_wptr - mp->b_rptr) <=
275		    (uintptr_t)INT_MAX);
276		len = (int)(mp->b_wptr - mp->b_rptr);
277		if (len > 0)
278			break;
279		mp1 = mp;
280		mp = mp->b_cont;
281		freeb(mp1);
282		if (mp == NULL) {
283			return;
284		}
285	}
286
287	/* If we are the first on the list ... */
288	if (tcp->tcp_xmit_head == NULL) {
289		tcp->tcp_xmit_head = mp;
290		tcp->tcp_xmit_tail = mp;
291		tcp->tcp_xmit_tail_unsent = len;
292	} else {
293		/* If tiny tx and room in txq tail, pullup to save mblks. */
294		struct datab *dp;
295
296		mp1 = tcp->tcp_xmit_last;
297		if (len < tcp_tx_pull_len &&
298		    (dp = mp1->b_datap)->db_ref == 1 &&
299		    dp->db_lim - mp1->b_wptr >= len) {
300			ASSERT(len > 0);
301			ASSERT(!mp1->b_cont);
302			if (len == 1) {
303				*mp1->b_wptr++ = *mp->b_rptr;
304			} else {
305				bcopy(mp->b_rptr, mp1->b_wptr, len);
306				mp1->b_wptr += len;
307			}
308			if (mp1 == tcp->tcp_xmit_tail)
309				tcp->tcp_xmit_tail_unsent += len;
310			mp1->b_cont = mp->b_cont;
311			if (tcp->tcp_snd_zcopy_aware &&
312			    (mp->b_datap->db_struioflag & STRUIO_ZCNOTIFY))
313				mp1->b_datap->db_struioflag |= STRUIO_ZCNOTIFY;
314			freeb(mp);
315			mp = mp1;
316		} else {
317			tcp->tcp_xmit_last->b_cont = mp;
318		}
319		len += tcp->tcp_unsent;
320	}
321
322	/* Tack on however many more positive length mblks we have */
323	if ((mp1 = mp->b_cont) != NULL) {
324		do {
325			int tlen;
326			ASSERT((uintptr_t)(mp1->b_wptr - mp1->b_rptr) <=
327			    (uintptr_t)INT_MAX);
328			tlen = (int)(mp1->b_wptr - mp1->b_rptr);
329			if (tlen <= 0) {
330				mp->b_cont = mp1->b_cont;
331				freeb(mp1);
332			} else {
333				len += tlen;
334				mp = mp1;
335			}
336		} while ((mp1 = mp->b_cont) != NULL);
337	}
338	tcp->tcp_xmit_last = mp;
339	tcp->tcp_unsent = len;
340
341	if (urgent)
342		usable = 1;
343
344data_null:
345	snxt = tcp->tcp_snxt;
346	xmit_tail = tcp->tcp_xmit_tail;
347	tail_unsent = tcp->tcp_xmit_tail_unsent;
348
349	/*
350	 * Note that tcp_mss has been adjusted to take into account the
351	 * timestamp option if applicable.  Because SACK options do not
352	 * appear in every TCP segments and they are of variable lengths,
353	 * they cannot be included in tcp_mss.  Thus we need to calculate
354	 * the actual segment length when we need to send a segment which
355	 * includes SACK options.
356	 */
357	if (tcp->tcp_snd_sack_ok && tcp->tcp_num_sack_blk > 0) {
358		int32_t	opt_len;
359
360		num_sack_blk = MIN(tcp->tcp_max_sack_blk,
361		    tcp->tcp_num_sack_blk);
362		opt_len = num_sack_blk * sizeof (sack_blk_t) + TCPOPT_NOP_LEN *
363		    2 + TCPOPT_HEADER_LEN;
364		mss = tcp->tcp_mss - opt_len;
365		total_hdr_len = connp->conn_ht_iphc_len + opt_len;
366		tcp_hdr_len = connp->conn_ht_ulp_len + opt_len;
367	} else {
368		mss = tcp->tcp_mss;
369		total_hdr_len = connp->conn_ht_iphc_len;
370		tcp_hdr_len = connp->conn_ht_ulp_len;
371	}
372
373	if ((tcp->tcp_suna == snxt) && !tcp->tcp_localnet &&
374	    (TICK_TO_MSEC(now - tcp->tcp_last_recv_time) >= tcp->tcp_rto)) {
375		TCP_SET_INIT_CWND(tcp, mss, tcps->tcps_slow_start_after_idle);
376	}
377	if (tcpstate == TCPS_SYN_RCVD) {
378		/*
379		 * The three-way connection establishment handshake is not
380		 * complete yet. We want to queue the data for transmission
381		 * after entering ESTABLISHED state (RFC793). A jump to
382		 * "done" label effectively leaves data on the queue.
383		 */
384		goto done;
385	} else {
386		int usable_r;
387
388		/*
389		 * In the special case when cwnd is zero, which can only
390		 * happen if the connection is ECN capable, return now.
391		 * New segments is sent using tcp_timer().  The timer
392		 * is set in tcp_input_data().
393		 */
394		if (tcp->tcp_cwnd == 0) {
395			/*
396			 * Note that tcp_cwnd is 0 before 3-way handshake is
397			 * finished.
398			 */
399			ASSERT(tcp->tcp_ecn_ok ||
400			    tcp->tcp_state < TCPS_ESTABLISHED);
401			return;
402		}
403
404		/* NOTE: trouble if xmitting while SYN not acked? */
405		usable_r = snxt - tcp->tcp_suna;
406		usable_r = tcp->tcp_swnd - usable_r;
407
408		/*
409		 * Check if the receiver has shrunk the window.  If
410		 * tcp_wput_data() with NULL mp is called, tcp_fin_sent
411		 * cannot be set as there is unsent data, so FIN cannot
412		 * be sent out.  Otherwise, we need to take into account
413		 * of FIN as it consumes an "invisible" sequence number.
414		 */
415		ASSERT(tcp->tcp_fin_sent == 0);
416		if (usable_r < 0) {
417			/*
418			 * The receiver has shrunk the window and we have sent
419			 * -usable_r date beyond the window, re-adjust.
420			 *
421			 * If TCP window scaling is enabled, there can be
422			 * round down error as the advertised receive window
423			 * is actually right shifted n bits.  This means that
424			 * the lower n bits info is wiped out.  It will look
425			 * like the window is shrunk.  Do a check here to
426			 * see if the shrunk amount is actually within the
427			 * error in window calculation.  If it is, just
428			 * return.  Note that this check is inside the
429			 * shrunk window check.  This makes sure that even
430			 * though tcp_process_shrunk_swnd() is not called,
431			 * we will stop further processing.
432			 */
433			if ((-usable_r >> tcp->tcp_snd_ws) > 0) {
434				tcp_process_shrunk_swnd(tcp, -usable_r);
435			}
436			return;
437		}
438
439		/* usable = MIN(swnd, cwnd) - unacked_bytes */
440		if (tcp->tcp_swnd > tcp->tcp_cwnd)
441			usable_r -= tcp->tcp_swnd - tcp->tcp_cwnd;
442
443		/* usable = MIN(usable, unsent) */
444		if (usable_r > len)
445			usable_r = len;
446
447		/* usable = MAX(usable, {1 for urgent, 0 for data}) */
448		if (usable_r > 0) {
449			usable = usable_r;
450		} else {
451			/* Bypass all other unnecessary processing. */
452			goto done;
453		}
454	}
455
456	local_time = (mblk_t *)now;
457
458	/*
459	 * "Our" Nagle Algorithm.  This is not the same as in the old
460	 * BSD.  This is more in line with the true intent of Nagle.
461	 *
462	 * The conditions are:
463	 * 1. The amount of unsent data (or amount of data which can be
464	 *    sent, whichever is smaller) is less than Nagle limit.
465	 * 2. The last sent size is also less than Nagle limit.
466	 * 3. There is unack'ed data.
467	 * 4. Urgent pointer is not set.  Send urgent data ignoring the
468	 *    Nagle algorithm.  This reduces the probability that urgent
469	 *    bytes get "merged" together.
470	 * 5. The app has not closed the connection.  This eliminates the
471	 *    wait time of the receiving side waiting for the last piece of
472	 *    (small) data.
473	 *
474	 * If all are satisified, exit without sending anything.  Note
475	 * that Nagle limit can be smaller than 1 MSS.  Nagle limit is
476	 * the smaller of 1 MSS and global tcp_naglim_def (default to be
477	 * 4095).
478	 */
479	if (usable < (int)tcp->tcp_naglim &&
480	    tcp->tcp_naglim > tcp->tcp_last_sent_len &&
481	    snxt != tcp->tcp_suna &&
482	    !(tcp->tcp_valid_bits & TCP_URG_VALID) &&
483	    !(tcp->tcp_valid_bits & TCP_FSS_VALID)) {
484		goto done;
485	}
486
487	/*
488	 * If tcp_zero_win_probe is not set and the tcp->tcp_cork option
489	 * is set, then we have to force TCP not to send partial segment
490	 * (smaller than MSS bytes). We are calculating the usable now
491	 * based on full mss and will save the rest of remaining data for
492	 * later. When tcp_zero_win_probe is set, TCP needs to send out
493	 * something to do zero window probe.
494	 */
495	if (tcp->tcp_cork && !tcp->tcp_zero_win_probe) {
496		if (usable < mss)
497			goto done;
498		usable = (usable / mss) * mss;
499	}
500
501	/* Update the latest receive window size in TCP header. */
502	tcp->tcp_tcpha->tha_win = htons(tcp->tcp_rwnd >> tcp->tcp_rcv_ws);
503
504	/* Send the packet. */
505	rc = tcp_send(tcp, mss, total_hdr_len, tcp_hdr_len,
506	    num_sack_blk, &usable, &snxt, &tail_unsent, &xmit_tail,
507	    local_time);
508
509	/* Pretend that all we were trying to send really got sent */
510	if (rc < 0 && tail_unsent < 0) {
511		do {
512			xmit_tail = xmit_tail->b_cont;
513			xmit_tail->b_prev = local_time;
514			ASSERT((uintptr_t)(xmit_tail->b_wptr -
515			    xmit_tail->b_rptr) <= (uintptr_t)INT_MAX);
516			tail_unsent += (int)(xmit_tail->b_wptr -
517			    xmit_tail->b_rptr);
518		} while (tail_unsent < 0);
519	}
520done:;
521	tcp->tcp_xmit_tail = xmit_tail;
522	tcp->tcp_xmit_tail_unsent = tail_unsent;
523	len = tcp->tcp_snxt - snxt;
524	if (len) {
525		/*
526		 * If new data was sent, need to update the notsack
527		 * list, which is, afterall, data blocks that have
528		 * not been sack'ed by the receiver.  New data is
529		 * not sack'ed.
530		 */
531		if (tcp->tcp_snd_sack_ok && tcp->tcp_notsack_list != NULL) {
532			/* len is a negative value. */
533			tcp->tcp_pipe -= len;
534			tcp_notsack_update(&(tcp->tcp_notsack_list),
535			    tcp->tcp_snxt, snxt,
536			    &(tcp->tcp_num_notsack_blk),
537			    &(tcp->tcp_cnt_notsack_list));
538		}
539		tcp->tcp_snxt = snxt + tcp->tcp_fin_sent;
540		tcp->tcp_rack = tcp->tcp_rnxt;
541		tcp->tcp_rack_cnt = 0;
542		if ((snxt + len) == tcp->tcp_suna) {
543			TCP_TIMER_RESTART(tcp, tcp->tcp_rto);
544		}
545	} else if (snxt == tcp->tcp_suna && tcp->tcp_swnd == 0) {
546		/*
547		 * Didn't send anything. Make sure the timer is running
548		 * so that we will probe a zero window.
549		 */
550		TCP_TIMER_RESTART(tcp, tcp->tcp_rto);
551	}
552	/* Note that len is the amount we just sent but with a negative sign */
553	tcp->tcp_unsent += len;
554	mutex_enter(&tcp->tcp_non_sq_lock);
555	if (tcp->tcp_flow_stopped) {
556		if (TCP_UNSENT_BYTES(tcp) <= connp->conn_sndlowat) {
557			tcp_clrqfull(tcp);
558		}
559	} else if (TCP_UNSENT_BYTES(tcp) >= connp->conn_sndbuf) {
560		if (!(tcp->tcp_detached))
561			tcp_setqfull(tcp);
562	}
563	mutex_exit(&tcp->tcp_non_sq_lock);
564}
565
566/*
567 * Initial STREAMS write side put() procedure for sockets. It tries to
568 * handle the T_CAPABILITY_REQ which sockfs sends down while setting
569 * up the socket without using the squeue. Non T_CAPABILITY_REQ messages
570 * are handled by tcp_wput() as usual.
571 *
572 * All further messages will also be handled by tcp_wput() because we cannot
573 * be sure that the above short cut is safe later.
574 */
575void
576tcp_wput_sock(queue_t *wq, mblk_t *mp)
577{
578	conn_t			*connp = Q_TO_CONN(wq);
579	tcp_t			*tcp = connp->conn_tcp;
580	struct T_capability_req	*car = (struct T_capability_req *)mp->b_rptr;
581
582	ASSERT(wq->q_qinfo == &tcp_sock_winit);
583	wq->q_qinfo = &tcp_winit;
584
585	ASSERT(IPCL_IS_TCP(connp));
586	ASSERT(TCP_IS_SOCKET(tcp));
587
588	if (DB_TYPE(mp) == M_PCPROTO &&
589	    MBLKL(mp) == sizeof (struct T_capability_req) &&
590	    car->PRIM_type == T_CAPABILITY_REQ) {
591		tcp_capability_req(tcp, mp);
592		return;
593	}
594
595	tcp_wput(wq, mp);
596}
597
598/* ARGSUSED */
599void
600tcp_wput_fallback(queue_t *wq, mblk_t *mp)
601{
602#ifdef DEBUG
603	cmn_err(CE_CONT, "tcp_wput_fallback: Message during fallback \n");
604#endif
605	freemsg(mp);
606}
607
608/*
609 * Call by tcp_wput() to handle misc non M_DATA messages.
610 */
611/* ARGSUSED */
612static void
613tcp_wput_nondata(void *arg, mblk_t *mp, void *arg2, ip_recv_attr_t *dummy)
614{
615	conn_t	*connp = (conn_t *)arg;
616	tcp_t	*tcp = connp->conn_tcp;
617
618	ASSERT(DB_TYPE(mp) != M_IOCTL);
619	/*
620	 * TCP is D_MP and qprocsoff() is done towards the end of the tcp_close.
621	 * Once the close starts, streamhead and sockfs will not let any data
622	 * packets come down (close ensures that there are no threads using the
623	 * queue and no new threads will come down) but since qprocsoff()
624	 * hasn't happened yet, a M_FLUSH or some non data message might
625	 * get reflected back (in response to our own FLUSHRW) and get
626	 * processed after tcp_close() is done. The conn would still be valid
627	 * because a ref would have added but we need to check the state
628	 * before actually processing the packet.
629	 */
630	if (TCP_IS_DETACHED(tcp) || (tcp->tcp_state == TCPS_CLOSED)) {
631		freemsg(mp);
632		return;
633	}
634
635	switch (DB_TYPE(mp)) {
636	case M_IOCDATA:
637		tcp_wput_iocdata(tcp, mp);
638		break;
639	case M_FLUSH:
640		tcp_wput_flush(tcp, mp);
641		break;
642	default:
643		ip_wput_nondata(connp->conn_wq, mp);
644		break;
645	}
646}
647
648/* tcp_wput_flush is called by tcp_wput_nondata to handle M_FLUSH messages. */
649static void
650tcp_wput_flush(tcp_t *tcp, mblk_t *mp)
651{
652	uchar_t	fval = *mp->b_rptr;
653	mblk_t	*tail;
654	conn_t	*connp = tcp->tcp_connp;
655	queue_t	*q = connp->conn_wq;
656
657	/* TODO: How should flush interact with urgent data? */
658	if ((fval & FLUSHW) && tcp->tcp_xmit_head != NULL &&
659	    !(tcp->tcp_valid_bits & TCP_URG_VALID)) {
660		/*
661		 * Flush only data that has not yet been put on the wire.  If
662		 * we flush data that we have already transmitted, life, as we
663		 * know it, may come to an end.
664		 */
665		tail = tcp->tcp_xmit_tail;
666		tail->b_wptr -= tcp->tcp_xmit_tail_unsent;
667		tcp->tcp_xmit_tail_unsent = 0;
668		tcp->tcp_unsent = 0;
669		if (tail->b_wptr != tail->b_rptr)
670			tail = tail->b_cont;
671		if (tail) {
672			mblk_t **excess = &tcp->tcp_xmit_head;
673			for (;;) {
674				mblk_t *mp1 = *excess;
675				if (mp1 == tail)
676					break;
677				tcp->tcp_xmit_tail = mp1;
678				tcp->tcp_xmit_last = mp1;
679				excess = &mp1->b_cont;
680			}
681			*excess = NULL;
682			tcp_close_mpp(&tail);
683			if (tcp->tcp_snd_zcopy_aware)
684				tcp_zcopy_notify(tcp);
685		}
686		/*
687		 * We have no unsent data, so unsent must be less than
688		 * conn_sndlowat, so re-enable flow.
689		 */
690		mutex_enter(&tcp->tcp_non_sq_lock);
691		if (tcp->tcp_flow_stopped) {
692			tcp_clrqfull(tcp);
693		}
694		mutex_exit(&tcp->tcp_non_sq_lock);
695	}
696	/*
697	 * TODO: you can't just flush these, you have to increase rwnd for one
698	 * thing.  For another, how should urgent data interact?
699	 */
700	if (fval & FLUSHR) {
701		*mp->b_rptr = fval & ~FLUSHW;
702		/* XXX */
703		qreply(q, mp);
704		return;
705	}
706	freemsg(mp);
707}
708
709/*
710 * tcp_wput_iocdata is called by tcp_wput_nondata to handle all M_IOCDATA
711 * messages.
712 */
713static void
714tcp_wput_iocdata(tcp_t *tcp, mblk_t *mp)
715{
716	mblk_t		*mp1;
717	struct iocblk	*iocp = (struct iocblk *)mp->b_rptr;
718	STRUCT_HANDLE(strbuf, sb);
719	uint_t		addrlen;
720	conn_t		*connp = tcp->tcp_connp;
721	queue_t 	*q = connp->conn_wq;
722
723	/* Make sure it is one of ours. */
724	switch (iocp->ioc_cmd) {
725	case TI_GETMYNAME:
726	case TI_GETPEERNAME:
727		break;
728	default:
729		/*
730		 * If the conn is closing, then error the ioctl here. Otherwise
731		 * use the CONN_IOCTLREF_* macros to hold off tcp_close until
732		 * we're done here.
733		 */
734		mutex_enter(&connp->conn_lock);
735		if (connp->conn_state_flags & CONN_CLOSING) {
736			mutex_exit(&connp->conn_lock);
737			iocp->ioc_error = EINVAL;
738			mp->b_datap->db_type = M_IOCNAK;
739			iocp->ioc_count = 0;
740			qreply(q, mp);
741			return;
742		}
743
744		CONN_INC_IOCTLREF_LOCKED(connp);
745		ip_wput_nondata(q, mp);
746		CONN_DEC_IOCTLREF(connp);
747		return;
748	}
749	switch (mi_copy_state(q, mp, &mp1)) {
750	case -1:
751		return;
752	case MI_COPY_CASE(MI_COPY_IN, 1):
753		break;
754	case MI_COPY_CASE(MI_COPY_OUT, 1):
755		/* Copy out the strbuf. */
756		mi_copyout(q, mp);
757		return;
758	case MI_COPY_CASE(MI_COPY_OUT, 2):
759		/* All done. */
760		mi_copy_done(q, mp, 0);
761		return;
762	default:
763		mi_copy_done(q, mp, EPROTO);
764		return;
765	}
766	/* Check alignment of the strbuf */
767	if (!OK_32PTR(mp1->b_rptr)) {
768		mi_copy_done(q, mp, EINVAL);
769		return;
770	}
771
772	STRUCT_SET_HANDLE(sb, iocp->ioc_flag, (void *)mp1->b_rptr);
773
774	if (connp->conn_family == AF_INET)
775		addrlen = sizeof (sin_t);
776	else
777		addrlen = sizeof (sin6_t);
778
779	if (STRUCT_FGET(sb, maxlen) < addrlen) {
780		mi_copy_done(q, mp, EINVAL);
781		return;
782	}
783
784	switch (iocp->ioc_cmd) {
785	case TI_GETMYNAME:
786		break;
787	case TI_GETPEERNAME:
788		if (tcp->tcp_state < TCPS_SYN_RCVD) {
789			mi_copy_done(q, mp, ENOTCONN);
790			return;
791		}
792		break;
793	}
794	mp1 = mi_copyout_alloc(q, mp, STRUCT_FGETP(sb, buf), addrlen, B_TRUE);
795	if (!mp1)
796		return;
797
798	STRUCT_FSET(sb, len, addrlen);
799	switch (((struct iocblk *)mp->b_rptr)->ioc_cmd) {
800	case TI_GETMYNAME:
801		(void) conn_getsockname(connp, (struct sockaddr *)mp1->b_wptr,
802		    &addrlen);
803		break;
804	case TI_GETPEERNAME:
805		(void) conn_getpeername(connp, (struct sockaddr *)mp1->b_wptr,
806		    &addrlen);
807		break;
808	}
809	mp1->b_wptr += addrlen;
810	/* Copy out the address */
811	mi_copyout(q, mp);
812}
813
814/*
815 * tcp_wput_ioctl is called by tcp_wput_nondata() to handle all M_IOCTL
816 * messages.
817 */
818/* ARGSUSED */
819static void
820tcp_wput_ioctl(void *arg, mblk_t *mp, void *arg2, ip_recv_attr_t *dummy)
821{
822	conn_t 		*connp = (conn_t *)arg;
823	tcp_t		*tcp = connp->conn_tcp;
824	queue_t		*q = connp->conn_wq;
825	struct iocblk	*iocp;
826
827	ASSERT(DB_TYPE(mp) == M_IOCTL);
828	/*
829	 * Try and ASSERT the minimum possible references on the
830	 * conn early enough. Since we are executing on write side,
831	 * the connection is obviously not detached and that means
832	 * there is a ref each for TCP and IP. Since we are behind
833	 * the squeue, the minimum references needed are 3. If the
834	 * conn is in classifier hash list, there should be an
835	 * extra ref for that (we check both the possibilities).
836	 */
837	ASSERT((connp->conn_fanout != NULL && connp->conn_ref >= 4) ||
838	    (connp->conn_fanout == NULL && connp->conn_ref >= 3));
839
840	iocp = (struct iocblk *)mp->b_rptr;
841	switch (iocp->ioc_cmd) {
842	case _SIOCSOCKFALLBACK:
843		/*
844		 * Either sockmod is about to be popped and the socket
845		 * would now be treated as a plain stream, or a module
846		 * is about to be pushed so we could no longer use read-
847		 * side synchronous streams for fused loopback tcp.
848		 * Drain any queued data and disable direct sockfs
849		 * interface from now on.
850		 */
851		if (!tcp->tcp_issocket) {
852			DB_TYPE(mp) = M_IOCNAK;
853			iocp->ioc_error = EINVAL;
854		} else {
855			tcp_use_pure_tpi(tcp);
856			DB_TYPE(mp) = M_IOCACK;
857			iocp->ioc_error = 0;
858		}
859		iocp->ioc_count = 0;
860		iocp->ioc_rval = 0;
861		qreply(q, mp);
862		return;
863	}
864
865	/*
866	 * If the conn is closing, then error the ioctl here. Otherwise bump the
867	 * conn_ioctlref to hold off tcp_close until we're done here.
868	 */
869	mutex_enter(&(connp)->conn_lock);
870	if ((connp)->conn_state_flags & CONN_CLOSING) {
871		mutex_exit(&(connp)->conn_lock);
872		iocp->ioc_error = EINVAL;
873		mp->b_datap->db_type = M_IOCNAK;
874		iocp->ioc_count = 0;
875		qreply(q, mp);
876		return;
877	}
878
879	CONN_INC_IOCTLREF_LOCKED(connp);
880	ip_wput_nondata(q, mp);
881	CONN_DEC_IOCTLREF(connp);
882}
883
884/*
885 * This routine is called by tcp_wput() to handle all TPI requests.
886 */
887/* ARGSUSED */
888static void
889tcp_wput_proto(void *arg, mblk_t *mp, void *arg2, ip_recv_attr_t *dummy)
890{
891	conn_t		*connp = (conn_t *)arg;
892	tcp_t		*tcp = connp->conn_tcp;
893	union T_primitives *tprim = (union T_primitives *)mp->b_rptr;
894	uchar_t		*rptr;
895	t_scalar_t	type;
896	cred_t		*cr;
897
898	/*
899	 * Try and ASSERT the minimum possible references on the
900	 * conn early enough. Since we are executing on write side,
901	 * the connection is obviously not detached and that means
902	 * there is a ref each for TCP and IP. Since we are behind
903	 * the squeue, the minimum references needed are 3. If the
904	 * conn is in classifier hash list, there should be an
905	 * extra ref for that (we check both the possibilities).
906	 */
907	ASSERT((connp->conn_fanout != NULL && connp->conn_ref >= 4) ||
908	    (connp->conn_fanout == NULL && connp->conn_ref >= 3));
909
910	rptr = mp->b_rptr;
911	ASSERT((uintptr_t)(mp->b_wptr - rptr) <= (uintptr_t)INT_MAX);
912	if ((mp->b_wptr - rptr) >= sizeof (t_scalar_t)) {
913		type = ((union T_primitives *)rptr)->type;
914		if (type == T_EXDATA_REQ) {
915			tcp_output_urgent(connp, mp, arg2, NULL);
916		} else if (type != T_DATA_REQ) {
917			goto non_urgent_data;
918		} else {
919			/* TODO: options, flags, ... from user */
920			/* Set length to zero for reclamation below */
921			tcp_wput_data(tcp, mp->b_cont, B_TRUE);
922			freeb(mp);
923		}
924		return;
925	} else {
926		if (connp->conn_debug) {
927			(void) strlog(TCP_MOD_ID, 0, 1, SL_ERROR|SL_TRACE,
928			    "tcp_wput_proto, dropping one...");
929		}
930		freemsg(mp);
931		return;
932	}
933
934non_urgent_data:
935
936	switch ((int)tprim->type) {
937	case T_SSL_PROXY_BIND_REQ:	/* an SSL proxy endpoint bind request */
938		/*
939		 * save the kssl_ent_t from the next block, and convert this
940		 * back to a normal bind_req.
941		 */
942		if (mp->b_cont != NULL) {
943			ASSERT(MBLKL(mp->b_cont) >= sizeof (kssl_ent_t));
944
945			if (tcp->tcp_kssl_ent != NULL) {
946				kssl_release_ent(tcp->tcp_kssl_ent, NULL,
947				    KSSL_NO_PROXY);
948				tcp->tcp_kssl_ent = NULL;
949			}
950			bcopy(mp->b_cont->b_rptr, &tcp->tcp_kssl_ent,
951			    sizeof (kssl_ent_t));
952			kssl_hold_ent(tcp->tcp_kssl_ent);
953			freemsg(mp->b_cont);
954			mp->b_cont = NULL;
955		}
956		tprim->type = T_BIND_REQ;
957
958	/* FALLTHROUGH */
959	case O_T_BIND_REQ:	/* bind request */
960	case T_BIND_REQ:	/* new semantics bind request */
961		tcp_tpi_bind(tcp, mp);
962		break;
963	case T_UNBIND_REQ:	/* unbind request */
964		tcp_tpi_unbind(tcp, mp);
965		break;
966	case O_T_CONN_RES:	/* old connection response XXX */
967	case T_CONN_RES:	/* connection response */
968		tcp_tli_accept(tcp, mp);
969		break;
970	case T_CONN_REQ:	/* connection request */
971		tcp_tpi_connect(tcp, mp);
972		break;
973	case T_DISCON_REQ:	/* disconnect request */
974		tcp_disconnect(tcp, mp);
975		break;
976	case T_CAPABILITY_REQ:
977		tcp_capability_req(tcp, mp);	/* capability request */
978		break;
979	case T_INFO_REQ:	/* information request */
980		tcp_info_req(tcp, mp);
981		break;
982	case T_SVR4_OPTMGMT_REQ:	/* manage options req */
983	case T_OPTMGMT_REQ:
984		/*
985		 * Note:  no support for snmpcom_req() through new
986		 * T_OPTMGMT_REQ. See comments in ip.c
987		 */
988
989		/*
990		 * All Solaris components should pass a db_credp
991		 * for this TPI message, hence we ASSERT.
992		 * But in case there is some other M_PROTO that looks
993		 * like a TPI message sent by some other kernel
994		 * component, we check and return an error.
995		 */
996		cr = msg_getcred(mp, NULL);
997		ASSERT(cr != NULL);
998		if (cr == NULL) {
999			tcp_err_ack(tcp, mp, TSYSERR, EINVAL);
1000			return;
1001		}
1002		/*
1003		 * If EINPROGRESS is returned, the request has been queued
1004		 * for subsequent processing by ip_restart_optmgmt(), which
1005		 * will do the CONN_DEC_REF().
1006		 */
1007		if ((int)tprim->type == T_SVR4_OPTMGMT_REQ) {
1008			svr4_optcom_req(connp->conn_wq, mp, cr, &tcp_opt_obj);
1009		} else {
1010			tpi_optcom_req(connp->conn_wq, mp, cr, &tcp_opt_obj);
1011		}
1012		break;
1013
1014	case T_UNITDATA_REQ:	/* unitdata request */
1015		tcp_err_ack(tcp, mp, TNOTSUPPORT, 0);
1016		break;
1017	case T_ORDREL_REQ:	/* orderly release req */
1018		freemsg(mp);
1019
1020		if (tcp->tcp_fused)
1021			tcp_unfuse(tcp);
1022
1023		if (tcp_xmit_end(tcp) != 0) {
1024			/*
1025			 * We were crossing FINs and got a reset from
1026			 * the other side. Just ignore it.
1027			 */
1028			if (connp->conn_debug) {
1029				(void) strlog(TCP_MOD_ID, 0, 1,
1030				    SL_ERROR|SL_TRACE,
1031				    "tcp_wput_proto, T_ORDREL_REQ out of "
1032				    "state %s",
1033				    tcp_display(tcp, NULL,
1034				    DISP_ADDR_AND_PORT));
1035			}
1036		}
1037		break;
1038	case T_ADDR_REQ:
1039		tcp_addr_req(tcp, mp);
1040		break;
1041	default:
1042		if (connp->conn_debug) {
1043			(void) strlog(TCP_MOD_ID, 0, 1, SL_ERROR|SL_TRACE,
1044			    "tcp_wput_proto, bogus TPI msg, type %d",
1045			    tprim->type);
1046		}
1047		/*
1048		 * We used to M_ERROR.  Sending TNOTSUPPORT gives the user
1049		 * to recover.
1050		 */
1051		tcp_err_ack(tcp, mp, TNOTSUPPORT, 0);
1052		break;
1053	}
1054}
1055
1056/*
1057 * Handle special out-of-band ioctl requests (see PSARC/2008/265).
1058 */
1059static void
1060tcp_wput_cmdblk(queue_t *q, mblk_t *mp)
1061{
1062	void	*data;
1063	mblk_t	*datamp = mp->b_cont;
1064	conn_t	*connp = Q_TO_CONN(q);
1065	tcp_t	*tcp = connp->conn_tcp;
1066	cmdblk_t *cmdp = (cmdblk_t *)mp->b_rptr;
1067
1068	if (datamp == NULL || MBLKL(datamp) < cmdp->cb_len) {
1069		cmdp->cb_error = EPROTO;
1070		qreply(q, mp);
1071		return;
1072	}
1073
1074	data = datamp->b_rptr;
1075
1076	switch (cmdp->cb_cmd) {
1077	case TI_GETPEERNAME:
1078		if (tcp->tcp_state < TCPS_SYN_RCVD)
1079			cmdp->cb_error = ENOTCONN;
1080		else
1081			cmdp->cb_error = conn_getpeername(connp, data,
1082			    &cmdp->cb_len);
1083		break;
1084	case TI_GETMYNAME:
1085		cmdp->cb_error = conn_getsockname(connp, data, &cmdp->cb_len);
1086		break;
1087	default:
1088		cmdp->cb_error = EINVAL;
1089		break;
1090	}
1091
1092	qreply(q, mp);
1093}
1094
1095/*
1096 * The TCP fast path write put procedure.
1097 * NOTE: the logic of the fast path is duplicated from tcp_wput_data()
1098 */
1099/* ARGSUSED */
1100void
1101tcp_output(void *arg, mblk_t *mp, void *arg2, ip_recv_attr_t *dummy)
1102{
1103	int		len;
1104	int		hdrlen;
1105	int		plen;
1106	mblk_t		*mp1;
1107	uchar_t		*rptr;
1108	uint32_t	snxt;
1109	tcpha_t		*tcpha;
1110	struct datab	*db;
1111	uint32_t	suna;
1112	uint32_t	mss;
1113	ipaddr_t	*dst;
1114	ipaddr_t	*src;
1115	uint32_t	sum;
1116	int		usable;
1117	conn_t		*connp = (conn_t *)arg;
1118	tcp_t		*tcp = connp->conn_tcp;
1119	uint32_t	msize;
1120	tcp_stack_t	*tcps = tcp->tcp_tcps;
1121	ip_xmit_attr_t	*ixa;
1122	clock_t		now;
1123
1124	/*
1125	 * Try and ASSERT the minimum possible references on the
1126	 * conn early enough. Since we are executing on write side,
1127	 * the connection is obviously not detached and that means
1128	 * there is a ref each for TCP and IP. Since we are behind
1129	 * the squeue, the minimum references needed are 3. If the
1130	 * conn is in classifier hash list, there should be an
1131	 * extra ref for that (we check both the possibilities).
1132	 */
1133	ASSERT((connp->conn_fanout != NULL && connp->conn_ref >= 4) ||
1134	    (connp->conn_fanout == NULL && connp->conn_ref >= 3));
1135
1136	ASSERT(DB_TYPE(mp) == M_DATA);
1137	msize = (mp->b_cont == NULL) ? MBLKL(mp) : msgdsize(mp);
1138
1139	mutex_enter(&tcp->tcp_non_sq_lock);
1140	tcp->tcp_squeue_bytes -= msize;
1141	mutex_exit(&tcp->tcp_non_sq_lock);
1142
1143	/* Bypass tcp protocol for fused tcp loopback */
1144	if (tcp->tcp_fused && tcp_fuse_output(tcp, mp, msize))
1145		return;
1146
1147	mss = tcp->tcp_mss;
1148	/*
1149	 * If ZEROCOPY has turned off, try not to send any zero-copy message
1150	 * down. Do backoff, now.
1151	 */
1152	if (tcp->tcp_snd_zcopy_aware && !tcp->tcp_snd_zcopy_on)
1153		mp = tcp_zcopy_backoff(tcp, mp, B_FALSE);
1154
1155
1156	ASSERT((uintptr_t)(mp->b_wptr - mp->b_rptr) <= (uintptr_t)INT_MAX);
1157	len = (int)(mp->b_wptr - mp->b_rptr);
1158
1159	/*
1160	 * Criteria for fast path:
1161	 *
1162	 *   1. no unsent data
1163	 *   2. single mblk in request
1164	 *   3. connection established
1165	 *   4. data in mblk
1166	 *   5. len <= mss
1167	 *   6. no tcp_valid bits
1168	 */
1169	if ((tcp->tcp_unsent != 0) ||
1170	    (tcp->tcp_cork) ||
1171	    (mp->b_cont != NULL) ||
1172	    (tcp->tcp_state != TCPS_ESTABLISHED) ||
1173	    (len == 0) ||
1174	    (len > mss) ||
1175	    (tcp->tcp_valid_bits != 0)) {
1176		tcp_wput_data(tcp, mp, B_FALSE);
1177		return;
1178	}
1179
1180	ASSERT(tcp->tcp_xmit_tail_unsent == 0);
1181	ASSERT(tcp->tcp_fin_sent == 0);
1182
1183	/* queue new packet onto retransmission queue */
1184	if (tcp->tcp_xmit_head == NULL) {
1185		tcp->tcp_xmit_head = mp;
1186	} else {
1187		tcp->tcp_xmit_last->b_cont = mp;
1188	}
1189	tcp->tcp_xmit_last = mp;
1190	tcp->tcp_xmit_tail = mp;
1191
1192	/* find out how much we can send */
1193	/* BEGIN CSTYLED */
1194	/*
1195	 *    un-acked	   usable
1196	 *  |--------------|-----------------|
1197	 *  tcp_suna       tcp_snxt	  tcp_suna+tcp_swnd
1198	 */
1199	/* END CSTYLED */
1200
1201	/* start sending from tcp_snxt */
1202	snxt = tcp->tcp_snxt;
1203
1204	/*
1205	 * Check to see if this connection has been idled for some
1206	 * time and no ACK is expected.  If it is, we need to slow
1207	 * start again to get back the connection's "self-clock" as
1208	 * described in VJ's paper.
1209	 *
1210	 * Reinitialize tcp_cwnd after idle.
1211	 */
1212	now = LBOLT_FASTPATH;
1213	if ((tcp->tcp_suna == snxt) && !tcp->tcp_localnet &&
1214	    (TICK_TO_MSEC(now - tcp->tcp_last_recv_time) >= tcp->tcp_rto)) {
1215		TCP_SET_INIT_CWND(tcp, mss, tcps->tcps_slow_start_after_idle);
1216	}
1217
1218	usable = tcp->tcp_swnd;		/* tcp window size */
1219	if (usable > tcp->tcp_cwnd)
1220		usable = tcp->tcp_cwnd;	/* congestion window smaller */
1221	usable -= snxt;		/* subtract stuff already sent */
1222	suna = tcp->tcp_suna;
1223	usable += suna;
1224	/* usable can be < 0 if the congestion window is smaller */
1225	if (len > usable) {
1226		/* Can't send complete M_DATA in one shot */
1227		goto slow;
1228	}
1229
1230	mutex_enter(&tcp->tcp_non_sq_lock);
1231	if (tcp->tcp_flow_stopped &&
1232	    TCP_UNSENT_BYTES(tcp) <= connp->conn_sndlowat) {
1233		tcp_clrqfull(tcp);
1234	}
1235	mutex_exit(&tcp->tcp_non_sq_lock);
1236
1237	/*
1238	 * determine if anything to send (Nagle).
1239	 *
1240	 *   1. len < tcp_mss (i.e. small)
1241	 *   2. unacknowledged data present
1242	 *   3. len < nagle limit
1243	 *   4. last packet sent < nagle limit (previous packet sent)
1244	 */
1245	if ((len < mss) && (snxt != suna) &&
1246	    (len < (int)tcp->tcp_naglim) &&
1247	    (tcp->tcp_last_sent_len < tcp->tcp_naglim)) {
1248		/*
1249		 * This was the first unsent packet and normally
1250		 * mss < xmit_hiwater so there is no need to worry
1251		 * about flow control. The next packet will go
1252		 * through the flow control check in tcp_wput_data().
1253		 */
1254		/* leftover work from above */
1255		tcp->tcp_unsent = len;
1256		tcp->tcp_xmit_tail_unsent = len;
1257
1258		return;
1259	}
1260
1261	/*
1262	 * len <= tcp->tcp_mss && len == unsent so no sender silly window.  Can
1263	 * send now.
1264	 */
1265
1266	if (snxt == suna) {
1267		TCP_TIMER_RESTART(tcp, tcp->tcp_rto);
1268	}
1269
1270	/* we have always sent something */
1271	tcp->tcp_rack_cnt = 0;
1272
1273	tcp->tcp_snxt = snxt + len;
1274	tcp->tcp_rack = tcp->tcp_rnxt;
1275
1276	if ((mp1 = dupb(mp)) == 0)
1277		goto no_memory;
1278	mp->b_prev = (mblk_t *)(uintptr_t)now;
1279	mp->b_next = (mblk_t *)(uintptr_t)snxt;
1280
1281	/* adjust tcp header information */
1282	tcpha = tcp->tcp_tcpha;
1283	tcpha->tha_flags = (TH_ACK|TH_PUSH);
1284
1285	sum = len + connp->conn_ht_ulp_len + connp->conn_sum;
1286	sum = (sum >> 16) + (sum & 0xFFFF);
1287	tcpha->tha_sum = htons(sum);
1288
1289	tcpha->tha_seq = htonl(snxt);
1290
1291	TCPS_BUMP_MIB(tcps, tcpOutDataSegs);
1292	TCPS_UPDATE_MIB(tcps, tcpOutDataBytes, len);
1293	BUMP_LOCAL(tcp->tcp_obsegs);
1294
1295	/* Update the latest receive window size in TCP header. */
1296	tcpha->tha_win = htons(tcp->tcp_rwnd >> tcp->tcp_rcv_ws);
1297
1298	tcp->tcp_last_sent_len = (ushort_t)len;
1299
1300	plen = len + connp->conn_ht_iphc_len;
1301
1302	ixa = connp->conn_ixa;
1303	ixa->ixa_pktlen = plen;
1304
1305	if (ixa->ixa_flags & IXAF_IS_IPV4) {
1306		tcp->tcp_ipha->ipha_length = htons(plen);
1307	} else {
1308		tcp->tcp_ip6h->ip6_plen = htons(plen - IPV6_HDR_LEN);
1309	}
1310
1311	/* see if we need to allocate a mblk for the headers */
1312	hdrlen = connp->conn_ht_iphc_len;
1313	rptr = mp1->b_rptr - hdrlen;
1314	db = mp1->b_datap;
1315	if ((db->db_ref != 2) || rptr < db->db_base ||
1316	    (!OK_32PTR(rptr))) {
1317		/* NOTE: we assume allocb returns an OK_32PTR */
1318		mp = allocb(hdrlen + tcps->tcps_wroff_xtra, BPRI_MED);
1319		if (!mp) {
1320			freemsg(mp1);
1321			goto no_memory;
1322		}
1323		mp->b_cont = mp1;
1324		mp1 = mp;
1325		/* Leave room for Link Level header */
1326		rptr = &mp1->b_rptr[tcps->tcps_wroff_xtra];
1327		mp1->b_wptr = &rptr[hdrlen];
1328	}
1329	mp1->b_rptr = rptr;
1330
1331	/* Fill in the timestamp option. */
1332	if (tcp->tcp_snd_ts_ok) {
1333		uint32_t llbolt = (uint32_t)LBOLT_FASTPATH;
1334
1335		U32_TO_BE32(llbolt,
1336		    (char *)tcpha + TCP_MIN_HEADER_LENGTH+4);
1337		U32_TO_BE32(tcp->tcp_ts_recent,
1338		    (char *)tcpha + TCP_MIN_HEADER_LENGTH+8);
1339	} else {
1340		ASSERT(connp->conn_ht_ulp_len == TCP_MIN_HEADER_LENGTH);
1341	}
1342
1343	/* copy header into outgoing packet */
1344	dst = (ipaddr_t *)rptr;
1345	src = (ipaddr_t *)connp->conn_ht_iphc;
1346	dst[0] = src[0];
1347	dst[1] = src[1];
1348	dst[2] = src[2];
1349	dst[3] = src[3];
1350	dst[4] = src[4];
1351	dst[5] = src[5];
1352	dst[6] = src[6];
1353	dst[7] = src[7];
1354	dst[8] = src[8];
1355	dst[9] = src[9];
1356	if (hdrlen -= 40) {
1357		hdrlen >>= 2;
1358		dst += 10;
1359		src += 10;
1360		do {
1361			*dst++ = *src++;
1362		} while (--hdrlen);
1363	}
1364
1365	/*
1366	 * Set the ECN info in the TCP header.  Note that this
1367	 * is not the template header.
1368	 */
1369	if (tcp->tcp_ecn_ok) {
1370		TCP_SET_ECT(tcp, rptr);
1371
1372		tcpha = (tcpha_t *)(rptr + ixa->ixa_ip_hdr_length);
1373		if (tcp->tcp_ecn_echo_on)
1374			tcpha->tha_flags |= TH_ECE;
1375		if (tcp->tcp_cwr && !tcp->tcp_ecn_cwr_sent) {
1376			tcpha->tha_flags |= TH_CWR;
1377			tcp->tcp_ecn_cwr_sent = B_TRUE;
1378		}
1379	}
1380
1381	if (tcp->tcp_ip_forward_progress) {
1382		tcp->tcp_ip_forward_progress = B_FALSE;
1383		connp->conn_ixa->ixa_flags |= IXAF_REACH_CONF;
1384	} else {
1385		connp->conn_ixa->ixa_flags &= ~IXAF_REACH_CONF;
1386	}
1387	tcp_send_data(tcp, mp1);
1388	return;
1389
1390	/*
1391	 * If we ran out of memory, we pretend to have sent the packet
1392	 * and that it was lost on the wire.
1393	 */
1394no_memory:
1395	return;
1396
1397slow:
1398	/* leftover work from above */
1399	tcp->tcp_unsent = len;
1400	tcp->tcp_xmit_tail_unsent = len;
1401	tcp_wput_data(tcp, NULL, B_FALSE);
1402}
1403
1404/* ARGSUSED2 */
1405void
1406tcp_output_urgent(void *arg, mblk_t *mp, void *arg2, ip_recv_attr_t *dummy)
1407{
1408	int len;
1409	uint32_t msize;
1410	conn_t *connp = (conn_t *)arg;
1411	tcp_t *tcp = connp->conn_tcp;
1412
1413	msize = msgdsize(mp);
1414
1415	len = msize - 1;
1416	if (len < 0) {
1417		freemsg(mp);
1418		return;
1419	}
1420
1421	/*
1422	 * Try to force urgent data out on the wire. Even if we have unsent
1423	 * data this will at least send the urgent flag.
1424	 * XXX does not handle more flag correctly.
1425	 */
1426	len += tcp->tcp_unsent;
1427	len += tcp->tcp_snxt;
1428	tcp->tcp_urg = len;
1429	tcp->tcp_valid_bits |= TCP_URG_VALID;
1430
1431	/* Bypass tcp protocol for fused tcp loopback */
1432	if (tcp->tcp_fused && tcp_fuse_output(tcp, mp, msize))
1433		return;
1434
1435	/* Strip off the T_EXDATA_REQ if the data is from TPI */
1436	if (DB_TYPE(mp) != M_DATA) {
1437		mblk_t *mp1 = mp;
1438		ASSERT(!IPCL_IS_NONSTR(connp));
1439		mp = mp->b_cont;
1440		freeb(mp1);
1441	}
1442	tcp_wput_data(tcp, mp, B_TRUE);
1443}
1444
1445/*
1446 * Called by streams close routine via squeues when our client blows off her
1447 * descriptor, we take this to mean: "close the stream state NOW, close the tcp
1448 * connection politely" When SO_LINGER is set (with a non-zero linger time and
1449 * it is not a nonblocking socket) then this routine sleeps until the FIN is
1450 * acked.
1451 *
1452 * NOTE: tcp_close potentially returns error when lingering.
1453 * However, the stream head currently does not pass these errors
1454 * to the application. 4.4BSD only returns EINTR and EWOULDBLOCK
1455 * errors to the application (from tsleep()) and not errors
1456 * like ECONNRESET caused by receiving a reset packet.
1457 */
1458
1459/* ARGSUSED */
1460void
1461tcp_close_output(void *arg, mblk_t *mp, void *arg2, ip_recv_attr_t *dummy)
1462{
1463	char	*msg;
1464	conn_t	*connp = (conn_t *)arg;
1465	tcp_t	*tcp = connp->conn_tcp;
1466	clock_t	delta = 0;
1467	tcp_stack_t	*tcps = tcp->tcp_tcps;
1468
1469	ASSERT((connp->conn_fanout != NULL && connp->conn_ref >= 4) ||
1470	    (connp->conn_fanout == NULL && connp->conn_ref >= 3));
1471
1472	mutex_enter(&tcp->tcp_eager_lock);
1473	if (tcp->tcp_conn_req_cnt_q0 != 0 || tcp->tcp_conn_req_cnt_q != 0) {
1474		/* Cleanup for listener */
1475		tcp_eager_cleanup(tcp, 0);
1476		tcp->tcp_wait_for_eagers = 1;
1477	}
1478	mutex_exit(&tcp->tcp_eager_lock);
1479
1480	tcp->tcp_lso = B_FALSE;
1481
1482	msg = NULL;
1483	switch (tcp->tcp_state) {
1484	case TCPS_CLOSED:
1485	case TCPS_IDLE:
1486	case TCPS_BOUND:
1487	case TCPS_LISTEN:
1488		break;
1489	case TCPS_SYN_SENT:
1490		msg = "tcp_close, during connect";
1491		break;
1492	case TCPS_SYN_RCVD:
1493		/*
1494		 * Close during the connect 3-way handshake
1495		 * but here there may or may not be pending data
1496		 * already on queue. Process almost same as in
1497		 * the ESTABLISHED state.
1498		 */
1499		/* FALLTHRU */
1500	default:
1501		if (tcp->tcp_fused)
1502			tcp_unfuse(tcp);
1503
1504		/*
1505		 * If SO_LINGER has set a zero linger time, abort the
1506		 * connection with a reset.
1507		 */
1508		if (connp->conn_linger && connp->conn_lingertime == 0) {
1509			msg = "tcp_close, zero lingertime";
1510			break;
1511		}
1512
1513		/*
1514		 * Abort connection if there is unread data queued.
1515		 */
1516		if (tcp->tcp_rcv_list || tcp->tcp_reass_head) {
1517			msg = "tcp_close, unread data";
1518			break;
1519		}
1520		/*
1521		 * We have done a qwait() above which could have possibly
1522		 * drained more messages in turn causing transition to a
1523		 * different state. Check whether we have to do the rest
1524		 * of the processing or not.
1525		 */
1526		if (tcp->tcp_state <= TCPS_LISTEN)
1527			break;
1528
1529		/*
1530		 * Transmit the FIN before detaching the tcp_t.
1531		 * After tcp_detach returns this queue/perimeter
1532		 * no longer owns the tcp_t thus others can modify it.
1533		 */
1534		(void) tcp_xmit_end(tcp);
1535
1536		/*
1537		 * If lingering on close then wait until the fin is acked,
1538		 * the SO_LINGER time passes, or a reset is sent/received.
1539		 */
1540		if (connp->conn_linger && connp->conn_lingertime > 0 &&
1541		    !(tcp->tcp_fin_acked) &&
1542		    tcp->tcp_state >= TCPS_ESTABLISHED) {
1543			if (tcp->tcp_closeflags & (FNDELAY|FNONBLOCK)) {
1544				tcp->tcp_client_errno = EWOULDBLOCK;
1545			} else if (tcp->tcp_client_errno == 0) {
1546
1547				ASSERT(tcp->tcp_linger_tid == 0);
1548
1549				tcp->tcp_linger_tid = TCP_TIMER(tcp,
1550				    tcp_close_linger_timeout,
1551				    connp->conn_lingertime * hz);
1552
1553				/* tcp_close_linger_timeout will finish close */
1554				if (tcp->tcp_linger_tid == 0)
1555					tcp->tcp_client_errno = ENOSR;
1556				else
1557					return;
1558			}
1559
1560			/*
1561			 * Check if we need to detach or just close
1562			 * the instance.
1563			 */
1564			if (tcp->tcp_state <= TCPS_LISTEN)
1565				break;
1566		}
1567
1568		/*
1569		 * Make sure that no other thread will access the conn_rq of
1570		 * this instance (through lookups etc.) as conn_rq will go
1571		 * away shortly.
1572		 */
1573		tcp_acceptor_hash_remove(tcp);
1574
1575		mutex_enter(&tcp->tcp_non_sq_lock);
1576		if (tcp->tcp_flow_stopped) {
1577			tcp_clrqfull(tcp);
1578		}
1579		mutex_exit(&tcp->tcp_non_sq_lock);
1580
1581		if (tcp->tcp_timer_tid != 0) {
1582			delta = TCP_TIMER_CANCEL(tcp, tcp->tcp_timer_tid);
1583			tcp->tcp_timer_tid = 0;
1584		}
1585		/*
1586		 * Need to cancel those timers which will not be used when
1587		 * TCP is detached.  This has to be done before the conn_wq
1588		 * is set to NULL.
1589		 */
1590		tcp_timers_stop(tcp);
1591
1592		tcp->tcp_detached = B_TRUE;
1593		if (tcp->tcp_state == TCPS_TIME_WAIT) {
1594			tcp_time_wait_append(tcp);
1595			TCP_DBGSTAT(tcps, tcp_detach_time_wait);
1596			ASSERT(connp->conn_ref >= 3);
1597			goto finish;
1598		}
1599
1600		/*
1601		 * If delta is zero the timer event wasn't executed and was
1602		 * successfully canceled. In this case we need to restart it
1603		 * with the minimal delta possible.
1604		 */
1605		if (delta >= 0)
1606			tcp->tcp_timer_tid = TCP_TIMER(tcp, tcp_timer,
1607			    delta ? delta : 1);
1608
1609		ASSERT(connp->conn_ref >= 3);
1610		goto finish;
1611	}
1612
1613	/* Detach did not complete. Still need to remove q from stream. */
1614	if (msg) {
1615		if (tcp->tcp_state == TCPS_ESTABLISHED ||
1616		    tcp->tcp_state == TCPS_CLOSE_WAIT)
1617			TCPS_BUMP_MIB(tcps, tcpEstabResets);
1618		if (tcp->tcp_state == TCPS_SYN_SENT ||
1619		    tcp->tcp_state == TCPS_SYN_RCVD)
1620			TCPS_BUMP_MIB(tcps, tcpAttemptFails);
1621		tcp_xmit_ctl(msg, tcp,  tcp->tcp_snxt, 0, TH_RST);
1622	}
1623
1624	tcp_closei_local(tcp);
1625	CONN_DEC_REF(connp);
1626	ASSERT(connp->conn_ref >= 2);
1627
1628finish:
1629	mutex_enter(&tcp->tcp_closelock);
1630	/*
1631	 * Don't change the queues in the case of a listener that has
1632	 * eagers in its q or q0. It could surprise the eagers.
1633	 * Instead wait for the eagers outside the squeue.
1634	 */
1635	if (!tcp->tcp_wait_for_eagers) {
1636		tcp->tcp_detached = B_TRUE;
1637		connp->conn_rq = NULL;
1638		connp->conn_wq = NULL;
1639	}
1640
1641	/* Signal tcp_close() to finish closing. */
1642	tcp->tcp_closed = 1;
1643	cv_signal(&tcp->tcp_closecv);
1644	mutex_exit(&tcp->tcp_closelock);
1645}
1646
1647/* ARGSUSED */
1648void
1649tcp_shutdown_output(void *arg, mblk_t *mp, void *arg2, ip_recv_attr_t *dummy)
1650{
1651	conn_t 	*connp = (conn_t *)arg;
1652	tcp_t	*tcp = connp->conn_tcp;
1653
1654	freemsg(mp);
1655
1656	if (tcp->tcp_fused)
1657		tcp_unfuse(tcp);
1658
1659	if (tcp_xmit_end(tcp) != 0) {
1660		/*
1661		 * We were crossing FINs and got a reset from
1662		 * the other side. Just ignore it.
1663		 */
1664		if (connp->conn_debug) {
1665			(void) strlog(TCP_MOD_ID, 0, 1,
1666			    SL_ERROR|SL_TRACE,
1667			    "tcp_shutdown_output() out of state %s",
1668			    tcp_display(tcp, NULL, DISP_ADDR_AND_PORT));
1669		}
1670	}
1671}
1672
1673#pragma inline(tcp_send_data)
1674
1675void
1676tcp_send_data(tcp_t *tcp, mblk_t *mp)
1677{
1678	conn_t		*connp = tcp->tcp_connp;
1679
1680	/*
1681	 * Check here to avoid sending zero-copy message down to IP when
1682	 * ZEROCOPY capability has turned off. We only need to deal with
1683	 * the race condition between sockfs and the notification here.
1684	 * Since we have tried to backoff the tcp_xmit_head when turning
1685	 * zero-copy off and new messages in tcp_output(), we simply drop
1686	 * the dup'ed packet here and let tcp retransmit, if tcp_xmit_zc_clean
1687	 * is not true.
1688	 */
1689	if (tcp->tcp_snd_zcopy_aware && !tcp->tcp_snd_zcopy_on &&
1690	    !tcp->tcp_xmit_zc_clean) {
1691		ip_drop_output("TCP ZC was disabled but not clean", mp, NULL);
1692		freemsg(mp);
1693		return;
1694	}
1695
1696	ASSERT(connp->conn_ixa->ixa_notify_cookie == connp->conn_tcp);
1697	(void) conn_ip_output(mp, connp->conn_ixa);
1698}
1699
1700/* ARGSUSED2 */
1701void
1702tcp_send_synack(void *arg, mblk_t *mp, void *arg2, ip_recv_attr_t *dummy)
1703{
1704	conn_t	*econnp = (conn_t *)arg;
1705	tcp_t	*tcp = econnp->conn_tcp;
1706
1707	/* Guard against a RST having blown it away while on the squeue */
1708	if (tcp->tcp_state == TCPS_CLOSED) {
1709		freemsg(mp);
1710		return;
1711	}
1712
1713	(void) conn_ip_output(mp, econnp->conn_ixa);
1714}
1715
1716/*
1717 * tcp_send() is called by tcp_wput_data() and returns one of the following:
1718 *
1719 * -1 = failed allocation.
1720 *  0 = success; burst count reached, or usable send window is too small,
1721 *      and that we'd rather wait until later before sending again.
1722 */
1723static int
1724tcp_send(tcp_t *tcp, const int mss, const int total_hdr_len,
1725    const int tcp_hdr_len, const int num_sack_blk, int *usable,
1726    uint_t *snxt, int *tail_unsent, mblk_t **xmit_tail, mblk_t *local_time)
1727{
1728	int		num_burst_seg = tcp->tcp_snd_burst;
1729	int		num_lso_seg = 1;
1730	uint_t		lso_usable;
1731	boolean_t	do_lso_send = B_FALSE;
1732	tcp_stack_t	*tcps = tcp->tcp_tcps;
1733	conn_t		*connp = tcp->tcp_connp;
1734	ip_xmit_attr_t	*ixa = connp->conn_ixa;
1735
1736	/*
1737	 * Check LSO possibility. The value of tcp->tcp_lso indicates whether
1738	 * the underlying connection is LSO capable. Will check whether having
1739	 * enough available data to initiate LSO transmission in the for(){}
1740	 * loops.
1741	 */
1742	if (tcp->tcp_lso && (tcp->tcp_valid_bits & ~TCP_FSS_VALID) == 0)
1743		do_lso_send = B_TRUE;
1744
1745	for (;;) {
1746		struct datab	*db;
1747		tcpha_t		*tcpha;
1748		uint32_t	sum;
1749		mblk_t		*mp, *mp1;
1750		uchar_t		*rptr;
1751		int		len;
1752
1753		/*
1754		 * Burst count reached, return successfully.
1755		 */
1756		if (num_burst_seg == 0)
1757			break;
1758
1759		/*
1760		 * Calculate the maximum payload length we can send at one
1761		 * time.
1762		 */
1763		if (do_lso_send) {
1764			/*
1765			 * Check whether be able to to do LSO for the current
1766			 * available data.
1767			 */
1768			if (num_burst_seg >= 2 && (*usable - 1) / mss >= 1) {
1769				lso_usable = MIN(tcp->tcp_lso_max, *usable);
1770				lso_usable = MIN(lso_usable,
1771				    num_burst_seg * mss);
1772
1773				num_lso_seg = lso_usable / mss;
1774				if (lso_usable % mss) {
1775					num_lso_seg++;
1776					tcp->tcp_last_sent_len = (ushort_t)
1777					    (lso_usable % mss);
1778				} else {
1779					tcp->tcp_last_sent_len = (ushort_t)mss;
1780				}
1781			} else {
1782				do_lso_send = B_FALSE;
1783				num_lso_seg = 1;
1784				lso_usable = mss;
1785			}
1786		}
1787
1788		ASSERT(num_lso_seg <= IP_MAXPACKET / mss + 1);
1789#ifdef DEBUG
1790		DTRACE_PROBE2(tcp_send_lso, int, num_lso_seg, boolean_t,
1791		    do_lso_send);
1792#endif
1793		/*
1794		 * Adjust num_burst_seg here.
1795		 */
1796		num_burst_seg -= num_lso_seg;
1797
1798		len = mss;
1799		if (len > *usable) {
1800			ASSERT(do_lso_send == B_FALSE);
1801
1802			len = *usable;
1803			if (len <= 0) {
1804				/* Terminate the loop */
1805				break;	/* success; too small */
1806			}
1807			/*
1808			 * Sender silly-window avoidance.
1809			 * Ignore this if we are going to send a
1810			 * zero window probe out.
1811			 *
1812			 * TODO: force data into microscopic window?
1813			 *	==> (!pushed || (unsent > usable))
1814			 */
1815			if (len < (tcp->tcp_max_swnd >> 1) &&
1816			    (tcp->tcp_unsent - (*snxt - tcp->tcp_snxt)) > len &&
1817			    !((tcp->tcp_valid_bits & TCP_URG_VALID) &&
1818			    len == 1) && (! tcp->tcp_zero_win_probe)) {
1819				/*
1820				 * If the retransmit timer is not running
1821				 * we start it so that we will retransmit
1822				 * in the case when the receiver has
1823				 * decremented the window.
1824				 */
1825				if (*snxt == tcp->tcp_snxt &&
1826				    *snxt == tcp->tcp_suna) {
1827					/*
1828					 * We are not supposed to send
1829					 * anything.  So let's wait a little
1830					 * bit longer before breaking SWS
1831					 * avoidance.
1832					 *
1833					 * What should the value be?
1834					 * Suggestion: MAX(init rexmit time,
1835					 * tcp->tcp_rto)
1836					 */
1837					TCP_TIMER_RESTART(tcp, tcp->tcp_rto);
1838				}
1839				break;	/* success; too small */
1840			}
1841		}
1842
1843		tcpha = tcp->tcp_tcpha;
1844
1845		/*
1846		 * The reason to adjust len here is that we need to set flags
1847		 * and calculate checksum.
1848		 */
1849		if (do_lso_send)
1850			len = lso_usable;
1851
1852		*usable -= len; /* Approximate - can be adjusted later */
1853		if (*usable > 0)
1854			tcpha->tha_flags = TH_ACK;
1855		else
1856			tcpha->tha_flags = (TH_ACK | TH_PUSH);
1857
1858		/*
1859		 * Prime pump for IP's checksumming on our behalf.
1860		 * Include the adjustment for a source route if any.
1861		 * In case of LSO, the partial pseudo-header checksum should
1862		 * exclusive TCP length, so zero tha_sum before IP calculate
1863		 * pseudo-header checksum for partial checksum offload.
1864		 */
1865		if (do_lso_send) {
1866			sum = 0;
1867		} else {
1868			sum = len + tcp_hdr_len + connp->conn_sum;
1869			sum = (sum >> 16) + (sum & 0xFFFF);
1870		}
1871		tcpha->tha_sum = htons(sum);
1872		tcpha->tha_seq = htonl(*snxt);
1873
1874		/*
1875		 * Branch off to tcp_xmit_mp() if any of the VALID bits is
1876		 * set.  For the case when TCP_FSS_VALID is the only valid
1877		 * bit (normal active close), branch off only when we think
1878		 * that the FIN flag needs to be set.  Note for this case,
1879		 * that (snxt + len) may not reflect the actual seg_len,
1880		 * as len may be further reduced in tcp_xmit_mp().  If len
1881		 * gets modified, we will end up here again.
1882		 */
1883		if (tcp->tcp_valid_bits != 0 &&
1884		    (tcp->tcp_valid_bits != TCP_FSS_VALID ||
1885		    ((*snxt + len) == tcp->tcp_fss))) {
1886			uchar_t		*prev_rptr;
1887			uint32_t	prev_snxt = tcp->tcp_snxt;
1888
1889			if (*tail_unsent == 0) {
1890				ASSERT((*xmit_tail)->b_cont != NULL);
1891				*xmit_tail = (*xmit_tail)->b_cont;
1892				prev_rptr = (*xmit_tail)->b_rptr;
1893				*tail_unsent = (int)((*xmit_tail)->b_wptr -
1894				    (*xmit_tail)->b_rptr);
1895			} else {
1896				prev_rptr = (*xmit_tail)->b_rptr;
1897				(*xmit_tail)->b_rptr = (*xmit_tail)->b_wptr -
1898				    *tail_unsent;
1899			}
1900			mp = tcp_xmit_mp(tcp, *xmit_tail, len, NULL, NULL,
1901			    *snxt, B_FALSE, (uint32_t *)&len, B_FALSE);
1902			/* Restore tcp_snxt so we get amount sent right. */
1903			tcp->tcp_snxt = prev_snxt;
1904			if (prev_rptr == (*xmit_tail)->b_rptr) {
1905				/*
1906				 * If the previous timestamp is still in use,
1907				 * don't stomp on it.
1908				 */
1909				if ((*xmit_tail)->b_next == NULL) {
1910					(*xmit_tail)->b_prev = local_time;
1911					(*xmit_tail)->b_next =
1912					    (mblk_t *)(uintptr_t)(*snxt);
1913				}
1914			} else
1915				(*xmit_tail)->b_rptr = prev_rptr;
1916
1917			if (mp == NULL) {
1918				return (-1);
1919			}
1920			mp1 = mp->b_cont;
1921
1922			if (len <= mss) /* LSO is unusable (!do_lso_send) */
1923				tcp->tcp_last_sent_len = (ushort_t)len;
1924			while (mp1->b_cont) {
1925				*xmit_tail = (*xmit_tail)->b_cont;
1926				(*xmit_tail)->b_prev = local_time;
1927				(*xmit_tail)->b_next =
1928				    (mblk_t *)(uintptr_t)(*snxt);
1929				mp1 = mp1->b_cont;
1930			}
1931			*snxt += len;
1932			*tail_unsent = (*xmit_tail)->b_wptr - mp1->b_wptr;
1933			BUMP_LOCAL(tcp->tcp_obsegs);
1934			TCPS_BUMP_MIB(tcps, tcpOutDataSegs);
1935			TCPS_UPDATE_MIB(tcps, tcpOutDataBytes, len);
1936			tcp_send_data(tcp, mp);
1937			continue;
1938		}
1939
1940		*snxt += len;	/* Adjust later if we don't send all of len */
1941		TCPS_BUMP_MIB(tcps, tcpOutDataSegs);
1942		TCPS_UPDATE_MIB(tcps, tcpOutDataBytes, len);
1943
1944		if (*tail_unsent) {
1945			/* Are the bytes above us in flight? */
1946			rptr = (*xmit_tail)->b_wptr - *tail_unsent;
1947			if (rptr != (*xmit_tail)->b_rptr) {
1948				*tail_unsent -= len;
1949				if (len <= mss) /* LSO is unusable */
1950					tcp->tcp_last_sent_len = (ushort_t)len;
1951				len += total_hdr_len;
1952				ixa->ixa_pktlen = len;
1953
1954				if (ixa->ixa_flags & IXAF_IS_IPV4) {
1955					tcp->tcp_ipha->ipha_length = htons(len);
1956				} else {
1957					tcp->tcp_ip6h->ip6_plen =
1958					    htons(len - IPV6_HDR_LEN);
1959				}
1960
1961				mp = dupb(*xmit_tail);
1962				if (mp == NULL) {
1963					return (-1);	/* out_of_mem */
1964				}
1965				mp->b_rptr = rptr;
1966				/*
1967				 * If the old timestamp is no longer in use,
1968				 * sample a new timestamp now.
1969				 */
1970				if ((*xmit_tail)->b_next == NULL) {
1971					(*xmit_tail)->b_prev = local_time;
1972					(*xmit_tail)->b_next =
1973					    (mblk_t *)(uintptr_t)(*snxt-len);
1974				}
1975				goto must_alloc;
1976			}
1977		} else {
1978			*xmit_tail = (*xmit_tail)->b_cont;
1979			ASSERT((uintptr_t)((*xmit_tail)->b_wptr -
1980			    (*xmit_tail)->b_rptr) <= (uintptr_t)INT_MAX);
1981			*tail_unsent = (int)((*xmit_tail)->b_wptr -
1982			    (*xmit_tail)->b_rptr);
1983		}
1984
1985		(*xmit_tail)->b_prev = local_time;
1986		(*xmit_tail)->b_next = (mblk_t *)(uintptr_t)(*snxt - len);
1987
1988		*tail_unsent -= len;
1989		if (len <= mss) /* LSO is unusable (!do_lso_send) */
1990			tcp->tcp_last_sent_len = (ushort_t)len;
1991
1992		len += total_hdr_len;
1993		ixa->ixa_pktlen = len;
1994
1995		if (ixa->ixa_flags & IXAF_IS_IPV4) {
1996			tcp->tcp_ipha->ipha_length = htons(len);
1997		} else {
1998			tcp->tcp_ip6h->ip6_plen = htons(len - IPV6_HDR_LEN);
1999		}
2000
2001		mp = dupb(*xmit_tail);
2002		if (mp == NULL) {
2003			return (-1);	/* out_of_mem */
2004		}
2005
2006		len = total_hdr_len;
2007		/*
2008		 * There are four reasons to allocate a new hdr mblk:
2009		 *  1) The bytes above us are in use by another packet
2010		 *  2) We don't have good alignment
2011		 *  3) The mblk is being shared
2012		 *  4) We don't have enough room for a header
2013		 */
2014		rptr = mp->b_rptr - len;
2015		if (!OK_32PTR(rptr) ||
2016		    ((db = mp->b_datap), db->db_ref != 2) ||
2017		    rptr < db->db_base) {
2018			/* NOTE: we assume allocb returns an OK_32PTR */
2019
2020		must_alloc:;
2021			mp1 = allocb(connp->conn_ht_iphc_allocated +
2022			    tcps->tcps_wroff_xtra, BPRI_MED);
2023			if (mp1 == NULL) {
2024				freemsg(mp);
2025				return (-1);	/* out_of_mem */
2026			}
2027			mp1->b_cont = mp;
2028			mp = mp1;
2029			/* Leave room for Link Level header */
2030			len = total_hdr_len;
2031			rptr = &mp->b_rptr[tcps->tcps_wroff_xtra];
2032			mp->b_wptr = &rptr[len];
2033		}
2034
2035		/*
2036		 * Fill in the header using the template header, and add
2037		 * options such as time-stamp, ECN and/or SACK, as needed.
2038		 */
2039		tcp_fill_header(tcp, rptr, (clock_t)local_time, num_sack_blk);
2040
2041		mp->b_rptr = rptr;
2042
2043		if (*tail_unsent) {
2044			int spill = *tail_unsent;
2045
2046			mp1 = mp->b_cont;
2047			if (mp1 == NULL)
2048				mp1 = mp;
2049
2050			/*
2051			 * If we're a little short, tack on more mblks until
2052			 * there is no more spillover.
2053			 */
2054			while (spill < 0) {
2055				mblk_t *nmp;
2056				int nmpsz;
2057
2058				nmp = (*xmit_tail)->b_cont;
2059				nmpsz = MBLKL(nmp);
2060
2061				/*
2062				 * Excess data in mblk; can we split it?
2063				 * If LSO is enabled for the connection,
2064				 * keep on splitting as this is a transient
2065				 * send path.
2066				 */
2067				if (!do_lso_send && (spill + nmpsz > 0)) {
2068					/*
2069					 * Don't split if stream head was
2070					 * told to break up larger writes
2071					 * into smaller ones.
2072					 */
2073					if (tcp->tcp_maxpsz_multiplier > 0)
2074						break;
2075
2076					/*
2077					 * Next mblk is less than SMSS/2
2078					 * rounded up to nearest 64-byte;
2079					 * let it get sent as part of the
2080					 * next segment.
2081					 */
2082					if (tcp->tcp_localnet &&
2083					    !tcp->tcp_cork &&
2084					    (nmpsz < roundup((mss >> 1), 64)))
2085						break;
2086				}
2087
2088				*xmit_tail = nmp;
2089				ASSERT((uintptr_t)nmpsz <= (uintptr_t)INT_MAX);
2090				/* Stash for rtt use later */
2091				(*xmit_tail)->b_prev = local_time;
2092				(*xmit_tail)->b_next =
2093				    (mblk_t *)(uintptr_t)(*snxt - len);
2094				mp1->b_cont = dupb(*xmit_tail);
2095				mp1 = mp1->b_cont;
2096
2097				spill += nmpsz;
2098				if (mp1 == NULL) {
2099					*tail_unsent = spill;
2100					freemsg(mp);
2101					return (-1);	/* out_of_mem */
2102				}
2103			}
2104
2105			/* Trim back any surplus on the last mblk */
2106			if (spill >= 0) {
2107				mp1->b_wptr -= spill;
2108				*tail_unsent = spill;
2109			} else {
2110				/*
2111				 * We did not send everything we could in
2112				 * order to remain within the b_cont limit.
2113				 */
2114				*usable -= spill;
2115				*snxt += spill;
2116				tcp->tcp_last_sent_len += spill;
2117				TCPS_UPDATE_MIB(tcps, tcpOutDataBytes, spill);
2118				/*
2119				 * Adjust the checksum
2120				 */
2121				tcpha = (tcpha_t *)(rptr +
2122				    ixa->ixa_ip_hdr_length);
2123				sum += spill;
2124				sum = (sum >> 16) + (sum & 0xFFFF);
2125				tcpha->tha_sum = htons(sum);
2126				if (connp->conn_ipversion == IPV4_VERSION) {
2127					sum = ntohs(
2128					    ((ipha_t *)rptr)->ipha_length) +
2129					    spill;
2130					((ipha_t *)rptr)->ipha_length =
2131					    htons(sum);
2132				} else {
2133					sum = ntohs(
2134					    ((ip6_t *)rptr)->ip6_plen) +
2135					    spill;
2136					((ip6_t *)rptr)->ip6_plen =
2137					    htons(sum);
2138				}
2139				ixa->ixa_pktlen += spill;
2140				*tail_unsent = 0;
2141			}
2142		}
2143		if (tcp->tcp_ip_forward_progress) {
2144			tcp->tcp_ip_forward_progress = B_FALSE;
2145			ixa->ixa_flags |= IXAF_REACH_CONF;
2146		} else {
2147			ixa->ixa_flags &= ~IXAF_REACH_CONF;
2148		}
2149
2150		if (do_lso_send) {
2151			/* Append LSO information to the mp. */
2152			lso_info_set(mp, mss, HW_LSO);
2153			ixa->ixa_fragsize = IP_MAXPACKET;
2154			ixa->ixa_extra_ident = num_lso_seg - 1;
2155
2156			DTRACE_PROBE2(tcp_send_lso, int, num_lso_seg,
2157			    boolean_t, B_TRUE);
2158
2159			tcp_send_data(tcp, mp);
2160
2161			/*
2162			 * Restore values of ixa_fragsize and ixa_extra_ident.
2163			 */
2164			ixa->ixa_fragsize = ixa->ixa_pmtu;
2165			ixa->ixa_extra_ident = 0;
2166			tcp->tcp_obsegs += num_lso_seg;
2167			TCP_STAT(tcps, tcp_lso_times);
2168			TCP_STAT_UPDATE(tcps, tcp_lso_pkt_out, num_lso_seg);
2169		} else {
2170			/*
2171			 * Make sure to clean up LSO information. Wherever a
2172			 * new mp uses the prepended header room after dupb(),
2173			 * lso_info_cleanup() should be called.
2174			 */
2175			lso_info_cleanup(mp);
2176			tcp_send_data(tcp, mp);
2177			BUMP_LOCAL(tcp->tcp_obsegs);
2178		}
2179	}
2180
2181	return (0);
2182}
2183
2184/*
2185 * Initiate closedown sequence on an active connection.  (May be called as
2186 * writer.)  Return value zero for OK return, non-zero for error return.
2187 */
2188static int
2189tcp_xmit_end(tcp_t *tcp)
2190{
2191	mblk_t		*mp;
2192	tcp_stack_t	*tcps = tcp->tcp_tcps;
2193	iulp_t		uinfo;
2194	ip_stack_t	*ipst = tcps->tcps_netstack->netstack_ip;
2195	conn_t		*connp = tcp->tcp_connp;
2196
2197	if (tcp->tcp_state < TCPS_SYN_RCVD ||
2198	    tcp->tcp_state > TCPS_CLOSE_WAIT) {
2199		/*
2200		 * Invalid state, only states TCPS_SYN_RCVD,
2201		 * TCPS_ESTABLISHED and TCPS_CLOSE_WAIT are valid
2202		 */
2203		return (-1);
2204	}
2205
2206	tcp->tcp_fss = tcp->tcp_snxt + tcp->tcp_unsent;
2207	tcp->tcp_valid_bits |= TCP_FSS_VALID;
2208	/*
2209	 * If there is nothing more unsent, send the FIN now.
2210	 * Otherwise, it will go out with the last segment.
2211	 */
2212	if (tcp->tcp_unsent == 0) {
2213		mp = tcp_xmit_mp(tcp, NULL, 0, NULL, NULL,
2214		    tcp->tcp_fss, B_FALSE, NULL, B_FALSE);
2215
2216		if (mp) {
2217			tcp_send_data(tcp, mp);
2218		} else {
2219			/*
2220			 * Couldn't allocate msg.  Pretend we got it out.
2221			 * Wait for rexmit timeout.
2222			 */
2223			tcp->tcp_snxt = tcp->tcp_fss + 1;
2224			TCP_TIMER_RESTART(tcp, tcp->tcp_rto);
2225		}
2226
2227		/*
2228		 * If needed, update tcp_rexmit_snxt as tcp_snxt is
2229		 * changed.
2230		 */
2231		if (tcp->tcp_rexmit && tcp->tcp_rexmit_nxt == tcp->tcp_fss) {
2232			tcp->tcp_rexmit_nxt = tcp->tcp_snxt;
2233		}
2234	} else {
2235		/*
2236		 * If tcp->tcp_cork is set, then the data will not get sent,
2237		 * so we have to check that and unset it first.
2238		 */
2239		if (tcp->tcp_cork)
2240			tcp->tcp_cork = B_FALSE;
2241		tcp_wput_data(tcp, NULL, B_FALSE);
2242	}
2243
2244	/*
2245	 * If TCP does not get enough samples of RTT or tcp_rtt_updates
2246	 * is 0, don't update the cache.
2247	 */
2248	if (tcps->tcps_rtt_updates == 0 ||
2249	    tcp->tcp_rtt_update < tcps->tcps_rtt_updates)
2250		return (0);
2251
2252	/*
2253	 * We do not have a good algorithm to update ssthresh at this time.
2254	 * So don't do any update.
2255	 */
2256	bzero(&uinfo, sizeof (uinfo));
2257	uinfo.iulp_rtt = tcp->tcp_rtt_sa;
2258	uinfo.iulp_rtt_sd = tcp->tcp_rtt_sd;
2259
2260	/*
2261	 * Note that uinfo is kept for conn_faddr in the DCE. Could update even
2262	 * if source routed but we don't.
2263	 */
2264	if (connp->conn_ipversion == IPV4_VERSION) {
2265		if (connp->conn_faddr_v4 !=  tcp->tcp_ipha->ipha_dst) {
2266			return (0);
2267		}
2268		(void) dce_update_uinfo_v4(connp->conn_faddr_v4, &uinfo, ipst);
2269	} else {
2270		uint_t ifindex;
2271
2272		if (!(IN6_ARE_ADDR_EQUAL(&connp->conn_faddr_v6,
2273		    &tcp->tcp_ip6h->ip6_dst))) {
2274			return (0);
2275		}
2276		ifindex = 0;
2277		if (IN6_IS_ADDR_LINKSCOPE(&connp->conn_faddr_v6)) {
2278			ip_xmit_attr_t *ixa = connp->conn_ixa;
2279
2280			/*
2281			 * If we are going to create a DCE we'd better have
2282			 * an ifindex
2283			 */
2284			if (ixa->ixa_nce != NULL) {
2285				ifindex = ixa->ixa_nce->nce_common->ncec_ill->
2286				    ill_phyint->phyint_ifindex;
2287			} else {
2288				return (0);
2289			}
2290		}
2291
2292		(void) dce_update_uinfo(&connp->conn_faddr_v6, ifindex, &uinfo,
2293		    ipst);
2294	}
2295	return (0);
2296}
2297
2298/*
2299 * Send out a control packet on the tcp connection specified.  This routine
2300 * is typically called where we need a simple ACK or RST generated.
2301 */
2302void
2303tcp_xmit_ctl(char *str, tcp_t *tcp, uint32_t seq, uint32_t ack, int ctl)
2304{
2305	uchar_t		*rptr;
2306	tcpha_t		*tcpha;
2307	ipha_t		*ipha = NULL;
2308	ip6_t		*ip6h = NULL;
2309	uint32_t	sum;
2310	int		total_hdr_len;
2311	int		ip_hdr_len;
2312	mblk_t		*mp;
2313	tcp_stack_t	*tcps = tcp->tcp_tcps;
2314	conn_t		*connp = tcp->tcp_connp;
2315	ip_xmit_attr_t	*ixa = connp->conn_ixa;
2316
2317	/*
2318	 * Save sum for use in source route later.
2319	 */
2320	sum = connp->conn_ht_ulp_len + connp->conn_sum;
2321	total_hdr_len = connp->conn_ht_iphc_len;
2322	ip_hdr_len = ixa->ixa_ip_hdr_length;
2323
2324	/* If a text string is passed in with the request, pass it to strlog. */
2325	if (str != NULL && connp->conn_debug) {
2326		(void) strlog(TCP_MOD_ID, 0, 1, SL_TRACE,
2327		    "tcp_xmit_ctl: '%s', seq 0x%x, ack 0x%x, ctl 0x%x",
2328		    str, seq, ack, ctl);
2329	}
2330	mp = allocb(connp->conn_ht_iphc_allocated + tcps->tcps_wroff_xtra,
2331	    BPRI_MED);
2332	if (mp == NULL) {
2333		return;
2334	}
2335	rptr = &mp->b_rptr[tcps->tcps_wroff_xtra];
2336	mp->b_rptr = rptr;
2337	mp->b_wptr = &rptr[total_hdr_len];
2338	bcopy(connp->conn_ht_iphc, rptr, total_hdr_len);
2339
2340	ixa->ixa_pktlen = total_hdr_len;
2341
2342	if (ixa->ixa_flags & IXAF_IS_IPV4) {
2343		ipha = (ipha_t *)rptr;
2344		ipha->ipha_length = htons(total_hdr_len);
2345	} else {
2346		ip6h = (ip6_t *)rptr;
2347		ip6h->ip6_plen = htons(total_hdr_len - IPV6_HDR_LEN);
2348	}
2349	tcpha = (tcpha_t *)&rptr[ip_hdr_len];
2350	tcpha->tha_flags = (uint8_t)ctl;
2351	if (ctl & TH_RST) {
2352		TCPS_BUMP_MIB(tcps, tcpOutRsts);
2353		TCPS_BUMP_MIB(tcps, tcpOutControl);
2354		/*
2355		 * Don't send TSopt w/ TH_RST packets per RFC 1323.
2356		 */
2357		if (tcp->tcp_snd_ts_ok &&
2358		    tcp->tcp_state > TCPS_SYN_SENT) {
2359			mp->b_wptr = &rptr[total_hdr_len - TCPOPT_REAL_TS_LEN];
2360			*(mp->b_wptr) = TCPOPT_EOL;
2361
2362			ixa->ixa_pktlen = total_hdr_len - TCPOPT_REAL_TS_LEN;
2363
2364			if (connp->conn_ipversion == IPV4_VERSION) {
2365				ipha->ipha_length = htons(total_hdr_len -
2366				    TCPOPT_REAL_TS_LEN);
2367			} else {
2368				ip6h->ip6_plen = htons(total_hdr_len -
2369				    IPV6_HDR_LEN - TCPOPT_REAL_TS_LEN);
2370			}
2371			tcpha->tha_offset_and_reserved -= (3 << 4);
2372			sum -= TCPOPT_REAL_TS_LEN;
2373		}
2374	}
2375	if (ctl & TH_ACK) {
2376		if (tcp->tcp_snd_ts_ok) {
2377			uint32_t llbolt = (uint32_t)LBOLT_FASTPATH;
2378
2379			U32_TO_BE32(llbolt,
2380			    (char *)tcpha + TCP_MIN_HEADER_LENGTH+4);
2381			U32_TO_BE32(tcp->tcp_ts_recent,
2382			    (char *)tcpha + TCP_MIN_HEADER_LENGTH+8);
2383		}
2384
2385		/* Update the latest receive window size in TCP header. */
2386		tcpha->tha_win = htons(tcp->tcp_rwnd >> tcp->tcp_rcv_ws);
2387		/* Track what we sent to the peer */
2388		tcp->tcp_tcpha->tha_win = tcpha->tha_win;
2389		tcp->tcp_rack = ack;
2390		tcp->tcp_rack_cnt = 0;
2391		TCPS_BUMP_MIB(tcps, tcpOutAck);
2392	}
2393	BUMP_LOCAL(tcp->tcp_obsegs);
2394	tcpha->tha_seq = htonl(seq);
2395	tcpha->tha_ack = htonl(ack);
2396	/*
2397	 * Include the adjustment for a source route if any.
2398	 */
2399	sum = (sum >> 16) + (sum & 0xFFFF);
2400	tcpha->tha_sum = htons(sum);
2401	tcp_send_data(tcp, mp);
2402}
2403
2404/*
2405 * Generate a reset based on an inbound packet, connp is set by caller
2406 * when RST is in response to an unexpected inbound packet for which
2407 * there is active tcp state in the system.
2408 *
2409 * IPSEC NOTE : Try to send the reply with the same protection as it came
2410 * in.  We have the ip_recv_attr_t which is reversed to form the ip_xmit_attr_t.
2411 * That way the packet will go out at the same level of protection as it
2412 * came in with.
2413 */
2414static void
2415tcp_xmit_early_reset(char *str, mblk_t *mp, uint32_t seq, uint32_t ack, int ctl,
2416    ip_recv_attr_t *ira, ip_stack_t *ipst, conn_t *connp)
2417{
2418	ipha_t		*ipha = NULL;
2419	ip6_t		*ip6h = NULL;
2420	ushort_t	len;
2421	tcpha_t		*tcpha;
2422	int		i;
2423	ipaddr_t	v4addr;
2424	in6_addr_t	v6addr;
2425	netstack_t	*ns = ipst->ips_netstack;
2426	tcp_stack_t	*tcps = ns->netstack_tcp;
2427	ip_xmit_attr_t	ixas, *ixa;
2428	uint_t		ip_hdr_len = ira->ira_ip_hdr_length;
2429	boolean_t	need_refrele = B_FALSE;		/* ixa_refrele(ixa) */
2430	ushort_t	port;
2431
2432	if (!tcp_send_rst_chk(tcps)) {
2433		TCP_STAT(tcps, tcp_rst_unsent);
2434		freemsg(mp);
2435		return;
2436	}
2437
2438	/*
2439	 * If connp != NULL we use conn_ixa to keep IP_NEXTHOP and other
2440	 * options from the listener. In that case the caller must ensure that
2441	 * we are running on the listener = connp squeue.
2442	 *
2443	 * We get a safe copy of conn_ixa so we don't need to restore anything
2444	 * we or ip_output_simple might change in the ixa.
2445	 */
2446	if (connp != NULL) {
2447		ASSERT(connp->conn_on_sqp);
2448
2449		ixa = conn_get_ixa_exclusive(connp);
2450		if (ixa == NULL) {
2451			TCP_STAT(tcps, tcp_rst_unsent);
2452			freemsg(mp);
2453			return;
2454		}
2455		need_refrele = B_TRUE;
2456	} else {
2457		bzero(&ixas, sizeof (ixas));
2458		ixa = &ixas;
2459		/*
2460		 * IXAF_VERIFY_SOURCE is overkill since we know the
2461		 * packet was for us.
2462		 */
2463		ixa->ixa_flags |= IXAF_SET_ULP_CKSUM | IXAF_VERIFY_SOURCE;
2464		ixa->ixa_protocol = IPPROTO_TCP;
2465		ixa->ixa_zoneid = ira->ira_zoneid;
2466		ixa->ixa_ifindex = 0;
2467		ixa->ixa_ipst = ipst;
2468		ixa->ixa_cred = kcred;
2469		ixa->ixa_cpid = NOPID;
2470	}
2471
2472	if (str && tcps->tcps_dbg) {
2473		(void) strlog(TCP_MOD_ID, 0, 1, SL_TRACE,
2474		    "tcp_xmit_early_reset: '%s', seq 0x%x, ack 0x%x, "
2475		    "flags 0x%x",
2476		    str, seq, ack, ctl);
2477	}
2478	if (mp->b_datap->db_ref != 1) {
2479		mblk_t *mp1 = copyb(mp);
2480		freemsg(mp);
2481		mp = mp1;
2482		if (mp == NULL)
2483			goto done;
2484	} else if (mp->b_cont) {
2485		freemsg(mp->b_cont);
2486		mp->b_cont = NULL;
2487		DB_CKSUMFLAGS(mp) = 0;
2488	}
2489	/*
2490	 * We skip reversing source route here.
2491	 * (for now we replace all IP options with EOL)
2492	 */
2493	if (IPH_HDR_VERSION(mp->b_rptr) == IPV4_VERSION) {
2494		ipha = (ipha_t *)mp->b_rptr;
2495		for (i = IP_SIMPLE_HDR_LENGTH; i < (int)ip_hdr_len; i++)
2496			mp->b_rptr[i] = IPOPT_EOL;
2497		/*
2498		 * Make sure that src address isn't flagrantly invalid.
2499		 * Not all broadcast address checking for the src address
2500		 * is possible, since we don't know the netmask of the src
2501		 * addr.  No check for destination address is done, since
2502		 * IP will not pass up a packet with a broadcast dest
2503		 * address to TCP.  Similar checks are done below for IPv6.
2504		 */
2505		if (ipha->ipha_src == 0 || ipha->ipha_src == INADDR_BROADCAST ||
2506		    CLASSD(ipha->ipha_src)) {
2507			BUMP_MIB(&ipst->ips_ip_mib, ipIfStatsInDiscards);
2508			ip_drop_input("ipIfStatsInDiscards", mp, NULL);
2509			freemsg(mp);
2510			goto done;
2511		}
2512	} else {
2513		ip6h = (ip6_t *)mp->b_rptr;
2514
2515		if (IN6_IS_ADDR_UNSPECIFIED(&ip6h->ip6_src) ||
2516		    IN6_IS_ADDR_MULTICAST(&ip6h->ip6_src)) {
2517			BUMP_MIB(&ipst->ips_ip6_mib, ipIfStatsInDiscards);
2518			ip_drop_input("ipIfStatsInDiscards", mp, NULL);
2519			freemsg(mp);
2520			goto done;
2521		}
2522
2523		/* Remove any extension headers assuming partial overlay */
2524		if (ip_hdr_len > IPV6_HDR_LEN) {
2525			uint8_t *to;
2526
2527			to = mp->b_rptr + ip_hdr_len - IPV6_HDR_LEN;
2528			ovbcopy(ip6h, to, IPV6_HDR_LEN);
2529			mp->b_rptr += ip_hdr_len - IPV6_HDR_LEN;
2530			ip_hdr_len = IPV6_HDR_LEN;
2531			ip6h = (ip6_t *)mp->b_rptr;
2532			ip6h->ip6_nxt = IPPROTO_TCP;
2533		}
2534	}
2535	tcpha = (tcpha_t *)&mp->b_rptr[ip_hdr_len];
2536	if (tcpha->tha_flags & TH_RST) {
2537		freemsg(mp);
2538		goto done;
2539	}
2540	tcpha->tha_offset_and_reserved = (5 << 4);
2541	len = ip_hdr_len + sizeof (tcpha_t);
2542	mp->b_wptr = &mp->b_rptr[len];
2543	if (IPH_HDR_VERSION(mp->b_rptr) == IPV4_VERSION) {
2544		ipha->ipha_length = htons(len);
2545		/* Swap addresses */
2546		v4addr = ipha->ipha_src;
2547		ipha->ipha_src = ipha->ipha_dst;
2548		ipha->ipha_dst = v4addr;
2549		ipha->ipha_ident = 0;
2550		ipha->ipha_ttl = (uchar_t)tcps->tcps_ipv4_ttl;
2551		ixa->ixa_flags |= IXAF_IS_IPV4;
2552		ixa->ixa_ip_hdr_length = ip_hdr_len;
2553	} else {
2554		ip6h->ip6_plen = htons(len - IPV6_HDR_LEN);
2555		/* Swap addresses */
2556		v6addr = ip6h->ip6_src;
2557		ip6h->ip6_src = ip6h->ip6_dst;
2558		ip6h->ip6_dst = v6addr;
2559		ip6h->ip6_hops = (uchar_t)tcps->tcps_ipv6_hoplimit;
2560		ixa->ixa_flags &= ~IXAF_IS_IPV4;
2561
2562		if (IN6_IS_ADDR_LINKSCOPE(&ip6h->ip6_dst)) {
2563			ixa->ixa_flags |= IXAF_SCOPEID_SET;
2564			ixa->ixa_scopeid = ira->ira_ruifindex;
2565		}
2566		ixa->ixa_ip_hdr_length = IPV6_HDR_LEN;
2567	}
2568	ixa->ixa_pktlen = len;
2569
2570	/* Swap the ports */
2571	port = tcpha->tha_fport;
2572	tcpha->tha_fport = tcpha->tha_lport;
2573	tcpha->tha_lport = port;
2574
2575	tcpha->tha_ack = htonl(ack);
2576	tcpha->tha_seq = htonl(seq);
2577	tcpha->tha_win = 0;
2578	tcpha->tha_sum = htons(sizeof (tcpha_t));
2579	tcpha->tha_flags = (uint8_t)ctl;
2580	if (ctl & TH_RST) {
2581		TCPS_BUMP_MIB(tcps, tcpOutRsts);
2582		TCPS_BUMP_MIB(tcps, tcpOutControl);
2583	}
2584
2585	/* Discard any old label */
2586	if (ixa->ixa_free_flags & IXA_FREE_TSL) {
2587		ASSERT(ixa->ixa_tsl != NULL);
2588		label_rele(ixa->ixa_tsl);
2589		ixa->ixa_free_flags &= ~IXA_FREE_TSL;
2590	}
2591	ixa->ixa_tsl = ira->ira_tsl;	/* Behave as a multi-level responder */
2592
2593	if (ira->ira_flags & IRAF_IPSEC_SECURE) {
2594		/*
2595		 * Apply IPsec based on how IPsec was applied to
2596		 * the packet that caused the RST.
2597		 */
2598		if (!ipsec_in_to_out(ira, ixa, mp, ipha, ip6h)) {
2599			BUMP_MIB(&ipst->ips_ip_mib, ipIfStatsOutDiscards);
2600			/* Note: mp already consumed and ip_drop_packet done */
2601			goto done;
2602		}
2603	} else {
2604		/*
2605		 * This is in clear. The RST message we are building
2606		 * here should go out in clear, independent of our policy.
2607		 */
2608		ixa->ixa_flags |= IXAF_NO_IPSEC;
2609	}
2610
2611	/*
2612	 * NOTE:  one might consider tracing a TCP packet here, but
2613	 * this function has no active TCP state and no tcp structure
2614	 * that has a trace buffer.  If we traced here, we would have
2615	 * to keep a local trace buffer in tcp_record_trace().
2616	 */
2617
2618	(void) ip_output_simple(mp, ixa);
2619done:
2620	ixa_cleanup(ixa);
2621	if (need_refrele) {
2622		ASSERT(ixa != &ixas);
2623		ixa_refrele(ixa);
2624	}
2625}
2626
2627/*
2628 * Generate a "no listener here" RST in response to an "unknown" segment.
2629 * connp is set by caller when RST is in response to an unexpected
2630 * inbound packet for which there is active tcp state in the system.
2631 * Note that we are reusing the incoming mp to construct the outgoing RST.
2632 */
2633void
2634tcp_xmit_listeners_reset(mblk_t *mp, ip_recv_attr_t *ira, ip_stack_t *ipst,
2635    conn_t *connp)
2636{
2637	uchar_t		*rptr;
2638	uint32_t	seg_len;
2639	tcpha_t		*tcpha;
2640	uint32_t	seg_seq;
2641	uint32_t	seg_ack;
2642	uint_t		flags;
2643	ipha_t 		*ipha;
2644	ip6_t 		*ip6h;
2645	boolean_t	policy_present;
2646	netstack_t	*ns = ipst->ips_netstack;
2647	tcp_stack_t	*tcps = ns->netstack_tcp;
2648	ipsec_stack_t	*ipss = tcps->tcps_netstack->netstack_ipsec;
2649	uint_t		ip_hdr_len = ira->ira_ip_hdr_length;
2650
2651	TCP_STAT(tcps, tcp_no_listener);
2652
2653	if (IPH_HDR_VERSION(mp->b_rptr) == IPV4_VERSION) {
2654		policy_present = ipss->ipsec_inbound_v4_policy_present;
2655		ipha = (ipha_t *)mp->b_rptr;
2656		ip6h = NULL;
2657	} else {
2658		policy_present = ipss->ipsec_inbound_v6_policy_present;
2659		ipha = NULL;
2660		ip6h = (ip6_t *)mp->b_rptr;
2661	}
2662
2663	if (policy_present) {
2664		/*
2665		 * The conn_t parameter is NULL because we already know
2666		 * nobody's home.
2667		 */
2668		mp = ipsec_check_global_policy(mp, (conn_t *)NULL, ipha, ip6h,
2669		    ira, ns);
2670		if (mp == NULL)
2671			return;
2672	}
2673	if (is_system_labeled() && !tsol_can_reply_error(mp, ira)) {
2674		DTRACE_PROBE2(
2675		    tx__ip__log__error__nolistener__tcp,
2676		    char *, "Could not reply with RST to mp(1)",
2677		    mblk_t *, mp);
2678		ip2dbg(("tcp_xmit_listeners_reset: not permitted to reply\n"));
2679		freemsg(mp);
2680		return;
2681	}
2682
2683	rptr = mp->b_rptr;
2684
2685	tcpha = (tcpha_t *)&rptr[ip_hdr_len];
2686	seg_seq = ntohl(tcpha->tha_seq);
2687	seg_ack = ntohl(tcpha->tha_ack);
2688	flags = tcpha->tha_flags;
2689
2690	seg_len = msgdsize(mp) - (TCP_HDR_LENGTH(tcpha) + ip_hdr_len);
2691	if (flags & TH_RST) {
2692		freemsg(mp);
2693	} else if (flags & TH_ACK) {
2694		tcp_xmit_early_reset("no tcp, reset", mp, seg_ack, 0, TH_RST,
2695		    ira, ipst, connp);
2696	} else {
2697		if (flags & TH_SYN) {
2698			seg_len++;
2699		} else {
2700			/*
2701			 * Here we violate the RFC.  Note that a normal
2702			 * TCP will never send a segment without the ACK
2703			 * flag, except for RST or SYN segment.  This
2704			 * segment is neither.  Just drop it on the
2705			 * floor.
2706			 */
2707			freemsg(mp);
2708			TCP_STAT(tcps, tcp_rst_unsent);
2709			return;
2710		}
2711
2712		tcp_xmit_early_reset("no tcp, reset/ack", mp, 0,
2713		    seg_seq + seg_len, TH_RST | TH_ACK, ira, ipst, connp);
2714	}
2715}
2716
2717/*
2718 * tcp_xmit_mp is called to return a pointer to an mblk chain complete with
2719 * ip and tcp header ready to pass down to IP.  If the mp passed in is
2720 * non-NULL, then up to max_to_send bytes of data will be dup'ed off that
2721 * mblk. (If sendall is not set the dup'ing will stop at an mblk boundary
2722 * otherwise it will dup partial mblks.)
2723 * Otherwise, an appropriate ACK packet will be generated.  This
2724 * routine is not usually called to send new data for the first time.  It
2725 * is mostly called out of the timer for retransmits, and to generate ACKs.
2726 *
2727 * If offset is not NULL, the returned mblk chain's first mblk's b_rptr will
2728 * be adjusted by *offset.  And after dupb(), the offset and the ending mblk
2729 * of the original mblk chain will be returned in *offset and *end_mp.
2730 */
2731mblk_t *
2732tcp_xmit_mp(tcp_t *tcp, mblk_t *mp, int32_t max_to_send, int32_t *offset,
2733    mblk_t **end_mp, uint32_t seq, boolean_t sendall, uint32_t *seg_len,
2734    boolean_t rexmit)
2735{
2736	int	data_length;
2737	int32_t	off = 0;
2738	uint_t	flags;
2739	mblk_t	*mp1;
2740	mblk_t	*mp2;
2741	uchar_t	*rptr;
2742	tcpha_t	*tcpha;
2743	int32_t	num_sack_blk = 0;
2744	int32_t	sack_opt_len = 0;
2745	tcp_stack_t	*tcps = tcp->tcp_tcps;
2746	conn_t		*connp = tcp->tcp_connp;
2747	ip_xmit_attr_t	*ixa = connp->conn_ixa;
2748
2749	/* Allocate for our maximum TCP header + link-level */
2750	mp1 = allocb(connp->conn_ht_iphc_allocated + tcps->tcps_wroff_xtra,
2751	    BPRI_MED);
2752	if (!mp1)
2753		return (NULL);
2754	data_length = 0;
2755
2756	/*
2757	 * Note that tcp_mss has been adjusted to take into account the
2758	 * timestamp option if applicable.  Because SACK options do not
2759	 * appear in every TCP segments and they are of variable lengths,
2760	 * they cannot be included in tcp_mss.  Thus we need to calculate
2761	 * the actual segment length when we need to send a segment which
2762	 * includes SACK options.
2763	 */
2764	if (tcp->tcp_snd_sack_ok && tcp->tcp_num_sack_blk > 0) {
2765		num_sack_blk = MIN(tcp->tcp_max_sack_blk,
2766		    tcp->tcp_num_sack_blk);
2767		sack_opt_len = num_sack_blk * sizeof (sack_blk_t) +
2768		    TCPOPT_NOP_LEN * 2 + TCPOPT_HEADER_LEN;
2769		if (max_to_send + sack_opt_len > tcp->tcp_mss)
2770			max_to_send -= sack_opt_len;
2771	}
2772
2773	if (offset != NULL) {
2774		off = *offset;
2775		/* We use offset as an indicator that end_mp is not NULL. */
2776		*end_mp = NULL;
2777	}
2778	for (mp2 = mp1; mp && data_length != max_to_send; mp = mp->b_cont) {
2779		/* This could be faster with cooperation from downstream */
2780		if (mp2 != mp1 && !sendall &&
2781		    data_length + (int)(mp->b_wptr - mp->b_rptr) >
2782		    max_to_send)
2783			/*
2784			 * Don't send the next mblk since the whole mblk
2785			 * does not fit.
2786			 */
2787			break;
2788		mp2->b_cont = dupb(mp);
2789		mp2 = mp2->b_cont;
2790		if (!mp2) {
2791			freemsg(mp1);
2792			return (NULL);
2793		}
2794		mp2->b_rptr += off;
2795		ASSERT((uintptr_t)(mp2->b_wptr - mp2->b_rptr) <=
2796		    (uintptr_t)INT_MAX);
2797
2798		data_length += (int)(mp2->b_wptr - mp2->b_rptr);
2799		if (data_length > max_to_send) {
2800			mp2->b_wptr -= data_length - max_to_send;
2801			data_length = max_to_send;
2802			off = mp2->b_wptr - mp->b_rptr;
2803			break;
2804		} else {
2805			off = 0;
2806		}
2807	}
2808	if (offset != NULL) {
2809		*offset = off;
2810		*end_mp = mp;
2811	}
2812	if (seg_len != NULL) {
2813		*seg_len = data_length;
2814	}
2815
2816	/* Update the latest receive window size in TCP header. */
2817	tcp->tcp_tcpha->tha_win = htons(tcp->tcp_rwnd >> tcp->tcp_rcv_ws);
2818
2819	rptr = mp1->b_rptr + tcps->tcps_wroff_xtra;
2820	mp1->b_rptr = rptr;
2821	mp1->b_wptr = rptr + connp->conn_ht_iphc_len + sack_opt_len;
2822	bcopy(connp->conn_ht_iphc, rptr, connp->conn_ht_iphc_len);
2823	tcpha = (tcpha_t *)&rptr[ixa->ixa_ip_hdr_length];
2824	tcpha->tha_seq = htonl(seq);
2825
2826	/*
2827	 * Use tcp_unsent to determine if the PUSH bit should be used assumes
2828	 * that this function was called from tcp_wput_data. Thus, when called
2829	 * to retransmit data the setting of the PUSH bit may appear some
2830	 * what random in that it might get set when it should not. This
2831	 * should not pose any performance issues.
2832	 */
2833	if (data_length != 0 && (tcp->tcp_unsent == 0 ||
2834	    tcp->tcp_unsent == data_length)) {
2835		flags = TH_ACK | TH_PUSH;
2836	} else {
2837		flags = TH_ACK;
2838	}
2839
2840	if (tcp->tcp_ecn_ok) {
2841		if (tcp->tcp_ecn_echo_on)
2842			flags |= TH_ECE;
2843
2844		/*
2845		 * Only set ECT bit and ECN_CWR if a segment contains new data.
2846		 * There is no TCP flow control for non-data segments, and
2847		 * only data segment is transmitted reliably.
2848		 */
2849		if (data_length > 0 && !rexmit) {
2850			TCP_SET_ECT(tcp, rptr);
2851			if (tcp->tcp_cwr && !tcp->tcp_ecn_cwr_sent) {
2852				flags |= TH_CWR;
2853				tcp->tcp_ecn_cwr_sent = B_TRUE;
2854			}
2855		}
2856	}
2857
2858	if (tcp->tcp_valid_bits) {
2859		uint32_t u1;
2860
2861		if ((tcp->tcp_valid_bits & TCP_ISS_VALID) &&
2862		    seq == tcp->tcp_iss) {
2863			uchar_t	*wptr;
2864
2865			/*
2866			 * If TCP_ISS_VALID and the seq number is tcp_iss,
2867			 * TCP can only be in SYN-SENT, SYN-RCVD or
2868			 * FIN-WAIT-1 state.  It can be FIN-WAIT-1 if
2869			 * our SYN is not ack'ed but the app closes this
2870			 * TCP connection.
2871			 */
2872			ASSERT(tcp->tcp_state == TCPS_SYN_SENT ||
2873			    tcp->tcp_state == TCPS_SYN_RCVD ||
2874			    tcp->tcp_state == TCPS_FIN_WAIT_1);
2875
2876			/*
2877			 * Tack on the MSS option.  It is always needed
2878			 * for both active and passive open.
2879			 *
2880			 * MSS option value should be interface MTU - MIN
2881			 * TCP/IP header according to RFC 793 as it means
2882			 * the maximum segment size TCP can receive.  But
2883			 * to get around some broken middle boxes/end hosts
2884			 * out there, we allow the option value to be the
2885			 * same as the MSS option size on the peer side.
2886			 * In this way, the other side will not send
2887			 * anything larger than they can receive.
2888			 *
2889			 * Note that for SYN_SENT state, the ndd param
2890			 * tcp_use_smss_as_mss_opt has no effect as we
2891			 * don't know the peer's MSS option value. So
2892			 * the only case we need to take care of is in
2893			 * SYN_RCVD state, which is done later.
2894			 */
2895			wptr = mp1->b_wptr;
2896			wptr[0] = TCPOPT_MAXSEG;
2897			wptr[1] = TCPOPT_MAXSEG_LEN;
2898			wptr += 2;
2899			u1 = tcp->tcp_initial_pmtu -
2900			    (connp->conn_ipversion == IPV4_VERSION ?
2901			    IP_SIMPLE_HDR_LENGTH : IPV6_HDR_LEN) -
2902			    TCP_MIN_HEADER_LENGTH;
2903			U16_TO_BE16(u1, wptr);
2904			mp1->b_wptr = wptr + 2;
2905			/* Update the offset to cover the additional word */
2906			tcpha->tha_offset_and_reserved += (1 << 4);
2907
2908			/*
2909			 * Note that the following way of filling in
2910			 * TCP options are not optimal.  Some NOPs can
2911			 * be saved.  But there is no need at this time
2912			 * to optimize it.  When it is needed, we will
2913			 * do it.
2914			 */
2915			switch (tcp->tcp_state) {
2916			case TCPS_SYN_SENT:
2917				flags = TH_SYN;
2918
2919				if (tcp->tcp_snd_ts_ok) {
2920					uint32_t llbolt =
2921					    (uint32_t)LBOLT_FASTPATH;
2922
2923					wptr = mp1->b_wptr;
2924					wptr[0] = TCPOPT_NOP;
2925					wptr[1] = TCPOPT_NOP;
2926					wptr[2] = TCPOPT_TSTAMP;
2927					wptr[3] = TCPOPT_TSTAMP_LEN;
2928					wptr += 4;
2929					U32_TO_BE32(llbolt, wptr);
2930					wptr += 4;
2931					ASSERT(tcp->tcp_ts_recent == 0);
2932					U32_TO_BE32(0L, wptr);
2933					mp1->b_wptr += TCPOPT_REAL_TS_LEN;
2934					tcpha->tha_offset_and_reserved +=
2935					    (3 << 4);
2936				}
2937
2938				/*
2939				 * Set up all the bits to tell other side
2940				 * we are ECN capable.
2941				 */
2942				if (tcp->tcp_ecn_ok) {
2943					flags |= (TH_ECE | TH_CWR);
2944				}
2945				break;
2946			case TCPS_SYN_RCVD:
2947				flags |= TH_SYN;
2948
2949				/*
2950				 * Reset the MSS option value to be SMSS
2951				 * We should probably add back the bytes
2952				 * for timestamp option and IPsec.  We
2953				 * don't do that as this is a workaround
2954				 * for broken middle boxes/end hosts, it
2955				 * is better for us to be more cautious.
2956				 * They may not take these things into
2957				 * account in their SMSS calculation.  Thus
2958				 * the peer's calculated SMSS may be smaller
2959				 * than what it can be.  This should be OK.
2960				 */
2961				if (tcps->tcps_use_smss_as_mss_opt) {
2962					u1 = tcp->tcp_mss;
2963					U16_TO_BE16(u1, wptr);
2964				}
2965
2966				/*
2967				 * If the other side is ECN capable, reply
2968				 * that we are also ECN capable.
2969				 */
2970				if (tcp->tcp_ecn_ok)
2971					flags |= TH_ECE;
2972				break;
2973			default:
2974				/*
2975				 * The above ASSERT() makes sure that this
2976				 * must be FIN-WAIT-1 state.  Our SYN has
2977				 * not been ack'ed so retransmit it.
2978				 */
2979				flags |= TH_SYN;
2980				break;
2981			}
2982
2983			if (tcp->tcp_snd_ws_ok) {
2984				wptr = mp1->b_wptr;
2985				wptr[0] =  TCPOPT_NOP;
2986				wptr[1] =  TCPOPT_WSCALE;
2987				wptr[2] =  TCPOPT_WS_LEN;
2988				wptr[3] = (uchar_t)tcp->tcp_rcv_ws;
2989				mp1->b_wptr += TCPOPT_REAL_WS_LEN;
2990				tcpha->tha_offset_and_reserved += (1 << 4);
2991			}
2992
2993			if (tcp->tcp_snd_sack_ok) {
2994				wptr = mp1->b_wptr;
2995				wptr[0] = TCPOPT_NOP;
2996				wptr[1] = TCPOPT_NOP;
2997				wptr[2] = TCPOPT_SACK_PERMITTED;
2998				wptr[3] = TCPOPT_SACK_OK_LEN;
2999				mp1->b_wptr += TCPOPT_REAL_SACK_OK_LEN;
3000				tcpha->tha_offset_and_reserved += (1 << 4);
3001			}
3002
3003			/* allocb() of adequate mblk assures space */
3004			ASSERT((uintptr_t)(mp1->b_wptr - mp1->b_rptr) <=
3005			    (uintptr_t)INT_MAX);
3006			u1 = (int)(mp1->b_wptr - mp1->b_rptr);
3007			/*
3008			 * Get IP set to checksum on our behalf
3009			 * Include the adjustment for a source route if any.
3010			 */
3011			u1 += connp->conn_sum;
3012			u1 = (u1 >> 16) + (u1 & 0xFFFF);
3013			tcpha->tha_sum = htons(u1);
3014			TCPS_BUMP_MIB(tcps, tcpOutControl);
3015		}
3016		if ((tcp->tcp_valid_bits & TCP_FSS_VALID) &&
3017		    (seq + data_length) == tcp->tcp_fss) {
3018			if (!tcp->tcp_fin_acked) {
3019				flags |= TH_FIN;
3020				TCPS_BUMP_MIB(tcps, tcpOutControl);
3021			}
3022			if (!tcp->tcp_fin_sent) {
3023				tcp->tcp_fin_sent = B_TRUE;
3024				switch (tcp->tcp_state) {
3025				case TCPS_SYN_RCVD:
3026				case TCPS_ESTABLISHED:
3027					tcp->tcp_state = TCPS_FIN_WAIT_1;
3028					break;
3029				case TCPS_CLOSE_WAIT:
3030					tcp->tcp_state = TCPS_LAST_ACK;
3031					break;
3032				}
3033				if (tcp->tcp_suna == tcp->tcp_snxt)
3034					TCP_TIMER_RESTART(tcp, tcp->tcp_rto);
3035				tcp->tcp_snxt = tcp->tcp_fss + 1;
3036			}
3037		}
3038		/*
3039		 * Note the trick here.  u1 is unsigned.  When tcp_urg
3040		 * is smaller than seq, u1 will become a very huge value.
3041		 * So the comparison will fail.  Also note that tcp_urp
3042		 * should be positive, see RFC 793 page 17.
3043		 */
3044		u1 = tcp->tcp_urg - seq + TCP_OLD_URP_INTERPRETATION;
3045		if ((tcp->tcp_valid_bits & TCP_URG_VALID) && u1 != 0 &&
3046		    u1 < (uint32_t)(64 * 1024)) {
3047			flags |= TH_URG;
3048			TCPS_BUMP_MIB(tcps, tcpOutUrg);
3049			tcpha->tha_urp = htons(u1);
3050		}
3051	}
3052	tcpha->tha_flags = (uchar_t)flags;
3053	tcp->tcp_rack = tcp->tcp_rnxt;
3054	tcp->tcp_rack_cnt = 0;
3055
3056	if (tcp->tcp_snd_ts_ok) {
3057		if (tcp->tcp_state != TCPS_SYN_SENT) {
3058			uint32_t llbolt = (uint32_t)LBOLT_FASTPATH;
3059
3060			U32_TO_BE32(llbolt,
3061			    (char *)tcpha + TCP_MIN_HEADER_LENGTH+4);
3062			U32_TO_BE32(tcp->tcp_ts_recent,
3063			    (char *)tcpha + TCP_MIN_HEADER_LENGTH+8);
3064		}
3065	}
3066
3067	if (num_sack_blk > 0) {
3068		uchar_t *wptr = (uchar_t *)tcpha + connp->conn_ht_ulp_len;
3069		sack_blk_t *tmp;
3070		int32_t	i;
3071
3072		wptr[0] = TCPOPT_NOP;
3073		wptr[1] = TCPOPT_NOP;
3074		wptr[2] = TCPOPT_SACK;
3075		wptr[3] = TCPOPT_HEADER_LEN + num_sack_blk *
3076		    sizeof (sack_blk_t);
3077		wptr += TCPOPT_REAL_SACK_LEN;
3078
3079		tmp = tcp->tcp_sack_list;
3080		for (i = 0; i < num_sack_blk; i++) {
3081			U32_TO_BE32(tmp[i].begin, wptr);
3082			wptr += sizeof (tcp_seq);
3083			U32_TO_BE32(tmp[i].end, wptr);
3084			wptr += sizeof (tcp_seq);
3085		}
3086		tcpha->tha_offset_and_reserved += ((num_sack_blk * 2 + 1) << 4);
3087	}
3088	ASSERT((uintptr_t)(mp1->b_wptr - rptr) <= (uintptr_t)INT_MAX);
3089	data_length += (int)(mp1->b_wptr - rptr);
3090
3091	ixa->ixa_pktlen = data_length;
3092
3093	if (ixa->ixa_flags & IXAF_IS_IPV4) {
3094		((ipha_t *)rptr)->ipha_length = htons(data_length);
3095	} else {
3096		ip6_t *ip6 = (ip6_t *)rptr;
3097
3098		ip6->ip6_plen = htons(data_length - IPV6_HDR_LEN);
3099	}
3100
3101	/*
3102	 * Prime pump for IP
3103	 * Include the adjustment for a source route if any.
3104	 */
3105	data_length -= ixa->ixa_ip_hdr_length;
3106	data_length += connp->conn_sum;
3107	data_length = (data_length >> 16) + (data_length & 0xFFFF);
3108	tcpha->tha_sum = htons(data_length);
3109	if (tcp->tcp_ip_forward_progress) {
3110		tcp->tcp_ip_forward_progress = B_FALSE;
3111		connp->conn_ixa->ixa_flags |= IXAF_REACH_CONF;
3112	} else {
3113		connp->conn_ixa->ixa_flags &= ~IXAF_REACH_CONF;
3114	}
3115	return (mp1);
3116}
3117
3118/*
3119 * If this routine returns B_TRUE, TCP can generate a RST in response
3120 * to a segment.  If it returns B_FALSE, TCP should not respond.
3121 */
3122static boolean_t
3123tcp_send_rst_chk(tcp_stack_t *tcps)
3124{
3125	int64_t	now;
3126
3127	/*
3128	 * TCP needs to protect itself from generating too many RSTs.
3129	 * This can be a DoS attack by sending us random segments
3130	 * soliciting RSTs.
3131	 *
3132	 * What we do here is to have a limit of tcp_rst_sent_rate RSTs
3133	 * in each 1 second interval.  In this way, TCP still generate
3134	 * RSTs in normal cases but when under attack, the impact is
3135	 * limited.
3136	 */
3137	if (tcps->tcps_rst_sent_rate_enabled != 0) {
3138		now = ddi_get_lbolt64();
3139		if (TICK_TO_MSEC(now - tcps->tcps_last_rst_intrvl) >
3140		    1*SECONDS) {
3141			tcps->tcps_last_rst_intrvl = now;
3142			tcps->tcps_rst_cnt = 1;
3143		} else if (++tcps->tcps_rst_cnt > tcps->tcps_rst_sent_rate) {
3144			return (B_FALSE);
3145		}
3146	}
3147	return (B_TRUE);
3148}
3149
3150/*
3151 * This function handles all retransmissions if SACK is enabled for this
3152 * connection.  First it calculates how many segments can be retransmitted
3153 * based on tcp_pipe.  Then it goes thru the notsack list to find eligible
3154 * segments.  A segment is eligible if sack_cnt for that segment is greater
3155 * than or equal tcp_dupack_fast_retransmit.  After it has retransmitted
3156 * all eligible segments, it checks to see if TCP can send some new segments
3157 * (fast recovery).  If it can, set the appropriate flag for tcp_input_data().
3158 *
3159 * Parameters:
3160 *	tcp_t *tcp: the tcp structure of the connection.
3161 *	uint_t *flags: in return, appropriate value will be set for
3162 *	tcp_input_data().
3163 */
3164void
3165tcp_sack_rexmit(tcp_t *tcp, uint_t *flags)
3166{
3167	notsack_blk_t	*notsack_blk;
3168	int32_t		usable_swnd;
3169	int32_t		mss;
3170	uint32_t	seg_len;
3171	mblk_t		*xmit_mp;
3172	tcp_stack_t	*tcps = tcp->tcp_tcps;
3173
3174	ASSERT(tcp->tcp_sack_info != NULL);
3175	ASSERT(tcp->tcp_notsack_list != NULL);
3176	ASSERT(tcp->tcp_rexmit == B_FALSE);
3177
3178	/* Defensive coding in case there is a bug... */
3179	if (tcp->tcp_notsack_list == NULL) {
3180		return;
3181	}
3182	notsack_blk = tcp->tcp_notsack_list;
3183	mss = tcp->tcp_mss;
3184
3185	/*
3186	 * Limit the num of outstanding data in the network to be
3187	 * tcp_cwnd_ssthresh, which is half of the original congestion wnd.
3188	 */
3189	usable_swnd = tcp->tcp_cwnd_ssthresh - tcp->tcp_pipe;
3190
3191	/* At least retransmit 1 MSS of data. */
3192	if (usable_swnd <= 0) {
3193		usable_swnd = mss;
3194	}
3195
3196	/* Make sure no new RTT samples will be taken. */
3197	tcp->tcp_csuna = tcp->tcp_snxt;
3198
3199	notsack_blk = tcp->tcp_notsack_list;
3200	while (usable_swnd > 0) {
3201		mblk_t		*snxt_mp, *tmp_mp;
3202		tcp_seq		begin = tcp->tcp_sack_snxt;
3203		tcp_seq		end;
3204		int32_t		off;
3205
3206		for (; notsack_blk != NULL; notsack_blk = notsack_blk->next) {
3207			if (SEQ_GT(notsack_blk->end, begin) &&
3208			    (notsack_blk->sack_cnt >=
3209			    tcps->tcps_dupack_fast_retransmit)) {
3210				end = notsack_blk->end;
3211				if (SEQ_LT(begin, notsack_blk->begin)) {
3212					begin = notsack_blk->begin;
3213				}
3214				break;
3215			}
3216		}
3217		/*
3218		 * All holes are filled.  Manipulate tcp_cwnd to send more
3219		 * if we can.  Note that after the SACK recovery, tcp_cwnd is
3220		 * set to tcp_cwnd_ssthresh.
3221		 */
3222		if (notsack_blk == NULL) {
3223			usable_swnd = tcp->tcp_cwnd_ssthresh - tcp->tcp_pipe;
3224			if (usable_swnd <= 0 || tcp->tcp_unsent == 0) {
3225				tcp->tcp_cwnd = tcp->tcp_snxt - tcp->tcp_suna;
3226				ASSERT(tcp->tcp_cwnd > 0);
3227				return;
3228			} else {
3229				usable_swnd = usable_swnd / mss;
3230				tcp->tcp_cwnd = tcp->tcp_snxt - tcp->tcp_suna +
3231				    MAX(usable_swnd * mss, mss);
3232				*flags |= TH_XMIT_NEEDED;
3233				return;
3234			}
3235		}
3236
3237		/*
3238		 * Note that we may send more than usable_swnd allows here
3239		 * because of round off, but no more than 1 MSS of data.
3240		 */
3241		seg_len = end - begin;
3242		if (seg_len > mss)
3243			seg_len = mss;
3244		snxt_mp = tcp_get_seg_mp(tcp, begin, &off);
3245		ASSERT(snxt_mp != NULL);
3246		/* This should not happen.  Defensive coding again... */
3247		if (snxt_mp == NULL) {
3248			return;
3249		}
3250
3251		xmit_mp = tcp_xmit_mp(tcp, snxt_mp, seg_len, &off,
3252		    &tmp_mp, begin, B_TRUE, &seg_len, B_TRUE);
3253		if (xmit_mp == NULL)
3254			return;
3255
3256		usable_swnd -= seg_len;
3257		tcp->tcp_pipe += seg_len;
3258		tcp->tcp_sack_snxt = begin + seg_len;
3259
3260		tcp_send_data(tcp, xmit_mp);
3261
3262		/*
3263		 * Update the send timestamp to avoid false retransmission.
3264		 */
3265		snxt_mp->b_prev = (mblk_t *)ddi_get_lbolt();
3266
3267		TCPS_BUMP_MIB(tcps, tcpRetransSegs);
3268		TCPS_UPDATE_MIB(tcps, tcpRetransBytes, seg_len);
3269		TCPS_BUMP_MIB(tcps, tcpOutSackRetransSegs);
3270		/*
3271		 * Update tcp_rexmit_max to extend this SACK recovery phase.
3272		 * This happens when new data sent during fast recovery is
3273		 * also lost.  If TCP retransmits those new data, it needs
3274		 * to extend SACK recover phase to avoid starting another
3275		 * fast retransmit/recovery unnecessarily.
3276		 */
3277		if (SEQ_GT(tcp->tcp_sack_snxt, tcp->tcp_rexmit_max)) {
3278			tcp->tcp_rexmit_max = tcp->tcp_sack_snxt;
3279		}
3280	}
3281}
3282
3283/*
3284 * tcp_ss_rexmit() is called to do slow start retransmission after a timeout
3285 * or ICMP errors.
3286 *
3287 * To limit the number of duplicate segments, we limit the number of segment
3288 * to be sent in one time to tcp_snd_burst, the burst variable.
3289 */
3290void
3291tcp_ss_rexmit(tcp_t *tcp)
3292{
3293	uint32_t	snxt;
3294	uint32_t	smax;
3295	int32_t		win;
3296	int32_t		mss;
3297	int32_t		off;
3298	int32_t		burst = tcp->tcp_snd_burst;
3299	mblk_t		*snxt_mp;
3300	tcp_stack_t	*tcps = tcp->tcp_tcps;
3301
3302	/*
3303	 * Note that tcp_rexmit can be set even though TCP has retransmitted
3304	 * all unack'ed segments.
3305	 */
3306	if (SEQ_LT(tcp->tcp_rexmit_nxt, tcp->tcp_rexmit_max)) {
3307		smax = tcp->tcp_rexmit_max;
3308		snxt = tcp->tcp_rexmit_nxt;
3309		if (SEQ_LT(snxt, tcp->tcp_suna)) {
3310			snxt = tcp->tcp_suna;
3311		}
3312		win = MIN(tcp->tcp_cwnd, tcp->tcp_swnd);
3313		win -= snxt - tcp->tcp_suna;
3314		mss = tcp->tcp_mss;
3315		snxt_mp = tcp_get_seg_mp(tcp, snxt, &off);
3316
3317		while (SEQ_LT(snxt, smax) && (win > 0) &&
3318		    (burst > 0) && (snxt_mp != NULL)) {
3319			mblk_t	*xmit_mp;
3320			mblk_t	*old_snxt_mp = snxt_mp;
3321			uint32_t cnt = mss;
3322
3323			if (win < cnt) {
3324				cnt = win;
3325			}
3326			if (SEQ_GT(snxt + cnt, smax)) {
3327				cnt = smax - snxt;
3328			}
3329			xmit_mp = tcp_xmit_mp(tcp, snxt_mp, cnt, &off,
3330			    &snxt_mp, snxt, B_TRUE, &cnt, B_TRUE);
3331			if (xmit_mp == NULL)
3332				return;
3333
3334			tcp_send_data(tcp, xmit_mp);
3335
3336			snxt += cnt;
3337			win -= cnt;
3338			/*
3339			 * Update the send timestamp to avoid false
3340			 * retransmission.
3341			 */
3342			old_snxt_mp->b_prev = (mblk_t *)ddi_get_lbolt();
3343			TCPS_BUMP_MIB(tcps, tcpRetransSegs);
3344			TCPS_UPDATE_MIB(tcps, tcpRetransBytes, cnt);
3345
3346			tcp->tcp_rexmit_nxt = snxt;
3347			burst--;
3348		}
3349		/*
3350		 * If we have transmitted all we have at the time
3351		 * we started the retranmission, we can leave
3352		 * the rest of the job to tcp_wput_data().  But we
3353		 * need to check the send window first.  If the
3354		 * win is not 0, go on with tcp_wput_data().
3355		 */
3356		if (SEQ_LT(snxt, smax) || win == 0) {
3357			return;
3358		}
3359	}
3360	/* Only call tcp_wput_data() if there is data to be sent. */
3361	if (tcp->tcp_unsent) {
3362		tcp_wput_data(tcp, NULL, B_FALSE);
3363	}
3364}
3365
3366/*
3367 * Do slow start retransmission after ICMP errors of PMTU changes.
3368 */
3369void
3370tcp_rexmit_after_error(tcp_t *tcp)
3371{
3372	/*
3373	 * All sent data has been acknowledged or no data left to send, just
3374	 * to return.
3375	 */
3376	if (!SEQ_LT(tcp->tcp_suna, tcp->tcp_snxt) ||
3377	    (tcp->tcp_xmit_head == NULL))
3378		return;
3379
3380	if ((tcp->tcp_valid_bits & TCP_FSS_VALID) && (tcp->tcp_unsent == 0))
3381		tcp->tcp_rexmit_max = tcp->tcp_fss;
3382	else
3383		tcp->tcp_rexmit_max = tcp->tcp_snxt;
3384
3385	tcp->tcp_rexmit_nxt = tcp->tcp_suna;
3386	tcp->tcp_rexmit = B_TRUE;
3387	tcp->tcp_dupack_cnt = 0;
3388	tcp->tcp_snd_burst = TCP_CWND_SS;
3389	tcp_ss_rexmit(tcp);
3390}
3391
3392/*
3393 * tcp_get_seg_mp() is called to get the pointer to a segment in the
3394 * send queue which starts at the given sequence number. If the given
3395 * sequence number is equal to last valid sequence number (tcp_snxt), the
3396 * returned mblk is the last valid mblk, and off is set to the length of
3397 * that mblk.
3398 *
3399 * send queue which starts at the given seq. no.
3400 *
3401 * Parameters:
3402 *	tcp_t *tcp: the tcp instance pointer.
3403 *	uint32_t seq: the starting seq. no of the requested segment.
3404 *	int32_t *off: after the execution, *off will be the offset to
3405 *		the returned mblk which points to the requested seq no.
3406 *		It is the caller's responsibility to send in a non-null off.
3407 *
3408 * Return:
3409 *	A mblk_t pointer pointing to the requested segment in send queue.
3410 */
3411static mblk_t *
3412tcp_get_seg_mp(tcp_t *tcp, uint32_t seq, int32_t *off)
3413{
3414	int32_t	cnt;
3415	mblk_t	*mp;
3416
3417	/* Defensive coding.  Make sure we don't send incorrect data. */
3418	if (SEQ_LT(seq, tcp->tcp_suna) || SEQ_GT(seq, tcp->tcp_snxt))
3419		return (NULL);
3420
3421	cnt = seq - tcp->tcp_suna;
3422	mp = tcp->tcp_xmit_head;
3423	while (cnt > 0 && mp != NULL) {
3424		cnt -= mp->b_wptr - mp->b_rptr;
3425		if (cnt <= 0) {
3426			cnt += mp->b_wptr - mp->b_rptr;
3427			break;
3428		}
3429		mp = mp->b_cont;
3430	}
3431	ASSERT(mp != NULL);
3432	*off = cnt;
3433	return (mp);
3434}
3435
3436/*
3437 * This routine adjusts next-to-send sequence number variables, in the
3438 * case where the reciever has shrunk it's window.
3439 */
3440void
3441tcp_update_xmit_tail(tcp_t *tcp, uint32_t snxt)
3442{
3443	mblk_t *xmit_tail;
3444	int32_t offset;
3445
3446	tcp->tcp_snxt = snxt;
3447
3448	/* Get the mblk, and the offset in it, as per the shrunk window */
3449	xmit_tail = tcp_get_seg_mp(tcp, snxt, &offset);
3450	ASSERT(xmit_tail != NULL);
3451	tcp->tcp_xmit_tail = xmit_tail;
3452	tcp->tcp_xmit_tail_unsent = xmit_tail->b_wptr -
3453	    xmit_tail->b_rptr - offset;
3454}
3455
3456/*
3457 * This handles the case when the receiver has shrunk its win. Per RFC 1122
3458 * if the receiver shrinks the window, i.e. moves the right window to the
3459 * left, the we should not send new data, but should retransmit normally the
3460 * old unacked data between suna and suna + swnd. We might has sent data
3461 * that is now outside the new window, pretend that we didn't send  it.
3462 */
3463static void
3464tcp_process_shrunk_swnd(tcp_t *tcp, uint32_t shrunk_count)
3465{
3466	uint32_t	snxt = tcp->tcp_snxt;
3467
3468	ASSERT(shrunk_count > 0);
3469
3470	if (!tcp->tcp_is_wnd_shrnk) {
3471		tcp->tcp_snxt_shrunk = snxt;
3472		tcp->tcp_is_wnd_shrnk = B_TRUE;
3473	} else if (SEQ_GT(snxt, tcp->tcp_snxt_shrunk)) {
3474		tcp->tcp_snxt_shrunk = snxt;
3475	}
3476
3477	/* Pretend we didn't send the data outside the window */
3478	snxt -= shrunk_count;
3479
3480	/* Reset all the values per the now shrunk window */
3481	tcp_update_xmit_tail(tcp, snxt);
3482	tcp->tcp_unsent += shrunk_count;
3483
3484	/*
3485	 * If the SACK option is set, delete the entire list of
3486	 * notsack'ed blocks.
3487	 */
3488	if (tcp->tcp_sack_info != NULL) {
3489		if (tcp->tcp_notsack_list != NULL)
3490			TCP_NOTSACK_REMOVE_ALL(tcp->tcp_notsack_list, tcp);
3491	}
3492
3493	if (tcp->tcp_suna == tcp->tcp_snxt && tcp->tcp_swnd == 0)
3494		/*
3495		 * Make sure the timer is running so that we will probe a zero
3496		 * window.
3497		 */
3498		TCP_TIMER_RESTART(tcp, tcp->tcp_rto);
3499}
3500
3501/*
3502 * tcp_fill_header is called by tcp_send() to fill the outgoing TCP header
3503 * with the template header, as well as other options such as time-stamp,
3504 * ECN and/or SACK.
3505 */
3506static void
3507tcp_fill_header(tcp_t *tcp, uchar_t *rptr, clock_t now, int num_sack_blk)
3508{
3509	tcpha_t *tcp_tmpl, *tcpha;
3510	uint32_t *dst, *src;
3511	int hdrlen;
3512	conn_t *connp = tcp->tcp_connp;
3513
3514	ASSERT(OK_32PTR(rptr));
3515
3516	/* Template header */
3517	tcp_tmpl = tcp->tcp_tcpha;
3518
3519	/* Header of outgoing packet */
3520	tcpha = (tcpha_t *)(rptr + connp->conn_ixa->ixa_ip_hdr_length);
3521
3522	/* dst and src are opaque 32-bit fields, used for copying */
3523	dst = (uint32_t *)rptr;
3524	src = (uint32_t *)connp->conn_ht_iphc;
3525	hdrlen = connp->conn_ht_iphc_len;
3526
3527	/* Fill time-stamp option if needed */
3528	if (tcp->tcp_snd_ts_ok) {
3529		U32_TO_BE32((uint32_t)now,
3530		    (char *)tcp_tmpl + TCP_MIN_HEADER_LENGTH + 4);
3531		U32_TO_BE32(tcp->tcp_ts_recent,
3532		    (char *)tcp_tmpl + TCP_MIN_HEADER_LENGTH + 8);
3533	} else {
3534		ASSERT(connp->conn_ht_ulp_len == TCP_MIN_HEADER_LENGTH);
3535	}
3536
3537	/*
3538	 * Copy the template header; is this really more efficient than
3539	 * calling bcopy()?  For simple IPv4/TCP, it may be the case,
3540	 * but perhaps not for other scenarios.
3541	 */
3542	dst[0] = src[0];
3543	dst[1] = src[1];
3544	dst[2] = src[2];
3545	dst[3] = src[3];
3546	dst[4] = src[4];
3547	dst[5] = src[5];
3548	dst[6] = src[6];
3549	dst[7] = src[7];
3550	dst[8] = src[8];
3551	dst[9] = src[9];
3552	if (hdrlen -= 40) {
3553		hdrlen >>= 2;
3554		dst += 10;
3555		src += 10;
3556		do {
3557			*dst++ = *src++;
3558		} while (--hdrlen);
3559	}
3560
3561	/*
3562	 * Set the ECN info in the TCP header if it is not a zero
3563	 * window probe.  Zero window probe is only sent in
3564	 * tcp_wput_data() and tcp_timer().
3565	 */
3566	if (tcp->tcp_ecn_ok && !tcp->tcp_zero_win_probe) {
3567		TCP_SET_ECT(tcp, rptr);
3568
3569		if (tcp->tcp_ecn_echo_on)
3570			tcpha->tha_flags |= TH_ECE;
3571		if (tcp->tcp_cwr && !tcp->tcp_ecn_cwr_sent) {
3572			tcpha->tha_flags |= TH_CWR;
3573			tcp->tcp_ecn_cwr_sent = B_TRUE;
3574		}
3575	}
3576
3577	/* Fill in SACK options */
3578	if (num_sack_blk > 0) {
3579		uchar_t *wptr = rptr + connp->conn_ht_iphc_len;
3580		sack_blk_t *tmp;
3581		int32_t	i;
3582
3583		wptr[0] = TCPOPT_NOP;
3584		wptr[1] = TCPOPT_NOP;
3585		wptr[2] = TCPOPT_SACK;
3586		wptr[3] = TCPOPT_HEADER_LEN + num_sack_blk *
3587		    sizeof (sack_blk_t);
3588		wptr += TCPOPT_REAL_SACK_LEN;
3589
3590		tmp = tcp->tcp_sack_list;
3591		for (i = 0; i < num_sack_blk; i++) {
3592			U32_TO_BE32(tmp[i].begin, wptr);
3593			wptr += sizeof (tcp_seq);
3594			U32_TO_BE32(tmp[i].end, wptr);
3595			wptr += sizeof (tcp_seq);
3596		}
3597		tcpha->tha_offset_and_reserved +=
3598		    ((num_sack_blk * 2 + 1) << 4);
3599	}
3600}
3601