VirtualBox

source: vbox/trunk/src/VBox/NetworkServices/NAT/pxtcp.c@ 60080

Last change on this file since 60080 was 60080, checked in by vboxsync, 9 years ago

NAT/Net: pxtcp_pmgr_pump - dissimilate two DPRINTFs.

  • Property svn:eol-style set to native
  • Property svn:keywords set to Id Revision
File size: 67.6 KB
Line 
1/* $Id: pxtcp.c 60080 2016-03-17 15:36:35Z vboxsync $ */
2/** @file
3 * NAT Network - TCP proxy.
4 */
5
6/*
7 * Copyright (C) 2013-2015 Oracle Corporation
8 *
9 * This file is part of VirtualBox Open Source Edition (OSE), as
10 * available from http://www.virtualbox.org. This file is free software;
11 * you can redistribute it and/or modify it under the terms of the GNU
12 * General Public License (GPL) as published by the Free Software
13 * Foundation, in version 2 as it comes in the "COPYING" file of the
14 * VirtualBox OSE distribution. VirtualBox OSE is distributed in the
15 * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
16 */
17
18#define LOG_GROUP LOG_GROUP_NAT_SERVICE
19
20#include "winutils.h"
21
22#include "pxtcp.h"
23
24#include "proxy.h"
25#include "proxy_pollmgr.h"
26#include "pxremap.h"
27#include "portfwd.h" /* fwspec */
28
29#ifndef RT_OS_WINDOWS
30#include <sys/types.h>
31#include <sys/socket.h>
32#include <sys/ioctl.h>
33#ifdef RT_OS_SOLARIS
34#include <sys/filio.h> /* FIONREAD is BSD'ism */
35#endif
36#include <stdlib.h>
37#include <stdint.h>
38#include <stdio.h>
39#include <string.h>
40#include <poll.h>
41
42#include <err.h> /* BSD'ism */
43#else
44#include <stdlib.h>
45#include <stdio.h>
46#include <string.h>
47
48#include <iprt/stdint.h>
49#include "winpoll.h"
50#endif
51
52#include "lwip/opt.h"
53
54#include "lwip/sys.h"
55#include "lwip/tcpip.h"
56#include "lwip/netif.h"
57#include "lwip/tcp_impl.h" /* XXX: to access tcp_abandon() */
58#include "lwip/icmp.h"
59#include "lwip/icmp6.h"
60
61/*
62 * Different OSes have different quirks in reporting POLLHUP for TCP
63 * sockets.
64 *
65 * Using shutdown(2) "how" values here would be more readable, but
66 * since SHUT_RD is 0, we can't use 0 for "none", unfortunately.
67 */
68#if defined(RT_OS_NETBSD) || defined(RT_OS_SOLARIS)
69# define HAVE_TCP_POLLHUP 0 /* not reported */
70#elif defined(RT_OS_DARWIN) || defined(RT_OS_WINDOWS)
71# define HAVE_TCP_POLLHUP POLLIN /* reported when remote closes */
72#else
73# define HAVE_TCP_POLLHUP (POLLIN|POLLOUT) /* reported when both directions are closed */
74#endif
75
76
77/**
78 * Ring buffer for inbound data. Filled with data from the host
79 * socket on poll manager thread. Data consumed by scheduling
80 * tcp_write() to the pcb on the lwip thread.
81 *
82 * NB: There is actually third party present, the lwip stack itself.
83 * Thus the buffer doesn't have dual free vs. data split, but rather
84 * three-way free / send and unACKed data / unsent data split.
85 */
86struct ringbuf {
87 char *buf;
88 size_t bufsize;
89
90 /*
91 * Start of free space, producer writes here (up till "unacked").
92 */
93 volatile size_t vacant;
94
95 /*
96 * Start of sent but unacknowledged data. The data are "owned" by
97 * the stack as it may need to retransmit. This is the free space
98 * limit for producer.
99 */
100 volatile size_t unacked;
101
102 /*
103 * Start of unsent data, consumer reads/sends from here (up till
104 * "vacant"). Not declared volatile since it's only accessed from
105 * the consumer thread.
106 */
107 size_t unsent;
108};
109
110
111/**
112 */
113struct pxtcp {
114 /**
115 * Our poll manager handler. Must be first, strong/weak
116 * references depend on this "inheritance".
117 */
118 struct pollmgr_handler pmhdl;
119
120 /**
121 * lwIP (internal/guest) side of the proxied connection.
122 */
123 struct tcp_pcb *pcb;
124
125 /**
126 * Host (external) side of the proxied connection.
127 */
128 SOCKET sock;
129
130 /**
131 * Socket events we are currently polling for.
132 */
133 int events;
134
135 /**
136 * Socket error. Currently used to save connect(2) errors so that
137 * we can decide if we need to send ICMP error.
138 */
139 int sockerr;
140
141 /**
142 * Interface that we have got the SYN from. Needed to send ICMP
143 * with correct source address.
144 */
145 struct netif *netif;
146
147 /**
148 * For tentatively accepted connections for which we are in
149 * process of connecting to the real destination this is the
150 * initial pbuf that we might need to build ICMP error.
151 *
152 * When connection is established this is used to hold outbound
153 * pbuf chain received by pxtcp_pcb_recv() but not yet completely
154 * forwarded over the socket. We cannot "return" it to lwIP since
155 * the head of the chain is already sent and freed.
156 */
157 struct pbuf *unsent;
158
159 /**
160 * Guest has closed its side. Reported to pxtcp_pcb_recv() only
161 * once and we might not be able to forward it immediately if we
162 * have unsent pbuf.
163 */
164 int outbound_close;
165
166 /**
167 * Outbound half-close has been done on the socket.
168 */
169 int outbound_close_done;
170
171 /**
172 * External has closed its side. We might not be able to forward
173 * it immediately if we have unforwarded data.
174 */
175 int inbound_close;
176
177 /**
178 * Inbound half-close has been done on the pcb.
179 */
180 int inbound_close_done;
181
182 /**
183 * On systems that report POLLHUP as soon as the final FIN is
184 * received on a socket we cannot continue polling for the rest of
185 * input, so we have to read (pull) last data from the socket on
186 * the lwIP thread instead of polling/pushing it from the poll
187 * manager thread. See comment in pxtcp_pmgr_pump() POLLHUP case.
188 */
189 int inbound_pull;
190
191
192 /**
193 * When poll manager schedules delete we may not be able to delete
194 * a pxtcp immediately if not all inbound data has been acked by
195 * the guest: lwIP may need to resend and the data are in pxtcp's
196 * inbuf::buf. We defer delete until all data are acked to
197 * pxtcp_pcb_sent().
198 */
199 int deferred_delete;
200
201 /**
202 * Ring-buffer for inbound data.
203 */
204 struct ringbuf inbuf;
205
206 /**
207 * lwIP thread's strong reference to us.
208 */
209 struct pollmgr_refptr *rp;
210
211
212 /*
213 * We use static messages to call functions on the lwIP thread to
214 * void malloc/free overhead.
215 */
216 struct tcpip_msg msg_delete; /* delete pxtcp */
217 struct tcpip_msg msg_reset; /* reset connection and delete pxtcp */
218 struct tcpip_msg msg_accept; /* confirm accept of proxied connection */
219 struct tcpip_msg msg_outbound; /* trigger send of outbound data */
220 struct tcpip_msg msg_inbound; /* trigger send of inbound data */
221 struct tcpip_msg msg_inpull; /* trigger pull of last inbound data */
222};
223
224
225
226static struct pxtcp *pxtcp_allocate(void);
227static void pxtcp_free(struct pxtcp *);
228
229static void pxtcp_pcb_associate(struct pxtcp *, struct tcp_pcb *);
230static void pxtcp_pcb_dissociate(struct pxtcp *);
231
232/* poll manager callbacks for pxtcp related channels */
233static int pxtcp_pmgr_chan_add(struct pollmgr_handler *, SOCKET, int);
234static int pxtcp_pmgr_chan_pollout(struct pollmgr_handler *, SOCKET, int);
235static int pxtcp_pmgr_chan_pollin(struct pollmgr_handler *, SOCKET, int);
236#if !(HAVE_TCP_POLLHUP & POLLOUT)
237static int pxtcp_pmgr_chan_del(struct pollmgr_handler *, SOCKET, int);
238#endif
239static int pxtcp_pmgr_chan_reset(struct pollmgr_handler *, SOCKET, int);
240
241/* helper functions for sending/receiving pxtcp over poll manager channels */
242static ssize_t pxtcp_chan_send(enum pollmgr_slot_t, struct pxtcp *);
243static ssize_t pxtcp_chan_send_weak(enum pollmgr_slot_t, struct pxtcp *);
244static struct pxtcp *pxtcp_chan_recv(struct pollmgr_handler *, SOCKET, int);
245static struct pxtcp *pxtcp_chan_recv_strong(struct pollmgr_handler *, SOCKET, int);
246
247/* poll manager callbacks for individual sockets */
248static int pxtcp_pmgr_connect(struct pollmgr_handler *, SOCKET, int);
249static int pxtcp_pmgr_pump(struct pollmgr_handler *, SOCKET, int);
250
251/* get incoming traffic into ring buffer */
252static ssize_t pxtcp_sock_read(struct pxtcp *, int *);
253static ssize_t pxtcp_sock_recv(struct pxtcp *, IOVEC *, size_t); /* default */
254
255/* convenience functions for poll manager callbacks */
256static int pxtcp_schedule_delete(struct pxtcp *);
257static int pxtcp_schedule_reset(struct pxtcp *);
258static int pxtcp_schedule_reject(struct pxtcp *);
259
260/* lwip thread callbacks called via proxy_lwip_post() */
261static void pxtcp_pcb_delete_pxtcp(void *);
262static void pxtcp_pcb_reset_pxtcp(void *);
263static void pxtcp_pcb_accept_refuse(void *);
264static void pxtcp_pcb_accept_confirm(void *);
265static void pxtcp_pcb_write_outbound(void *);
266static void pxtcp_pcb_write_inbound(void *);
267static void pxtcp_pcb_pull_inbound(void *);
268
269/* tcp pcb callbacks */
270static err_t pxtcp_pcb_heard(void *, struct tcp_pcb *, struct pbuf *); /* global */
271static err_t pxtcp_pcb_accept(void *, struct tcp_pcb *, err_t);
272static err_t pxtcp_pcb_connected(void *, struct tcp_pcb *, err_t);
273static err_t pxtcp_pcb_recv(void *, struct tcp_pcb *, struct pbuf *, err_t);
274static err_t pxtcp_pcb_sent(void *, struct tcp_pcb *, u16_t);
275static err_t pxtcp_pcb_poll(void *, struct tcp_pcb *);
276static void pxtcp_pcb_err(void *, err_t);
277
278static err_t pxtcp_pcb_forward_outbound(struct pxtcp *, struct pbuf *);
279static void pxtcp_pcb_forward_outbound_close(struct pxtcp *);
280
281static ssize_t pxtcp_sock_send(struct pxtcp *, IOVEC *, size_t);
282
283static void pxtcp_pcb_forward_inbound(struct pxtcp *);
284static void pxtcp_pcb_forward_inbound_close(struct pxtcp *);
285DECLINLINE(int) pxtcp_pcb_forward_inbound_done(const struct pxtcp *);
286static void pxtcp_pcb_schedule_poll(struct pxtcp *);
287static void pxtcp_pcb_cancel_poll(struct pxtcp *);
288
289static void pxtcp_pcb_reject(struct tcp_pcb *, int, struct netif *, struct pbuf *);
290DECLINLINE(void) pxtcp_pcb_maybe_deferred_delete(struct pxtcp *);
291
292/* poll manager handlers for pxtcp channels */
293static struct pollmgr_handler pxtcp_pmgr_chan_add_hdl;
294static struct pollmgr_handler pxtcp_pmgr_chan_pollout_hdl;
295static struct pollmgr_handler pxtcp_pmgr_chan_pollin_hdl;
296#if !(HAVE_TCP_POLLHUP & POLLOUT)
297static struct pollmgr_handler pxtcp_pmgr_chan_del_hdl;
298#endif
299static struct pollmgr_handler pxtcp_pmgr_chan_reset_hdl;
300
301
302/**
303 * Init PXTCP - must be run when neither lwIP tcpip thread, nor poll
304 * manager threads haven't been created yet.
305 */
306void
307pxtcp_init(void)
308{
309 /*
310 * Create channels.
311 */
312#define CHANNEL(SLOT, NAME) do { \
313 NAME##_hdl.callback = NAME; \
314 NAME##_hdl.data = NULL; \
315 NAME##_hdl.slot = -1; \
316 pollmgr_add_chan(SLOT, &NAME##_hdl); \
317 } while (0)
318
319 CHANNEL(POLLMGR_CHAN_PXTCP_ADD, pxtcp_pmgr_chan_add);
320 CHANNEL(POLLMGR_CHAN_PXTCP_POLLIN, pxtcp_pmgr_chan_pollin);
321 CHANNEL(POLLMGR_CHAN_PXTCP_POLLOUT, pxtcp_pmgr_chan_pollout);
322#if !(HAVE_TCP_POLLHUP & POLLOUT)
323 CHANNEL(POLLMGR_CHAN_PXTCP_DEL, pxtcp_pmgr_chan_del);
324#endif
325 CHANNEL(POLLMGR_CHAN_PXTCP_RESET, pxtcp_pmgr_chan_reset);
326
327#undef CHANNEL
328
329 /*
330 * Listen to outgoing connection from guest(s).
331 */
332 tcp_proxy_accept(pxtcp_pcb_heard);
333}
334
335
336/**
337 * Syntactic sugar for sending pxtcp pointer over poll manager
338 * channel. Used by lwip thread functions.
339 */
340static ssize_t
341pxtcp_chan_send(enum pollmgr_slot_t slot, struct pxtcp *pxtcp)
342{
343 return pollmgr_chan_send(slot, &pxtcp, sizeof(pxtcp));
344}
345
346
347/**
348 * Syntactic sugar for sending weak reference to pxtcp over poll
349 * manager channel. Used by lwip thread functions.
350 */
351static ssize_t
352pxtcp_chan_send_weak(enum pollmgr_slot_t slot, struct pxtcp *pxtcp)
353{
354 pollmgr_refptr_weak_ref(pxtcp->rp);
355 return pollmgr_chan_send(slot, &pxtcp->rp, sizeof(pxtcp->rp));
356}
357
358
359/**
360 * Counterpart of pxtcp_chan_send().
361 */
362static struct pxtcp *
363pxtcp_chan_recv(struct pollmgr_handler *handler, SOCKET fd, int revents)
364{
365 struct pxtcp *pxtcp;
366
367 pxtcp = (struct pxtcp *)pollmgr_chan_recv_ptr(handler, fd, revents);
368 return pxtcp;
369}
370
371
372/**
373 * Counterpart of pxtcp_chan_send_weak().
374 */
375static struct pxtcp *
376pxtcp_chan_recv_strong(struct pollmgr_handler *handler, SOCKET fd, int revents)
377{
378 struct pollmgr_refptr *rp;
379 struct pollmgr_handler *base;
380 struct pxtcp *pxtcp;
381
382 rp = (struct pollmgr_refptr *)pollmgr_chan_recv_ptr(handler, fd, revents);
383 base = (struct pollmgr_handler *)pollmgr_refptr_get(rp);
384 pxtcp = (struct pxtcp *)base;
385
386 return pxtcp;
387}
388
389
390/**
391 * Register pxtcp with poll manager.
392 *
393 * Used for POLLMGR_CHAN_PXTCP_ADD and by port-forwarding. Since
394 * error handling is different in these two cases, we leave it up to
395 * the caller.
396 */
397int
398pxtcp_pmgr_add(struct pxtcp *pxtcp)
399{
400 int status;
401
402 LWIP_ASSERT1(pxtcp != NULL);
403 LWIP_ASSERT1(pxtcp->sock >= 0);
404 LWIP_ASSERT1(pxtcp->pmhdl.callback != NULL);
405 LWIP_ASSERT1(pxtcp->pmhdl.data == (void *)pxtcp);
406 LWIP_ASSERT1(pxtcp->pmhdl.slot < 0);
407
408 status = pollmgr_add(&pxtcp->pmhdl, pxtcp->sock, pxtcp->events);
409 return status;
410}
411
412
413/**
414 * Unregister pxtcp with poll manager.
415 *
416 * Used for POLLMGR_CHAN_PXTCP_RESET and by port-forwarding (on error
417 * leg).
418 */
419void
420pxtcp_pmgr_del(struct pxtcp *pxtcp)
421{
422 LWIP_ASSERT1(pxtcp != NULL);
423
424 pollmgr_del_slot(pxtcp->pmhdl.slot);
425}
426
427
428/**
429 * POLLMGR_CHAN_PXTCP_ADD handler.
430 *
431 * Get new pxtcp from lwip thread and start polling its socket.
432 */
433static int
434pxtcp_pmgr_chan_add(struct pollmgr_handler *handler, SOCKET fd, int revents)
435{
436 struct pxtcp *pxtcp;
437 int status;
438
439 pxtcp = pxtcp_chan_recv(handler, fd, revents);
440 DPRINTF0(("pxtcp_add: new pxtcp %p; pcb %p; sock %d\n",
441 (void *)pxtcp, (void *)pxtcp->pcb, pxtcp->sock));
442
443 status = pxtcp_pmgr_add(pxtcp);
444 if (status < 0) {
445 (void) pxtcp_schedule_reset(pxtcp);
446 }
447
448 return POLLIN;
449}
450
451
452/**
453 * POLLMGR_CHAN_PXTCP_POLLOUT handler.
454 *
455 * pxtcp_pcb_forward_outbound() on the lwIP thread tried to send data
456 * and failed, it now requests us to poll the socket for POLLOUT and
457 * schedule pxtcp_pcb_forward_outbound() when sock is writable again.
458 */
459static int
460pxtcp_pmgr_chan_pollout(struct pollmgr_handler *handler, SOCKET fd, int revents)
461{
462 struct pxtcp *pxtcp;
463
464 pxtcp = pxtcp_chan_recv_strong(handler, fd, revents);
465 DPRINTF0(("pxtcp_pollout: pxtcp %p\n", (void *)pxtcp));
466
467 if (pxtcp == NULL) {
468 return POLLIN;
469 }
470
471 LWIP_ASSERT1(pxtcp->pmhdl.data == (void *)pxtcp);
472 LWIP_ASSERT1(pxtcp->pmhdl.slot > 0);
473
474 pxtcp->events |= POLLOUT;
475 pollmgr_update_events(pxtcp->pmhdl.slot, pxtcp->events);
476
477 return POLLIN;
478}
479
480
481/**
482 * POLLMGR_CHAN_PXTCP_POLLIN handler.
483 */
484static int
485pxtcp_pmgr_chan_pollin(struct pollmgr_handler *handler, SOCKET fd, int revents)
486{
487 struct pxtcp *pxtcp;
488
489 pxtcp = pxtcp_chan_recv_strong(handler, fd, revents);
490 DPRINTF2(("pxtcp_pollin: pxtcp %p\n", (void *)pxtcp));
491
492 if (pxtcp == NULL) {
493 return POLLIN;
494 }
495
496 LWIP_ASSERT1(pxtcp->pmhdl.data == (void *)pxtcp);
497 LWIP_ASSERT1(pxtcp->pmhdl.slot > 0);
498
499 if (pxtcp->inbound_close) {
500 return POLLIN;
501 }
502
503 pxtcp->events |= POLLIN;
504 pollmgr_update_events(pxtcp->pmhdl.slot, pxtcp->events);
505
506 return POLLIN;
507}
508
509
510#if !(HAVE_TCP_POLLHUP & POLLOUT)
511/**
512 * POLLMGR_CHAN_PXTCP_DEL handler.
513 *
514 * Schedule pxtcp deletion. We only need this if host system doesn't
515 * report POLLHUP for fully closed tcp sockets.
516 */
517static int
518pxtcp_pmgr_chan_del(struct pollmgr_handler *handler, SOCKET fd, int revents)
519{
520 struct pxtcp *pxtcp;
521
522 pxtcp = pxtcp_chan_recv_strong(handler, fd, revents);
523 if (pxtcp == NULL) {
524 return POLLIN;
525 }
526
527 DPRINTF(("PXTCP_DEL: pxtcp %p; pcb %p; sock %d\n",
528 (void *)pxtcp, (void *)pxtcp->pcb, pxtcp->sock));
529
530 LWIP_ASSERT1(pxtcp->pmhdl.callback != NULL);
531 LWIP_ASSERT1(pxtcp->pmhdl.data == (void *)pxtcp);
532
533 LWIP_ASSERT1(pxtcp->inbound_close); /* EOF read */
534 LWIP_ASSERT1(pxtcp->outbound_close_done); /* EOF sent */
535
536 pxtcp_pmgr_del(pxtcp);
537 (void) pxtcp_schedule_delete(pxtcp);
538
539 return POLLIN;
540}
541#endif /* !(HAVE_TCP_POLLHUP & POLLOUT) */
542
543
544/**
545 * POLLMGR_CHAN_PXTCP_RESET handler.
546 *
547 * Close the socket with RST and delete pxtcp.
548 */
549static int
550pxtcp_pmgr_chan_reset(struct pollmgr_handler *handler, SOCKET fd, int revents)
551{
552 struct pxtcp *pxtcp;
553
554 pxtcp = pxtcp_chan_recv_strong(handler, fd, revents);
555 if (pxtcp == NULL) {
556 return POLLIN;
557 }
558
559 DPRINTF0(("PXTCP_RESET: pxtcp %p; pcb %p; sock %d\n",
560 (void *)pxtcp, (void *)pxtcp->pcb, pxtcp->sock));
561
562 LWIP_ASSERT1(pxtcp->pmhdl.callback != NULL);
563 LWIP_ASSERT1(pxtcp->pmhdl.data == (void *)pxtcp);
564
565 pxtcp_pmgr_del(pxtcp);
566
567 proxy_reset_socket(pxtcp->sock);
568 pxtcp->sock = INVALID_SOCKET;
569
570 (void) pxtcp_schedule_reset(pxtcp);
571
572 return POLLIN;
573}
574
575
576static struct pxtcp *
577pxtcp_allocate(void)
578{
579 struct pxtcp *pxtcp;
580
581 pxtcp = (struct pxtcp *)malloc(sizeof(*pxtcp));
582 if (pxtcp == NULL) {
583 return NULL;
584 }
585
586 pxtcp->pmhdl.callback = NULL;
587 pxtcp->pmhdl.data = (void *)pxtcp;
588 pxtcp->pmhdl.slot = -1;
589
590 pxtcp->pcb = NULL;
591 pxtcp->sock = INVALID_SOCKET;
592 pxtcp->events = 0;
593 pxtcp->sockerr = 0;
594 pxtcp->netif = NULL;
595 pxtcp->unsent = NULL;
596 pxtcp->outbound_close = 0;
597 pxtcp->outbound_close_done = 0;
598 pxtcp->inbound_close = 0;
599 pxtcp->inbound_close_done = 0;
600 pxtcp->inbound_pull = 0;
601 pxtcp->deferred_delete = 0;
602
603 pxtcp->inbuf.bufsize = 64 * 1024;
604 pxtcp->inbuf.buf = (char *)malloc(pxtcp->inbuf.bufsize);
605 if (pxtcp->inbuf.buf == NULL) {
606 free(pxtcp);
607 return NULL;
608 }
609 pxtcp->inbuf.vacant = 0;
610 pxtcp->inbuf.unacked = 0;
611 pxtcp->inbuf.unsent = 0;
612
613 pxtcp->rp = pollmgr_refptr_create(&pxtcp->pmhdl);
614 if (pxtcp->rp == NULL) {
615 free(pxtcp->inbuf.buf);
616 free(pxtcp);
617 return NULL;
618 }
619
620#define CALLBACK_MSG(MSG, FUNC) \
621 do { \
622 pxtcp->MSG.type = TCPIP_MSG_CALLBACK_STATIC; \
623 pxtcp->MSG.sem = NULL; \
624 pxtcp->MSG.msg.cb.function = FUNC; \
625 pxtcp->MSG.msg.cb.ctx = (void *)pxtcp; \
626 } while (0)
627
628 CALLBACK_MSG(msg_delete, pxtcp_pcb_delete_pxtcp);
629 CALLBACK_MSG(msg_reset, pxtcp_pcb_reset_pxtcp);
630 CALLBACK_MSG(msg_accept, pxtcp_pcb_accept_confirm);
631 CALLBACK_MSG(msg_outbound, pxtcp_pcb_write_outbound);
632 CALLBACK_MSG(msg_inbound, pxtcp_pcb_write_inbound);
633 CALLBACK_MSG(msg_inpull, pxtcp_pcb_pull_inbound);
634
635#undef CALLBACK_MSG
636
637 return pxtcp;
638}
639
640
641/**
642 * Exported to fwtcp to create pxtcp for incoming port-forwarded
643 * connections. Completed with pcb in pxtcp_pcb_connect().
644 */
645struct pxtcp *
646pxtcp_create_forwarded(SOCKET sock)
647{
648 struct pxtcp *pxtcp;
649
650 pxtcp = pxtcp_allocate();
651 if (pxtcp == NULL) {
652 return NULL;
653 }
654
655 pxtcp->sock = sock;
656 pxtcp->pmhdl.callback = pxtcp_pmgr_pump;
657 pxtcp->events = 0;
658
659 return pxtcp;
660}
661
662
663static void
664pxtcp_pcb_associate(struct pxtcp *pxtcp, struct tcp_pcb *pcb)
665{
666 LWIP_ASSERT1(pxtcp != NULL);
667 LWIP_ASSERT1(pcb != NULL);
668
669 pxtcp->pcb = pcb;
670
671 tcp_arg(pcb, pxtcp);
672
673 tcp_recv(pcb, pxtcp_pcb_recv);
674 tcp_sent(pcb, pxtcp_pcb_sent);
675 tcp_poll(pcb, NULL, 255);
676 tcp_err(pcb, pxtcp_pcb_err);
677}
678
679
680static void
681pxtcp_free(struct pxtcp *pxtcp)
682{
683 if (pxtcp->unsent != NULL) {
684 pbuf_free(pxtcp->unsent);
685 }
686 if (pxtcp->inbuf.buf != NULL) {
687 free(pxtcp->inbuf.buf);
688 }
689 free(pxtcp);
690}
691
692
693/**
694 * Counterpart to pxtcp_create_forwarded() to destruct pxtcp that
695 * fwtcp failed to register with poll manager to post to lwip thread
696 * for doing connect.
697 */
698void
699pxtcp_cancel_forwarded(struct pxtcp *pxtcp)
700{
701 LWIP_ASSERT1(pxtcp->pcb == NULL);
702 pxtcp_pcb_reset_pxtcp(pxtcp);
703}
704
705
706static void
707pxtcp_pcb_dissociate(struct pxtcp *pxtcp)
708{
709 if (pxtcp == NULL || pxtcp->pcb == NULL) {
710 return;
711 }
712
713 DPRINTF(("%s: pxtcp %p <-> pcb %p\n",
714 __func__, (void *)pxtcp, (void *)pxtcp->pcb));
715
716 /*
717 * We must have dissociated from a fully closed pcb immediately
718 * since lwip recycles them and we don't wan't to mess with what
719 * would be someone else's pcb that we happen to have a stale
720 * pointer to.
721 */
722 LWIP_ASSERT1(pxtcp->pcb->callback_arg == pxtcp);
723
724 tcp_recv(pxtcp->pcb, NULL);
725 tcp_sent(pxtcp->pcb, NULL);
726 tcp_poll(pxtcp->pcb, NULL, 255);
727 tcp_err(pxtcp->pcb, NULL);
728 tcp_arg(pxtcp->pcb, NULL);
729 pxtcp->pcb = NULL;
730}
731
732
733/**
734 * Lwip thread callback invoked via pxtcp::msg_delete
735 *
736 * Since we use static messages to communicate to the lwip thread, we
737 * cannot delete pxtcp without making sure there are no unprocessed
738 * messages in the lwip thread mailbox.
739 *
740 * The easiest way to ensure that is to send this "delete" message as
741 * the last one and when it's processed we know there are no more and
742 * it's safe to delete pxtcp.
743 *
744 * Poll manager handlers should use pxtcp_schedule_delete()
745 * convenience function.
746 */
747static void
748pxtcp_pcb_delete_pxtcp(void *ctx)
749{
750 struct pxtcp *pxtcp = (struct pxtcp *)ctx;
751
752 DPRINTF(("%s: pxtcp %p, pcb %p, sock %d%s\n",
753 __func__, (void *)pxtcp, (void *)pxtcp->pcb, pxtcp->sock,
754 (pxtcp->deferred_delete && !pxtcp->inbound_pull
755 ? " (was deferred)" : "")));
756
757 LWIP_ASSERT1(pxtcp != NULL);
758 LWIP_ASSERT1(pxtcp->pmhdl.slot < 0);
759 LWIP_ASSERT1(pxtcp->outbound_close_done);
760 LWIP_ASSERT1(pxtcp->inbound_close); /* not necessarily done */
761
762
763 /*
764 * pxtcp is no longer registered with poll manager, so it's safe
765 * to close the socket.
766 */
767 if (pxtcp->sock != INVALID_SOCKET) {
768 closesocket(pxtcp->sock);
769 pxtcp->sock = INVALID_SOCKET;
770 }
771
772 /*
773 * We might have already dissociated from a fully closed pcb, or
774 * guest might have sent us a reset while msg_delete was in
775 * transit. If there's no pcb, we are done.
776 */
777 if (pxtcp->pcb == NULL) {
778 pollmgr_refptr_unref(pxtcp->rp);
779 pxtcp_free(pxtcp);
780 return;
781 }
782
783 /*
784 * Have we completely forwarded all inbound traffic to the guest?
785 *
786 * We may still be waiting for ACKs. We may have failed to send
787 * some of the data (tcp_write() failed with ERR_MEM). We may
788 * have failed to send the FIN (tcp_shutdown() failed with
789 * ERR_MEM).
790 */
791 if (pxtcp_pcb_forward_inbound_done(pxtcp)) {
792 pxtcp_pcb_dissociate(pxtcp);
793 pollmgr_refptr_unref(pxtcp->rp);
794 pxtcp_free(pxtcp);
795 }
796 else {
797 DPRINTF2(("delete: pxtcp %p; pcb %p:"
798 " unacked %d, unsent %d, vacant %d, %s - DEFER!\n",
799 (void *)pxtcp, (void *)pxtcp->pcb,
800 (int)pxtcp->inbuf.unacked,
801 (int)pxtcp->inbuf.unsent,
802 (int)pxtcp->inbuf.vacant,
803 pxtcp->inbound_close_done ? "FIN sent" : "FIN is NOT sent"));
804
805 LWIP_ASSERT1(!pxtcp->deferred_delete);
806 pxtcp->deferred_delete = 1;
807 }
808}
809
810
811/**
812 * If we couldn't delete pxtcp right away in the msg_delete callback
813 * from the poll manager thread, we repeat the check at the end of
814 * relevant pcb callbacks.
815 */
816DECLINLINE(void)
817pxtcp_pcb_maybe_deferred_delete(struct pxtcp *pxtcp)
818{
819 if (pxtcp->deferred_delete && pxtcp_pcb_forward_inbound_done(pxtcp)) {
820 pxtcp_pcb_delete_pxtcp(pxtcp);
821 }
822}
823
824
825/**
826 * Poll manager callbacks should use this convenience wrapper to
827 * schedule pxtcp deletion on the lwip thread and to deregister from
828 * the poll manager.
829 */
830static int
831pxtcp_schedule_delete(struct pxtcp *pxtcp)
832{
833 /*
834 * If pollmgr_refptr_get() is called by any channel before
835 * scheduled deletion happens, let them know we are gone.
836 */
837 pxtcp->pmhdl.slot = -1;
838
839 /*
840 * Schedule deletion. Since poll manager thread may be pre-empted
841 * right after we send the message, the deletion may actually
842 * happen on the lwip thread before we return from this function,
843 * so it's not safe to refer to pxtcp after this call.
844 */
845 proxy_lwip_post(&pxtcp->msg_delete);
846
847 /* tell poll manager to deregister us */
848 return -1;
849}
850
851
852/**
853 * Lwip thread callback invoked via pxtcp::msg_reset
854 *
855 * Like pxtcp_pcb_delete(), but sends RST to the guest before
856 * deleting this pxtcp.
857 */
858static void
859pxtcp_pcb_reset_pxtcp(void *ctx)
860{
861 struct pxtcp *pxtcp = (struct pxtcp *)ctx;
862 LWIP_ASSERT1(pxtcp != NULL);
863
864 DPRINTF0(("%s: pxtcp %p, pcb %p, sock %d\n",
865 __func__, (void *)pxtcp, (void *)pxtcp->pcb, pxtcp->sock));
866
867 if (pxtcp->sock != INVALID_SOCKET) {
868 proxy_reset_socket(pxtcp->sock);
869 pxtcp->sock = INVALID_SOCKET;
870 }
871
872 if (pxtcp->pcb != NULL) {
873 struct tcp_pcb *pcb = pxtcp->pcb;
874 pxtcp_pcb_dissociate(pxtcp);
875 tcp_abort(pcb);
876 }
877
878 pollmgr_refptr_unref(pxtcp->rp);
879 pxtcp_free(pxtcp);
880}
881
882
883
884/**
885 * Poll manager callbacks should use this convenience wrapper to
886 * schedule pxtcp reset and deletion on the lwip thread and to
887 * deregister from the poll manager.
888 *
889 * See pxtcp_schedule_delete() for additional comments.
890 */
891static int
892pxtcp_schedule_reset(struct pxtcp *pxtcp)
893{
894 pxtcp->pmhdl.slot = -1;
895 proxy_lwip_post(&pxtcp->msg_reset);
896 return -1;
897}
898
899
900/**
901 * Reject proxy connection attempt. Depending on the cause (sockerr)
902 * we may just drop the pcb silently, generate an ICMP datagram or
903 * send TCP reset.
904 */
905static void
906pxtcp_pcb_reject(struct tcp_pcb *pcb, int sockerr,
907 struct netif *netif, struct pbuf *p)
908{
909 int reset = 0;
910
911 if (sockerr == ECONNREFUSED) {
912 reset = 1;
913 }
914 else if (p != NULL) {
915 struct netif *oif;
916
917 LWIP_ASSERT1(netif != NULL);
918
919 oif = ip_current_netif();
920 ip_current_netif() = netif;
921
922 if (PCB_ISIPV6(pcb)) {
923 if (sockerr == EHOSTDOWN) {
924 icmp6_dest_unreach(p, ICMP6_DUR_ADDRESS); /* XXX: ??? */
925 }
926 else if (sockerr == EHOSTUNREACH
927 || sockerr == ENETDOWN
928 || sockerr == ENETUNREACH)
929 {
930 icmp6_dest_unreach(p, ICMP6_DUR_NO_ROUTE);
931 }
932 }
933 else {
934 if (sockerr == EHOSTDOWN
935 || sockerr == EHOSTUNREACH
936 || sockerr == ENETDOWN
937 || sockerr == ENETUNREACH)
938 {
939 icmp_dest_unreach(p, ICMP_DUR_HOST);
940 }
941 }
942
943 ip_current_netif() = oif;
944 }
945
946 tcp_abandon(pcb, reset);
947}
948
949
950/**
951 * Called from poll manager thread via pxtcp::msg_accept when proxy
952 * failed to connect to the destination. Also called when we failed
953 * to register pxtcp with poll manager.
954 *
955 * This is like pxtcp_pcb_reset_pxtcp() but is more discriminate in
956 * how this unestablished connection is terminated.
957 */
958static void
959pxtcp_pcb_accept_refuse(void *ctx)
960{
961 struct pxtcp *pxtcp = (struct pxtcp *)ctx;
962
963 DPRINTF0(("%s: pxtcp %p, pcb %p, sock %d: %R[sockerr]\n",
964 __func__, (void *)pxtcp, (void *)pxtcp->pcb,
965 pxtcp->sock, pxtcp->sockerr));
966
967 LWIP_ASSERT1(pxtcp != NULL);
968 LWIP_ASSERT1(pxtcp->sock == INVALID_SOCKET);
969
970 if (pxtcp->pcb != NULL) {
971 struct tcp_pcb *pcb = pxtcp->pcb;
972 pxtcp_pcb_dissociate(pxtcp);
973 pxtcp_pcb_reject(pcb, pxtcp->sockerr, pxtcp->netif, pxtcp->unsent);
974 }
975
976 pollmgr_refptr_unref(pxtcp->rp);
977 pxtcp_free(pxtcp);
978}
979
980
981/**
982 * Convenience wrapper for poll manager connect callback to reject
983 * connection attempt.
984 *
985 * Like pxtcp_schedule_reset(), but the callback is more discriminate
986 * in how this unestablished connection is terminated.
987 */
988static int
989pxtcp_schedule_reject(struct pxtcp *pxtcp)
990{
991 pxtcp->msg_accept.msg.cb.function = pxtcp_pcb_accept_refuse;
992 pxtcp->pmhdl.slot = -1;
993 proxy_lwip_post(&pxtcp->msg_accept);
994 return -1;
995}
996
997
998/**
999 * Global tcp_proxy_accept() callback for proxied outgoing TCP
1000 * connections from guest(s).
1001 */
1002static err_t
1003pxtcp_pcb_heard(void *arg, struct tcp_pcb *newpcb, struct pbuf *syn)
1004{
1005 LWIP_UNUSED_ARG(arg);
1006
1007 return pxtcp_pcb_accept_outbound(newpcb, syn,
1008 PCB_ISIPV6(newpcb), &newpcb->local_ip, newpcb->local_port);
1009}
1010
1011
1012err_t
1013pxtcp_pcb_accept_outbound(struct tcp_pcb *newpcb, struct pbuf *p,
1014 int is_ipv6, ipX_addr_t *dst_addr, u16_t dst_port)
1015{
1016 struct pxtcp *pxtcp;
1017 ipX_addr_t mapped_dst_addr;
1018 int sdom;
1019 SOCKET sock;
1020 ssize_t nsent;
1021 int sockerr = 0;
1022
1023 /*
1024 * TCP first calls accept callback when it receives the first SYN
1025 * and "tentatively accepts" new proxied connection attempt. When
1026 * proxy "confirms" the SYN and sends SYN|ACK and the guest
1027 * replies with ACK the accept callback is called again, this time
1028 * with the established connection.
1029 */
1030 LWIP_ASSERT1(newpcb->state == SYN_RCVD_0);
1031 tcp_accept(newpcb, pxtcp_pcb_accept);
1032 tcp_arg(newpcb, NULL);
1033
1034 tcp_setprio(newpcb, TCP_PRIO_MAX);
1035
1036 pxremap_outbound_ipX(is_ipv6, &mapped_dst_addr, dst_addr);
1037
1038 sdom = is_ipv6 ? PF_INET6 : PF_INET;
1039 sock = proxy_connected_socket(sdom, SOCK_STREAM,
1040 &mapped_dst_addr, dst_port);
1041 if (sock == INVALID_SOCKET) {
1042 sockerr = SOCKERRNO();
1043 goto abort;
1044 }
1045
1046 pxtcp = pxtcp_allocate();
1047 if (pxtcp == NULL) {
1048 proxy_reset_socket(sock);
1049 goto abort;
1050 }
1051
1052 /* save initial datagram in case we need to reply with ICMP */
1053 if (p != NULL) {
1054 pbuf_ref(p);
1055 pxtcp->unsent = p;
1056 pxtcp->netif = ip_current_netif();
1057 }
1058
1059 pxtcp_pcb_associate(pxtcp, newpcb);
1060 pxtcp->sock = sock;
1061
1062 pxtcp->pmhdl.callback = pxtcp_pmgr_connect;
1063 pxtcp->events = POLLOUT;
1064
1065 nsent = pxtcp_chan_send(POLLMGR_CHAN_PXTCP_ADD, pxtcp);
1066 if (nsent < 0) {
1067 pxtcp->sock = INVALID_SOCKET;
1068 proxy_reset_socket(sock);
1069 pxtcp_pcb_accept_refuse(pxtcp);
1070 return ERR_ABRT;
1071 }
1072
1073 return ERR_OK;
1074
1075 abort:
1076 DPRINTF0(("%s: pcb %p, sock %d: %R[sockerr]\n",
1077 __func__, (void *)newpcb, sock, sockerr));
1078 pxtcp_pcb_reject(newpcb, sockerr, ip_current_netif(), p);
1079 return ERR_ABRT;
1080}
1081
1082
1083/**
1084 * tcp_proxy_accept() callback for accepted proxied outgoing TCP
1085 * connections from guest(s). This is "real" accept with three-way
1086 * handshake completed.
1087 */
1088static err_t
1089pxtcp_pcb_accept(void *arg, struct tcp_pcb *pcb, err_t error)
1090{
1091 struct pxtcp *pxtcp = (struct pxtcp *)arg;
1092
1093 LWIP_UNUSED_ARG(pcb); /* used only in asserts */
1094 LWIP_UNUSED_ARG(error); /* always ERR_OK */
1095
1096 LWIP_ASSERT1(pxtcp != NULL);
1097 LWIP_ASSERT1(pxtcp->pcb = pcb);
1098 LWIP_ASSERT1(pcb->callback_arg == pxtcp);
1099
1100 /* send any inbound data that are already queued */
1101 pxtcp_pcb_forward_inbound(pxtcp);
1102 return ERR_OK;
1103}
1104
1105
1106/**
1107 * Initial poll manager callback for proxied outgoing TCP connections.
1108 * pxtcp_pcb_accept() sets pxtcp::pmhdl::callback to this.
1109 *
1110 * Waits for connect(2) to the destination to complete. On success
1111 * replaces itself with pxtcp_pmgr_pump() callback common to all
1112 * established TCP connections.
1113 */
1114static int
1115pxtcp_pmgr_connect(struct pollmgr_handler *handler, SOCKET fd, int revents)
1116{
1117 struct pxtcp *pxtcp;
1118
1119 pxtcp = (struct pxtcp *)handler->data;
1120 LWIP_ASSERT1(handler == &pxtcp->pmhdl);
1121 LWIP_ASSERT1(fd == pxtcp->sock);
1122 LWIP_ASSERT1(pxtcp->sockerr == 0);
1123
1124 if (revents & POLLNVAL) {
1125 pxtcp->sock = INVALID_SOCKET;
1126 pxtcp->sockerr = ETIMEDOUT;
1127 return pxtcp_schedule_reject(pxtcp);
1128 }
1129
1130 /*
1131 * Solaris and NetBSD don't report either POLLERR or POLLHUP when
1132 * connect(2) fails, just POLLOUT. In that case we always need to
1133 * check SO_ERROR.
1134 */
1135#if defined(RT_OS_SOLARIS) || defined(RT_OS_NETBSD)
1136# define CONNECT_CHECK_ERROR POLLOUT
1137#else
1138# define CONNECT_CHECK_ERROR (POLLERR | POLLHUP)
1139#endif
1140
1141 /*
1142 * Check the cause of the failure so that pxtcp_pcb_reject() may
1143 * behave accordingly.
1144 */
1145 if (revents & CONNECT_CHECK_ERROR) {
1146 socklen_t optlen = (socklen_t)sizeof(pxtcp->sockerr);
1147 int status;
1148 SOCKET s;
1149
1150 status = getsockopt(pxtcp->sock, SOL_SOCKET, SO_ERROR,
1151 (char *)&pxtcp->sockerr, &optlen);
1152 if (RT_UNLIKELY(status == SOCKET_ERROR)) { /* should not happen */
1153 DPRINTF(("%s: sock %d: SO_ERROR failed: %R[sockerr]\n",
1154 __func__, fd, SOCKERRNO()));
1155 pxtcp->sockerr = ETIMEDOUT;
1156 }
1157 else {
1158 /* don't spam this log on successful connect(2) */
1159 if ((revents & (POLLERR | POLLHUP)) /* we were told it's failed */
1160 || pxtcp->sockerr != 0) /* we determined it's failed */
1161 {
1162 DPRINTF(("%s: sock %d: connect: %R[sockerr]\n",
1163 __func__, fd, pxtcp->sockerr));
1164 }
1165
1166 if ((revents & (POLLERR | POLLHUP))
1167 && RT_UNLIKELY(pxtcp->sockerr == 0))
1168 {
1169 /* if we're told it's failed, make sure it's marked as such */
1170 pxtcp->sockerr = ETIMEDOUT;
1171 }
1172 }
1173
1174 if (pxtcp->sockerr != 0) {
1175 s = pxtcp->sock;
1176 pxtcp->sock = INVALID_SOCKET;
1177 closesocket(s);
1178 return pxtcp_schedule_reject(pxtcp);
1179 }
1180 }
1181
1182 if (revents & POLLOUT) { /* connect is successful */
1183 /* confirm accept to the guest */
1184 proxy_lwip_post(&pxtcp->msg_accept);
1185
1186 /*
1187 * Switch to common callback used for all established proxied
1188 * connections.
1189 */
1190 pxtcp->pmhdl.callback = pxtcp_pmgr_pump;
1191
1192 /*
1193 * Initially we poll for incoming traffic only. Outgoing
1194 * traffic is fast-forwarded by pxtcp_pcb_recv(); if it fails
1195 * it will ask us to poll for POLLOUT too.
1196 */
1197 pxtcp->events = POLLIN;
1198 return pxtcp->events;
1199 }
1200
1201 /* should never get here */
1202 DPRINTF0(("%s: pxtcp %p, sock %d: unexpected revents 0x%x\n",
1203 __func__, (void *)pxtcp, fd, revents));
1204 return pxtcp_schedule_reset(pxtcp);
1205}
1206
1207
1208/**
1209 * Called from poll manager thread via pxtcp::msg_accept when proxy
1210 * connected to the destination. Finalize accept by sending SYN|ACK
1211 * to the guest.
1212 */
1213static void
1214pxtcp_pcb_accept_confirm(void *ctx)
1215{
1216 struct pxtcp *pxtcp = (struct pxtcp *)ctx;
1217 err_t error;
1218
1219 LWIP_ASSERT1(pxtcp != NULL);
1220 if (pxtcp->pcb == NULL) {
1221 return;
1222 }
1223
1224 /* we are not going to reply with ICMP, so we can drop initial pbuf */
1225 if (pxtcp->unsent != NULL) {
1226 pbuf_free(pxtcp->unsent);
1227 pxtcp->unsent = NULL;
1228 }
1229
1230 error = tcp_proxy_accept_confirm(pxtcp->pcb);
1231
1232 /*
1233 * If lwIP failed to enqueue SYN|ACK because it's out of pbufs it
1234 * abandons the pcb. Retrying that is not very easy, since it
1235 * would require keeping "fractional state". From guest's point
1236 * of view there is no reply to its SYN so it will either resend
1237 * the SYN (effetively triggering full connection retry for us),
1238 * or it will eventually time out.
1239 */
1240 if (error == ERR_ABRT) {
1241 pxtcp->pcb = NULL; /* pcb is gone */
1242 pxtcp_chan_send_weak(POLLMGR_CHAN_PXTCP_RESET, pxtcp);
1243 }
1244
1245 /*
1246 * else if (error != ERR_OK): even if tcp_output() failed with
1247 * ERR_MEM - don't give up, that SYN|ACK is enqueued and will be
1248 * retransmitted eventually.
1249 */
1250}
1251
1252
1253/**
1254 * Entry point for port-forwarding.
1255 *
1256 * fwtcp accepts new incoming connection, creates pxtcp for the socket
1257 * (with no pcb yet) and adds it to the poll manager (polling for
1258 * errors only). Then it calls this function to construct the pcb and
1259 * perform connection to the guest.
1260 */
1261void
1262pxtcp_pcb_connect(struct pxtcp *pxtcp, const struct fwspec *fwspec)
1263{
1264 struct sockaddr_storage ss;
1265 socklen_t sslen;
1266 struct tcp_pcb *pcb;
1267 ipX_addr_t src_addr, dst_addr;
1268 u16_t src_port, dst_port;
1269 int status;
1270 err_t error;
1271
1272 LWIP_ASSERT1(pxtcp != NULL);
1273 LWIP_ASSERT1(pxtcp->pcb == NULL);
1274 LWIP_ASSERT1(fwspec->stype == SOCK_STREAM);
1275
1276 pcb = tcp_new();
1277 if (pcb == NULL) {
1278 goto reset;
1279 }
1280
1281 tcp_setprio(pcb, TCP_PRIO_MAX);
1282 pxtcp_pcb_associate(pxtcp, pcb);
1283
1284 sslen = sizeof(ss);
1285 status = getpeername(pxtcp->sock, (struct sockaddr *)&ss, &sslen);
1286 if (status == SOCKET_ERROR) {
1287 goto reset;
1288 }
1289
1290 /* nit: compares PF and AF, but they are the same everywhere */
1291 LWIP_ASSERT1(ss.ss_family == fwspec->sdom);
1292
1293 status = fwany_ipX_addr_set_src(&src_addr, (const struct sockaddr *)&ss);
1294 if (status == PXREMAP_FAILED) {
1295 goto reset;
1296 }
1297
1298 if (ss.ss_family == PF_INET) {
1299 const struct sockaddr_in *peer4 = (const struct sockaddr_in *)&ss;
1300
1301 src_port = peer4->sin_port;
1302
1303 memcpy(&dst_addr.ip4, &fwspec->dst.sin.sin_addr, sizeof(ip_addr_t));
1304 dst_port = fwspec->dst.sin.sin_port;
1305 }
1306 else { /* PF_INET6 */
1307 const struct sockaddr_in6 *peer6 = (const struct sockaddr_in6 *)&ss;
1308 ip_set_v6(pcb, 1);
1309
1310 src_port = peer6->sin6_port;
1311
1312 memcpy(&dst_addr.ip6, &fwspec->dst.sin6.sin6_addr, sizeof(ip6_addr_t));
1313 dst_port = fwspec->dst.sin6.sin6_port;
1314 }
1315
1316 /* lwip port arguments are in host order */
1317 src_port = ntohs(src_port);
1318 dst_port = ntohs(dst_port);
1319
1320 error = tcp_proxy_bind(pcb, ipX_2_ip(&src_addr), src_port);
1321 if (error != ERR_OK) {
1322 goto reset;
1323 }
1324
1325 error = tcp_connect(pcb, ipX_2_ip(&dst_addr), dst_port,
1326 /* callback: */ pxtcp_pcb_connected);
1327 if (error != ERR_OK) {
1328 goto reset;
1329 }
1330
1331 return;
1332
1333 reset:
1334 pxtcp_chan_send_weak(POLLMGR_CHAN_PXTCP_RESET, pxtcp);
1335}
1336
1337
1338/**
1339 * Port-forwarded connection to guest is successful, pump data.
1340 */
1341static err_t
1342pxtcp_pcb_connected(void *arg, struct tcp_pcb *pcb, err_t error)
1343{
1344 struct pxtcp *pxtcp = (struct pxtcp *)arg;
1345
1346 LWIP_ASSERT1(error == ERR_OK); /* always called with ERR_OK */
1347 LWIP_UNUSED_ARG(error);
1348
1349 LWIP_ASSERT1(pxtcp != NULL);
1350 LWIP_ASSERT1(pxtcp->pcb == pcb);
1351 LWIP_ASSERT1(pcb->callback_arg == pxtcp);
1352 LWIP_UNUSED_ARG(pcb);
1353
1354 DPRINTF0(("%s: new pxtcp %p; pcb %p; sock %d\n",
1355 __func__, (void *)pxtcp, (void *)pxtcp->pcb, pxtcp->sock));
1356
1357 /* ACK on connection is like ACK on data in pxtcp_pcb_sent() */
1358 pxtcp_chan_send_weak(POLLMGR_CHAN_PXTCP_POLLIN, pxtcp);
1359
1360 return ERR_OK;
1361}
1362
1363
1364/**
1365 * tcp_recv() callback.
1366 */
1367static err_t
1368pxtcp_pcb_recv(void *arg, struct tcp_pcb *pcb, struct pbuf *p, err_t error)
1369{
1370 struct pxtcp *pxtcp = (struct pxtcp *)arg;
1371
1372 LWIP_ASSERT1(error == ERR_OK); /* always called with ERR_OK */
1373 LWIP_UNUSED_ARG(error);
1374
1375 LWIP_ASSERT1(pxtcp != NULL);
1376 LWIP_ASSERT1(pxtcp->pcb == pcb);
1377 LWIP_ASSERT1(pcb->callback_arg == pxtcp);
1378 LWIP_UNUSED_ARG(pcb);
1379
1380
1381 /*
1382 * Have we done sending previous batch?
1383 */
1384 if (pxtcp->unsent != NULL) {
1385 if (p != NULL) {
1386 /*
1387 * Return an error to tell TCP to hold onto that pbuf.
1388 * It will be presented to us later from tcp_fasttmr().
1389 */
1390 return ERR_WOULDBLOCK;
1391 }
1392 else {
1393 /*
1394 * Unlike data, p == NULL indicating orderly shutdown is
1395 * NOT presented to us again
1396 */
1397 pxtcp->outbound_close = 1;
1398 return ERR_OK;
1399 }
1400 }
1401
1402
1403 /*
1404 * Guest closed?
1405 */
1406 if (p == NULL) {
1407 pxtcp->outbound_close = 1;
1408 pxtcp_pcb_forward_outbound_close(pxtcp);
1409 return ERR_OK;
1410 }
1411
1412
1413 /*
1414 * Got data, send what we can without blocking.
1415 */
1416 return pxtcp_pcb_forward_outbound(pxtcp, p);
1417}
1418
1419
1420/**
1421 * Guest half-closed its TX side of the connection.
1422 *
1423 * Called either immediately from pxtcp_pcb_recv() when it gets NULL,
1424 * or from pxtcp_pcb_forward_outbound() when it finishes forwarding
1425 * previously unsent data and sees pxtcp::outbound_close flag saved by
1426 * pxtcp_pcb_recv().
1427 */
1428static void
1429pxtcp_pcb_forward_outbound_close(struct pxtcp *pxtcp)
1430{
1431 struct tcp_pcb *pcb;
1432
1433 LWIP_ASSERT1(pxtcp != NULL);
1434 LWIP_ASSERT1(pxtcp->outbound_close);
1435 LWIP_ASSERT1(!pxtcp->outbound_close_done);
1436
1437 pcb = pxtcp->pcb;
1438 LWIP_ASSERT1(pcb != NULL);
1439
1440 DPRINTF(("outbound_close: pxtcp %p; pcb %p %s\n",
1441 (void *)pxtcp, (void *)pcb, tcp_debug_state_str(pcb->state)));
1442
1443
1444 /* set the flag first, since shutdown() may trigger POLLHUP */
1445 pxtcp->outbound_close_done = 1;
1446 shutdown(pxtcp->sock, SHUT_WR); /* half-close the socket */
1447
1448#if !(HAVE_TCP_POLLHUP & POLLOUT)
1449 /*
1450 * We need to nudge poll manager manually, since OS will not
1451 * report POLLHUP.
1452 */
1453 if (pxtcp->inbound_close) {
1454 pxtcp_chan_send_weak(POLLMGR_CHAN_PXTCP_DEL, pxtcp);
1455 }
1456#endif
1457
1458
1459 /* no more outbound data coming to us */
1460 tcp_recv(pcb, NULL);
1461
1462 /*
1463 * If we have already done inbound close previously (active close
1464 * on the pcb), then we must not hold onto a pcb in TIME_WAIT
1465 * state since those will be recycled by lwip when it runs out of
1466 * free pcbs in the pool.
1467 *
1468 * The test is true also for a pcb in CLOSING state that waits
1469 * just for the ACK of its FIN (to transition to TIME_WAIT).
1470 */
1471 if (pxtcp_pcb_forward_inbound_done(pxtcp)) {
1472 pxtcp_pcb_dissociate(pxtcp);
1473 }
1474}
1475
1476
1477/**
1478 * Forward outbound data from pcb to socket.
1479 *
1480 * Called by pxtcp_pcb_recv() to forward new data and by callout
1481 * triggered by POLLOUT on the socket to send previously unsent data.
1482 *
1483 * (Re)scehdules one-time callout if not all data are sent.
1484 */
1485static err_t
1486pxtcp_pcb_forward_outbound(struct pxtcp *pxtcp, struct pbuf *p)
1487{
1488 struct pbuf *qs, *q;
1489 size_t qoff;
1490 size_t forwarded;
1491 int sockerr;
1492
1493 LWIP_ASSERT1(pxtcp->unsent == NULL || pxtcp->unsent == p);
1494
1495 forwarded = 0;
1496 sockerr = 0;
1497
1498 q = NULL;
1499 qoff = 0;
1500
1501 qs = p;
1502 while (qs != NULL) {
1503 IOVEC iov[8];
1504 const size_t iovsize = sizeof(iov)/sizeof(iov[0]);
1505 size_t fwd1;
1506 ssize_t nsent;
1507 size_t i;
1508
1509 fwd1 = 0;
1510 for (i = 0, q = qs; i < iovsize && q != NULL; ++i, q = q->next) {
1511 LWIP_ASSERT1(q->len > 0);
1512 IOVEC_SET_BASE(iov[i], q->payload);
1513 IOVEC_SET_LEN(iov[i], q->len);
1514 fwd1 += q->len;
1515 }
1516
1517 /*
1518 * TODO: This is where application-level proxy can hook into
1519 * to process outbound traffic.
1520 */
1521 nsent = pxtcp_sock_send(pxtcp, iov, i);
1522
1523 if (nsent == (ssize_t)fwd1) {
1524 /* successfully sent this chain fragment completely */
1525 forwarded += nsent;
1526 qs = q;
1527 }
1528 else if (nsent >= 0) {
1529 /* successfully sent only some data */
1530 forwarded += nsent;
1531
1532 /* find the first pbuf that was not completely forwarded */
1533 qoff = nsent;
1534 for (i = 0, q = qs; i < iovsize && q != NULL; ++i, q = q->next) {
1535 if (qoff < q->len) {
1536 break;
1537 }
1538 qoff -= q->len;
1539 }
1540 LWIP_ASSERT1(q != NULL);
1541 LWIP_ASSERT1(qoff < q->len);
1542 break;
1543 }
1544 else {
1545 sockerr = -nsent;
1546
1547 /*
1548 * Some errors are really not errors - if we get them,
1549 * it's not different from getting nsent == 0, so filter
1550 * them out here.
1551 */
1552 if (proxy_error_is_transient(sockerr)) {
1553 sockerr = 0;
1554 }
1555 q = qs;
1556 qoff = 0;
1557 break;
1558 }
1559 }
1560
1561 if (forwarded > 0) {
1562 DPRINTF2(("forward_outbound: pxtcp %p, pcb %p: sent %d bytes\n",
1563 (void *)pxtcp, (void *)pxtcp->pcb, (int)forwarded));
1564 tcp_recved(pxtcp->pcb, (u16_t)forwarded);
1565 }
1566
1567 if (q == NULL) { /* everything is forwarded? */
1568 LWIP_ASSERT1(sockerr == 0);
1569 LWIP_ASSERT1(forwarded == p->tot_len);
1570
1571 pxtcp->unsent = NULL;
1572 pbuf_free(p);
1573 if (pxtcp->outbound_close) {
1574 pxtcp_pcb_forward_outbound_close(pxtcp);
1575 }
1576 }
1577 else {
1578 if (q != p) {
1579 /* free forwarded pbufs at the beginning of the chain */
1580 pbuf_ref(q);
1581 pbuf_free(p);
1582 }
1583 if (qoff > 0) {
1584 /* advance payload pointer past the forwarded part */
1585 pbuf_header(q, -(s16_t)qoff);
1586 }
1587 pxtcp->unsent = q;
1588 DPRINTF2(("forward_outbound: pxtcp %p, pcb %p: kept %d bytes\n",
1589 (void *)pxtcp, (void *)pxtcp->pcb, (int)q->tot_len));
1590
1591 /*
1592 * Have sendmsg() failed?
1593 *
1594 * Connection reset will be detected by poll and
1595 * pxtcp_schedule_reset() will be called.
1596 *
1597 * Otherwise something *really* unexpected must have happened,
1598 * so we'd better abort.
1599 */
1600 if (sockerr != 0 && sockerr != ECONNRESET) {
1601 struct tcp_pcb *pcb = pxtcp->pcb;
1602 DPRINTF2(("forward_outbound: pxtcp %p, pcb %p: %R[sockerr]\n",
1603 (void *)pxtcp, (void *)pcb, sockerr));
1604
1605 pxtcp_pcb_dissociate(pxtcp);
1606
1607 tcp_abort(pcb);
1608
1609 /* call error callback manually since we've already dissociated */
1610 pxtcp_pcb_err((void *)pxtcp, ERR_ABRT);
1611 return ERR_ABRT;
1612 }
1613
1614 /* schedule one-shot POLLOUT on the socket */
1615 pxtcp_chan_send_weak(POLLMGR_CHAN_PXTCP_POLLOUT, pxtcp);
1616 }
1617 return ERR_OK;
1618}
1619
1620
1621#if !defined(RT_OS_WINDOWS)
1622static ssize_t
1623pxtcp_sock_send(struct pxtcp *pxtcp, IOVEC *iov, size_t iovlen)
1624{
1625 struct msghdr mh;
1626 ssize_t nsent;
1627
1628#ifdef MSG_NOSIGNAL
1629 const int send_flags = MSG_NOSIGNAL;
1630#else
1631 const int send_flags = 0;
1632#endif
1633
1634 memset(&mh, 0, sizeof(mh));
1635
1636 mh.msg_iov = iov;
1637 mh.msg_iovlen = iovlen;
1638
1639 nsent = sendmsg(pxtcp->sock, &mh, send_flags);
1640 if (nsent < 0) {
1641 nsent = -SOCKERRNO();
1642 }
1643
1644 return nsent;
1645}
1646#else /* RT_OS_WINDOWS */
1647static ssize_t
1648pxtcp_sock_send(struct pxtcp *pxtcp, IOVEC *iov, size_t iovlen)
1649{
1650 DWORD nsent;
1651 int status;
1652
1653 status = WSASend(pxtcp->sock, iov, (DWORD)iovlen, &nsent,
1654 0, NULL, NULL);
1655 if (status == SOCKET_ERROR) {
1656 return -SOCKERRNO();
1657 }
1658
1659 return nsent;
1660}
1661#endif /* RT_OS_WINDOWS */
1662
1663
1664/**
1665 * Callback from poll manager (on POLLOUT) to send data from
1666 * pxtcp::unsent pbuf to socket.
1667 */
1668static void
1669pxtcp_pcb_write_outbound(void *ctx)
1670{
1671 struct pxtcp *pxtcp = (struct pxtcp *)ctx;
1672 LWIP_ASSERT1(pxtcp != NULL);
1673
1674 if (pxtcp->pcb == NULL) {
1675 return;
1676 }
1677
1678 pxtcp_pcb_forward_outbound(pxtcp, pxtcp->unsent);
1679}
1680
1681
1682/**
1683 * Common poll manager callback used by both outgoing and incoming
1684 * (port-forwarded) connections that has connected socket.
1685 */
1686static int
1687pxtcp_pmgr_pump(struct pollmgr_handler *handler, SOCKET fd, int revents)
1688{
1689 struct pxtcp *pxtcp;
1690 int status;
1691 int sockerr;
1692
1693 pxtcp = (struct pxtcp *)handler->data;
1694 LWIP_ASSERT1(handler == &pxtcp->pmhdl);
1695 LWIP_ASSERT1(fd == pxtcp->sock);
1696
1697 if (revents & POLLNVAL) {
1698 pxtcp->sock = INVALID_SOCKET;
1699 return pxtcp_schedule_reset(pxtcp);
1700 }
1701
1702 if (revents & POLLERR) {
1703 socklen_t optlen = (socklen_t)sizeof(sockerr);
1704
1705 status = getsockopt(pxtcp->sock, SOL_SOCKET, SO_ERROR,
1706 (char *)&sockerr, &optlen);
1707 if (status == SOCKET_ERROR) { /* should not happen */
1708 DPRINTF(("sock %d: POLLERR: SO_ERROR failed: %R[sockerr]\n",
1709 fd, SOCKERRNO()));
1710 }
1711 else {
1712 DPRINTF0(("sock %d: POLLERR: %R[sockerr]\n", fd, sockerr));
1713 }
1714 return pxtcp_schedule_reset(pxtcp);
1715 }
1716
1717 if (revents & POLLOUT) {
1718 pxtcp->events &= ~POLLOUT;
1719 proxy_lwip_post(&pxtcp->msg_outbound);
1720 }
1721
1722 if (revents & POLLIN) {
1723 ssize_t nread;
1724 int stop_pollin;
1725
1726 nread = pxtcp_sock_read(pxtcp, &stop_pollin);
1727 if (nread < 0) {
1728 sockerr = -(int)nread;
1729 DPRINTF0(("sock %d: POLLIN: %R[sockerr]\n", fd, sockerr));
1730 return pxtcp_schedule_reset(pxtcp);
1731 }
1732
1733 if (stop_pollin) {
1734 pxtcp->events &= ~POLLIN;
1735 }
1736
1737 if (nread > 0) {
1738 proxy_lwip_post(&pxtcp->msg_inbound);
1739#if !HAVE_TCP_POLLHUP
1740 /*
1741 * If host does not report POLLHUP for closed sockets
1742 * (e.g. NetBSD) we should check for full close manually.
1743 */
1744 if (pxtcp->inbound_close && pxtcp->outbound_close_done) {
1745 LWIP_ASSERT1((revents & POLLHUP) == 0);
1746 return pxtcp_schedule_delete(pxtcp);
1747 }
1748#endif
1749 }
1750 }
1751
1752#if !HAVE_TCP_POLLHUP
1753 LWIP_ASSERT1((revents & POLLHUP) == 0);
1754#else
1755 if (revents & POLLHUP) {
1756 DPRINTF(("sock %d: HUP\n", fd));
1757#if HAVE_TCP_POLLHUP == POLLIN
1758 /*
1759 * Remote closed inbound.
1760 */
1761 if (!pxtcp->outbound_close_done) {
1762 /*
1763 * We might still need to poll for POLLOUT, but we can not
1764 * poll for POLLIN anymore (even if not all data are read)
1765 * because we will be spammed by POLLHUP.
1766 */
1767 pxtcp->events &= ~POLLIN;
1768 if (!pxtcp->inbound_close) {
1769 /* the rest of the input has to be pulled */
1770 proxy_lwip_post(&pxtcp->msg_inpull);
1771 }
1772 }
1773 else
1774#endif
1775 /*
1776 * Both directions are closed.
1777 */
1778 {
1779 LWIP_ASSERT1(pxtcp->outbound_close_done);
1780
1781 if (pxtcp->inbound_close) {
1782 /* there's no unread data, we are done */
1783 return pxtcp_schedule_delete(pxtcp);
1784 }
1785 else {
1786 /* pull the rest of the input first (deferred_delete) */
1787 pxtcp->pmhdl.slot = -1;
1788 proxy_lwip_post(&pxtcp->msg_inpull);
1789 return -1;
1790 }
1791 /* NOTREACHED */
1792 }
1793
1794 }
1795#endif /* HAVE_TCP_POLLHUP */
1796
1797 return pxtcp->events;
1798}
1799
1800
1801/**
1802 * Read data from socket to ringbuf. This may be used both on lwip
1803 * and poll manager threads.
1804 *
1805 * Flag pointed to by pstop is set when further reading is impossible,
1806 * either temporary when buffer is full, or permanently when EOF is
1807 * received.
1808 *
1809 * Returns number of bytes read. NB: EOF is reported as 1!
1810 *
1811 * Returns zero if nothing was read, either because buffer is full, or
1812 * if no data is available (EWOULDBLOCK, EINTR &c).
1813 *
1814 * Returns -errno on real socket errors.
1815 */
1816static ssize_t
1817pxtcp_sock_read(struct pxtcp *pxtcp, int *pstop)
1818{
1819 IOVEC iov[2];
1820 size_t iovlen;
1821 ssize_t nread;
1822
1823 const size_t sz = pxtcp->inbuf.bufsize;
1824 size_t beg, lim, wrnew;
1825
1826 *pstop = 0;
1827
1828 beg = pxtcp->inbuf.vacant;
1829 IOVEC_SET_BASE(iov[0], &pxtcp->inbuf.buf[beg]);
1830
1831 /* lim is the index we can NOT write to */
1832 lim = pxtcp->inbuf.unacked;
1833 if (lim == 0) {
1834 lim = sz - 1; /* empty slot at the end */
1835 }
1836 else if (lim == 1 && beg != 0) {
1837 lim = sz; /* empty slot at the beginning */
1838 }
1839 else {
1840 --lim;
1841 }
1842
1843 if (beg == lim) {
1844 /*
1845 * Buffer is full, stop polling for POLLIN.
1846 *
1847 * pxtcp_pcb_sent() will re-enable POLLIN when guest ACKs
1848 * data, freeing space in the ring buffer.
1849 */
1850 *pstop = 1;
1851 return 0;
1852 }
1853
1854 if (beg < lim) {
1855 /* free space in one chunk */
1856 iovlen = 1;
1857 IOVEC_SET_LEN(iov[0], lim - beg);
1858 }
1859 else {
1860 /* free space in two chunks */
1861 iovlen = 2;
1862 IOVEC_SET_LEN(iov[0], sz - beg);
1863 IOVEC_SET_BASE(iov[1], &pxtcp->inbuf.buf[0]);
1864 IOVEC_SET_LEN(iov[1], lim);
1865 }
1866
1867 /*
1868 * TODO: This is where application-level proxy can hook into to
1869 * process inbound traffic.
1870 */
1871 nread = pxtcp_sock_recv(pxtcp, iov, iovlen);
1872
1873 if (nread > 0) {
1874 wrnew = beg + nread;
1875 if (wrnew >= sz) {
1876 wrnew -= sz;
1877 }
1878 pxtcp->inbuf.vacant = wrnew;
1879 DPRINTF2(("pxtcp %p: sock %d read %d bytes\n",
1880 (void *)pxtcp, pxtcp->sock, (int)nread));
1881 return nread;
1882 }
1883 else if (nread == 0) {
1884 *pstop = 1;
1885 pxtcp->inbound_close = 1;
1886 DPRINTF2(("pxtcp %p: sock %d read EOF\n",
1887 (void *)pxtcp, pxtcp->sock));
1888 return 1;
1889 }
1890 else {
1891 int sockerr = -nread;
1892
1893 if (proxy_error_is_transient(sockerr)) {
1894 /* haven't read anything, just return */
1895 DPRINTF2(("pxtcp %p: sock %d read cancelled\n",
1896 (void *)pxtcp, pxtcp->sock));
1897 return 0;
1898 }
1899 else {
1900 /* socket error! */
1901 DPRINTF0(("pxtcp %p: sock %d read: %R[sockerr]\n",
1902 (void *)pxtcp, pxtcp->sock, sockerr));
1903 return -sockerr;
1904 }
1905 }
1906}
1907
1908
1909#if !defined(RT_OS_WINDOWS)
1910static ssize_t
1911pxtcp_sock_recv(struct pxtcp *pxtcp, IOVEC *iov, size_t iovlen)
1912{
1913 struct msghdr mh;
1914 ssize_t nread;
1915
1916 memset(&mh, 0, sizeof(mh));
1917
1918 mh.msg_iov = iov;
1919 mh.msg_iovlen = iovlen;
1920
1921 nread = recvmsg(pxtcp->sock, &mh, 0);
1922 if (nread < 0) {
1923 nread = -SOCKERRNO();
1924 }
1925
1926 return nread;
1927}
1928#else /* RT_OS_WINDOWS */
1929static ssize_t
1930pxtcp_sock_recv(struct pxtcp *pxtcp, IOVEC *iov, size_t iovlen)
1931{
1932 DWORD flags;
1933 DWORD nread;
1934 int status;
1935
1936 flags = 0;
1937 status = WSARecv(pxtcp->sock, iov, (DWORD)iovlen, &nread,
1938 &flags, NULL, NULL);
1939 if (status == SOCKET_ERROR) {
1940 return -SOCKERRNO();
1941 }
1942
1943 return (ssize_t)nread;
1944}
1945#endif /* RT_OS_WINDOWS */
1946
1947
1948/**
1949 * Callback from poll manager (pxtcp::msg_inbound) to trigger output
1950 * from ringbuf to guest.
1951 */
1952static void
1953pxtcp_pcb_write_inbound(void *ctx)
1954{
1955 struct pxtcp *pxtcp = (struct pxtcp *)ctx;
1956 LWIP_ASSERT1(pxtcp != NULL);
1957
1958 if (pxtcp->pcb == NULL) {
1959 return;
1960 }
1961
1962 pxtcp_pcb_forward_inbound(pxtcp);
1963}
1964
1965
1966/**
1967 * tcp_poll() callback
1968 *
1969 * We swtich it on when tcp_write() or tcp_shutdown() fail with
1970 * ERR_MEM to prevent connection from stalling. If there are ACKs or
1971 * more inbound data then pxtcp_pcb_forward_inbound() will be
1972 * triggered again, but if neither happens, tcp_poll() comes to the
1973 * rescue.
1974 */
1975static err_t
1976pxtcp_pcb_poll(void *arg, struct tcp_pcb *pcb)
1977{
1978 struct pxtcp *pxtcp = (struct pxtcp *)arg;
1979 LWIP_UNUSED_ARG(pcb);
1980
1981 DPRINTF2(("%s: pxtcp %p; pcb %p\n",
1982 __func__, (void *)pxtcp, (void *)pxtcp->pcb));
1983
1984 pxtcp_pcb_forward_inbound(pxtcp);
1985
1986 /*
1987 * If the last thing holding up deletion of the pxtcp was failed
1988 * tcp_shutdown() and it succeeded, we may be the last callback.
1989 */
1990 pxtcp_pcb_maybe_deferred_delete(pxtcp);
1991
1992 return ERR_OK;
1993}
1994
1995
1996static void
1997pxtcp_pcb_schedule_poll(struct pxtcp *pxtcp)
1998{
1999 tcp_poll(pxtcp->pcb, pxtcp_pcb_poll, 0);
2000}
2001
2002
2003static void
2004pxtcp_pcb_cancel_poll(struct pxtcp *pxtcp)
2005{
2006 tcp_poll(pxtcp->pcb, NULL, 255);
2007}
2008
2009
2010/**
2011 * Forward inbound data from ring buffer to the guest.
2012 *
2013 * Scheduled by poll manager thread after it receives more data into
2014 * the ring buffer (we have more data to send).
2015
2016 * Also called from tcp_sent() callback when guest ACKs some data,
2017 * increasing pcb->snd_buf (we are permitted to send more data).
2018 *
2019 * Also called from tcp_poll() callback if previous attempt to forward
2020 * inbound data failed with ERR_MEM (we need to try again).
2021 */
2022static void
2023pxtcp_pcb_forward_inbound(struct pxtcp *pxtcp)
2024{
2025 struct tcp_pcb *pcb;
2026 size_t sndbuf;
2027 size_t beg, lim, sndlim;
2028 size_t toeob, tolim;
2029 size_t nsent;
2030 err_t error;
2031
2032 LWIP_ASSERT1(pxtcp != NULL);
2033 pcb = pxtcp->pcb;
2034 if (pcb == NULL) {
2035 return;
2036 }
2037
2038 if (/* __predict_false */ pcb->state < ESTABLISHED) {
2039 /*
2040 * If we have just confirmed accept of this connection, the
2041 * pcb is in SYN_RCVD state and we still haven't received the
2042 * ACK of our SYN. It's only in SYN_RCVD -> ESTABLISHED
2043 * transition that lwip decrements pcb->acked so that that ACK
2044 * is not reported to pxtcp_pcb_sent(). If we send something
2045 * now and immediately close (think "daytime", e.g.) while
2046 * still in SYN_RCVD state, we will move directly to
2047 * FIN_WAIT_1 and when our confirming SYN is ACK'ed lwip will
2048 * report it to pxtcp_pcb_sent().
2049 */
2050 DPRINTF2(("forward_inbound: pxtcp %p; pcb %p %s - later...\n",
2051 (void *)pxtcp, (void *)pcb, tcp_debug_state_str(pcb->state)));
2052 return;
2053 }
2054
2055
2056 beg = pxtcp->inbuf.unsent; /* private to lwip thread */
2057 lim = pxtcp->inbuf.vacant;
2058
2059 if (beg == lim) {
2060 if (pxtcp->inbound_close && !pxtcp->inbound_close_done) {
2061 pxtcp_pcb_forward_inbound_close(pxtcp);
2062 tcp_output(pcb);
2063 return;
2064 }
2065
2066 /*
2067 * Else, there's no data to send.
2068 *
2069 * If there is free space in the buffer, producer will
2070 * reschedule us as it receives more data and vacant (lim)
2071 * advances.
2072 *
2073 * If buffer is full when all data have been passed to
2074 * tcp_write() but not yet acknowledged, we will advance
2075 * unacked on ACK, freeing some space for producer to write to
2076 * (then see above).
2077 */
2078 return;
2079 }
2080
2081 sndbuf = tcp_sndbuf(pcb);
2082 if (sndbuf == 0) {
2083 /*
2084 * Can't send anything now. As guest ACKs some data, TCP will
2085 * call pxtcp_pcb_sent() callback and we will come here again.
2086 */
2087 return;
2088 }
2089
2090 nsent = 0;
2091
2092 /*
2093 * We have three limits to consider:
2094 * - how much data we have in the ringbuf
2095 * - how much data we are allowed to send
2096 * - ringbuf size
2097 */
2098 toeob = pxtcp->inbuf.bufsize - beg;
2099 if (lim < beg) { /* lim wrapped */
2100 if (sndbuf < toeob) { /* but we are limited by sndbuf */
2101 /* so beg is not going to wrap, treat sndbuf as lim */
2102 lim = beg + sndbuf; /* ... and proceed to the simple case */
2103 }
2104 else { /* we are limited by the end of the buffer, beg will wrap */
2105 u8_t maybemore;
2106 if (toeob == sndbuf || lim == 0) {
2107 maybemore = 0;
2108 }
2109 else {
2110 maybemore = TCP_WRITE_FLAG_MORE;
2111 }
2112
2113 error = tcp_write(pcb, &pxtcp->inbuf.buf[beg], toeob, maybemore);
2114 if (error != ERR_OK) {
2115 goto writeerr;
2116 }
2117 nsent += toeob;
2118 pxtcp->inbuf.unsent = 0; /* wrap */
2119
2120 if (maybemore) {
2121 beg = 0;
2122 sndbuf -= toeob;
2123 }
2124 else {
2125 /* we are done sending, but ... */
2126 goto check_inbound_close;
2127 }
2128 }
2129 }
2130
2131 LWIP_ASSERT1(beg < lim);
2132 sndlim = beg + sndbuf;
2133 if (lim > sndlim) {
2134 lim = sndlim;
2135 }
2136 tolim = lim - beg;
2137 if (tolim > 0) {
2138 error = tcp_write(pcb, &pxtcp->inbuf.buf[beg], (u16_t)tolim, 0);
2139 if (error != ERR_OK) {
2140 goto writeerr;
2141 }
2142 nsent += tolim;
2143 pxtcp->inbuf.unsent = lim;
2144 }
2145
2146 check_inbound_close:
2147 if (pxtcp->inbound_close && pxtcp->inbuf.unsent == pxtcp->inbuf.vacant) {
2148 pxtcp_pcb_forward_inbound_close(pxtcp);
2149 }
2150
2151 DPRINTF2(("forward_inbound: pxtcp %p, pcb %p: sent %d bytes\n",
2152 (void *)pxtcp, (void *)pcb, (int)nsent));
2153 tcp_output(pcb);
2154 pxtcp_pcb_cancel_poll(pxtcp);
2155 return;
2156
2157 writeerr:
2158 if (error == ERR_MEM) {
2159 if (nsent > 0) { /* first write succeeded, second failed */
2160 DPRINTF2(("forward_inbound: pxtcp %p, pcb %p: sent %d bytes only\n",
2161 (void *)pxtcp, (void *)pcb, (int)nsent));
2162 tcp_output(pcb);
2163 }
2164 DPRINTF(("forward_inbound: pxtcp %p, pcb %p: ERR_MEM\n",
2165 (void *)pxtcp, (void *)pcb));
2166 pxtcp_pcb_schedule_poll(pxtcp);
2167 }
2168 else {
2169 DPRINTF(("forward_inbound: pxtcp %p, pcb %p: %s\n",
2170 (void *)pxtcp, (void *)pcb, proxy_lwip_strerr(error)));
2171
2172 /* XXX: We shouldn't get ERR_ARG. Check ERR_CONN conditions early? */
2173 LWIP_ASSERT1(error == ERR_MEM);
2174 }
2175}
2176
2177
2178static void
2179pxtcp_pcb_forward_inbound_close(struct pxtcp *pxtcp)
2180{
2181 struct tcp_pcb *pcb;
2182 err_t error;
2183
2184 LWIP_ASSERT1(pxtcp != NULL);
2185 LWIP_ASSERT1(pxtcp->inbound_close);
2186 LWIP_ASSERT1(!pxtcp->inbound_close_done);
2187 LWIP_ASSERT1(pxtcp->inbuf.unsent == pxtcp->inbuf.vacant);
2188
2189 pcb = pxtcp->pcb;
2190 LWIP_ASSERT1(pcb != NULL);
2191
2192 DPRINTF(("inbound_close: pxtcp %p; pcb %p: %s\n",
2193 (void *)pxtcp, (void *)pcb, tcp_debug_state_str(pcb->state)));
2194
2195 error = tcp_shutdown(pcb, /*RX*/ 0, /*TX*/ 1);
2196 if (error != ERR_OK) {
2197 DPRINTF(("inbound_close: pxtcp %p; pcb %p:"
2198 " tcp_shutdown: error=%s\n",
2199 (void *)pxtcp, (void *)pcb, proxy_lwip_strerr(error)));
2200 pxtcp_pcb_schedule_poll(pxtcp);
2201 return;
2202 }
2203
2204 pxtcp_pcb_cancel_poll(pxtcp);
2205 pxtcp->inbound_close_done = 1;
2206
2207
2208 /*
2209 * If we have already done outbound close previously (passive
2210 * close on the pcb), then we must not hold onto a pcb in LAST_ACK
2211 * state since those will be deleted by lwip when that last ack
2212 * comes from the guest.
2213 *
2214 * NB: We do NOT check for deferred delete here, even though we
2215 * have just set one of its conditions, inbound_close_done. We
2216 * let pcb callbacks that called us do that. It's simpler and
2217 * cleaner that way.
2218 */
2219 if (pxtcp->outbound_close_done && pxtcp_pcb_forward_inbound_done(pxtcp)) {
2220 pxtcp_pcb_dissociate(pxtcp);
2221 }
2222}
2223
2224
2225/**
2226 * Check that all forwarded inbound data is sent and acked, and that
2227 * inbound close is scheduled (we aren't called back when it's acked).
2228 */
2229DECLINLINE(int)
2230pxtcp_pcb_forward_inbound_done(const struct pxtcp *pxtcp)
2231{
2232 return (pxtcp->inbound_close_done /* also implies that all data forwarded */
2233 && pxtcp->inbuf.unacked == pxtcp->inbuf.unsent);
2234}
2235
2236
2237/**
2238 * tcp_sent() callback - guest acknowledged len bytes.
2239 *
2240 * We can advance inbuf::unacked index, making more free space in the
2241 * ringbuf and wake up producer on poll manager thread.
2242 *
2243 * We can also try to send more data if we have any since pcb->snd_buf
2244 * was increased and we are now permitted to send more.
2245 */
2246static err_t
2247pxtcp_pcb_sent(void *arg, struct tcp_pcb *pcb, u16_t len)
2248{
2249 struct pxtcp *pxtcp = (struct pxtcp *)arg;
2250 size_t unacked;
2251
2252 LWIP_ASSERT1(pxtcp != NULL);
2253 LWIP_ASSERT1(pxtcp->pcb == pcb);
2254 LWIP_ASSERT1(pcb->callback_arg == pxtcp);
2255 LWIP_UNUSED_ARG(pcb); /* only in assert */
2256
2257 DPRINTF2(("%s: pxtcp %p; pcb %p: +%d ACKed:"
2258 " unacked %d, unsent %d, vacant %d\n",
2259 __func__, (void *)pxtcp, (void *)pcb, (int)len,
2260 (int)pxtcp->inbuf.unacked,
2261 (int)pxtcp->inbuf.unsent,
2262 (int)pxtcp->inbuf.vacant));
2263
2264 if (/* __predict_false */ len == 0) {
2265 /* we are notified to start pulling */
2266 LWIP_ASSERT1(!pxtcp->inbound_close);
2267 LWIP_ASSERT1(pxtcp->inbound_pull);
2268
2269 unacked = pxtcp->inbuf.unacked;
2270 }
2271 else {
2272 /*
2273 * Advance unacked index. Guest acknowledged the data, so it
2274 * won't be needed again for potential retransmits.
2275 */
2276 unacked = pxtcp->inbuf.unacked + len;
2277 if (unacked > pxtcp->inbuf.bufsize) {
2278 unacked -= pxtcp->inbuf.bufsize;
2279 }
2280 pxtcp->inbuf.unacked = unacked;
2281 }
2282
2283 /* arrange for more inbound data */
2284 if (!pxtcp->inbound_close) {
2285 if (!pxtcp->inbound_pull) {
2286 /* wake up producer, in case it has stopped polling for POLLIN */
2287 pxtcp_chan_send_weak(POLLMGR_CHAN_PXTCP_POLLIN, pxtcp);
2288#ifdef RT_OS_WINDOWS
2289 /**
2290 * We have't got enought room in ring buffer to read atm,
2291 * but we don't want to lose notification from WSAW4ME when
2292 * space would be available, so we reset event with empty recv
2293 */
2294 recv(pxtcp->sock, NULL, 0, 0);
2295#endif
2296 }
2297 else {
2298 ssize_t nread;
2299 int stop_pollin; /* ignored */
2300
2301 nread = pxtcp_sock_read(pxtcp, &stop_pollin);
2302
2303 if (nread < 0) {
2304 int sockerr = -(int)nread;
2305 LWIP_UNUSED_ARG(sockerr);
2306 DPRINTF0(("%s: sock %d: %R[sockerr]\n",
2307 __func__, pxtcp->sock, sockerr));
2308
2309#if HAVE_TCP_POLLHUP == POLLIN /* see counterpart in pxtcp_pmgr_pump() */
2310 /*
2311 * It may still be registered with poll manager for POLLOUT.
2312 */
2313 pxtcp_chan_send_weak(POLLMGR_CHAN_PXTCP_RESET, pxtcp);
2314 return ERR_OK;
2315#else
2316 /*
2317 * It is no longer registered with poll manager so we
2318 * can kill it directly.
2319 */
2320 pxtcp_pcb_reset_pxtcp(pxtcp);
2321 return ERR_ABRT;
2322#endif
2323 }
2324 }
2325 }
2326
2327 /* forward more data if we can */
2328 if (!pxtcp->inbound_close_done) {
2329 pxtcp_pcb_forward_inbound(pxtcp);
2330
2331 /*
2332 * NB: we might have dissociated from a pcb that transitioned
2333 * to LAST_ACK state, so don't refer to pcb below.
2334 */
2335 }
2336
2337
2338 /* have we got all the acks? */
2339 if (pxtcp->inbound_close /* no more new data */
2340 && pxtcp->inbuf.unsent == pxtcp->inbuf.vacant /* all data is sent */
2341 && unacked == pxtcp->inbuf.unsent) /* ... and is acked */
2342 {
2343 char *buf;
2344
2345 DPRINTF(("%s: pxtcp %p; pcb %p; all data ACKed\n",
2346 __func__, (void *)pxtcp, (void *)pxtcp->pcb));
2347
2348 /* no more retransmits, so buf is not needed */
2349 buf = pxtcp->inbuf.buf;
2350 pxtcp->inbuf.buf = NULL;
2351 free(buf);
2352
2353 /* no more acks, so no more callbacks */
2354 if (pxtcp->pcb != NULL) {
2355 tcp_sent(pxtcp->pcb, NULL);
2356 }
2357
2358 /*
2359 * We may be the last callback for this pcb if we have also
2360 * successfully forwarded inbound_close.
2361 */
2362 pxtcp_pcb_maybe_deferred_delete(pxtcp);
2363 }
2364
2365 return ERR_OK;
2366}
2367
2368
2369/**
2370 * Callback from poll manager (pxtcp::msg_inpull) to switch
2371 * pxtcp_pcb_sent() to actively pull the last bits of input. See
2372 * POLLHUP comment in pxtcp_pmgr_pump().
2373 *
2374 * pxtcp::sock is deregistered from poll manager after this callback
2375 * is scheduled.
2376 */
2377static void
2378pxtcp_pcb_pull_inbound(void *ctx)
2379{
2380 struct pxtcp *pxtcp = (struct pxtcp *)ctx;
2381 LWIP_ASSERT1(pxtcp != NULL);
2382
2383 if (pxtcp->pcb == NULL) {
2384 DPRINTF(("%s: pxtcp %p: PCB IS GONE\n", __func__, (void *)pxtcp));
2385 pxtcp_pcb_reset_pxtcp(pxtcp);
2386 return;
2387 }
2388
2389 pxtcp->inbound_pull = 1;
2390 if (pxtcp->outbound_close_done) {
2391 DPRINTF(("%s: pxtcp %p: pcb %p (deferred delete)\n",
2392 __func__, (void *)pxtcp, (void *)pxtcp->pcb));
2393 pxtcp->deferred_delete = 1;
2394 }
2395 else {
2396 DPRINTF(("%s: pxtcp %p: pcb %p\n",
2397 __func__, (void *)pxtcp, (void *)pxtcp->pcb));
2398 }
2399
2400 pxtcp_pcb_sent(pxtcp, pxtcp->pcb, 0);
2401}
2402
2403
2404/**
2405 * tcp_err() callback.
2406 *
2407 * pcb is not passed to this callback since it may be already
2408 * deallocated by the stack, but we can't do anything useful with it
2409 * anyway since connection is gone.
2410 */
2411static void
2412pxtcp_pcb_err(void *arg, err_t error)
2413{
2414 struct pxtcp *pxtcp = (struct pxtcp *)arg;
2415 LWIP_ASSERT1(pxtcp != NULL);
2416
2417 /*
2418 * ERR_CLSD is special - it is reported here when:
2419 *
2420 * . guest has already half-closed
2421 * . we send FIN to guest when external half-closes
2422 * . guest acks that FIN
2423 *
2424 * Since connection is closed but receive has been already closed
2425 * lwip can only report this via tcp_err. At this point the pcb
2426 * is still alive, so we can peek at it if need be.
2427 *
2428 * The interesting twist is when the ACK from guest that akcs our
2429 * FIN also acks some data. In this scenario lwip will NOT call
2430 * tcp_sent() callback with the ACK for that last bit of data but
2431 * instead will call tcp_err with ERR_CLSD right away. Since that
2432 * ACK also acknowledges all the data, we should run some of
2433 * pxtcp_pcb_sent() logic here.
2434 */
2435 if (error == ERR_CLSD) {
2436 struct tcp_pcb *pcb = pxtcp->pcb; /* still alive */
2437
2438 DPRINTF2(("ERR_CLSD: pxtcp %p; pcb %p:"
2439 " pcb->acked %d;"
2440 " unacked %d, unsent %d, vacant %d\n",
2441 (void *)pxtcp, (void *)pcb,
2442 pcb->acked,
2443 (int)pxtcp->inbuf.unacked,
2444 (int)pxtcp->inbuf.unsent,
2445 (int)pxtcp->inbuf.vacant));
2446
2447 LWIP_ASSERT1(pxtcp->pcb == pcb);
2448 LWIP_ASSERT1(pcb->callback_arg == pxtcp);
2449
2450 if (pcb->acked > 0) {
2451 pxtcp_pcb_sent(pxtcp, pcb, pcb->acked);
2452 }
2453 return;
2454 }
2455
2456 DPRINTF0(("tcp_err: pxtcp=%p, error=%s\n",
2457 (void *)pxtcp, proxy_lwip_strerr(error)));
2458
2459 pxtcp->pcb = NULL; /* pcb is gone */
2460 if (pxtcp->deferred_delete) {
2461 pxtcp_pcb_reset_pxtcp(pxtcp);
2462 }
2463 else {
2464 pxtcp_chan_send_weak(POLLMGR_CHAN_PXTCP_RESET, pxtcp);
2465 }
2466}
Note: See TracBrowser for help on using the repository browser.

© 2024 Oracle Support Privacy / Do Not Sell My Info Terms of Use Trademark Policy Automated Access Etiquette