mirror of
https://github.com/yrutschle/sslh.git
synced 2025-04-08 13:20:04 +03:00
617 lines
18 KiB
C
617 lines
18 KiB
C
/*
|
|
sslh-select: mono-processus server
|
|
|
|
# Copyright (C) 2007-2021 Yves Rutschle
|
|
#
|
|
# This program is free software; you can redistribute it
|
|
# and/or modify it under the terms of the GNU General Public
|
|
# License as published by the Free Software Foundation; either
|
|
# version 2 of the License, or (at your option) any later
|
|
# version.
|
|
#
|
|
# This program is distributed in the hope that it will be
|
|
# useful, but WITHOUT ANY WARRANTY; without even the implied
|
|
# warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR
|
|
# PURPOSE. See the GNU General Public License for more
|
|
# details.
|
|
#
|
|
# The full text for the General Public License is here:
|
|
# http://www.gnu.org/licenses/gpl.html
|
|
|
|
*/
|
|
|
|
/* Why use select(2) rather than poll(2)?
|
|
* No real reason except that's how it was written at first. This article:
|
|
* https://daniel.haxx.se/docs/poll-vs-select.html suggests that over a few
|
|
* hundred file descriptors, both become very slow, so there is little
|
|
* incentive to move to poll() to support more than FD_SETSIZE (which is 1024
|
|
* on many Linux. To support large numbers of descriptors, either use the fork
|
|
* version, or we'll have to write a new version based on libev. */
|
|
|
|
#define __LINUX__
|
|
|
|
#include <limits.h>
|
|
|
|
#include "common.h"
|
|
#include "probe.h"
|
|
#include "udp-listener.h"
|
|
#include "collection.h"
|
|
#include "gap.h"
|
|
|
|
static int debug = 0;
|
|
|
|
const char* server_type = "sslh-select";
|
|
|
|
/* Global state for a select() loop */
|
|
struct select_info {
|
|
int max_fd; /* Highest fd number to pass to select() */
|
|
|
|
int num_probing; /* Number of connections currently probing
|
|
* We use this to know if we need to time out of
|
|
* select() */
|
|
gap_array* probing_list; /* Pointers to cnx that are in probing mode */
|
|
|
|
fd_set fds_r, fds_w; /* reference fd sets (used to init working copies) */
|
|
cnx_collection* collection; /* Collection of connections linked to this loop */
|
|
|
|
time_t next_timeout; /* time at which next UDP connection times out */
|
|
};
|
|
|
|
|
|
|
|
static int tidy_connection(struct connection *cnx, struct select_info* fd_info)
|
|
{
|
|
int i;
|
|
fd_set* fds = &fd_info->fds_r;
|
|
fd_set* fds2 = &fd_info->fds_w;
|
|
|
|
for (i = 0; i < 2; i++) {
|
|
if (cnx->q[i].fd != -1) {
|
|
if (cfg.verbose)
|
|
fprintf(stderr, "closing fd %d\n", cnx->q[i].fd);
|
|
|
|
FD_CLR(cnx->q[i].fd, fds);
|
|
FD_CLR(cnx->q[i].fd, fds2);
|
|
close(cnx->q[i].fd);
|
|
if (cnx->q[i].deferred_data)
|
|
free(cnx->q[i].deferred_data);
|
|
}
|
|
}
|
|
collection_remove_cnx(fd_info->collection, cnx);
|
|
return 0;
|
|
}
|
|
|
|
/* if fd becomes higher than FD_SETSIZE, things won't work so well with FD_SET
|
|
* and FD_CLR. Need to drop connections if we go above that limit */
|
|
static int fd_is_in_range(int fd) {
|
|
if (fd >= FD_SETSIZE) {
|
|
log_message(LOG_ERR, "too many open file descriptor to monitor them all -- dropping connection\n");
|
|
return 0;
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
/* Accepts a connection from the main socket and assigns it to an empty slot.
|
|
* If no slots are available, allocate another few. If that fails, drop the
|
|
* connexion */
|
|
static struct connection* accept_new_connection(int listen_socket, struct cnx_collection *collection)
|
|
{
|
|
int in_socket, res;
|
|
|
|
|
|
if (cfg.verbose) fprintf(stderr, "accepting from %d\n", listen_socket);
|
|
|
|
in_socket = accept(listen_socket, 0, 0);
|
|
CHECK_RES_RETURN(in_socket, "accept", NULL);
|
|
|
|
if (!fd_is_in_range(in_socket)) {
|
|
close(in_socket);
|
|
return NULL;
|
|
}
|
|
|
|
res = set_nonblock(in_socket);
|
|
if (res == -1) {
|
|
close(in_socket);
|
|
return NULL;
|
|
}
|
|
|
|
struct connection* cnx = collection_alloc_cnx_from_fd(collection, in_socket);
|
|
if (!cnx) {
|
|
close(in_socket);
|
|
return NULL;
|
|
}
|
|
|
|
return cnx;
|
|
}
|
|
|
|
|
|
/* Connect queue 1 of connection to SSL; returns new file descriptor */
|
|
static int connect_queue(struct connection* cnx,
|
|
struct select_info* fd_info)
|
|
{
|
|
struct queue *q = &cnx->q[1];
|
|
|
|
q->fd = connect_addr(cnx, cnx->q[0].fd, NON_BLOCKING);
|
|
if ((q->fd != -1) && fd_is_in_range(q->fd)) {
|
|
log_connection(NULL, cnx);
|
|
flush_deferred(q);
|
|
if (q->deferred_data) {
|
|
FD_SET(q->fd, &fd_info->fds_w);
|
|
FD_CLR(cnx->q[0].fd, &fd_info->fds_r);
|
|
}
|
|
FD_SET(q->fd, &fd_info->fds_r);
|
|
collection_add_fd(fd_info->collection, cnx, q->fd);
|
|
return q->fd;
|
|
} else {
|
|
tidy_connection(cnx, fd_info);
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
/* shovels data from active fd to the other
|
|
returns after one socket closed or operation would block
|
|
*/
|
|
static void shovel(struct connection *cnx, int active_fd, struct select_info* fd_info)
|
|
{
|
|
struct queue *read_q, *write_q;
|
|
|
|
read_q = &cnx->q[active_fd];
|
|
write_q = &cnx->q[1-active_fd];
|
|
|
|
if (cfg.verbose)
|
|
fprintf(stderr, "activity on fd%d\n", read_q->fd);
|
|
|
|
switch(fd2fd(write_q, read_q)) {
|
|
case -1:
|
|
case FD_CNXCLOSED:
|
|
tidy_connection(cnx, fd_info);
|
|
break;
|
|
|
|
case FD_STALLED:
|
|
FD_SET(write_q->fd, &fd_info->fds_w);
|
|
FD_CLR(read_q->fd, &fd_info->fds_r);
|
|
break;
|
|
|
|
default: /* Nothing */
|
|
break;
|
|
}
|
|
}
|
|
|
|
/* shovels data from one fd to the other and vice-versa
|
|
returns after one socket closed
|
|
*/
|
|
static void shovel_single(struct connection *cnx)
|
|
{
|
|
fd_set fds_r, fds_w;
|
|
int res, i;
|
|
int max_fd = MAX(cnx->q[0].fd, cnx->q[1].fd) + 1;
|
|
|
|
FD_ZERO(&fds_r);
|
|
FD_ZERO(&fds_w);
|
|
while (1) {
|
|
for (i = 0; i < 2; i++) {
|
|
if (cnx->q[i].deferred_data_size) {
|
|
FD_SET(cnx->q[i].fd, &fds_w);
|
|
FD_CLR(cnx->q[1-i].fd, &fds_r);
|
|
} else {
|
|
FD_CLR(cnx->q[i].fd, &fds_w);
|
|
FD_SET(cnx->q[1-i].fd, &fds_r);
|
|
}
|
|
}
|
|
|
|
res = select(
|
|
max_fd,
|
|
&fds_r,
|
|
&fds_w,
|
|
NULL,
|
|
NULL
|
|
);
|
|
CHECK_RES_DIE(res, "select");
|
|
|
|
for (i = 0; i < 2; i++) {
|
|
if (FD_ISSET(cnx->q[i].fd, &fds_w)) {
|
|
res = flush_deferred(&cnx->q[i]);
|
|
if ((res == -1) && ((errno == EPIPE) || (errno == ECONNRESET))) {
|
|
if (cfg.verbose)
|
|
fprintf(stderr, "%s socket closed\n", i ? "server" : "client");
|
|
return;
|
|
}
|
|
}
|
|
if (FD_ISSET(cnx->q[i].fd, &fds_r)) {
|
|
res = fd2fd(&cnx->q[1-i], &cnx->q[i]);
|
|
if (!res) {
|
|
if (cfg.verbose)
|
|
fprintf(stderr, "socket closed\n");
|
|
return;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Child process that makes internal connection and proxies
|
|
*/
|
|
static void connect_proxy(struct connection *cnx)
|
|
{
|
|
int in_socket;
|
|
int out_socket;
|
|
|
|
/* Minimize the file descriptor value to help select() */
|
|
in_socket = dup(cnx->q[0].fd);
|
|
if (in_socket == -1) {
|
|
in_socket = cnx->q[0].fd;
|
|
} else {
|
|
close(cnx->q[0].fd);
|
|
cnx->q[0].fd = in_socket;
|
|
}
|
|
|
|
/* Connect the target socket */
|
|
out_socket = connect_addr(cnx, in_socket, BLOCKING);
|
|
CHECK_RES_DIE(out_socket, "connect");
|
|
|
|
cnx->q[1].fd = out_socket;
|
|
|
|
log_connection(NULL, cnx);
|
|
|
|
shovel_single(cnx);
|
|
|
|
close(in_socket);
|
|
close(out_socket);
|
|
|
|
if (cfg.verbose)
|
|
fprintf(stderr, "connection closed down\n");
|
|
|
|
exit(0);
|
|
}
|
|
|
|
/* Removes cnx from probing list */
|
|
static void remove_probing_cnx(struct select_info* fd_info, struct connection* cnx)
|
|
{
|
|
gap_remove_ptr(fd_info->probing_list, cnx, fd_info->num_probing);
|
|
fd_info->num_probing--;
|
|
}
|
|
|
|
static void add_probing_cnx(struct select_info* fd_info, struct connection* cnx)
|
|
{
|
|
gap_set(fd_info->probing_list, fd_info->num_probing, cnx);
|
|
fd_info->num_probing++;
|
|
}
|
|
|
|
|
|
/* Process read activity on a socket in probe state
|
|
* IN/OUT cnx: connection data, updated if connected
|
|
* IN/OUT info: updated if connected
|
|
* */
|
|
|
|
static void probing_read_process(struct connection* cnx,
|
|
struct select_info* fd_info)
|
|
{
|
|
int res;
|
|
|
|
/* If timed out it's SSH, otherwise the client sent
|
|
* data so probe the protocol */
|
|
if ((cnx->probe_timeout < time(NULL))) {
|
|
cnx->proto = timeout_protocol();
|
|
if (cfg.verbose)
|
|
log_message(LOG_INFO,
|
|
"timed out, connect to %s\n",
|
|
cnx->proto->name);
|
|
} else {
|
|
res = probe_client_protocol(cnx);
|
|
if (res == PROBE_AGAIN)
|
|
return;
|
|
}
|
|
|
|
remove_probing_cnx(fd_info, cnx);
|
|
cnx->state = ST_SHOVELING;
|
|
|
|
/* libwrap check if required for this protocol */
|
|
if (cnx->proto->service &&
|
|
check_access_rights(cnx->q[0].fd, cnx->proto->service)) {
|
|
tidy_connection(cnx, fd_info);
|
|
res = -1;
|
|
} else if (cnx->proto->fork) {
|
|
switch (fork()) {
|
|
case 0: /* child */
|
|
/* TODO: close all file descriptors except 2 */
|
|
/* free(cnx); */
|
|
connect_proxy(cnx);
|
|
exit(0);
|
|
case -1: log_message(LOG_ERR, "fork failed: err %d: %s\n", errno, strerror(errno));
|
|
break;
|
|
default: /* parent */
|
|
break;
|
|
}
|
|
tidy_connection(cnx, fd_info);
|
|
res = -1;
|
|
} else {
|
|
res = connect_queue(cnx, fd_info);
|
|
}
|
|
|
|
if (res >= fd_info->max_fd)
|
|
fd_info->max_fd = res + 1;;
|
|
}
|
|
|
|
|
|
/* Returns the queue index that contains the specified file descriptor */
|
|
int active_queue(struct connection* cnx, int fd)
|
|
{
|
|
if (cnx->q[0].fd == fd) return 0;
|
|
if (cnx->q[1].fd == fd) return 1;
|
|
|
|
log_message(LOG_ERR, "file descriptor %d not found in connection object\n", fd);
|
|
return -1;
|
|
}
|
|
|
|
/* Process a connection that is active in read */
|
|
static void tcp_read_process(struct select_info* fd_info,
|
|
int fd)
|
|
{
|
|
if (debug) fprintf(stderr, "cnx_read_process fd %d\n", fd);
|
|
|
|
cnx_collection* collection = fd_info->collection;
|
|
struct connection* cnx = collection_get_cnx_from_fd(collection, fd);
|
|
/* Determine active queue (0 or 1): if fd is that of q[1], active_q = 1,
|
|
* otherwise it's 0 */
|
|
int active_q = active_queue(cnx, fd);
|
|
|
|
switch (cnx->state) {
|
|
|
|
case ST_PROBING:
|
|
if (active_q == 1) {
|
|
fprintf(stderr, "Activity on fd2 while probing, impossible\n");
|
|
dump_connection(cnx);
|
|
exit(1);
|
|
}
|
|
|
|
probing_read_process(cnx, fd_info);
|
|
|
|
break;
|
|
|
|
case ST_SHOVELING:
|
|
shovel(cnx, active_q, fd_info);
|
|
break;
|
|
|
|
default: /* illegal */
|
|
log_message(LOG_ERR, "Illegal connection state %d\n", cnx->state);
|
|
dump_connection(cnx);
|
|
exit(1);
|
|
}
|
|
}
|
|
|
|
static void cnx_read_process(struct select_info* fd_info, int fd)
|
|
{
|
|
cnx_collection* collection = fd_info->collection;
|
|
struct connection* cnx = collection_get_cnx_from_fd(collection, fd);
|
|
switch (cnx->type) {
|
|
case SOCK_STREAM:
|
|
tcp_read_process(fd_info, fd);
|
|
break;
|
|
|
|
case SOCK_DGRAM:
|
|
udp_s2c_forward(cnx);
|
|
break;
|
|
|
|
default:
|
|
log_message(LOG_ERR, "cnx_read_process: Illegal connection type %d\n", cnx->type);
|
|
dump_connection(cnx);
|
|
exit(1);
|
|
}
|
|
|
|
}
|
|
|
|
/* Process a connection that is active in write */
|
|
static void cnx_write_process(struct select_info* fd_info, int fd)
|
|
{
|
|
if (debug) fprintf(stderr, "cnx_write_process fd %d\n", fd);
|
|
|
|
struct connection* cnx = collection_get_cnx_from_fd(fd_info->collection, fd);
|
|
int res;
|
|
int queue = active_queue(cnx, fd);
|
|
|
|
res = flush_deferred(&cnx->q[queue]);
|
|
if ((res == -1) && ((errno == EPIPE) || (errno == ECONNRESET))) {
|
|
if (cnx->state == ST_PROBING) remove_probing_cnx(fd_info, cnx);
|
|
tidy_connection(cnx, fd_info);
|
|
} else {
|
|
/* If no deferred data is left, stop monitoring the fd
|
|
* for write, and restart monitoring the other one for reads*/
|
|
if (!cnx->q[queue].deferred_data_size) {
|
|
FD_CLR(cnx->q[queue].fd, &fd_info->fds_w);
|
|
FD_SET(cnx->q[1-queue].fd, &fd_info->fds_r);
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Process a connection that accepts a socket
|
|
* (For UDP, this means all traffic coming from remote clients)
|
|
* */
|
|
void cnx_accept_process(struct select_info* fd_info, struct listen_endpoint* listen_socket)
|
|
{
|
|
int fd = listen_socket->socketfd;
|
|
int type = listen_socket->type;
|
|
struct connection* cnx;
|
|
int new_fd;
|
|
|
|
if (debug) fprintf(stderr, "cnx_accept_process fd %d\n", fd);
|
|
|
|
switch (type) {
|
|
case SOCK_STREAM:
|
|
cnx = accept_new_connection(fd, fd_info->collection);
|
|
|
|
if (cnx) {
|
|
add_probing_cnx(fd_info, cnx);
|
|
new_fd = cnx->q[0].fd;
|
|
}
|
|
break;
|
|
|
|
case SOCK_DGRAM:
|
|
new_fd = udp_c2s_forward(fd, fd_info->collection, fd_info->max_fd);
|
|
fprintf(stderr, "new_fd %d\n", new_fd);
|
|
if (new_fd == -1)
|
|
return;
|
|
break;
|
|
|
|
default:
|
|
log_message(LOG_ERR, "Inconsistent cnx type: %d\n", type);
|
|
exit(1);
|
|
return;
|
|
}
|
|
|
|
FD_SET(new_fd, &fd_info->fds_r);
|
|
if (new_fd >= fd_info->max_fd)
|
|
fd_info->max_fd = new_fd + 1;
|
|
|
|
}
|
|
|
|
/* Check all connections to see if a UDP connections has timed out, then free
|
|
* it. At the same time, keep track of the closest, next timeout. Only do the
|
|
* search through connections if that timeout actually happened. If the
|
|
* connection that would have timed out has had activity, it doesn't matter: we
|
|
* go through connections to find the next timeout, which was needed anyway. */
|
|
static void udp_timeouts(struct select_info* fd_info)
|
|
{
|
|
time_t now = time(NULL);
|
|
|
|
if (now < fd_info->next_timeout) return;
|
|
|
|
time_t next_timeout = INT_MAX;
|
|
|
|
for (int i = 0; i < fd_info->max_fd; i++) {
|
|
/* if it's either in read or write set, there is a connection
|
|
* behind that file descriptor */
|
|
if (FD_ISSET(i, &fd_info->fds_r) || FD_ISSET(i, &fd_info->fds_w)) {
|
|
struct connection* cnx = collection_get_cnx_from_fd(fd_info->collection, i);
|
|
if (cnx) {
|
|
time_t timeout = udp_timeout(cnx);
|
|
if (!timeout) continue; /* Not a UDP connection */
|
|
if (cnx && (timeout <= now)) {
|
|
if (cfg.verbose > 3)
|
|
fprintf(stderr, "timed out UDP %d\n", cnx->target_sock);
|
|
close(cnx->target_sock);
|
|
FD_CLR(i, &fd_info->fds_r);
|
|
FD_CLR(i, &fd_info->fds_w);
|
|
collection_remove_cnx(fd_info->collection, cnx);
|
|
} else {
|
|
if (timeout < next_timeout) next_timeout = timeout;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
if (next_timeout != INT_MAX)
|
|
fd_info->next_timeout = next_timeout;
|
|
}
|
|
|
|
/* Main loop: the idea is as follow:
|
|
* - fds_r and fds_w contain the file descriptors to monitor in read and write
|
|
* - When a file descriptor goes off, process it: read from it, write the data
|
|
* to its corresponding pair.
|
|
* - When a file descriptor blocks when writing, remove the read fd from fds_r,
|
|
* move the data to a deferred buffer, and add the write fd to fds_w. Deferred
|
|
* buffer is allocated dynamically.
|
|
* - When we can write to a file descriptor that has deferred data, we try to
|
|
* write as much as we can. Once all data is written, remove the fd from fds_w
|
|
* and add its corresponding pair to fds_r, free the buffer.
|
|
*
|
|
* That way, each pair of file descriptor (read from one, write to the other)
|
|
* is monitored either for read or for write, but never for both.
|
|
*/
|
|
void main_loop(struct listen_endpoint listen_sockets[], int num_addr_listen)
|
|
{
|
|
struct select_info fd_info = {0};
|
|
fd_set readfds, writefds; /* working read and write fd sets */
|
|
struct timeval tv;
|
|
int i, res;
|
|
|
|
fd_info.num_probing = 0;
|
|
FD_ZERO(&fd_info.fds_r);
|
|
FD_ZERO(&fd_info.fds_w);
|
|
fd_info.probing_list = gap_init(0);
|
|
|
|
for (i = 0; i < num_addr_listen; i++) {
|
|
FD_SET(listen_sockets[i].socketfd, &fd_info.fds_r);
|
|
set_nonblock(listen_sockets[i].socketfd);
|
|
}
|
|
fd_info.max_fd = listen_sockets[num_addr_listen-1].socketfd + 1;
|
|
|
|
fd_info.collection = collection_init(fd_info.max_fd);
|
|
|
|
while (1)
|
|
{
|
|
memset(&tv, 0, sizeof(tv));
|
|
tv.tv_sec = cfg.timeout;
|
|
|
|
memcpy(&readfds, &fd_info.fds_r, sizeof(readfds));
|
|
memcpy(&writefds, &fd_info.fds_w, sizeof(writefds));
|
|
|
|
if (cfg.verbose)
|
|
fprintf(stderr, "selecting... max_fd=%d num_probing=%d\n",
|
|
fd_info.max_fd, fd_info.num_probing);
|
|
res = select(fd_info.max_fd, &readfds, &writefds,
|
|
NULL, fd_info.num_probing ? &tv : NULL);
|
|
if (res < 0)
|
|
perror("select");
|
|
|
|
|
|
/* UDP timeouts: clear out connections after some idle time */
|
|
udp_timeouts(&fd_info);
|
|
|
|
/* Check main socket for new connections */
|
|
for (i = 0; i < num_addr_listen; i++) {
|
|
if (FD_ISSET(listen_sockets[i].socketfd, &readfds)) {
|
|
cnx_accept_process(&fd_info, &listen_sockets[i]);
|
|
|
|
/* don't also process it as a read socket */
|
|
FD_CLR(listen_sockets[i].socketfd, &readfds);
|
|
}
|
|
}
|
|
|
|
/* Check all sockets for write activity */
|
|
for (i = 0; i < fd_info.max_fd; i++) {
|
|
if (FD_ISSET(i, &writefds)) {
|
|
cnx_write_process(&fd_info, i);
|
|
}
|
|
}
|
|
|
|
/* Check sockets in probing state for timeouts */
|
|
for (i = 0; i < fd_info.num_probing; i++) {
|
|
struct connection* cnx = gap_get(fd_info.probing_list, i);
|
|
if (!cnx || cnx->state != ST_PROBING) {
|
|
log_message(LOG_ERR, "Inconsistent probing: cnx=%0xp\n", cnx);
|
|
if (cnx)
|
|
log_message(LOG_ERR, "Inconsistent probing: state=%d\n", cnx);
|
|
exit(1);
|
|
}
|
|
if (cnx->probe_timeout < time(NULL)) {
|
|
if (cfg.verbose)
|
|
fprintf(stderr, "timeout slot %d\n", i);
|
|
probing_read_process(cnx, &fd_info);
|
|
}
|
|
}
|
|
|
|
/* Check all sockets for read activity */
|
|
for (i = 0; i < fd_info.max_fd; i++) {
|
|
/* Check if it's active AND currently monitored (if a connection
|
|
* died, it gets tidied, which closes both sockets, but readfs does
|
|
* not know about that */
|
|
if (FD_ISSET(i, &readfds) && FD_ISSET(i, &fd_info.fds_r)) {
|
|
cnx_read_process(&fd_info, i);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
void start_shoveler(int listen_socket) {
|
|
fprintf(stderr, "inetd mode is not supported in select mode\n");
|
|
exit(1);
|
|
}
|
|
|
|
|
|
/* The actual main is in common.c: it's the same for both version of
|
|
* the server
|
|
*/
|
|
|
|
|