2003-10-08 04:04:08 +02:00
|
|
|
/* Copyright 2001,2002,2003 Roger Dingledine, Matej Pfajfar. */
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
/* See LICENSE for licensing information */
|
|
|
|
/* $Id$ */
|
2002-06-27 00:45:49 +02:00
|
|
|
|
|
|
|
#include "or.h"
|
|
|
|
|
2003-07-05 07:46:06 +02:00
|
|
|
/********* START PROTOTYPES **********/
|
|
|
|
|
2003-10-15 20:28:32 +02:00
|
|
|
static void dumpstats(int severity); /* log stats */
|
2003-07-05 07:46:06 +02:00
|
|
|
|
2002-06-27 00:45:49 +02:00
|
|
|
/********* START VARIABLES **********/
|
|
|
|
|
2003-08-11 22:22:48 +02:00
|
|
|
extern char *conn_type_to_string[];
|
2003-09-25 12:42:07 +02:00
|
|
|
extern char *conn_state_to_string[][_CONN_TYPE_MAX+1];
|
2003-08-11 22:22:48 +02:00
|
|
|
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
or_options_t options; /* command-line and config-file options */
|
2003-07-05 09:10:34 +02:00
|
|
|
int global_read_bucket; /* max number of bytes I can read this second */
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2003-10-02 22:00:38 +02:00
|
|
|
static int stats_prev_global_read_bucket;
|
|
|
|
static uint64_t stats_n_bytes_read = 0;
|
|
|
|
static long stats_n_seconds_reading = 0;
|
|
|
|
|
2002-07-05 08:27:23 +02:00
|
|
|
static connection_t *connection_array[MAXCONNECTIONS] =
|
2002-06-27 00:45:49 +02:00
|
|
|
{ NULL };
|
|
|
|
|
2002-09-03 20:44:24 +02:00
|
|
|
static struct pollfd poll_array[MAXCONNECTIONS];
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2002-07-05 08:27:23 +02:00
|
|
|
static int nfds=0; /* number of connections currently active */
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2003-08-12 10:04:31 +02:00
|
|
|
#ifndef MS_WINDOWS /* do signal stuff only on unix */
|
2002-09-22 00:41:48 +02:00
|
|
|
static int please_dumpstats=0; /* whether we should dump stats during the loop */
|
2003-09-21 08:15:43 +02:00
|
|
|
static int please_reset =0; /* whether we just got a sighup */
|
2003-08-12 08:41:53 +02:00
|
|
|
static int please_reap_children=0; /* whether we should waitpid for exited children*/
|
2003-08-12 10:04:31 +02:00
|
|
|
#endif /* signal stuff */
|
2002-09-22 00:41:48 +02:00
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
/* private keys */
|
|
|
|
static crypto_pk_env_t *onionkey=NULL;
|
|
|
|
static crypto_pk_env_t *linkkey=NULL;
|
|
|
|
static crypto_pk_env_t *identitykey=NULL;
|
2002-06-27 00:45:49 +02:00
|
|
|
|
|
|
|
/********* END VARIABLES ************/
|
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
void set_onion_key(crypto_pk_env_t *k) {
|
|
|
|
onionkey = k;
|
2002-09-28 02:52:59 +02:00
|
|
|
}
|
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
crypto_pk_env_t *get_onion_key(void) {
|
|
|
|
assert(onionkey);
|
|
|
|
return onionkey;
|
2002-09-28 02:52:59 +02:00
|
|
|
}
|
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
void set_link_key(crypto_pk_env_t *k)
|
|
|
|
{
|
|
|
|
linkkey = k;
|
|
|
|
}
|
|
|
|
|
|
|
|
crypto_pk_env_t *get_link_key(void)
|
|
|
|
{
|
|
|
|
assert(linkkey);
|
|
|
|
return linkkey;
|
2003-05-08 00:40:03 +02:00
|
|
|
}
|
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
void set_identity_key(crypto_pk_env_t *k) {
|
|
|
|
identitykey = k;
|
|
|
|
}
|
|
|
|
|
|
|
|
crypto_pk_env_t *get_identity_key(void) {
|
|
|
|
assert(identitykey);
|
|
|
|
return identitykey;
|
2003-05-08 00:40:03 +02:00
|
|
|
}
|
|
|
|
|
2002-07-05 08:27:23 +02:00
|
|
|
/****************************************************************************
|
|
|
|
*
|
|
|
|
* This section contains accessors and other methods on the connection_array
|
|
|
|
* and poll_array variables (which are global within this file and unavailable
|
|
|
|
* outside it).
|
|
|
|
*
|
|
|
|
****************************************************************************/
|
|
|
|
|
2002-06-27 00:45:49 +02:00
|
|
|
int connection_add(connection_t *conn) {
|
|
|
|
|
2002-09-03 20:36:40 +02:00
|
|
|
if(nfds >= options.MaxConn-1) {
|
2003-10-10 03:48:32 +02:00
|
|
|
log(LOG_WARN,"connection_add(): failing because nfds is too high.");
|
2002-06-27 00:45:49 +02:00
|
|
|
return -1;
|
|
|
|
}
|
2003-09-25 07:17:11 +02:00
|
|
|
|
2002-06-27 00:45:49 +02:00
|
|
|
conn->poll_index = nfds;
|
|
|
|
connection_set_poll_socket(conn);
|
|
|
|
connection_array[nfds] = conn;
|
|
|
|
|
|
|
|
/* zero these out here, because otherwise we'll inherit values from the previously freed one */
|
|
|
|
poll_array[nfds].events = 0;
|
|
|
|
poll_array[nfds].revents = 0;
|
|
|
|
|
|
|
|
nfds++;
|
|
|
|
|
2002-07-16 04:12:58 +02:00
|
|
|
log(LOG_INFO,"connection_add(): new conn type %d, socket %d, nfds %d.",conn->type, conn->s, nfds);
|
2002-06-27 00:45:49 +02:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
void connection_set_poll_socket(connection_t *conn) {
|
|
|
|
poll_array[conn->poll_index].fd = conn->s;
|
|
|
|
}
|
|
|
|
|
2003-10-10 03:48:32 +02:00
|
|
|
/* Remove the connection from the global list, and remove the
|
2003-10-09 20:45:14 +02:00
|
|
|
* corresponding poll entry. Calling this function will shift the last
|
|
|
|
* connection (if any) into the position occupied by conn.
|
|
|
|
*/
|
2002-06-27 00:45:49 +02:00
|
|
|
int connection_remove(connection_t *conn) {
|
|
|
|
int current_index;
|
|
|
|
|
|
|
|
assert(conn);
|
|
|
|
assert(nfds>0);
|
|
|
|
|
2002-07-22 06:08:37 +02:00
|
|
|
log(LOG_INFO,"connection_remove(): removing socket %d, nfds now %d",conn->s, nfds-1);
|
2003-10-10 03:48:32 +02:00
|
|
|
/* if it's an edge conn, remove it from the list
|
|
|
|
* of conn's on this circuit. If it's not on an edge,
|
|
|
|
* flush and send destroys for all circuits on this conn
|
|
|
|
*/
|
|
|
|
circuit_about_to_close_connection(conn);
|
2002-06-27 00:45:49 +02:00
|
|
|
|
|
|
|
current_index = conn->poll_index;
|
|
|
|
if(current_index == nfds-1) { /* this is the end */
|
|
|
|
nfds--;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* we replace this one with the one at the end, then free it */
|
|
|
|
nfds--;
|
|
|
|
poll_array[current_index].fd = poll_array[nfds].fd;
|
|
|
|
poll_array[current_index].events = poll_array[nfds].events;
|
|
|
|
poll_array[current_index].revents = poll_array[nfds].revents;
|
|
|
|
connection_array[current_index] = connection_array[nfds];
|
|
|
|
connection_array[current_index]->poll_index = current_index;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2003-09-30 21:06:22 +02:00
|
|
|
void get_connection_array(connection_t ***array, int *n) {
|
|
|
|
*array = connection_array;
|
|
|
|
*n = nfds;
|
2003-06-25 09:19:30 +02:00
|
|
|
}
|
|
|
|
|
2002-06-27 00:45:49 +02:00
|
|
|
void connection_watch_events(connection_t *conn, short events) {
|
|
|
|
|
|
|
|
assert(conn && conn->poll_index < nfds);
|
|
|
|
|
|
|
|
poll_array[conn->poll_index].events = events;
|
|
|
|
}
|
|
|
|
|
2003-09-07 12:24:40 +02:00
|
|
|
int connection_is_reading(connection_t *conn) {
|
|
|
|
return poll_array[conn->poll_index].events & POLLIN;
|
|
|
|
}
|
|
|
|
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
void connection_stop_reading(connection_t *conn) {
|
|
|
|
|
|
|
|
assert(conn && conn->poll_index < nfds);
|
|
|
|
|
2002-07-18 08:37:58 +02:00
|
|
|
log(LOG_DEBUG,"connection_stop_reading() called.");
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
if(poll_array[conn->poll_index].events & POLLIN)
|
|
|
|
poll_array[conn->poll_index].events -= POLLIN;
|
|
|
|
}
|
|
|
|
|
|
|
|
void connection_start_reading(connection_t *conn) {
|
|
|
|
|
|
|
|
assert(conn && conn->poll_index < nfds);
|
|
|
|
|
|
|
|
poll_array[conn->poll_index].events |= POLLIN;
|
|
|
|
}
|
|
|
|
|
2002-07-18 08:37:58 +02:00
|
|
|
void connection_stop_writing(connection_t *conn) {
|
|
|
|
|
|
|
|
assert(conn && conn->poll_index < nfds);
|
|
|
|
|
|
|
|
if(poll_array[conn->poll_index].events & POLLOUT)
|
|
|
|
poll_array[conn->poll_index].events -= POLLOUT;
|
|
|
|
}
|
|
|
|
|
|
|
|
void connection_start_writing(connection_t *conn) {
|
|
|
|
|
|
|
|
assert(conn && conn->poll_index < nfds);
|
|
|
|
|
|
|
|
poll_array[conn->poll_index].events |= POLLOUT;
|
|
|
|
}
|
|
|
|
|
2003-07-05 07:46:06 +02:00
|
|
|
static void conn_read(int i) {
|
2003-09-28 08:48:20 +02:00
|
|
|
connection_t *conn = connection_array[i];
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2003-09-28 08:48:20 +02:00
|
|
|
/* see http://www.greenend.org.uk/rjk/2001/06/poll.html for
|
|
|
|
* discussion of POLLIN vs POLLHUP */
|
2003-09-05 08:04:03 +02:00
|
|
|
if(!(poll_array[i].revents & (POLLIN|POLLHUP|POLLERR)))
|
2003-10-09 20:45:14 +02:00
|
|
|
if(!connection_is_reading(conn) ||
|
|
|
|
!connection_has_pending_tls_data(conn))
|
2003-09-28 08:48:20 +02:00
|
|
|
return; /* this conn should not read */
|
2003-08-14 19:13:52 +02:00
|
|
|
|
cleanups, bugfixes, more verbose logs
Fixed up the assert_*_ok funcs some (more work remains)
Changed config so it reads either /etc/torrc or the -f arg, never both
Finally tracked down a nasty bug with our use of tls:
It turns out that if you ask SSL_read() for no more than n bytes, it
will read the entire record from the network (and maybe part of the next
record, I'm not sure), give you n bytes of it, and keep the remaining
bytes internally. This is fine, except our poll-for-read looks at the
network, and there are no bytes pending on the network, so we never know
to ask SSL_read() for more bytes. Currently I've hacked it so if we ask
for n bytes and it returns n bytes, then it reads again right then. This
will interact poorly with our rate limiting; we need a cleaner solution.
svn:r481
2003-09-24 23:24:52 +02:00
|
|
|
log_fn(LOG_DEBUG,"socket %d wants to read.",conn->s);
|
2003-09-23 21:47:41 +02:00
|
|
|
|
|
|
|
assert_connection_ok(conn, time(NULL));
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
|
2003-09-05 08:04:03 +02:00
|
|
|
if(
|
|
|
|
/* XXX does POLLHUP also mean it's definitely broken? */
|
2003-08-14 19:13:52 +02:00
|
|
|
#ifdef MS_WINDOWS
|
2003-09-05 08:04:03 +02:00
|
|
|
(poll_array[i].revents & POLLERR) ||
|
2003-08-14 19:13:52 +02:00
|
|
|
#endif
|
2003-09-05 08:04:03 +02:00
|
|
|
connection_handle_read(conn) < 0)
|
|
|
|
{
|
|
|
|
/* this connection is broken. remove it */
|
|
|
|
log_fn(LOG_INFO,"%s connection broken, removing.", conn_type_to_string[conn->type]);
|
|
|
|
connection_remove(conn);
|
|
|
|
connection_free(conn);
|
|
|
|
if(i<nfds) { /* we just replaced the one at i with a new one. process it too. */
|
2003-07-05 07:46:06 +02:00
|
|
|
conn_read(i);
|
2003-09-05 08:04:03 +02:00
|
|
|
}
|
2003-09-23 21:47:41 +02:00
|
|
|
} else assert_connection_ok(conn, time(NULL));
|
2002-06-27 00:45:49 +02:00
|
|
|
}
|
|
|
|
|
2003-07-05 07:46:06 +02:00
|
|
|
static void conn_write(int i) {
|
2002-06-27 00:45:49 +02:00
|
|
|
connection_t *conn;
|
|
|
|
|
2003-09-05 08:04:03 +02:00
|
|
|
if(!(poll_array[i].revents & POLLOUT))
|
|
|
|
return; /* this conn doesn't want to write */
|
|
|
|
|
2003-07-05 07:46:06 +02:00
|
|
|
conn = connection_array[i];
|
cleanups, bugfixes, more verbose logs
Fixed up the assert_*_ok funcs some (more work remains)
Changed config so it reads either /etc/torrc or the -f arg, never both
Finally tracked down a nasty bug with our use of tls:
It turns out that if you ask SSL_read() for no more than n bytes, it
will read the entire record from the network (and maybe part of the next
record, I'm not sure), give you n bytes of it, and keep the remaining
bytes internally. This is fine, except our poll-for-read looks at the
network, and there are no bytes pending on the network, so we never know
to ask SSL_read() for more bytes. Currently I've hacked it so if we ask
for n bytes and it returns n bytes, then it reads again right then. This
will interact poorly with our rate limiting; we need a cleaner solution.
svn:r481
2003-09-24 23:24:52 +02:00
|
|
|
log_fn(LOG_DEBUG,"socket %d wants to write.",conn->s);
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2003-09-23 21:47:41 +02:00
|
|
|
assert_connection_ok(conn, time(NULL));
|
|
|
|
|
2003-09-05 08:04:03 +02:00
|
|
|
if(connection_handle_write(conn) < 0) { /* this connection is broken. remove it. */
|
2003-09-26 12:03:50 +02:00
|
|
|
log_fn(LOG_INFO,"%s connection broken, removing.", conn_type_to_string[conn->type]);
|
2003-07-05 07:46:06 +02:00
|
|
|
connection_remove(conn);
|
|
|
|
connection_free(conn);
|
2003-09-05 08:04:03 +02:00
|
|
|
if(i<nfds) { /* we just replaced the one at i with a new one. process it too. */
|
2003-07-05 07:46:06 +02:00
|
|
|
conn_write(i);
|
2002-06-27 00:45:49 +02:00
|
|
|
}
|
2003-09-23 21:47:41 +02:00
|
|
|
} else assert_connection_ok(conn, time(NULL));
|
2002-06-27 00:45:49 +02:00
|
|
|
}
|
|
|
|
|
2003-10-09 20:45:14 +02:00
|
|
|
static void conn_close_if_marked(int i) {
|
2002-06-27 00:45:49 +02:00
|
|
|
connection_t *conn;
|
|
|
|
|
|
|
|
conn = connection_array[i];
|
2003-09-23 21:47:41 +02:00
|
|
|
assert_connection_ok(conn, time(NULL));
|
2002-06-27 00:45:49 +02:00
|
|
|
if(conn->marked_for_close) {
|
2003-09-26 12:03:50 +02:00
|
|
|
log_fn(LOG_INFO,"Cleaning up connection (fd %d).",conn->s);
|
2003-09-18 10:11:31 +02:00
|
|
|
if(conn->s >= 0) { /* might be an incomplete edge connection */
|
Integrated onion proxy into or/
The 'or' process can now be told (by the global_role variable) what
roles this server should play -- connect to all ORs, listen for ORs,
listen for OPs, listen for APs, or any combination.
* everything in /src/op/ is now obsolete.
* connection_ap.c now handles all interactions with application proxies
* "port" is now or_port, op_port, ap_port. But routers are still always
referenced (say, in conn_get_by_addr_port()) by addr / or_port. We
should make routers.c actually read these new ports (currently I've
kludged it so op_port = or_port+10, ap_port=or_port+20)
* circuits currently know if they're at the beginning of the path because
circ->cpath is set. They use this instead for crypts (both ways),
if it's set.
* I still obey the "send a 0 back to the AP when you're ready" protocol,
but I think we should phase it out. I can simply not read from the AP
socket until I'm ready.
I need to do a lot of cleanup work here, but the code appears to work, so
now's a good time for a checkin.
svn:r22
2002-07-02 11:36:58 +02:00
|
|
|
/* FIXME there's got to be a better way to check for this -- and make other checks? */
|
2003-09-25 12:42:07 +02:00
|
|
|
if(connection_speaks_cells(conn)) {
|
|
|
|
if(conn->state == OR_CONN_STATE_OPEN)
|
|
|
|
flush_buf_tls(conn->tls, conn->outbuf, &conn->outbuf_flushlen);
|
|
|
|
} else {
|
2003-09-25 07:17:11 +02:00
|
|
|
flush_buf(conn->s, conn->outbuf, &conn->outbuf_flushlen);
|
2003-09-25 12:42:07 +02:00
|
|
|
}
|
2003-09-18 10:11:31 +02:00
|
|
|
if(connection_wants_to_flush(conn)) /* not done flushing */
|
2003-10-10 03:48:32 +02:00
|
|
|
log_fn(LOG_WARN,"Conn (socket %d) still wants to flush. Losing %d bytes!",conn->s, (int)buf_datalen(conn->inbuf));
|
Integrated onion proxy into or/
The 'or' process can now be told (by the global_role variable) what
roles this server should play -- connect to all ORs, listen for ORs,
listen for OPs, listen for APs, or any combination.
* everything in /src/op/ is now obsolete.
* connection_ap.c now handles all interactions with application proxies
* "port" is now or_port, op_port, ap_port. But routers are still always
referenced (say, in conn_get_by_addr_port()) by addr / or_port. We
should make routers.c actually read these new ports (currently I've
kludged it so op_port = or_port+10, ap_port=or_port+20)
* circuits currently know if they're at the beginning of the path because
circ->cpath is set. They use this instead for crypts (both ways),
if it's set.
* I still obey the "send a 0 back to the AP when you're ready" protocol,
but I think we should phase it out. I can simply not read from the AP
socket until I'm ready.
I need to do a lot of cleanup work here, but the code appears to work, so
now's a good time for a checkin.
svn:r22
2002-07-02 11:36:58 +02:00
|
|
|
}
|
2002-06-27 00:45:49 +02:00
|
|
|
connection_remove(conn);
|
|
|
|
connection_free(conn);
|
|
|
|
if(i<nfds) { /* we just replaced the one at i with a new one.
|
|
|
|
process it too. */
|
2003-10-09 20:45:14 +02:00
|
|
|
conn_close_if_marked(i);
|
2002-06-27 00:45:49 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2003-10-10 03:48:32 +02:00
|
|
|
/* Perform regular maintenance tasks for a single connection. This
|
2003-10-09 20:45:14 +02:00
|
|
|
* function gets run once per second per connection by run_housekeeping.
|
|
|
|
*/
|
|
|
|
static void run_connection_housekeeping(int i, time_t now) {
|
|
|
|
cell_t cell;
|
|
|
|
connection_t *conn = connection_array[i];
|
|
|
|
if(connection_receiver_bucket_should_increase(conn)) {
|
|
|
|
conn->receiver_bucket += conn->bandwidth;
|
|
|
|
// log_fn(LOG_DEBUG,"Receiver bucket %d now %d.", i, conn->receiver_bucket);
|
|
|
|
}
|
|
|
|
|
|
|
|
if(conn->wants_to_read == 1 /* it's marked to turn reading back on now */
|
|
|
|
&& global_read_bucket > 0 /* and we're allowed to read */
|
|
|
|
&& (!connection_speaks_cells(conn) || conn->receiver_bucket > 0)) {
|
|
|
|
/* and either a non-cell conn or a cell conn with non-empty bucket */
|
|
|
|
conn->wants_to_read = 0;
|
|
|
|
connection_start_reading(conn);
|
|
|
|
if(conn->wants_to_write == 1) {
|
|
|
|
conn->wants_to_write = 0;
|
|
|
|
connection_start_writing(conn);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* check connections to see whether we should send a keepalive, expire, or wait */
|
|
|
|
if(!connection_speaks_cells(conn))
|
|
|
|
return;
|
|
|
|
|
|
|
|
if(now >= conn->timestamp_lastwritten + options.KeepalivePeriod) {
|
|
|
|
if((!options.OnionRouter && !circuit_get_by_conn(conn)) ||
|
|
|
|
(!connection_state_is_open(conn))) {
|
|
|
|
/* we're an onion proxy, with no circuits; or our handshake has expired. kill it. */
|
|
|
|
log_fn(LOG_INFO,"Expiring connection to %d (%s:%d).",
|
|
|
|
i,conn->address, conn->port);
|
|
|
|
conn->marked_for_close = 1;
|
|
|
|
} else {
|
|
|
|
/* either a full router, or we've got a circuit. send a padding cell. */
|
|
|
|
log_fn(LOG_DEBUG,"Sending keepalive to (%s:%d)",
|
|
|
|
conn->address, conn->port);
|
|
|
|
memset(&cell,0,sizeof(cell_t));
|
|
|
|
cell.command = CELL_PADDING;
|
|
|
|
connection_or_write_cell_to_buf(&cell, conn);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Perform regular maintenance tasks. This function gets run once per
|
|
|
|
* second by prepare_for_poll.
|
|
|
|
*/
|
|
|
|
static void run_scheduled_events(time_t now) {
|
2002-10-02 01:37:31 +02:00
|
|
|
static long time_to_fetch_directory = 0;
|
2003-04-16 08:18:31 +02:00
|
|
|
static long time_to_new_circuit = 0;
|
|
|
|
circuit_t *circ;
|
2003-10-09 20:45:14 +02:00
|
|
|
int i;
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
|
2003-10-09 20:45:14 +02:00
|
|
|
/* 1. Every DirFetchPostPeriod seconds, we get a new directory and upload
|
|
|
|
* our descriptor (if any). */
|
|
|
|
if(time_to_fetch_directory < now) {
|
|
|
|
/* it's time to fetch a new directory and/or post our descriptor */
|
|
|
|
if(options.OnionRouter) {
|
|
|
|
router_rebuild_descriptor();
|
|
|
|
router_upload_desc_to_dirservers();
|
|
|
|
}
|
|
|
|
if(!options.DirPort) {
|
|
|
|
/* NOTE directory servers do not currently fetch directories.
|
|
|
|
* Hope this doesn't bite us later. */
|
|
|
|
directory_initiate_command(router_pick_directory_server(),
|
|
|
|
DIR_CONN_STATE_CONNECTING_FETCH);
|
|
|
|
}
|
|
|
|
time_to_fetch_directory = now + options.DirFetchPostPeriod;
|
|
|
|
}
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
|
2003-10-10 03:48:32 +02:00
|
|
|
/* 2. Every NewCircuitPeriod seconds, we expire old circuits and make a
|
2003-10-09 20:45:14 +02:00
|
|
|
* new one as needed.
|
|
|
|
*/
|
2003-10-21 11:48:17 +02:00
|
|
|
if(options.SocksPort && time_to_new_circuit < now) {
|
2003-10-09 20:45:14 +02:00
|
|
|
circuit_expire_unused_circuits();
|
|
|
|
circuit_launch_new(-1); /* tell it to forget about previous failures */
|
|
|
|
circ = circuit_get_newest_open();
|
|
|
|
if(!circ || circ->dirty) {
|
|
|
|
log_fn(LOG_INFO,"Youngest circuit %s; launching replacement.", circ ? "dirty" : "missing");
|
|
|
|
circuit_launch_new(0); /* make an onion and lay the circuit */
|
|
|
|
}
|
|
|
|
time_to_new_circuit = now + options.NewCircuitPeriod;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* 3. Every second, we check how much bandwidth we've consumed and
|
|
|
|
* increment global_read_bucket.
|
|
|
|
*/
|
|
|
|
stats_n_bytes_read += stats_prev_global_read_bucket-global_read_bucket;
|
|
|
|
if(global_read_bucket < 9*options.TotalBandwidth) {
|
|
|
|
global_read_bucket += options.TotalBandwidth;
|
|
|
|
log_fn(LOG_DEBUG,"global_read_bucket now %d.", global_read_bucket);
|
|
|
|
}
|
|
|
|
stats_prev_global_read_bucket = global_read_bucket;
|
|
|
|
|
2003-03-06 05:52:02 +01:00
|
|
|
|
2003-10-09 20:45:14 +02:00
|
|
|
/* 4. We do houskeeping for each connection... */
|
|
|
|
for(i=0;i<nfds;i++) {
|
|
|
|
run_connection_housekeeping(i, now);
|
|
|
|
}
|
2003-10-02 22:00:38 +02:00
|
|
|
|
2003-10-09 20:45:14 +02:00
|
|
|
/* 5. and blow away any connections that need to die. can't do this later
|
|
|
|
* because we might open up a circuit and not realize we're about to cull
|
|
|
|
* the connection it's running over.
|
|
|
|
*/
|
|
|
|
for(i=0;i<nfds;i++)
|
|
|
|
conn_close_if_marked(i);
|
|
|
|
}
|
2002-09-26 14:09:10 +02:00
|
|
|
|
2003-10-09 20:45:14 +02:00
|
|
|
static int prepare_for_poll(void) {
|
|
|
|
static long current_second = 0; /* from previous calls to gettimeofday */
|
|
|
|
connection_t *conn;
|
|
|
|
struct timeval now;
|
|
|
|
int i;
|
2003-04-16 08:18:31 +02:00
|
|
|
|
2003-10-09 20:45:14 +02:00
|
|
|
tor_gettimeofday(&now);
|
2003-07-05 09:10:34 +02:00
|
|
|
|
2003-10-09 20:45:14 +02:00
|
|
|
if(now.tv_sec > current_second) { /* the second has rolled over. check more stuff. */
|
2002-09-26 14:09:10 +02:00
|
|
|
|
2003-10-09 20:45:14 +02:00
|
|
|
++stats_n_seconds_reading;
|
|
|
|
run_scheduled_events(now.tv_sec);
|
2002-10-02 01:37:31 +02:00
|
|
|
|
2002-12-31 16:04:14 +01:00
|
|
|
current_second = now.tv_sec; /* remember which second it is, for next time */
|
2002-10-02 01:37:31 +02:00
|
|
|
}
|
|
|
|
|
2003-09-30 22:36:20 +02:00
|
|
|
for(i=0;i<nfds;i++) {
|
|
|
|
conn = connection_array[i];
|
2003-10-09 20:45:14 +02:00
|
|
|
if(connection_has_pending_tls_data(conn)) {
|
2003-09-30 22:36:20 +02:00
|
|
|
log_fn(LOG_DEBUG,"sock %d has pending bytes.",conn->s);
|
|
|
|
return 0; /* has pending bytes to read; don't let poll wait. */
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return (1000 - (now.tv_usec / 1000)); /* how many milliseconds til the next second? */
|
2003-03-06 05:52:02 +01:00
|
|
|
}
|
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
static crypto_pk_env_t *init_key_from_file(const char *fname)
|
|
|
|
{
|
|
|
|
crypto_pk_env_t *prkey = NULL;
|
|
|
|
int fd = -1;
|
|
|
|
FILE *file = NULL;
|
|
|
|
|
|
|
|
if (!(prkey = crypto_new_pk_env(CRYPTO_PK_RSA))) {
|
|
|
|
log(LOG_ERR, "Error creating crypto environment.");
|
|
|
|
goto error;
|
2002-06-27 00:45:49 +02:00
|
|
|
}
|
|
|
|
|
2003-09-26 20:27:35 +02:00
|
|
|
switch(file_status(fname)) {
|
2003-09-25 07:17:11 +02:00
|
|
|
case FN_DIR:
|
|
|
|
case FN_ERROR:
|
|
|
|
log(LOG_ERR, "Can't read key from %s", fname);
|
|
|
|
goto error;
|
|
|
|
case FN_NOENT:
|
|
|
|
log(LOG_INFO, "No key found in %s; generating fresh key.", fname);
|
|
|
|
if (crypto_pk_generate_key(prkey)) {
|
|
|
|
log(LOG_ERR, "Error generating key: %s", crypto_perror());
|
|
|
|
goto error;
|
2002-09-04 08:29:28 +02:00
|
|
|
}
|
2003-09-25 07:17:11 +02:00
|
|
|
if (crypto_pk_check_key(prkey) <= 0) {
|
|
|
|
log(LOG_ERR, "Generated key seems invalid");
|
|
|
|
goto error;
|
2003-05-08 00:40:03 +02:00
|
|
|
}
|
2003-09-25 07:17:11 +02:00
|
|
|
log(LOG_INFO, "Generated key seems valid");
|
2003-09-26 20:27:35 +02:00
|
|
|
if (crypto_pk_write_private_key_to_filename(prkey, fname)) {
|
|
|
|
log(LOG_ERR, "Couldn't write generated key to %s.", fname);
|
2003-09-25 07:17:11 +02:00
|
|
|
goto error;
|
|
|
|
}
|
|
|
|
return prkey;
|
|
|
|
case FN_FILE:
|
|
|
|
if (crypto_pk_read_private_key_from_filename(prkey, fname)) {
|
|
|
|
log(LOG_ERR, "Error loading private key.");
|
|
|
|
goto error;
|
2002-09-04 08:29:28 +02:00
|
|
|
}
|
2003-09-25 07:17:11 +02:00
|
|
|
return prkey;
|
|
|
|
default:
|
|
|
|
assert(0);
|
2002-06-27 00:45:49 +02:00
|
|
|
}
|
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
error:
|
|
|
|
if (prkey)
|
|
|
|
crypto_free_pk_env(prkey);
|
|
|
|
if (fd >= 0 && !file)
|
|
|
|
close(fd);
|
|
|
|
if (file)
|
|
|
|
fclose(file);
|
|
|
|
return NULL;
|
|
|
|
}
|
2003-09-08 08:26:38 +02:00
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
static int init_keys(void)
|
|
|
|
{
|
|
|
|
char keydir[512];
|
2003-09-26 20:27:35 +02:00
|
|
|
char fingerprint[FINGERPRINT_LEN+MAX_NICKNAME_LEN+3];
|
2003-09-25 07:17:11 +02:00
|
|
|
char *cp;
|
2003-09-30 10:18:10 +02:00
|
|
|
const char *tmp, *mydesc;
|
2003-09-25 07:17:11 +02:00
|
|
|
crypto_pk_env_t *prkey;
|
|
|
|
|
|
|
|
/* OP's don't need keys. Just initialize the TLS context.*/
|
2003-09-30 01:14:49 +02:00
|
|
|
if (!options.OnionRouter) {
|
|
|
|
assert(!options.DirPort);
|
2003-09-25 07:17:11 +02:00
|
|
|
if (tor_tls_context_new(NULL, 0, NULL)<0) {
|
|
|
|
log_fn(LOG_ERR, "Error creating TLS context for OP.");
|
2003-09-08 08:26:38 +02:00
|
|
|
return -1;
|
|
|
|
}
|
2003-09-25 07:17:11 +02:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
assert(options.DataDirectory);
|
|
|
|
if (strlen(options.DataDirectory) > (512-128)) {
|
|
|
|
log_fn(LOG_ERR, "DataDirectory is too long.");
|
|
|
|
return -1;
|
|
|
|
}
|
2003-09-30 01:14:49 +02:00
|
|
|
if (check_private_dir(options.DataDirectory, 1)) {
|
2003-09-25 07:17:11 +02:00
|
|
|
return -1;
|
|
|
|
}
|
2003-09-30 01:14:49 +02:00
|
|
|
sprintf(keydir,"%s/keys",options.DataDirectory);
|
2003-09-26 20:27:35 +02:00
|
|
|
if (check_private_dir(keydir, 1)) {
|
2003-09-25 07:17:11 +02:00
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
cp = keydir + strlen(keydir); /* End of string. */
|
|
|
|
|
|
|
|
/* 1. Read identity key. Make it if none is found. */
|
2003-09-30 01:14:49 +02:00
|
|
|
strcpy(cp, "/identity.key");
|
|
|
|
log_fn(LOG_INFO,"Reading/making identity key %s...",keydir);
|
2003-09-25 07:17:11 +02:00
|
|
|
prkey = init_key_from_file(keydir);
|
|
|
|
if (!prkey) return -1;
|
|
|
|
set_identity_key(prkey);
|
|
|
|
/* 2. Read onion key. Make it if none is found. */
|
2003-09-30 01:14:49 +02:00
|
|
|
strcpy(cp, "/onion.key");
|
|
|
|
log_fn(LOG_INFO,"Reading/making onion key %s...",keydir);
|
2003-09-25 07:17:11 +02:00
|
|
|
prkey = init_key_from_file(keydir);
|
|
|
|
if (!prkey) return -1;
|
|
|
|
set_onion_key(prkey);
|
|
|
|
|
|
|
|
/* 3. Initialize link key and TLS context. */
|
2003-09-30 01:14:49 +02:00
|
|
|
strcpy(cp, "/link.key");
|
|
|
|
log_fn(LOG_INFO,"Reading/making link key %s...",keydir);
|
2003-09-25 07:17:11 +02:00
|
|
|
prkey = init_key_from_file(keydir);
|
|
|
|
if (!prkey) return -1;
|
|
|
|
set_link_key(prkey);
|
|
|
|
if (tor_tls_context_new(prkey, 1, options.Nickname) < 0) {
|
|
|
|
log_fn(LOG_ERR, "Error initializing TLS context");
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
/* 4. Dump router descriptor to 'router.desc' */
|
|
|
|
/* Must be called after keys are initialized. */
|
2003-10-01 02:43:34 +02:00
|
|
|
if (!(router_get_my_descriptor())) {
|
2003-09-25 07:17:11 +02:00
|
|
|
log_fn(LOG_ERR, "Error initializing descriptor.");
|
|
|
|
return -1;
|
|
|
|
}
|
2003-09-30 21:27:54 +02:00
|
|
|
/* We need to add our own fingerprint so it gets recognized. */
|
|
|
|
if (dirserv_add_own_fingerprint(options.Nickname, get_identity_key())) {
|
|
|
|
log_fn(LOG_ERR, "Error adding own fingerprint to approved set");
|
|
|
|
return -1;
|
|
|
|
}
|
2003-09-30 10:18:10 +02:00
|
|
|
tmp = mydesc = router_get_my_descriptor();
|
2003-09-30 21:27:54 +02:00
|
|
|
if (dirserv_add_descriptor(&tmp)) {
|
|
|
|
log(LOG_ERR, "Unable to add own descriptor to directory.");
|
|
|
|
return -1;
|
|
|
|
}
|
2003-09-30 01:14:49 +02:00
|
|
|
sprintf(keydir,"%s/router.desc", options.DataDirectory);
|
|
|
|
log_fn(LOG_INFO,"Dumping descriptor to %s...",keydir);
|
2003-09-30 10:18:10 +02:00
|
|
|
if (write_str_to_file(keydir, mydesc)) {
|
2003-09-25 07:17:11 +02:00
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
/* 5. Dump fingerprint to 'fingerprint' */
|
2003-09-30 01:14:49 +02:00
|
|
|
sprintf(keydir,"%s/fingerprint", options.DataDirectory);
|
|
|
|
log_fn(LOG_INFO,"Dumping fingerprint to %s...",keydir);
|
2003-09-26 20:27:35 +02:00
|
|
|
assert(strlen(options.Nickname) <= MAX_NICKNAME_LEN);
|
|
|
|
strcpy(fingerprint, options.Nickname);
|
|
|
|
strcat(fingerprint, " ");
|
|
|
|
if (crypto_pk_get_fingerprint(get_identity_key(),
|
|
|
|
fingerprint+strlen(fingerprint))<0) {
|
2003-09-25 07:17:11 +02:00
|
|
|
log_fn(LOG_ERR, "Error computing fingerprint");
|
|
|
|
return -1;
|
|
|
|
}
|
2003-09-26 20:27:35 +02:00
|
|
|
strcat(fingerprint, "\n");
|
|
|
|
if (write_str_to_file(keydir, fingerprint))
|
|
|
|
return -1;
|
2003-09-30 01:14:49 +02:00
|
|
|
if(!options.DirPort)
|
|
|
|
return 0;
|
|
|
|
/* 6. [dirserver only] load approved-routers file */
|
|
|
|
sprintf(keydir,"%s/approved-routers", options.DataDirectory);
|
|
|
|
log_fn(LOG_INFO,"Loading approved fingerprints from %s...",keydir);
|
|
|
|
if(dirserv_parse_fingerprint_file(keydir) < 0) {
|
|
|
|
log_fn(LOG_ERR, "Error loading fingerprints");
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
/* 7. [dirserver only] load old directory, if it's there */
|
|
|
|
sprintf(keydir,"%s/cached-directory", options.DataDirectory);
|
|
|
|
log_fn(LOG_INFO,"Loading cached directory from %s...",keydir);
|
|
|
|
cp = read_file_to_str(keydir);
|
|
|
|
if(!cp) {
|
|
|
|
log_fn(LOG_INFO,"Cached directory %s not present. Ok.",keydir);
|
|
|
|
} else {
|
|
|
|
if(dirserv_init_from_directory_string(cp) < 0) {
|
|
|
|
log_fn(LOG_ERR, "Cached directory %s is corrupt", keydir);
|
|
|
|
free(cp);
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
free(cp);
|
|
|
|
}
|
|
|
|
/* success */
|
2003-09-25 07:17:11 +02:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int do_main_loop(void) {
|
|
|
|
int i;
|
|
|
|
int timeout;
|
|
|
|
int poll_result;
|
|
|
|
|
|
|
|
/* load the routers file */
|
|
|
|
if(router_get_list_from_file(options.RouterFile) < 0) {
|
|
|
|
log_fn(LOG_ERR,"Error loading router list.");
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* load the private keys, if we're supposed to have them, and set up the
|
|
|
|
* TLS context. */
|
|
|
|
if (init_keys() < 0) {
|
|
|
|
log_fn(LOG_ERR,"Error initializing keys; exiting");
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
if(options.OnionRouter) {
|
2003-09-25 12:42:07 +02:00
|
|
|
cpu_init(); /* launch cpuworkers. Need to do this *after* we've read the onion key. */
|
2003-09-27 00:02:02 +02:00
|
|
|
router_upload_desc_to_dirservers(); /* upload our descriptor to all dirservers */
|
2003-09-08 08:26:38 +02:00
|
|
|
}
|
2003-09-07 12:24:40 +02:00
|
|
|
|
2003-03-18 02:49:55 +01:00
|
|
|
/* start up the necessary connections based on which ports are
|
|
|
|
* non-zero. This is where we try to connect to all the other ORs,
|
2003-09-05 08:04:03 +02:00
|
|
|
* and start the listeners.
|
2003-03-18 02:49:55 +01:00
|
|
|
*/
|
2003-10-25 14:01:09 +02:00
|
|
|
if(retry_all_connections() < 0) {
|
|
|
|
log_fn(LOG_ERR,"Failed to bind one of the listener ports.");
|
|
|
|
return -1;
|
|
|
|
}
|
2002-06-27 00:45:49 +02:00
|
|
|
|
|
|
|
for(;;) {
|
2003-10-22 08:03:11 +02:00
|
|
|
#ifndef MS_WINDOWS /* do signal stuff only on unix */
|
2002-09-22 00:41:48 +02:00
|
|
|
if(please_dumpstats) {
|
2003-10-15 20:50:16 +02:00
|
|
|
/* prefer to log it at INFO, but make sure we always see it */
|
|
|
|
dumpstats(options.loglevel>LOG_INFO ? options.loglevel : LOG_INFO);
|
2002-09-28 03:40:11 +02:00
|
|
|
please_dumpstats = 0;
|
|
|
|
}
|
2003-09-21 08:15:43 +02:00
|
|
|
if(please_reset) {
|
|
|
|
/* fetch a new directory */
|
2003-03-18 02:49:55 +01:00
|
|
|
if(options.DirPort) {
|
2003-10-17 07:23:48 +02:00
|
|
|
|
|
|
|
/* reload the fingerprint file */
|
|
|
|
char keydir[512];
|
|
|
|
sprintf(keydir,"%s/approved-routers", options.DataDirectory);
|
|
|
|
log_fn(LOG_INFO,"Reloading approved fingerprints from %s...",keydir);
|
|
|
|
if(dirserv_parse_fingerprint_file(keydir) < 0) {
|
|
|
|
log_fn(LOG_WARN, "Error reloading fingerprints. Continuing with old list.");
|
|
|
|
}
|
|
|
|
|
2002-10-03 00:54:20 +02:00
|
|
|
if(router_get_list_from_file(options.RouterFile) < 0) {
|
2003-10-10 03:48:32 +02:00
|
|
|
log(LOG_WARN,"Error reloading router list. Continuing with old list.");
|
2002-09-28 03:40:11 +02:00
|
|
|
}
|
|
|
|
} else {
|
2003-09-21 08:15:43 +02:00
|
|
|
directory_initiate_command(router_pick_directory_server(), DIR_CONN_STATE_CONNECTING_FETCH);
|
2002-09-28 03:40:11 +02:00
|
|
|
}
|
2003-09-21 08:15:43 +02:00
|
|
|
|
2003-09-23 21:47:41 +02:00
|
|
|
/* close and reopen the log files */
|
|
|
|
reset_logs();
|
2003-09-21 08:15:43 +02:00
|
|
|
|
|
|
|
please_reset = 0;
|
2002-09-22 00:41:48 +02:00
|
|
|
}
|
2003-08-12 08:41:53 +02:00
|
|
|
if(please_reap_children) {
|
|
|
|
while(waitpid(-1,NULL,WNOHANG)) ; /* keep reaping until no more zombies */
|
|
|
|
please_reap_children = 0;
|
|
|
|
}
|
2003-08-12 10:04:31 +02:00
|
|
|
#endif /* signal stuff */
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
|
2003-09-05 08:04:03 +02:00
|
|
|
timeout = prepare_for_poll();
|
|
|
|
|
|
|
|
/* poll until we have an event, or the second ends */
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
poll_result = poll(poll_array, nfds, timeout);
|
|
|
|
|
2003-10-15 20:28:32 +02:00
|
|
|
/* let catch() handle things like ^c, and otherwise don't worry about it */
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
if(poll_result < 0) {
|
2003-10-15 20:28:32 +02:00
|
|
|
if(errno != EINTR) { /* let the program survive things like ^z */
|
|
|
|
log_fn(LOG_ERR,"poll failed.");
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
return -1;
|
2003-10-15 20:28:32 +02:00
|
|
|
} else {
|
2003-10-15 20:37:19 +02:00
|
|
|
log_fn(LOG_DEBUG,"poll interrupted.");
|
2003-10-15 20:28:32 +02:00
|
|
|
}
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
}
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2003-09-30 22:36:20 +02:00
|
|
|
/* do all the reads and errors first, so we can detect closed sockets */
|
|
|
|
for(i=0;i<nfds;i++)
|
|
|
|
conn_read(i); /* this also blows away broken connections */
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2003-09-30 22:36:20 +02:00
|
|
|
/* then do the writes */
|
|
|
|
for(i=0;i<nfds;i++)
|
|
|
|
conn_write(i);
|
|
|
|
|
|
|
|
/* any of the conns need to be closed now? */
|
|
|
|
for(i=0;i<nfds;i++)
|
2003-10-09 20:45:14 +02:00
|
|
|
conn_close_if_marked(i);
|
2002-06-27 00:45:49 +02:00
|
|
|
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
/* refilling buckets and sending cells happens at the beginning of the
|
|
|
|
* next iteration of the loop, inside prepare_for_poll()
|
|
|
|
*/
|
2002-06-27 00:45:49 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2002-09-28 03:40:11 +02:00
|
|
|
static void catch(int the_signal) {
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2003-10-22 08:03:11 +02:00
|
|
|
#ifndef MS_WINDOWS /* do signal stuff only on unix */
|
2002-09-28 03:40:11 +02:00
|
|
|
switch(the_signal) {
|
2002-11-24 09:45:54 +01:00
|
|
|
// case SIGABRT:
|
2002-09-28 03:40:11 +02:00
|
|
|
case SIGTERM:
|
|
|
|
case SIGINT:
|
2003-09-26 12:03:50 +02:00
|
|
|
log(LOG_ERR,"Catching signal %d, exiting cleanly.", the_signal);
|
2003-10-22 08:03:11 +02:00
|
|
|
/* we don't care if there was an error when we unlink, nothing
|
|
|
|
we could do about it anyways */
|
2003-10-08 10:54:52 +02:00
|
|
|
unlink(options.PidFile);
|
2002-09-28 03:40:11 +02:00
|
|
|
exit(0);
|
|
|
|
case SIGHUP:
|
2003-09-21 08:15:43 +02:00
|
|
|
please_reset = 1;
|
2002-09-28 03:40:11 +02:00
|
|
|
break;
|
|
|
|
case SIGUSR1:
|
|
|
|
please_dumpstats = 1;
|
|
|
|
break;
|
2003-08-12 08:41:53 +02:00
|
|
|
case SIGCHLD:
|
|
|
|
please_reap_children = 1;
|
2003-09-13 23:53:38 +02:00
|
|
|
break;
|
2002-09-28 03:40:11 +02:00
|
|
|
default:
|
2003-10-10 03:48:32 +02:00
|
|
|
log(LOG_WARN,"Caught signal %d that we can't handle??", the_signal);
|
2002-09-28 03:40:11 +02:00
|
|
|
}
|
2003-08-12 10:04:31 +02:00
|
|
|
#endif /* signal stuff */
|
2002-09-22 00:41:48 +02:00
|
|
|
}
|
|
|
|
|
2003-10-15 20:28:32 +02:00
|
|
|
static void dumpstats(int severity) {
|
2002-09-22 00:41:48 +02:00
|
|
|
int i;
|
|
|
|
connection_t *conn;
|
2003-10-04 05:29:09 +02:00
|
|
|
time_t now = time(NULL);
|
2002-09-22 00:41:48 +02:00
|
|
|
|
2003-10-15 20:28:32 +02:00
|
|
|
log(severity, "Dumping stats:");
|
2002-09-22 00:41:48 +02:00
|
|
|
|
|
|
|
for(i=0;i<nfds;i++) {
|
|
|
|
conn = connection_array[i];
|
2003-10-15 20:28:32 +02:00
|
|
|
log(severity, "Conn %d (socket %d) type %d (%s), state %d (%s), created %ld secs ago",
|
2002-09-22 00:41:48 +02:00
|
|
|
i, conn->s, conn->type, conn_type_to_string[conn->type],
|
2003-10-04 05:29:09 +02:00
|
|
|
conn->state, conn_state_to_string[conn->type][conn->state], now - conn->timestamp_created);
|
2002-09-22 00:41:48 +02:00
|
|
|
if(!connection_is_listener(conn)) {
|
2003-10-15 20:28:32 +02:00
|
|
|
log(severity,"Conn %d is to '%s:%d'.",i,conn->address, conn->port);
|
|
|
|
log(severity,"Conn %d: %d bytes waiting on inbuf (last read %ld secs ago)",i,
|
2003-09-25 07:17:11 +02:00
|
|
|
(int)buf_datalen(conn->inbuf),
|
2003-10-04 05:29:09 +02:00
|
|
|
now - conn->timestamp_lastread);
|
2003-10-15 20:28:32 +02:00
|
|
|
log(severity,"Conn %d: %d bytes waiting on outbuf (last written %ld secs ago)",i,
|
2003-10-04 05:29:09 +02:00
|
|
|
(int)buf_datalen(conn->outbuf), now - conn->timestamp_lastwritten);
|
2002-09-22 00:41:48 +02:00
|
|
|
}
|
2003-10-15 20:28:32 +02:00
|
|
|
circuit_dump_by_conn(conn, severity); /* dump info about all the circuits using this conn */
|
2002-09-22 00:41:48 +02:00
|
|
|
}
|
2003-10-15 20:28:32 +02:00
|
|
|
log(severity,
|
|
|
|
"Cells processed: %10lu padding\n"
|
2003-10-04 10:19:23 +02:00
|
|
|
" %10lu create\n"
|
|
|
|
" %10lu created\n"
|
|
|
|
" %10lu relay\n"
|
|
|
|
" (%10lu relayed)\n"
|
|
|
|
" (%10lu delivered)\n"
|
2003-10-15 20:28:32 +02:00
|
|
|
" %10lud destroy",
|
2003-10-02 22:00:38 +02:00
|
|
|
stats_n_padding_cells_processed,
|
|
|
|
stats_n_create_cells_processed,
|
|
|
|
stats_n_created_cells_processed,
|
|
|
|
stats_n_relay_cells_processed,
|
|
|
|
stats_n_relay_cells_relayed,
|
|
|
|
stats_n_relay_cells_delivered,
|
|
|
|
stats_n_destroy_cells_processed);
|
|
|
|
if (stats_n_data_cells_packaged)
|
2003-10-15 20:28:32 +02:00
|
|
|
log(severity,"Average outgoing cell fullness: %2.3f%%",
|
2003-10-02 22:00:38 +02:00
|
|
|
100*(((double)stats_n_data_bytes_packaged) /
|
|
|
|
(stats_n_data_cells_packaged*(CELL_PAYLOAD_SIZE-RELAY_HEADER_SIZE))) );
|
2003-10-15 21:25:28 +02:00
|
|
|
if (stats_n_data_cells_received)
|
2003-10-15 20:28:32 +02:00
|
|
|
log(severity,"Average incoming cell fullness: %2.3f%%",
|
2003-10-02 22:00:38 +02:00
|
|
|
100*(((double)stats_n_data_bytes_received) /
|
|
|
|
(stats_n_data_cells_received*(CELL_PAYLOAD_SIZE-RELAY_HEADER_SIZE))) );
|
|
|
|
|
|
|
|
if (stats_n_seconds_reading)
|
2003-10-15 20:28:32 +02:00
|
|
|
log(severity,"Average bandwidth used: %d bytes/sec",
|
2003-10-02 22:00:38 +02:00
|
|
|
(int) (stats_n_bytes_read/stats_n_seconds_reading));
|
2002-09-22 00:41:48 +02:00
|
|
|
}
|
|
|
|
|
2003-04-07 04:12:02 +02:00
|
|
|
int tor_main(int argc, char *argv[]) {
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2003-10-18 02:13:08 +02:00
|
|
|
/* give it somewhere to log to initially */
|
|
|
|
add_stream_log(LOG_INFO, "<stdout>", stdout);
|
2003-11-13 05:51:34 +01:00
|
|
|
log_fn(LOG_WARN,"Tor v%s. This is experimental software. Do not use it if you need anonymity.",VERSION);
|
2003-10-18 02:13:08 +02:00
|
|
|
|
2003-09-29 09:50:08 +02:00
|
|
|
if(getconfig(argc,argv,&options)) {
|
2003-10-20 03:19:54 +02:00
|
|
|
log_fn(LOG_ERR,"Reading config failed. For usage, try -h.");
|
2003-09-29 09:50:08 +02:00
|
|
|
return -1;
|
|
|
|
}
|
2003-10-15 20:50:16 +02:00
|
|
|
log_set_severity(options.loglevel); /* assign logging severity level from options */
|
2003-10-18 02:13:08 +02:00
|
|
|
close_logs(); /* close stdout, then open with correct loglevel if necessary */
|
|
|
|
if(!options.LogFile && !options.RunAsDaemon)
|
|
|
|
add_stream_log(options.loglevel, "<stdout>", stdout);
|
2003-10-15 20:50:16 +02:00
|
|
|
if(options.DebugLogFile)
|
|
|
|
add_file_log(LOG_DEBUG, options.DebugLogFile);
|
|
|
|
if(options.LogFile)
|
|
|
|
add_file_log(options.loglevel, options.LogFile);
|
|
|
|
|
2003-07-05 09:10:34 +02:00
|
|
|
global_read_bucket = options.TotalBandwidth; /* start it at 1 second of traffic */
|
2003-10-02 22:00:38 +02:00
|
|
|
stats_prev_global_read_bucket = global_read_bucket;
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2003-10-08 10:54:52 +02:00
|
|
|
/* write our pid to the pid file */
|
|
|
|
write_pidfile(options.PidFile);
|
|
|
|
|
2003-10-22 13:21:29 +02:00
|
|
|
/* now that we've written the pid file, we can switch the user and group. */
|
|
|
|
if(options.User || options.Group) {
|
|
|
|
if(switch_id(options.User, options.Group) != 0) {
|
2003-10-22 08:03:11 +02:00
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2003-10-15 20:50:16 +02:00
|
|
|
if(options.RunAsDaemon)
|
2003-03-17 03:41:36 +01:00
|
|
|
daemonize();
|
|
|
|
|
2003-05-28 04:03:25 +02:00
|
|
|
if(options.OnionRouter) { /* only spawn dns handlers if we're a router */
|
2003-06-17 16:31:05 +02:00
|
|
|
dns_init(); /* initialize the dns resolve tree, and spawn workers */
|
major overhaul: dns slave subsystem, topics
on startup, it forks off a master dns handler, which forks off dns
slaves (like the apache model). slaves as spawned as load increases,
and then reused. excess slaves are not ever killed, currently.
implemented topics. each topic has a receive window in each direction
at each edge of the circuit, and sends sendme's at the data level, as
per before. each circuit also has receive windows in each direction at
each hop; an edge sends a circuit-level sendme as soon as enough data
cells have arrived (regardless of whether the data cells were flushed
to the exit conns). removed the 'connected' cell type, since it's now
a topic command within data cells.
at the edge of the circuit, there can be multiple connections associated
with a single circuit. you find them via the linked list conn->next_topic.
currently each new ap connection starts its own circuit, so we ought
to see comparable performance to what we had before. but that's only
because i haven't written the code to reattach to old circuits. please
try to break it as-is, and then i'll make it reuse the same circuit and
we'll try to break that.
svn:r152
2003-01-26 10:02:24 +01:00
|
|
|
}
|
|
|
|
|
2003-08-12 10:04:31 +02:00
|
|
|
#ifndef MS_WINDOWS /* do signal stuff only on unix */
|
major overhaul: dns slave subsystem, topics
on startup, it forks off a master dns handler, which forks off dns
slaves (like the apache model). slaves as spawned as load increases,
and then reused. excess slaves are not ever killed, currently.
implemented topics. each topic has a receive window in each direction
at each edge of the circuit, and sends sendme's at the data level, as
per before. each circuit also has receive windows in each direction at
each hop; an edge sends a circuit-level sendme as soon as enough data
cells have arrived (regardless of whether the data cells were flushed
to the exit conns). removed the 'connected' cell type, since it's now
a topic command within data cells.
at the edge of the circuit, there can be multiple connections associated
with a single circuit. you find them via the linked list conn->next_topic.
currently each new ap connection starts its own circuit, so we ought
to see comparable performance to what we had before. but that's only
because i haven't written the code to reattach to old circuits. please
try to break it as-is, and then i'll make it reuse the same circuit and
we'll try to break that.
svn:r152
2003-01-26 10:02:24 +01:00
|
|
|
signal (SIGINT, catch); /* catch kills so we can exit cleanly */
|
|
|
|
signal (SIGTERM, catch);
|
2003-10-15 20:28:32 +02:00
|
|
|
signal (SIGUSR1, catch); /* to dump stats */
|
major overhaul: dns slave subsystem, topics
on startup, it forks off a master dns handler, which forks off dns
slaves (like the apache model). slaves as spawned as load increases,
and then reused. excess slaves are not ever killed, currently.
implemented topics. each topic has a receive window in each direction
at each edge of the circuit, and sends sendme's at the data level, as
per before. each circuit also has receive windows in each direction at
each hop; an edge sends a circuit-level sendme as soon as enough data
cells have arrived (regardless of whether the data cells were flushed
to the exit conns). removed the 'connected' cell type, since it's now
a topic command within data cells.
at the edge of the circuit, there can be multiple connections associated
with a single circuit. you find them via the linked list conn->next_topic.
currently each new ap connection starts its own circuit, so we ought
to see comparable performance to what we had before. but that's only
because i haven't written the code to reattach to old circuits. please
try to break it as-is, and then i'll make it reuse the same circuit and
we'll try to break that.
svn:r152
2003-01-26 10:02:24 +01:00
|
|
|
signal (SIGHUP, catch); /* to reload directory */
|
2003-08-12 08:41:53 +02:00
|
|
|
signal (SIGCHLD, catch); /* for exiting dns/cpu workers */
|
2003-08-12 10:04:31 +02:00
|
|
|
#endif /* signal stuff */
|
major overhaul: dns slave subsystem, topics
on startup, it forks off a master dns handler, which forks off dns
slaves (like the apache model). slaves as spawned as load increases,
and then reused. excess slaves are not ever killed, currently.
implemented topics. each topic has a receive window in each direction
at each edge of the circuit, and sends sendme's at the data level, as
per before. each circuit also has receive windows in each direction at
each hop; an edge sends a circuit-level sendme as soon as enough data
cells have arrived (regardless of whether the data cells were flushed
to the exit conns). removed the 'connected' cell type, since it's now
a topic command within data cells.
at the edge of the circuit, there can be multiple connections associated
with a single circuit. you find them via the linked list conn->next_topic.
currently each new ap connection starts its own circuit, so we ought
to see comparable performance to what we had before. but that's only
because i haven't written the code to reattach to old circuits. please
try to break it as-is, and then i'll make it reuse the same circuit and
we'll try to break that.
svn:r152
2003-01-26 10:02:24 +01:00
|
|
|
|
|
|
|
crypto_global_init();
|
2003-06-13 23:13:37 +02:00
|
|
|
crypto_seed_rng();
|
2003-09-29 09:50:08 +02:00
|
|
|
do_main_loop();
|
2002-08-22 09:30:03 +02:00
|
|
|
crypto_global_cleanup();
|
2003-09-29 09:50:08 +02:00
|
|
|
return -1;
|
2002-06-27 00:45:49 +02:00
|
|
|
}
|
|
|
|
|
2003-04-07 04:12:02 +02:00
|
|
|
/*
|
|
|
|
Local Variables:
|
|
|
|
mode:c
|
|
|
|
indent-tabs-mode:nil
|
|
|
|
c-basic-offset:2
|
|
|
|
End:
|
|
|
|
*/
|