2003-10-08 04:04:08 +02:00
|
|
|
/* Copyright 2001,2002,2003 Roger Dingledine, Matej Pfajfar. */
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
/* See LICENSE for licensing information */
|
|
|
|
/* $Id$ */
|
2002-06-27 00:45:49 +02:00
|
|
|
|
|
|
|
#include "or.h"
|
|
|
|
|
2004-02-29 02:31:33 +01:00
|
|
|
/********* PROTOTYPES **********/
|
2003-07-05 07:46:06 +02:00
|
|
|
|
2003-10-15 20:28:32 +02:00
|
|
|
static void dumpstats(int severity); /* log stats */
|
2003-11-13 07:49:25 +01:00
|
|
|
static int init_from_config(int argc, char **argv);
|
2003-07-05 07:46:06 +02:00
|
|
|
|
2002-06-27 00:45:49 +02:00
|
|
|
/********* START VARIABLES **********/
|
|
|
|
|
2003-08-11 22:22:48 +02:00
|
|
|
extern char *conn_type_to_string[];
|
2003-09-25 12:42:07 +02:00
|
|
|
extern char *conn_state_to_string[][_CONN_TYPE_MAX+1];
|
2003-08-11 22:22:48 +02:00
|
|
|
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
or_options_t options; /* command-line and config-file options */
|
2003-07-05 09:10:34 +02:00
|
|
|
int global_read_bucket; /* max number of bytes I can read this second */
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2003-10-02 22:00:38 +02:00
|
|
|
static int stats_prev_global_read_bucket;
|
|
|
|
static uint64_t stats_n_bytes_read = 0;
|
|
|
|
static long stats_n_seconds_reading = 0;
|
|
|
|
|
2002-07-05 08:27:23 +02:00
|
|
|
static connection_t *connection_array[MAXCONNECTIONS] =
|
2002-06-27 00:45:49 +02:00
|
|
|
{ NULL };
|
|
|
|
|
2002-09-03 20:44:24 +02:00
|
|
|
static struct pollfd poll_array[MAXCONNECTIONS];
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2002-07-05 08:27:23 +02:00
|
|
|
static int nfds=0; /* number of connections currently active */
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2003-08-12 10:04:31 +02:00
|
|
|
#ifndef MS_WINDOWS /* do signal stuff only on unix */
|
2002-09-22 00:41:48 +02:00
|
|
|
static int please_dumpstats=0; /* whether we should dump stats during the loop */
|
2003-11-13 07:49:25 +01:00
|
|
|
static int please_reset=0; /* whether we just got a sighup */
|
|
|
|
static int please_reap_children=0; /* whether we should waitpid for exited children */
|
2003-08-12 10:04:31 +02:00
|
|
|
#endif /* signal stuff */
|
2002-09-22 00:41:48 +02:00
|
|
|
|
2004-02-29 02:31:33 +01:00
|
|
|
int has_fetched_directory=0;
|
|
|
|
/* we set this to 1 when we've fetched a dir, to know whether to complain
|
|
|
|
* yet about unrecognized nicknames in entrynodes, exitnodes, etc. */
|
|
|
|
|
|
|
|
int has_completed_circuit=0;
|
|
|
|
/* we set this to 1 when we've opened a circuit, so we can print a log
|
|
|
|
* entry to inform the user that Tor is working. */
|
|
|
|
|
2002-06-27 00:45:49 +02:00
|
|
|
/********* END VARIABLES ************/
|
|
|
|
|
2002-07-05 08:27:23 +02:00
|
|
|
/****************************************************************************
|
|
|
|
*
|
|
|
|
* This section contains accessors and other methods on the connection_array
|
|
|
|
* and poll_array variables (which are global within this file and unavailable
|
|
|
|
* outside it).
|
|
|
|
*
|
|
|
|
****************************************************************************/
|
|
|
|
|
2002-06-27 00:45:49 +02:00
|
|
|
int connection_add(connection_t *conn) {
|
|
|
|
|
2002-09-03 20:36:40 +02:00
|
|
|
if(nfds >= options.MaxConn-1) {
|
2003-10-10 03:48:32 +02:00
|
|
|
log(LOG_WARN,"connection_add(): failing because nfds is too high.");
|
2002-06-27 00:45:49 +02:00
|
|
|
return -1;
|
|
|
|
}
|
2003-12-05 10:51:49 +01:00
|
|
|
|
2002-06-27 00:45:49 +02:00
|
|
|
conn->poll_index = nfds;
|
|
|
|
connection_set_poll_socket(conn);
|
|
|
|
connection_array[nfds] = conn;
|
|
|
|
|
|
|
|
/* zero these out here, because otherwise we'll inherit values from the previously freed one */
|
|
|
|
poll_array[nfds].events = 0;
|
|
|
|
poll_array[nfds].revents = 0;
|
|
|
|
|
|
|
|
nfds++;
|
|
|
|
|
2002-07-16 04:12:58 +02:00
|
|
|
log(LOG_INFO,"connection_add(): new conn type %d, socket %d, nfds %d.",conn->type, conn->s, nfds);
|
2002-06-27 00:45:49 +02:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
void connection_set_poll_socket(connection_t *conn) {
|
|
|
|
poll_array[conn->poll_index].fd = conn->s;
|
|
|
|
}
|
|
|
|
|
2003-10-10 03:48:32 +02:00
|
|
|
/* Remove the connection from the global list, and remove the
|
2003-10-09 20:45:14 +02:00
|
|
|
* corresponding poll entry. Calling this function will shift the last
|
|
|
|
* connection (if any) into the position occupied by conn.
|
|
|
|
*/
|
2002-06-27 00:45:49 +02:00
|
|
|
int connection_remove(connection_t *conn) {
|
|
|
|
int current_index;
|
|
|
|
|
|
|
|
assert(conn);
|
|
|
|
assert(nfds>0);
|
|
|
|
|
2002-07-22 06:08:37 +02:00
|
|
|
log(LOG_INFO,"connection_remove(): removing socket %d, nfds now %d",conn->s, nfds-1);
|
2003-10-10 03:48:32 +02:00
|
|
|
/* if it's an edge conn, remove it from the list
|
|
|
|
* of conn's on this circuit. If it's not on an edge,
|
|
|
|
* flush and send destroys for all circuits on this conn
|
|
|
|
*/
|
|
|
|
circuit_about_to_close_connection(conn);
|
2002-06-27 00:45:49 +02:00
|
|
|
|
|
|
|
current_index = conn->poll_index;
|
|
|
|
if(current_index == nfds-1) { /* this is the end */
|
|
|
|
nfds--;
|
|
|
|
return 0;
|
2003-12-05 10:51:49 +01:00
|
|
|
}
|
2002-06-27 00:45:49 +02:00
|
|
|
|
|
|
|
/* we replace this one with the one at the end, then free it */
|
|
|
|
nfds--;
|
2003-12-05 10:51:49 +01:00
|
|
|
poll_array[current_index].fd = poll_array[nfds].fd;
|
2002-06-27 00:45:49 +02:00
|
|
|
poll_array[current_index].events = poll_array[nfds].events;
|
|
|
|
poll_array[current_index].revents = poll_array[nfds].revents;
|
|
|
|
connection_array[current_index] = connection_array[nfds];
|
|
|
|
connection_array[current_index]->poll_index = current_index;
|
|
|
|
|
2003-12-05 10:51:49 +01:00
|
|
|
return 0;
|
2002-06-27 00:45:49 +02:00
|
|
|
}
|
|
|
|
|
2003-09-30 21:06:22 +02:00
|
|
|
void get_connection_array(connection_t ***array, int *n) {
|
|
|
|
*array = connection_array;
|
|
|
|
*n = nfds;
|
2003-06-25 09:19:30 +02:00
|
|
|
}
|
|
|
|
|
2002-06-27 00:45:49 +02:00
|
|
|
void connection_watch_events(connection_t *conn, short events) {
|
|
|
|
|
|
|
|
assert(conn && conn->poll_index < nfds);
|
|
|
|
|
|
|
|
poll_array[conn->poll_index].events = events;
|
|
|
|
}
|
|
|
|
|
2003-09-07 12:24:40 +02:00
|
|
|
int connection_is_reading(connection_t *conn) {
|
|
|
|
return poll_array[conn->poll_index].events & POLLIN;
|
|
|
|
}
|
|
|
|
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
void connection_stop_reading(connection_t *conn) {
|
|
|
|
|
|
|
|
assert(conn && conn->poll_index < nfds);
|
|
|
|
|
2002-07-18 08:37:58 +02:00
|
|
|
log(LOG_DEBUG,"connection_stop_reading() called.");
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
if(poll_array[conn->poll_index].events & POLLIN)
|
|
|
|
poll_array[conn->poll_index].events -= POLLIN;
|
|
|
|
}
|
|
|
|
|
|
|
|
void connection_start_reading(connection_t *conn) {
|
|
|
|
|
|
|
|
assert(conn && conn->poll_index < nfds);
|
|
|
|
|
|
|
|
poll_array[conn->poll_index].events |= POLLIN;
|
|
|
|
}
|
|
|
|
|
2004-02-27 05:42:14 +01:00
|
|
|
int connection_is_writing(connection_t *conn) {
|
|
|
|
return poll_array[conn->poll_index].events & POLLOUT;
|
|
|
|
}
|
|
|
|
|
2002-07-18 08:37:58 +02:00
|
|
|
void connection_stop_writing(connection_t *conn) {
|
|
|
|
|
|
|
|
assert(conn && conn->poll_index < nfds);
|
|
|
|
|
|
|
|
if(poll_array[conn->poll_index].events & POLLOUT)
|
|
|
|
poll_array[conn->poll_index].events -= POLLOUT;
|
|
|
|
}
|
|
|
|
|
|
|
|
void connection_start_writing(connection_t *conn) {
|
|
|
|
|
|
|
|
assert(conn && conn->poll_index < nfds);
|
|
|
|
|
|
|
|
poll_array[conn->poll_index].events |= POLLOUT;
|
|
|
|
}
|
|
|
|
|
2003-07-05 07:46:06 +02:00
|
|
|
static void conn_read(int i) {
|
2003-09-28 08:48:20 +02:00
|
|
|
connection_t *conn = connection_array[i];
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2004-02-28 05:44:48 +01:00
|
|
|
if (conn->marked_for_close)
|
|
|
|
return;
|
|
|
|
|
2003-09-28 08:48:20 +02:00
|
|
|
/* see http://www.greenend.org.uk/rjk/2001/06/poll.html for
|
|
|
|
* discussion of POLLIN vs POLLHUP */
|
2003-09-05 08:04:03 +02:00
|
|
|
if(!(poll_array[i].revents & (POLLIN|POLLHUP|POLLERR)))
|
2003-10-09 20:45:14 +02:00
|
|
|
if(!connection_is_reading(conn) ||
|
|
|
|
!connection_has_pending_tls_data(conn))
|
2003-09-28 08:48:20 +02:00
|
|
|
return; /* this conn should not read */
|
2003-08-14 19:13:52 +02:00
|
|
|
|
cleanups, bugfixes, more verbose logs
Fixed up the assert_*_ok funcs some (more work remains)
Changed config so it reads either /etc/torrc or the -f arg, never both
Finally tracked down a nasty bug with our use of tls:
It turns out that if you ask SSL_read() for no more than n bytes, it
will read the entire record from the network (and maybe part of the next
record, I'm not sure), give you n bytes of it, and keep the remaining
bytes internally. This is fine, except our poll-for-read looks at the
network, and there are no bytes pending on the network, so we never know
to ask SSL_read() for more bytes. Currently I've hacked it so if we ask
for n bytes and it returns n bytes, then it reads again right then. This
will interact poorly with our rate limiting; we need a cleaner solution.
svn:r481
2003-09-24 23:24:52 +02:00
|
|
|
log_fn(LOG_DEBUG,"socket %d wants to read.",conn->s);
|
2003-12-05 10:51:49 +01:00
|
|
|
|
2003-09-23 21:47:41 +02:00
|
|
|
assert_connection_ok(conn, time(NULL));
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
|
2003-09-05 08:04:03 +02:00
|
|
|
if(
|
|
|
|
/* XXX does POLLHUP also mean it's definitely broken? */
|
2003-08-14 19:13:52 +02:00
|
|
|
#ifdef MS_WINDOWS
|
2003-09-05 08:04:03 +02:00
|
|
|
(poll_array[i].revents & POLLERR) ||
|
2003-08-14 19:13:52 +02:00
|
|
|
#endif
|
2004-02-28 05:34:27 +01:00
|
|
|
connection_handle_read(conn) < 0) {
|
|
|
|
if (!conn->marked_for_close) {
|
|
|
|
/* this connection is broken. remove it */
|
|
|
|
/* XXX This shouldn't ever happen anymore. */
|
2004-02-28 12:48:22 +01:00
|
|
|
/* XXX but it'll clearly happen on MS_WINDOWS from POLLERR, right? */
|
2004-02-28 05:34:27 +01:00
|
|
|
log_fn(LOG_ERR,"Unhandled error on read for %s connection (fd %d); removing",
|
|
|
|
conn_type_to_string[conn->type], conn->s);
|
|
|
|
connection_mark_for_close(conn,0);
|
|
|
|
}
|
2004-02-28 05:11:53 +01:00
|
|
|
}
|
|
|
|
assert_connection_ok(conn, time(NULL));
|
2002-06-27 00:45:49 +02:00
|
|
|
}
|
|
|
|
|
2003-07-05 07:46:06 +02:00
|
|
|
static void conn_write(int i) {
|
2002-06-27 00:45:49 +02:00
|
|
|
connection_t *conn;
|
|
|
|
|
2003-09-05 08:04:03 +02:00
|
|
|
if(!(poll_array[i].revents & POLLOUT))
|
|
|
|
return; /* this conn doesn't want to write */
|
|
|
|
|
2003-07-05 07:46:06 +02:00
|
|
|
conn = connection_array[i];
|
2004-03-03 09:46:18 +01:00
|
|
|
log_fn(LOG_DEBUG,"socket %d wants to write.",conn->s);
|
2004-02-28 05:11:53 +01:00
|
|
|
if (conn->marked_for_close)
|
|
|
|
return;
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2003-09-23 21:47:41 +02:00
|
|
|
assert_connection_ok(conn, time(NULL));
|
|
|
|
|
2004-02-28 05:34:27 +01:00
|
|
|
if (connection_handle_write(conn) < 0) {
|
|
|
|
if (!conn->marked_for_close) {
|
|
|
|
/* this connection is broken. remove it. */
|
2004-02-28 05:44:48 +01:00
|
|
|
log_fn(LOG_WARN,"Unhandled error on read for %s connection (fd %d); removing",
|
2004-02-28 05:34:27 +01:00
|
|
|
conn_type_to_string[conn->type], conn->s);
|
2004-03-03 05:54:16 +01:00
|
|
|
conn->has_sent_end = 1; /* otherwise we cry wolf about duplicate close */
|
2004-02-28 05:34:27 +01:00
|
|
|
connection_mark_for_close(conn,0);
|
|
|
|
}
|
2004-02-28 05:11:53 +01:00
|
|
|
}
|
|
|
|
assert_connection_ok(conn, time(NULL));
|
2002-06-27 00:45:49 +02:00
|
|
|
}
|
|
|
|
|
2003-10-09 20:45:14 +02:00
|
|
|
static void conn_close_if_marked(int i) {
|
2002-06-27 00:45:49 +02:00
|
|
|
connection_t *conn;
|
2004-03-03 09:46:18 +01:00
|
|
|
int retval;
|
2002-06-27 00:45:49 +02:00
|
|
|
|
|
|
|
conn = connection_array[i];
|
2003-09-23 21:47:41 +02:00
|
|
|
assert_connection_ok(conn, time(NULL));
|
2004-03-03 09:46:18 +01:00
|
|
|
if(!conn->marked_for_close)
|
|
|
|
return; /* nothing to see here, move along */
|
|
|
|
|
|
|
|
log_fn(LOG_INFO,"Cleaning up connection (fd %d).",conn->s);
|
|
|
|
if(conn->s >= 0 && connection_wants_to_flush(conn)) {
|
|
|
|
/* -1 means it's an incomplete edge connection, or that the socket
|
|
|
|
* has already been closed as unflushable. */
|
|
|
|
if(!conn->hold_open_until_flushed)
|
2004-03-03 03:07:57 +01:00
|
|
|
log_fn(LOG_WARN,
|
2004-03-03 09:46:18 +01:00
|
|
|
"Conn (fd %d, type %d, state %d) marked, but wants to flush %d bytes. "
|
|
|
|
"(Marked at %s:%d)",
|
|
|
|
conn->s, conn->type, conn->state,
|
|
|
|
conn->outbuf_flushlen, conn->marked_for_close_file, conn->marked_for_close);
|
|
|
|
if(connection_speaks_cells(conn)) {
|
|
|
|
if(conn->state == OR_CONN_STATE_OPEN) {
|
|
|
|
retval = flush_buf_tls(conn->tls, conn->outbuf, &conn->outbuf_flushlen);
|
|
|
|
/* XXX actually, some non-zero results are maybe ok. which ones? */
|
|
|
|
} else
|
2004-03-06 02:43:37 +01:00
|
|
|
retval = -1; /* never flush non-open broken tls connections */
|
2004-03-03 09:46:18 +01:00
|
|
|
} else {
|
|
|
|
retval = flush_buf(conn->s, conn->outbuf, &conn->outbuf_flushlen);
|
|
|
|
}
|
2004-03-03 09:48:32 +01:00
|
|
|
if(retval >= 0 &&
|
2004-03-03 09:46:18 +01:00
|
|
|
conn->hold_open_until_flushed && connection_wants_to_flush(conn)) {
|
|
|
|
log_fn(LOG_INFO,"Holding conn (fd %d) open for more flushing.",conn->s);
|
|
|
|
/* XXX should we reset timestamp_lastwritten here? */
|
|
|
|
return;
|
Integrated onion proxy into or/
The 'or' process can now be told (by the global_role variable) what
roles this server should play -- connect to all ORs, listen for ORs,
listen for OPs, listen for APs, or any combination.
* everything in /src/op/ is now obsolete.
* connection_ap.c now handles all interactions with application proxies
* "port" is now or_port, op_port, ap_port. But routers are still always
referenced (say, in conn_get_by_addr_port()) by addr / or_port. We
should make routers.c actually read these new ports (currently I've
kludged it so op_port = or_port+10, ap_port=or_port+20)
* circuits currently know if they're at the beginning of the path because
circ->cpath is set. They use this instead for crypts (both ways),
if it's set.
* I still obey the "send a 0 back to the AP when you're ready" protocol,
but I think we should phase it out. I can simply not read from the AP
socket until I'm ready.
I need to do a lot of cleanup work here, but the code appears to work, so
now's a good time for a checkin.
svn:r22
2002-07-02 11:36:58 +02:00
|
|
|
}
|
2004-03-03 09:46:18 +01:00
|
|
|
if(connection_wants_to_flush(conn)) {
|
|
|
|
log_fn(LOG_WARN,"Conn (fd %d) still wants to flush. Losing %d bytes!",
|
|
|
|
conn->s, (int)buf_datalen(conn->outbuf));
|
2002-06-27 00:45:49 +02:00
|
|
|
}
|
|
|
|
}
|
2004-03-03 09:46:18 +01:00
|
|
|
connection_remove(conn);
|
|
|
|
connection_free(conn);
|
|
|
|
if(i<nfds) { /* we just replaced the one at i with a new one.
|
|
|
|
process it too. */
|
|
|
|
conn_close_if_marked(i);
|
|
|
|
}
|
2002-06-27 00:45:49 +02:00
|
|
|
}
|
|
|
|
|
2003-10-10 03:48:32 +02:00
|
|
|
/* Perform regular maintenance tasks for a single connection. This
|
2003-10-09 20:45:14 +02:00
|
|
|
* function gets run once per second per connection by run_housekeeping.
|
|
|
|
*/
|
|
|
|
static void run_connection_housekeeping(int i, time_t now) {
|
|
|
|
cell_t cell;
|
|
|
|
connection_t *conn = connection_array[i];
|
2003-12-05 10:51:49 +01:00
|
|
|
|
2003-10-09 20:45:14 +02:00
|
|
|
if(connection_receiver_bucket_should_increase(conn)) {
|
|
|
|
conn->receiver_bucket += conn->bandwidth;
|
|
|
|
// log_fn(LOG_DEBUG,"Receiver bucket %d now %d.", i, conn->receiver_bucket);
|
|
|
|
}
|
|
|
|
|
|
|
|
if(conn->wants_to_read == 1 /* it's marked to turn reading back on now */
|
|
|
|
&& global_read_bucket > 0 /* and we're allowed to read */
|
|
|
|
&& (!connection_speaks_cells(conn) || conn->receiver_bucket > 0)) {
|
|
|
|
/* and either a non-cell conn or a cell conn with non-empty bucket */
|
|
|
|
conn->wants_to_read = 0;
|
|
|
|
connection_start_reading(conn);
|
|
|
|
if(conn->wants_to_write == 1) {
|
|
|
|
conn->wants_to_write = 0;
|
|
|
|
connection_start_writing(conn);
|
|
|
|
}
|
|
|
|
}
|
2003-12-05 10:51:49 +01:00
|
|
|
|
2003-10-09 20:45:14 +02:00
|
|
|
/* check connections to see whether we should send a keepalive, expire, or wait */
|
|
|
|
if(!connection_speaks_cells(conn))
|
|
|
|
return;
|
|
|
|
|
|
|
|
if(now >= conn->timestamp_lastwritten + options.KeepalivePeriod) {
|
2003-11-20 18:49:45 +01:00
|
|
|
if((!options.ORPort && !circuit_get_by_conn(conn)) ||
|
2003-10-09 20:45:14 +02:00
|
|
|
(!connection_state_is_open(conn))) {
|
|
|
|
/* we're an onion proxy, with no circuits; or our handshake has expired. kill it. */
|
|
|
|
log_fn(LOG_INFO,"Expiring connection to %d (%s:%d).",
|
|
|
|
i,conn->address, conn->port);
|
2004-02-27 23:00:26 +01:00
|
|
|
connection_mark_for_close(conn,0); /* Suppress end ??? */
|
2004-02-28 00:23:33 +01:00
|
|
|
/* XXX there's no concept of 'suppressing end' here, because it's an OR
|
|
|
|
* connection, and there's no such thing as an end cell for an OR
|
|
|
|
* connection. -RD */
|
2003-10-09 20:45:14 +02:00
|
|
|
} else {
|
|
|
|
/* either a full router, or we've got a circuit. send a padding cell. */
|
|
|
|
log_fn(LOG_DEBUG,"Sending keepalive to (%s:%d)",
|
|
|
|
conn->address, conn->port);
|
|
|
|
memset(&cell,0,sizeof(cell_t));
|
|
|
|
cell.command = CELL_PADDING;
|
|
|
|
connection_or_write_cell_to_buf(&cell, conn);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Perform regular maintenance tasks. This function gets run once per
|
|
|
|
* second by prepare_for_poll.
|
|
|
|
*/
|
|
|
|
static void run_scheduled_events(time_t now) {
|
2002-10-02 01:37:31 +02:00
|
|
|
static long time_to_fetch_directory = 0;
|
2003-04-16 08:18:31 +02:00
|
|
|
static long time_to_new_circuit = 0;
|
|
|
|
circuit_t *circ;
|
2003-10-09 20:45:14 +02:00
|
|
|
int i;
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
|
2003-10-09 20:45:14 +02:00
|
|
|
/* 1. Every DirFetchPostPeriod seconds, we get a new directory and upload
|
|
|
|
* our descriptor (if any). */
|
|
|
|
if(time_to_fetch_directory < now) {
|
|
|
|
/* it's time to fetch a new directory and/or post our descriptor */
|
2003-11-20 18:49:45 +01:00
|
|
|
if(options.ORPort) {
|
2003-10-09 20:45:14 +02:00
|
|
|
router_rebuild_descriptor();
|
|
|
|
router_upload_desc_to_dirservers();
|
|
|
|
}
|
|
|
|
if(!options.DirPort) {
|
|
|
|
/* NOTE directory servers do not currently fetch directories.
|
|
|
|
* Hope this doesn't bite us later. */
|
|
|
|
directory_initiate_command(router_pick_directory_server(),
|
|
|
|
DIR_CONN_STATE_CONNECTING_FETCH);
|
|
|
|
}
|
|
|
|
time_to_fetch_directory = now + options.DirFetchPostPeriod;
|
|
|
|
}
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
|
2003-11-18 08:48:00 +01:00
|
|
|
/* 2. Every second, we examine pending circuits and prune the
|
2003-12-05 10:51:49 +01:00
|
|
|
* ones which have been pending for more than 3 seconds.
|
2003-11-18 08:48:00 +01:00
|
|
|
* We do this before step 3, so it can try building more if
|
|
|
|
* it's not comfortable with the number of available circuits.
|
|
|
|
*/
|
|
|
|
circuit_expire_building();
|
|
|
|
|
2004-01-20 10:21:46 +01:00
|
|
|
/* 2b. Also look at pending streams and prune the ones that 'began'
|
|
|
|
* a long time ago but haven't gotten a 'connected' yet.
|
|
|
|
* Do this before step 3, so we can put them back into pending
|
|
|
|
* state to be picked up by the new circuit.
|
|
|
|
*/
|
|
|
|
connection_ap_expire_beginning();
|
|
|
|
|
2004-03-03 06:08:01 +01:00
|
|
|
|
2004-03-03 07:26:34 +01:00
|
|
|
/* 2c. And expire connections that we've held open for too long.
|
2004-03-03 06:08:01 +01:00
|
|
|
*/
|
|
|
|
connection_expire_held_open();
|
|
|
|
|
2003-11-18 08:48:00 +01:00
|
|
|
/* 3. Every second, we try a new circuit if there are no valid
|
2003-11-16 22:49:52 +01:00
|
|
|
* circuits. Every NewCircuitPeriod seconds, we expire circuits
|
|
|
|
* that became dirty more than NewCircuitPeriod seconds ago,
|
|
|
|
* and we make a new circ if there are no clean circuits.
|
2003-10-09 20:45:14 +02:00
|
|
|
*/
|
2003-11-16 22:49:52 +01:00
|
|
|
if(options.SocksPort) {
|
2003-11-17 01:57:56 +01:00
|
|
|
|
|
|
|
/* launch a new circ for any pending streams that need one */
|
|
|
|
connection_ap_attach_pending();
|
|
|
|
|
|
|
|
circ = circuit_get_newest(NULL, 1);
|
2003-11-16 22:49:52 +01:00
|
|
|
if(time_to_new_circuit < now) {
|
|
|
|
client_dns_clean();
|
|
|
|
circuit_expire_unused_circuits();
|
2003-11-19 03:22:52 +01:00
|
|
|
circuit_reset_failure_count();
|
2003-11-16 22:49:52 +01:00
|
|
|
if(circ && circ->timestamp_dirty) {
|
|
|
|
log_fn(LOG_INFO,"Youngest circuit dirty; launching replacement.");
|
2003-11-19 03:22:52 +01:00
|
|
|
circuit_launch_new(); /* make a new circuit */
|
2003-11-16 22:49:52 +01:00
|
|
|
}
|
|
|
|
time_to_new_circuit = now + options.NewCircuitPeriod;
|
|
|
|
}
|
2004-01-30 20:31:39 +01:00
|
|
|
#define CIRCUIT_MIN_BUILDING 3
|
2003-11-19 03:22:52 +01:00
|
|
|
if(!circ && circuit_count_building() < CIRCUIT_MIN_BUILDING) {
|
2004-01-30 20:31:39 +01:00
|
|
|
/* if there's no open circ, and less than 3 are on the way,
|
2003-11-18 08:48:00 +01:00
|
|
|
* go ahead and try another.
|
|
|
|
*/
|
2003-11-19 03:22:52 +01:00
|
|
|
circuit_launch_new();
|
|
|
|
}
|
2003-10-09 20:45:14 +02:00
|
|
|
}
|
2003-11-16 22:49:52 +01:00
|
|
|
|
2003-12-05 10:51:49 +01:00
|
|
|
/* 4. Every second, we check how much bandwidth we've consumed and
|
2003-10-09 20:45:14 +02:00
|
|
|
* increment global_read_bucket.
|
|
|
|
*/
|
|
|
|
stats_n_bytes_read += stats_prev_global_read_bucket-global_read_bucket;
|
2004-01-11 00:40:38 +01:00
|
|
|
if(global_read_bucket < options.BandwidthBurst) {
|
|
|
|
global_read_bucket += options.BandwidthRate;
|
2003-10-09 20:45:14 +02:00
|
|
|
log_fn(LOG_DEBUG,"global_read_bucket now %d.", global_read_bucket);
|
|
|
|
}
|
|
|
|
stats_prev_global_read_bucket = global_read_bucket;
|
2003-03-06 05:52:02 +01:00
|
|
|
|
2003-11-18 08:48:00 +01:00
|
|
|
/* 5. We do housekeeping for each connection... */
|
2003-10-09 20:45:14 +02:00
|
|
|
for(i=0;i<nfds;i++) {
|
|
|
|
run_connection_housekeeping(i, now);
|
|
|
|
}
|
2003-10-02 22:00:38 +02:00
|
|
|
|
2004-03-02 18:48:17 +01:00
|
|
|
/* 6. And remove any marked circuits... */
|
|
|
|
circuit_close_all_marked();
|
|
|
|
|
|
|
|
/* 7. and blow away any connections that need to die. can't do this later
|
2003-10-09 20:45:14 +02:00
|
|
|
* because we might open up a circuit and not realize we're about to cull
|
|
|
|
* the connection it's running over.
|
2003-11-16 22:49:52 +01:00
|
|
|
* XXX we can remove this step once we audit circuit-building to make sure
|
|
|
|
* it doesn't pick a marked-for-close conn. -RD
|
2003-10-09 20:45:14 +02:00
|
|
|
*/
|
|
|
|
for(i=0;i<nfds;i++)
|
|
|
|
conn_close_if_marked(i);
|
|
|
|
}
|
2002-09-26 14:09:10 +02:00
|
|
|
|
2003-10-09 20:45:14 +02:00
|
|
|
static int prepare_for_poll(void) {
|
|
|
|
static long current_second = 0; /* from previous calls to gettimeofday */
|
|
|
|
connection_t *conn;
|
|
|
|
struct timeval now;
|
|
|
|
int i;
|
2003-04-16 08:18:31 +02:00
|
|
|
|
2003-10-09 20:45:14 +02:00
|
|
|
tor_gettimeofday(&now);
|
2003-07-05 09:10:34 +02:00
|
|
|
|
2003-10-09 20:45:14 +02:00
|
|
|
if(now.tv_sec > current_second) { /* the second has rolled over. check more stuff. */
|
2002-09-26 14:09:10 +02:00
|
|
|
|
2003-10-09 20:45:14 +02:00
|
|
|
++stats_n_seconds_reading;
|
|
|
|
run_scheduled_events(now.tv_sec);
|
2002-10-02 01:37:31 +02:00
|
|
|
|
2002-12-31 16:04:14 +01:00
|
|
|
current_second = now.tv_sec; /* remember which second it is, for next time */
|
2002-10-02 01:37:31 +02:00
|
|
|
}
|
|
|
|
|
2003-09-30 22:36:20 +02:00
|
|
|
for(i=0;i<nfds;i++) {
|
|
|
|
conn = connection_array[i];
|
2003-10-09 20:45:14 +02:00
|
|
|
if(connection_has_pending_tls_data(conn)) {
|
2003-09-30 22:36:20 +02:00
|
|
|
log_fn(LOG_DEBUG,"sock %d has pending bytes.",conn->s);
|
|
|
|
return 0; /* has pending bytes to read; don't let poll wait. */
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return (1000 - (now.tv_usec / 1000)); /* how many milliseconds til the next second? */
|
2003-03-06 05:52:02 +01:00
|
|
|
}
|
|
|
|
|
2003-11-13 07:49:25 +01:00
|
|
|
static int init_from_config(int argc, char **argv) {
|
|
|
|
if(getconfig(argc,argv,&options)) {
|
|
|
|
log_fn(LOG_ERR,"Reading config failed. For usage, try -h.");
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
close_logs(); /* we'll close, then open with correct loglevel if necessary */
|
2004-02-29 23:34:38 +01:00
|
|
|
|
|
|
|
if(options.User || options.Group) {
|
|
|
|
if(switch_id(options.User, options.Group) != 0) {
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (options.RunAsDaemon) {
|
|
|
|
start_daemon(options.DataDirectory);
|
|
|
|
}
|
|
|
|
|
2003-11-13 07:49:25 +01:00
|
|
|
if(!options.LogFile && !options.RunAsDaemon)
|
|
|
|
add_stream_log(options.loglevel, "<stdout>", stdout);
|
2004-01-05 06:22:04 +01:00
|
|
|
if(options.LogFile) {
|
2003-11-19 03:09:43 +01:00
|
|
|
if (add_file_log(options.loglevel, options.LogFile) != 0) {
|
|
|
|
/* opening the log file failed! Use stderr and log a warning */
|
|
|
|
add_stream_log(options.loglevel, "<stderr>", stderr);
|
|
|
|
log_fn(LOG_WARN, "Cannot write to LogFile '%s': %s.", options.LogFile, strerror(errno));
|
|
|
|
}
|
2004-01-05 06:22:04 +01:00
|
|
|
log_fn(LOG_WARN, "Successfully opened LogFile '%s', redirecting output.",
|
|
|
|
options.LogFile);
|
|
|
|
}
|
|
|
|
if(options.DebugLogFile) {
|
2003-11-19 03:09:43 +01:00
|
|
|
if (add_file_log(LOG_DEBUG, options.DebugLogFile) != 0)
|
2004-01-05 06:22:04 +01:00
|
|
|
log_fn(LOG_WARN, "Cannot write to DebugLogFile '%s': %s.", options.DebugLogFile, strerror(errno));
|
|
|
|
log_fn(LOG_DEBUG, "Successfully opened DebugLogFile '%s'.", options.DebugLogFile);
|
|
|
|
}
|
2003-11-13 07:49:25 +01:00
|
|
|
|
2004-01-11 00:40:38 +01:00
|
|
|
global_read_bucket = options.BandwidthBurst; /* start it at max traffic */
|
2003-11-13 07:49:25 +01:00
|
|
|
stats_prev_global_read_bucket = global_read_bucket;
|
|
|
|
|
2004-01-03 23:40:49 +01:00
|
|
|
if(options.RunAsDaemon) {
|
|
|
|
/* XXXX Can we delay this any more? */
|
|
|
|
finish_daemon();
|
2003-11-13 07:49:25 +01:00
|
|
|
}
|
|
|
|
|
2003-11-19 03:09:43 +01:00
|
|
|
/* write our pid to the pid file, if we do not have write permissions we will log a warning */
|
2003-11-19 23:45:06 +01:00
|
|
|
if(options.PidFile)
|
|
|
|
write_pidfile(options.PidFile);
|
2003-11-19 03:09:43 +01:00
|
|
|
|
2003-11-13 07:49:25 +01:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2004-02-26 23:30:44 +01:00
|
|
|
static int do_hup(void) {
|
|
|
|
char keydir[512];
|
|
|
|
|
|
|
|
log_fn(LOG_WARN,"Received sighup. Reloading config.");
|
2004-02-29 04:52:38 +01:00
|
|
|
has_completed_circuit=0;
|
2004-02-26 23:30:44 +01:00
|
|
|
/* first, reload config variables, in case they've changed */
|
|
|
|
/* no need to provide argc/v, they've been cached inside init_from_config */
|
|
|
|
if (init_from_config(0, NULL) < 0) {
|
|
|
|
exit(1);
|
|
|
|
}
|
|
|
|
if(retry_all_connections() < 0) {
|
|
|
|
log_fn(LOG_ERR,"Failed to bind one of the listener ports.");
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
if(options.DirPort) {
|
|
|
|
/* reload the approved-routers file */
|
|
|
|
sprintf(keydir,"%s/approved-routers", options.DataDirectory);
|
|
|
|
log_fn(LOG_INFO,"Reloading approved fingerprints from %s...",keydir);
|
|
|
|
if(dirserv_parse_fingerprint_file(keydir) < 0) {
|
|
|
|
log_fn(LOG_WARN, "Error reloading fingerprints. Continuing with old list.");
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
/* fetch a new directory */
|
|
|
|
directory_initiate_command(router_pick_directory_server(), DIR_CONN_STATE_CONNECTING_FETCH);
|
|
|
|
}
|
|
|
|
if(options.ORPort) {
|
|
|
|
router_rebuild_descriptor();
|
|
|
|
sprintf(keydir,"%s/router.desc", options.DataDirectory);
|
|
|
|
log_fn(LOG_INFO,"Dumping descriptor to %s...",keydir);
|
|
|
|
if (write_str_to_file(keydir, router_get_my_descriptor())) {
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
static int do_main_loop(void) {
|
|
|
|
int i;
|
|
|
|
int timeout;
|
|
|
|
int poll_result;
|
2003-12-05 10:51:49 +01:00
|
|
|
|
2003-09-25 07:17:11 +02:00
|
|
|
/* load the routers file */
|
2004-02-26 22:25:51 +01:00
|
|
|
if(options.RouterFile &&
|
|
|
|
router_set_routerlist_from_file(options.RouterFile) < 0) {
|
2003-09-25 07:17:11 +02:00
|
|
|
log_fn(LOG_ERR,"Error loading router list.");
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* load the private keys, if we're supposed to have them, and set up the
|
|
|
|
* TLS context. */
|
|
|
|
if (init_keys() < 0) {
|
|
|
|
log_fn(LOG_ERR,"Error initializing keys; exiting");
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
2003-11-20 18:49:45 +01:00
|
|
|
if(options.ORPort) {
|
2003-09-25 12:42:07 +02:00
|
|
|
cpu_init(); /* launch cpuworkers. Need to do this *after* we've read the onion key. */
|
2003-09-27 00:02:02 +02:00
|
|
|
router_upload_desc_to_dirservers(); /* upload our descriptor to all dirservers */
|
2003-09-08 08:26:38 +02:00
|
|
|
}
|
2003-09-07 12:24:40 +02:00
|
|
|
|
2003-03-18 02:49:55 +01:00
|
|
|
/* start up the necessary connections based on which ports are
|
|
|
|
* non-zero. This is where we try to connect to all the other ORs,
|
2003-09-05 08:04:03 +02:00
|
|
|
* and start the listeners.
|
2003-03-18 02:49:55 +01:00
|
|
|
*/
|
2003-10-25 14:01:09 +02:00
|
|
|
if(retry_all_connections() < 0) {
|
|
|
|
log_fn(LOG_ERR,"Failed to bind one of the listener ports.");
|
|
|
|
return -1;
|
|
|
|
}
|
2002-06-27 00:45:49 +02:00
|
|
|
|
|
|
|
for(;;) {
|
2003-10-22 08:03:11 +02:00
|
|
|
#ifndef MS_WINDOWS /* do signal stuff only on unix */
|
2002-09-22 00:41:48 +02:00
|
|
|
if(please_dumpstats) {
|
2003-10-15 20:50:16 +02:00
|
|
|
/* prefer to log it at INFO, but make sure we always see it */
|
|
|
|
dumpstats(options.loglevel>LOG_INFO ? options.loglevel : LOG_INFO);
|
2002-09-28 03:40:11 +02:00
|
|
|
please_dumpstats = 0;
|
|
|
|
}
|
2003-09-21 08:15:43 +02:00
|
|
|
if(please_reset) {
|
2004-02-26 23:30:44 +01:00
|
|
|
do_hup();
|
2003-09-21 08:15:43 +02:00
|
|
|
please_reset = 0;
|
2002-09-22 00:41:48 +02:00
|
|
|
}
|
2003-08-12 08:41:53 +02:00
|
|
|
if(please_reap_children) {
|
|
|
|
while(waitpid(-1,NULL,WNOHANG)) ; /* keep reaping until no more zombies */
|
|
|
|
please_reap_children = 0;
|
|
|
|
}
|
2003-08-12 10:04:31 +02:00
|
|
|
#endif /* signal stuff */
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
|
2003-09-05 08:04:03 +02:00
|
|
|
timeout = prepare_for_poll();
|
|
|
|
|
|
|
|
/* poll until we have an event, or the second ends */
|
2004-02-21 00:41:45 +01:00
|
|
|
poll_result = tor_poll(poll_array, nfds, timeout);
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
|
2003-10-15 20:28:32 +02:00
|
|
|
/* let catch() handle things like ^c, and otherwise don't worry about it */
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
if(poll_result < 0) {
|
2003-10-15 20:28:32 +02:00
|
|
|
if(errno != EINTR) { /* let the program survive things like ^z */
|
2004-02-28 00:23:33 +01:00
|
|
|
log_fn(LOG_ERR,"poll failed: %s",strerror(errno));
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
return -1;
|
2003-10-15 20:28:32 +02:00
|
|
|
} else {
|
2003-10-15 20:37:19 +02:00
|
|
|
log_fn(LOG_DEBUG,"poll interrupted.");
|
2003-10-15 20:28:32 +02:00
|
|
|
}
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
}
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2003-09-30 22:36:20 +02:00
|
|
|
/* do all the reads and errors first, so we can detect closed sockets */
|
|
|
|
for(i=0;i<nfds;i++)
|
2004-02-28 05:11:53 +01:00
|
|
|
conn_read(i); /* this also marks broken connections */
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2003-09-30 22:36:20 +02:00
|
|
|
/* then do the writes */
|
|
|
|
for(i=0;i<nfds;i++)
|
|
|
|
conn_write(i);
|
|
|
|
|
|
|
|
/* any of the conns need to be closed now? */
|
|
|
|
for(i=0;i<nfds;i++)
|
2003-12-05 10:51:49 +01:00
|
|
|
conn_close_if_marked(i);
|
2002-06-27 00:45:49 +02:00
|
|
|
|
Implemented link padding and receiver token buckets
Each socket reads at most 'bandwidth' bytes per second sustained, but
can handle bursts of up to 10*bandwidth bytes.
Cells are now sent out at evenly-spaced intervals, with padding sent
out otherwise. Set Linkpadding=0 in the rc file to send cells as soon
as they're available (and to never send padding cells).
Added license/copyrights statements at the top of most files.
router->min and router->max have been merged into a single 'bandwidth'
value. We should make the routerinfo_t reflect this (want to do that,
Mat?)
As the bandwidth increases, and we want to stop sleeping more and more
frequently to send a single cell, cpu usage goes up. At 128kB/s we're
pretty much calling poll with a timeout of 1ms or even 0ms. The current
code takes a timeout of 0-9ms and makes it 10ms. prepare_for_poll()
handles everything that should have happened in the past, so as long as
our buffers don't get too full in that 10ms, we're ok.
Speaking of too full, if you run three servers at 100kB/s with -l debug,
it spends too much time printing debugging messages to be able to keep
up with the cells. The outbuf ultimately fills up and it kills that
connection. If you run with -l err, it works fine up through 500kB/s and
probably beyond. Down the road we'll want to teach it to recognize when
an outbuf is getting full, and back off.
svn:r50
2002-07-16 03:12:15 +02:00
|
|
|
/* refilling buckets and sending cells happens at the beginning of the
|
|
|
|
* next iteration of the loop, inside prepare_for_poll()
|
|
|
|
*/
|
2002-06-27 00:45:49 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2002-09-28 03:40:11 +02:00
|
|
|
static void catch(int the_signal) {
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2003-10-22 08:03:11 +02:00
|
|
|
#ifndef MS_WINDOWS /* do signal stuff only on unix */
|
2002-09-28 03:40:11 +02:00
|
|
|
switch(the_signal) {
|
2002-11-24 09:45:54 +01:00
|
|
|
// case SIGABRT:
|
2002-09-28 03:40:11 +02:00
|
|
|
case SIGTERM:
|
|
|
|
case SIGINT:
|
2003-09-26 12:03:50 +02:00
|
|
|
log(LOG_ERR,"Catching signal %d, exiting cleanly.", the_signal);
|
2003-10-22 08:03:11 +02:00
|
|
|
/* we don't care if there was an error when we unlink, nothing
|
|
|
|
we could do about it anyways */
|
2003-11-19 23:45:06 +01:00
|
|
|
if(options.PidFile)
|
|
|
|
unlink(options.PidFile);
|
2002-09-28 03:40:11 +02:00
|
|
|
exit(0);
|
|
|
|
case SIGHUP:
|
2003-09-21 08:15:43 +02:00
|
|
|
please_reset = 1;
|
2002-09-28 03:40:11 +02:00
|
|
|
break;
|
|
|
|
case SIGUSR1:
|
|
|
|
please_dumpstats = 1;
|
|
|
|
break;
|
2003-08-12 08:41:53 +02:00
|
|
|
case SIGCHLD:
|
|
|
|
please_reap_children = 1;
|
2003-09-13 23:53:38 +02:00
|
|
|
break;
|
2002-09-28 03:40:11 +02:00
|
|
|
default:
|
2003-10-10 03:48:32 +02:00
|
|
|
log(LOG_WARN,"Caught signal %d that we can't handle??", the_signal);
|
2002-09-28 03:40:11 +02:00
|
|
|
}
|
2003-08-12 10:04:31 +02:00
|
|
|
#endif /* signal stuff */
|
2002-09-22 00:41:48 +02:00
|
|
|
}
|
|
|
|
|
2003-10-15 20:28:32 +02:00
|
|
|
static void dumpstats(int severity) {
|
2002-09-22 00:41:48 +02:00
|
|
|
int i;
|
|
|
|
connection_t *conn;
|
2003-10-04 05:29:09 +02:00
|
|
|
time_t now = time(NULL);
|
2002-09-22 00:41:48 +02:00
|
|
|
|
2003-10-15 20:28:32 +02:00
|
|
|
log(severity, "Dumping stats:");
|
2002-09-22 00:41:48 +02:00
|
|
|
|
|
|
|
for(i=0;i<nfds;i++) {
|
|
|
|
conn = connection_array[i];
|
2003-10-15 20:28:32 +02:00
|
|
|
log(severity, "Conn %d (socket %d) type %d (%s), state %d (%s), created %ld secs ago",
|
2002-09-22 00:41:48 +02:00
|
|
|
i, conn->s, conn->type, conn_type_to_string[conn->type],
|
2003-10-04 05:29:09 +02:00
|
|
|
conn->state, conn_state_to_string[conn->type][conn->state], now - conn->timestamp_created);
|
2002-09-22 00:41:48 +02:00
|
|
|
if(!connection_is_listener(conn)) {
|
2003-10-15 20:28:32 +02:00
|
|
|
log(severity,"Conn %d is to '%s:%d'.",i,conn->address, conn->port);
|
|
|
|
log(severity,"Conn %d: %d bytes waiting on inbuf (last read %ld secs ago)",i,
|
2003-09-25 07:17:11 +02:00
|
|
|
(int)buf_datalen(conn->inbuf),
|
2003-10-04 05:29:09 +02:00
|
|
|
now - conn->timestamp_lastread);
|
2003-10-15 20:28:32 +02:00
|
|
|
log(severity,"Conn %d: %d bytes waiting on outbuf (last written %ld secs ago)",i,
|
2003-10-04 05:29:09 +02:00
|
|
|
(int)buf_datalen(conn->outbuf), now - conn->timestamp_lastwritten);
|
2002-09-22 00:41:48 +02:00
|
|
|
}
|
2003-10-15 20:28:32 +02:00
|
|
|
circuit_dump_by_conn(conn, severity); /* dump info about all the circuits using this conn */
|
2002-09-22 00:41:48 +02:00
|
|
|
}
|
2003-10-15 20:28:32 +02:00
|
|
|
log(severity,
|
|
|
|
"Cells processed: %10lu padding\n"
|
2003-10-04 10:19:23 +02:00
|
|
|
" %10lu create\n"
|
|
|
|
" %10lu created\n"
|
|
|
|
" %10lu relay\n"
|
|
|
|
" (%10lu relayed)\n"
|
|
|
|
" (%10lu delivered)\n"
|
2004-01-16 09:27:17 +01:00
|
|
|
" %10lu destroy",
|
2003-10-02 22:00:38 +02:00
|
|
|
stats_n_padding_cells_processed,
|
|
|
|
stats_n_create_cells_processed,
|
|
|
|
stats_n_created_cells_processed,
|
|
|
|
stats_n_relay_cells_processed,
|
|
|
|
stats_n_relay_cells_relayed,
|
|
|
|
stats_n_relay_cells_delivered,
|
|
|
|
stats_n_destroy_cells_processed);
|
|
|
|
if (stats_n_data_cells_packaged)
|
2004-01-16 09:27:17 +01:00
|
|
|
log(severity,"Average packaged cell fullness: %2.3f%%",
|
2003-12-05 10:51:49 +01:00
|
|
|
100*(((double)stats_n_data_bytes_packaged) /
|
2003-12-23 08:42:01 +01:00
|
|
|
(stats_n_data_cells_packaged*RELAY_PAYLOAD_SIZE)) );
|
2003-10-15 21:25:28 +02:00
|
|
|
if (stats_n_data_cells_received)
|
2004-01-16 09:27:17 +01:00
|
|
|
log(severity,"Average delivered cell fullness: %2.3f%%",
|
2003-12-05 10:51:49 +01:00
|
|
|
100*(((double)stats_n_data_bytes_received) /
|
2003-12-23 08:42:01 +01:00
|
|
|
(stats_n_data_cells_received*RELAY_PAYLOAD_SIZE)) );
|
2003-12-05 10:51:49 +01:00
|
|
|
|
2003-10-02 22:00:38 +02:00
|
|
|
if (stats_n_seconds_reading)
|
2003-10-15 20:28:32 +02:00
|
|
|
log(severity,"Average bandwidth used: %d bytes/sec",
|
2003-10-02 22:00:38 +02:00
|
|
|
(int) (stats_n_bytes_read/stats_n_seconds_reading));
|
2002-09-22 00:41:48 +02:00
|
|
|
}
|
|
|
|
|
2003-04-07 04:12:02 +02:00
|
|
|
int tor_main(int argc, char *argv[]) {
|
2002-06-27 00:45:49 +02:00
|
|
|
|
2003-10-18 02:13:08 +02:00
|
|
|
/* give it somewhere to log to initially */
|
|
|
|
add_stream_log(LOG_INFO, "<stdout>", stdout);
|
2003-11-13 05:51:34 +01:00
|
|
|
log_fn(LOG_WARN,"Tor v%s. This is experimental software. Do not use it if you need anonymity.",VERSION);
|
2003-10-18 02:13:08 +02:00
|
|
|
|
2004-01-19 07:12:32 +01:00
|
|
|
if (init_from_config(argc,argv) < 0)
|
|
|
|
return -1;
|
|
|
|
|
2003-12-14 05:39:23 +01:00
|
|
|
#ifndef MS_WINDOWS
|
|
|
|
if(geteuid()==0)
|
|
|
|
log_fn(LOG_WARN,"You are running Tor as root. You don't need to, and you probably shouldn't.");
|
|
|
|
#endif
|
|
|
|
|
2003-11-20 18:49:45 +01:00
|
|
|
if(options.ORPort) { /* only spawn dns handlers if we're a router */
|
2003-06-17 16:31:05 +02:00
|
|
|
dns_init(); /* initialize the dns resolve tree, and spawn workers */
|
major overhaul: dns slave subsystem, topics
on startup, it forks off a master dns handler, which forks off dns
slaves (like the apache model). slaves as spawned as load increases,
and then reused. excess slaves are not ever killed, currently.
implemented topics. each topic has a receive window in each direction
at each edge of the circuit, and sends sendme's at the data level, as
per before. each circuit also has receive windows in each direction at
each hop; an edge sends a circuit-level sendme as soon as enough data
cells have arrived (regardless of whether the data cells were flushed
to the exit conns). removed the 'connected' cell type, since it's now
a topic command within data cells.
at the edge of the circuit, there can be multiple connections associated
with a single circuit. you find them via the linked list conn->next_topic.
currently each new ap connection starts its own circuit, so we ought
to see comparable performance to what we had before. but that's only
because i haven't written the code to reattach to old circuits. please
try to break it as-is, and then i'll make it reuse the same circuit and
we'll try to break that.
svn:r152
2003-01-26 10:02:24 +01:00
|
|
|
}
|
2003-11-16 18:00:02 +01:00
|
|
|
if(options.SocksPort) {
|
|
|
|
client_dns_init(); /* init the client dns cache */
|
|
|
|
}
|
major overhaul: dns slave subsystem, topics
on startup, it forks off a master dns handler, which forks off dns
slaves (like the apache model). slaves as spawned as load increases,
and then reused. excess slaves are not ever killed, currently.
implemented topics. each topic has a receive window in each direction
at each edge of the circuit, and sends sendme's at the data level, as
per before. each circuit also has receive windows in each direction at
each hop; an edge sends a circuit-level sendme as soon as enough data
cells have arrived (regardless of whether the data cells were flushed
to the exit conns). removed the 'connected' cell type, since it's now
a topic command within data cells.
at the edge of the circuit, there can be multiple connections associated
with a single circuit. you find them via the linked list conn->next_topic.
currently each new ap connection starts its own circuit, so we ought
to see comparable performance to what we had before. but that's only
because i haven't written the code to reattach to old circuits. please
try to break it as-is, and then i'll make it reuse the same circuit and
we'll try to break that.
svn:r152
2003-01-26 10:02:24 +01:00
|
|
|
|
2003-08-12 10:04:31 +02:00
|
|
|
#ifndef MS_WINDOWS /* do signal stuff only on unix */
|
major overhaul: dns slave subsystem, topics
on startup, it forks off a master dns handler, which forks off dns
slaves (like the apache model). slaves as spawned as load increases,
and then reused. excess slaves are not ever killed, currently.
implemented topics. each topic has a receive window in each direction
at each edge of the circuit, and sends sendme's at the data level, as
per before. each circuit also has receive windows in each direction at
each hop; an edge sends a circuit-level sendme as soon as enough data
cells have arrived (regardless of whether the data cells were flushed
to the exit conns). removed the 'connected' cell type, since it's now
a topic command within data cells.
at the edge of the circuit, there can be multiple connections associated
with a single circuit. you find them via the linked list conn->next_topic.
currently each new ap connection starts its own circuit, so we ought
to see comparable performance to what we had before. but that's only
because i haven't written the code to reattach to old circuits. please
try to break it as-is, and then i'll make it reuse the same circuit and
we'll try to break that.
svn:r152
2003-01-26 10:02:24 +01:00
|
|
|
signal (SIGINT, catch); /* catch kills so we can exit cleanly */
|
|
|
|
signal (SIGTERM, catch);
|
2003-10-15 20:28:32 +02:00
|
|
|
signal (SIGUSR1, catch); /* to dump stats */
|
major overhaul: dns slave subsystem, topics
on startup, it forks off a master dns handler, which forks off dns
slaves (like the apache model). slaves as spawned as load increases,
and then reused. excess slaves are not ever killed, currently.
implemented topics. each topic has a receive window in each direction
at each edge of the circuit, and sends sendme's at the data level, as
per before. each circuit also has receive windows in each direction at
each hop; an edge sends a circuit-level sendme as soon as enough data
cells have arrived (regardless of whether the data cells were flushed
to the exit conns). removed the 'connected' cell type, since it's now
a topic command within data cells.
at the edge of the circuit, there can be multiple connections associated
with a single circuit. you find them via the linked list conn->next_topic.
currently each new ap connection starts its own circuit, so we ought
to see comparable performance to what we had before. but that's only
because i haven't written the code to reattach to old circuits. please
try to break it as-is, and then i'll make it reuse the same circuit and
we'll try to break that.
svn:r152
2003-01-26 10:02:24 +01:00
|
|
|
signal (SIGHUP, catch); /* to reload directory */
|
2003-08-12 08:41:53 +02:00
|
|
|
signal (SIGCHLD, catch); /* for exiting dns/cpu workers */
|
2003-08-12 10:04:31 +02:00
|
|
|
#endif /* signal stuff */
|
major overhaul: dns slave subsystem, topics
on startup, it forks off a master dns handler, which forks off dns
slaves (like the apache model). slaves as spawned as load increases,
and then reused. excess slaves are not ever killed, currently.
implemented topics. each topic has a receive window in each direction
at each edge of the circuit, and sends sendme's at the data level, as
per before. each circuit also has receive windows in each direction at
each hop; an edge sends a circuit-level sendme as soon as enough data
cells have arrived (regardless of whether the data cells were flushed
to the exit conns). removed the 'connected' cell type, since it's now
a topic command within data cells.
at the edge of the circuit, there can be multiple connections associated
with a single circuit. you find them via the linked list conn->next_topic.
currently each new ap connection starts its own circuit, so we ought
to see comparable performance to what we had before. but that's only
because i haven't written the code to reattach to old circuits. please
try to break it as-is, and then i'll make it reuse the same circuit and
we'll try to break that.
svn:r152
2003-01-26 10:02:24 +01:00
|
|
|
|
|
|
|
crypto_global_init();
|
2003-06-13 23:13:37 +02:00
|
|
|
crypto_seed_rng();
|
2003-09-29 09:50:08 +02:00
|
|
|
do_main_loop();
|
2002-08-22 09:30:03 +02:00
|
|
|
crypto_global_cleanup();
|
2003-09-29 09:50:08 +02:00
|
|
|
return -1;
|
2002-06-27 00:45:49 +02:00
|
|
|
}
|
|
|
|
|
2003-04-07 04:12:02 +02:00
|
|
|
/*
|
|
|
|
Local Variables:
|
|
|
|
mode:c
|
|
|
|
indent-tabs-mode:nil
|
|
|
|
c-basic-offset:2
|
|
|
|
End:
|
|
|
|
*/
|