mirror of
https://gitlab.torproject.org/tpo/core/tor.git
synced 2024-11-11 13:43:47 +01:00
Send more descriptor requests per attempt when using tunneled connections
This commit is contained in:
parent
21d5dbd474
commit
5ed5ac185b
@ -1261,7 +1261,8 @@ directory_send_command(dir_connection_t *conn,
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (strlen(proxystring) + strlen(url) >= 4096) {
|
/* warn in the non-tunneled case */
|
||||||
|
if (direct && (strlen(proxystring) + strlen(url) >= 4096)) {
|
||||||
log_warn(LD_BUG,
|
log_warn(LD_BUG,
|
||||||
"Squid does not like URLs longer than 4095 bytes, and this "
|
"Squid does not like URLs longer than 4095 bytes, and this "
|
||||||
"one is %d bytes long: %s%s",
|
"one is %d bytes long: %s%s",
|
||||||
|
@ -4354,8 +4354,16 @@ MOCK_IMPL(STATIC void, initiate_descriptor_downloads,
|
|||||||
* 4058/41 (40 for the hash and 1 for the + that separates them) => 98
|
* 4058/41 (40 for the hash and 1 for the + that separates them) => 98
|
||||||
* So use 96 because it's a nice number.
|
* So use 96 because it's a nice number.
|
||||||
*/
|
*/
|
||||||
#define MAX_DL_PER_REQUEST 96
|
int
|
||||||
#define MAX_MICRODESC_DL_PER_REQUEST 92
|
max_dl_per_request(const or_options_t *options, int purpose)
|
||||||
|
{
|
||||||
|
int max = 96;
|
||||||
|
if (purpose == DIR_PURPOSE_FETCH_MICRODESC) {
|
||||||
|
max = options->TunnelDirConns ? 1000 : 92;
|
||||||
|
}
|
||||||
|
return max;
|
||||||
|
}
|
||||||
|
|
||||||
/** Don't split our requests so finely that we are requesting fewer than
|
/** Don't split our requests so finely that we are requesting fewer than
|
||||||
* this number per server. */
|
* this number per server. */
|
||||||
#define MIN_DL_PER_REQUEST 4
|
#define MIN_DL_PER_REQUEST 4
|
||||||
@ -4377,27 +4385,29 @@ launch_descriptor_downloads(int purpose,
|
|||||||
smartlist_t *downloadable,
|
smartlist_t *downloadable,
|
||||||
const routerstatus_t *source, time_t now)
|
const routerstatus_t *source, time_t now)
|
||||||
{
|
{
|
||||||
int should_delay = 0, n_downloadable;
|
|
||||||
const or_options_t *options = get_options();
|
const or_options_t *options = get_options();
|
||||||
const char *descname;
|
const char *descname;
|
||||||
|
const int fetch_microdesc = (purpose == DIR_PURPOSE_FETCH_MICRODESC);
|
||||||
|
|
||||||
tor_assert(purpose == DIR_PURPOSE_FETCH_SERVERDESC ||
|
tor_assert(fetch_microdesc || purpose == DIR_PURPOSE_FETCH_SERVERDESC);
|
||||||
purpose == DIR_PURPOSE_FETCH_MICRODESC);
|
descname = fetch_microdesc ? "microdesc" : "routerdesc";
|
||||||
|
|
||||||
descname = (purpose == DIR_PURPOSE_FETCH_SERVERDESC) ?
|
int n_downloadable = smartlist_len(downloadable);
|
||||||
"routerdesc" : "microdesc";
|
if (!n_downloadable)
|
||||||
|
return;
|
||||||
|
|
||||||
n_downloadable = smartlist_len(downloadable);
|
|
||||||
if (!directory_fetches_dir_info_early(options)) {
|
if (!directory_fetches_dir_info_early(options)) {
|
||||||
if (n_downloadable >= MAX_DL_TO_DELAY) {
|
if (n_downloadable >= MAX_DL_TO_DELAY) {
|
||||||
log_debug(LD_DIR,
|
log_debug(LD_DIR,
|
||||||
"There are enough downloadable %ss to launch requests.",
|
"There are enough downloadable %ss to launch requests.",
|
||||||
descname);
|
descname);
|
||||||
should_delay = 0;
|
|
||||||
} else {
|
} else {
|
||||||
should_delay = (last_descriptor_download_attempted +
|
|
||||||
options->TestingClientMaxIntervalWithoutRequest) > now;
|
/* should delay */
|
||||||
if (!should_delay && n_downloadable) {
|
if ((last_descriptor_download_attempted +
|
||||||
|
options->TestingClientMaxIntervalWithoutRequest) > now)
|
||||||
|
return;
|
||||||
|
|
||||||
if (last_descriptor_download_attempted) {
|
if (last_descriptor_download_attempted) {
|
||||||
log_info(LD_DIR,
|
log_info(LD_DIR,
|
||||||
"There are not many downloadable %ss, but we've "
|
"There are not many downloadable %ss, but we've "
|
||||||
@ -4410,39 +4420,35 @@ launch_descriptor_downloads(int purpose,
|
|||||||
"tried downloading descriptors recently. Downloading.",
|
"tried downloading descriptors recently. Downloading.",
|
||||||
descname);
|
descname);
|
||||||
}
|
}
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if (! should_delay && n_downloadable) {
|
int i, n_per_request, max_dl_per_req;
|
||||||
int i, n_per_request;
|
|
||||||
const char *req_plural = "", *rtr_plural = "";
|
const char *req_plural = "", *rtr_plural = "";
|
||||||
int pds_flags = PDS_RETRY_IF_NO_SERVERS;
|
int pds_flags = PDS_RETRY_IF_NO_SERVERS;
|
||||||
if (! authdir_mode_any_nonhidserv(options)) {
|
if (!authdir_mode_any_nonhidserv(options) || fetch_microdesc) {
|
||||||
/* If we wind up going to the authorities, we want to only open one
|
/* If we wind up going to the authorities, we want to only open one
|
||||||
* connection to each authority at a time, so that we don't overload
|
* connection to each authority at a time, so that we don't overload
|
||||||
* them. We do this by setting PDS_NO_EXISTING_SERVERDESC_FETCH
|
* them. We do this by setting PDS_NO_EXISTING_SERVERDESC_FETCH
|
||||||
* regardless of whether we're a cache or not; it gets ignored if we're
|
* regardless of whether we're a cache or not.
|
||||||
* not calling router_pick_trusteddirserver.
|
|
||||||
*
|
*
|
||||||
* Setting this flag can make initiate_descriptor_downloads() ignore
|
* Setting this flag can make initiate_descriptor_downloads() ignore
|
||||||
* requests. We need to make sure that we do in fact call
|
* requests. We need to make sure that we do in fact call
|
||||||
* update_router_descriptor_downloads() later on, once the connections
|
* update_router_descriptor_downloads() later on, once the connections
|
||||||
* have succeeded or failed.
|
* have succeeded or failed.
|
||||||
*/
|
*/
|
||||||
pds_flags |= (purpose == DIR_PURPOSE_FETCH_MICRODESC) ?
|
pds_flags |= fetch_microdesc ?
|
||||||
PDS_NO_EXISTING_MICRODESC_FETCH :
|
PDS_NO_EXISTING_MICRODESC_FETCH :
|
||||||
PDS_NO_EXISTING_SERVERDESC_FETCH;
|
PDS_NO_EXISTING_SERVERDESC_FETCH;
|
||||||
}
|
}
|
||||||
|
|
||||||
n_per_request = CEIL_DIV(n_downloadable, MIN_REQUESTS);
|
n_per_request = CEIL_DIV(n_downloadable, MIN_REQUESTS);
|
||||||
if (purpose == DIR_PURPOSE_FETCH_MICRODESC) {
|
max_dl_per_req = max_dl_per_request(options, purpose);
|
||||||
if (n_per_request > MAX_MICRODESC_DL_PER_REQUEST)
|
|
||||||
n_per_request = MAX_MICRODESC_DL_PER_REQUEST;
|
if (n_per_request > max_dl_per_req)
|
||||||
} else {
|
n_per_request = max_dl_per_req;
|
||||||
if (n_per_request > MAX_DL_PER_REQUEST)
|
|
||||||
n_per_request = MAX_DL_PER_REQUEST;
|
|
||||||
}
|
|
||||||
if (n_per_request < MIN_DL_PER_REQUEST)
|
if (n_per_request < MIN_DL_PER_REQUEST)
|
||||||
n_per_request = MIN_DL_PER_REQUEST;
|
n_per_request = MIN_DL_PER_REQUEST;
|
||||||
|
|
||||||
@ -4463,7 +4469,6 @@ launch_descriptor_downloads(int purpose,
|
|||||||
}
|
}
|
||||||
last_descriptor_download_attempted = now;
|
last_descriptor_download_attempted = now;
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
/** For any descriptor that we want that's currently listed in
|
/** For any descriptor that we want that's currently listed in
|
||||||
* <b>consensus</b>, download it as appropriate. */
|
* <b>consensus</b>, download it as appropriate. */
|
||||||
@ -4698,9 +4703,10 @@ update_extrainfo_downloads(time_t now)
|
|||||||
n_no_ei, n_have, n_delay, n_pending, smartlist_len(wanted));
|
n_no_ei, n_have, n_delay, n_pending, smartlist_len(wanted));
|
||||||
|
|
||||||
smartlist_shuffle(wanted);
|
smartlist_shuffle(wanted);
|
||||||
for (i = 0; i < smartlist_len(wanted); i += MAX_DL_PER_REQUEST) {
|
int max_dl_per_req = max_dl_per_request(options, DIR_PURPOSE_FETCH_EXTRAINFO);
|
||||||
|
for (i = 0; i < smartlist_len(wanted); i += max_dl_per_req) {
|
||||||
initiate_descriptor_downloads(NULL, DIR_PURPOSE_FETCH_EXTRAINFO,
|
initiate_descriptor_downloads(NULL, DIR_PURPOSE_FETCH_EXTRAINFO,
|
||||||
wanted, i, i + MAX_DL_PER_REQUEST,
|
wanted, i, i+max_dl_per_req,
|
||||||
PDS_RETRY_IF_NO_SERVERS|PDS_NO_EXISTING_SERVERDESC_FETCH);
|
PDS_RETRY_IF_NO_SERVERS|PDS_NO_EXISTING_SERVERDESC_FETCH);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user