mirror of
https://codeberg.org/anoncontributorxmr/monero.git
synced 2024-11-30 14:53:27 +01:00
22581a0441
Since connections from the ::connect method are now kept in a deque to be able to cancel them on exit, this leaks both memory and a file descriptor. Here, we clean those up after 30 seconds, to avoid this. 30 seconds is higher then the 5 second timeout used in the async code, so this should be safe. However, this is an assumption which would break if that async code was to start relying on longer timeouts.
322 lines
12 KiB
C++
322 lines
12 KiB
C++
/**
|
|
@file
|
|
@author from CrypoNote (see copyright below; Andrey N. Sabelnikov)
|
|
@monero rfree
|
|
@brief the connection templated-class for one peer connection
|
|
*/
|
|
// Copyright (c) 2006-2013, Andrey N. Sabelnikov, www.sabelnikov.net
|
|
// All rights reserved.
|
|
//
|
|
// Redistribution and use in source and binary forms, with or without
|
|
// modification, are permitted provided that the following conditions are met:
|
|
// * Redistributions of source code must retain the above copyright
|
|
// notice, this list of conditions and the following disclaimer.
|
|
// * Redistributions in binary form must reproduce the above copyright
|
|
// notice, this list of conditions and the following disclaimer in the
|
|
// documentation and/or other materials provided with the distribution.
|
|
// * Neither the name of the Andrey N. Sabelnikov nor the
|
|
// names of its contributors may be used to endorse or promote products
|
|
// derived from this software without specific prior written permission.
|
|
//
|
|
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND
|
|
// ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
|
|
// WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
|
// DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER BE LIABLE FOR ANY
|
|
// DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
|
|
// (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
|
|
// LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
|
|
// ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
|
|
// SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
//
|
|
|
|
|
|
|
|
#ifndef _ABSTRACT_TCP_SERVER2_H_
|
|
#define _ABSTRACT_TCP_SERVER2_H_
|
|
|
|
|
|
#include <boost/asio.hpp>
|
|
#include <string>
|
|
#include <vector>
|
|
#include <boost/noncopyable.hpp>
|
|
#include <boost/shared_ptr.hpp>
|
|
#include <atomic>
|
|
#include <map>
|
|
#include <memory>
|
|
|
|
#include <boost/asio.hpp>
|
|
#include <boost/array.hpp>
|
|
#include <boost/noncopyable.hpp>
|
|
#include <boost/shared_ptr.hpp>
|
|
#include <boost/enable_shared_from_this.hpp>
|
|
#include <boost/interprocess/detail/atomic.hpp>
|
|
#include <boost/thread/thread.hpp>
|
|
#include "net_utils_base.h"
|
|
#include "syncobj.h"
|
|
#include "../../../../src/p2p/connection_basic.hpp"
|
|
#include "../../../../contrib/otshell_utils/utils.hpp"
|
|
#include "../../../../src/p2p/network_throttle-detail.hpp"
|
|
|
|
#define ABSTRACT_SERVER_SEND_QUE_MAX_COUNT 1000
|
|
|
|
namespace epee
|
|
{
|
|
namespace net_utils
|
|
{
|
|
|
|
struct i_connection_filter
|
|
{
|
|
virtual bool is_remote_ip_allowed(uint32_t adress)=0;
|
|
protected:
|
|
virtual ~i_connection_filter(){}
|
|
};
|
|
|
|
|
|
/************************************************************************/
|
|
/* */
|
|
/************************************************************************/
|
|
/// Represents a single connection from a client.
|
|
template<class t_protocol_handler>
|
|
class connection
|
|
: public boost::enable_shared_from_this<connection<t_protocol_handler> >,
|
|
private boost::noncopyable,
|
|
public i_service_endpoint,
|
|
public connection_basic
|
|
{
|
|
public:
|
|
typedef typename t_protocol_handler::connection_context t_connection_context;
|
|
/// Construct a connection with the given io_service.
|
|
|
|
explicit connection( boost::asio::io_service& io_service,
|
|
typename t_protocol_handler::config_type& config,
|
|
std::atomic<long> &ref_sock_count, // the ++/-- counter
|
|
std::atomic<long> &sock_number, // the only increasing ++ number generator
|
|
i_connection_filter * &pfilter
|
|
,t_connection_type connection_type);
|
|
|
|
virtual ~connection();
|
|
/// Get the socket associated with the connection.
|
|
boost::asio::ip::tcp::socket& socket();
|
|
|
|
/// Start the first asynchronous operation for the connection.
|
|
bool start(bool is_income, bool is_multithreaded);
|
|
|
|
void get_context(t_connection_context& context_){context_ = context;}
|
|
|
|
void call_back_starter();
|
|
|
|
void save_dbg_log();
|
|
|
|
|
|
bool speed_limit_is_enabled() const; ///< tells us should we be sleeping here (e.g. do not sleep on RPC connections)
|
|
|
|
bool cancel();
|
|
|
|
private:
|
|
//----------------- i_service_endpoint ---------------------
|
|
virtual bool do_send(const void* ptr, size_t cb); ///< (see do_send from i_service_endpoint)
|
|
virtual bool do_send_chunk(const void* ptr, size_t cb); ///< will send (or queue) a part of data
|
|
virtual bool close();
|
|
virtual bool call_run_once_service_io();
|
|
virtual bool request_callback();
|
|
virtual boost::asio::io_service& get_io_service();
|
|
virtual bool add_ref();
|
|
virtual bool release();
|
|
//------------------------------------------------------
|
|
boost::shared_ptr<connection<t_protocol_handler> > safe_shared_from_this();
|
|
bool shutdown();
|
|
/// Handle completion of a read operation.
|
|
void handle_read(const boost::system::error_code& e,
|
|
std::size_t bytes_transferred);
|
|
|
|
/// Handle completion of a write operation.
|
|
void handle_write(const boost::system::error_code& e, size_t cb);
|
|
|
|
/// Buffer for incoming data.
|
|
boost::array<char, 8192> buffer_;
|
|
//boost::array<char, 1024> buffer_;
|
|
|
|
t_connection_context context;
|
|
i_connection_filter* &m_pfilter;
|
|
|
|
// TODO what do they mean about wait on destructor?? --rfree :
|
|
//this should be the last one, because it could be wait on destructor, while other activities possible on other threads
|
|
t_protocol_handler m_protocol_handler;
|
|
//typename t_protocol_handler::config_type m_dummy_config;
|
|
std::list<boost::shared_ptr<connection<t_protocol_handler> > > m_self_refs; // add_ref/release support
|
|
critical_section m_self_refs_lock;
|
|
critical_section m_chunking_lock; // held while we add small chunks of the big do_send() to small do_send_chunk()
|
|
|
|
t_connection_type m_connection_type;
|
|
|
|
// for calculate speed (last 60 sec)
|
|
network_throttle m_throttle_speed_in;
|
|
network_throttle m_throttle_speed_out;
|
|
std::mutex m_throttle_speed_in_mutex;
|
|
std::mutex m_throttle_speed_out_mutex;
|
|
|
|
public:
|
|
void setRpcStation();
|
|
};
|
|
|
|
|
|
/************************************************************************/
|
|
/* */
|
|
/************************************************************************/
|
|
template<class t_protocol_handler>
|
|
class boosted_tcp_server
|
|
: private boost::noncopyable
|
|
{
|
|
public:
|
|
typedef boost::shared_ptr<connection<t_protocol_handler> > connection_ptr;
|
|
typedef typename t_protocol_handler::connection_context t_connection_context;
|
|
/// Construct the server to listen on the specified TCP address and port, and
|
|
/// serve up files from the given directory.
|
|
|
|
boosted_tcp_server(t_connection_type connection_type);
|
|
explicit boosted_tcp_server(boost::asio::io_service& external_io_service, t_connection_type connection_type);
|
|
~boosted_tcp_server();
|
|
|
|
std::map<std::string, t_connection_type> server_type_map;
|
|
void create_server_type_map();
|
|
|
|
bool init_server(uint32_t port, const std::string address = "0.0.0.0");
|
|
bool init_server(const std::string port, const std::string& address = "0.0.0.0");
|
|
|
|
/// Run the server's io_service loop.
|
|
bool run_server(size_t threads_count, bool wait = true, const boost::thread::attributes& attrs = boost::thread::attributes());
|
|
|
|
/// wait for service workers stop
|
|
bool timed_wait_server_stop(uint64_t wait_mseconds);
|
|
|
|
/// Stop the server.
|
|
void send_stop_signal();
|
|
|
|
bool is_stop_signal_sent();
|
|
|
|
void set_threads_prefix(const std::string& prefix_name);
|
|
|
|
bool deinit_server(){return true;}
|
|
|
|
size_t get_threads_count(){return m_threads_count;}
|
|
|
|
void set_connection_filter(i_connection_filter* pfilter);
|
|
|
|
bool connect(const std::string& adr, const std::string& port, uint32_t conn_timeot, t_connection_context& cn, const std::string& bind_ip = "0.0.0.0");
|
|
template<class t_callback>
|
|
bool connect_async(const std::string& adr, const std::string& port, uint32_t conn_timeot, t_callback cb, const std::string& bind_ip = "0.0.0.0");
|
|
|
|
typename t_protocol_handler::config_type& get_config_object(){return m_config;}
|
|
|
|
int get_binded_port(){return m_port;}
|
|
|
|
boost::asio::io_service& get_io_service(){return io_service_;}
|
|
|
|
struct idle_callback_conext_base
|
|
{
|
|
virtual ~idle_callback_conext_base(){}
|
|
|
|
virtual bool call_handler(){return true;}
|
|
|
|
idle_callback_conext_base(boost::asio::io_service& io_serice):
|
|
m_timer(io_serice)
|
|
{}
|
|
boost::asio::deadline_timer m_timer;
|
|
uint64_t m_period;
|
|
};
|
|
|
|
template <class t_handler>
|
|
struct idle_callback_conext: public idle_callback_conext_base
|
|
{
|
|
idle_callback_conext(boost::asio::io_service& io_serice, t_handler& h, uint64_t period):
|
|
idle_callback_conext_base(io_serice),
|
|
m_handler(h)
|
|
{this->m_period = period;}
|
|
|
|
t_handler m_handler;
|
|
virtual bool call_handler()
|
|
{
|
|
return m_handler();
|
|
}
|
|
};
|
|
|
|
template<class t_handler>
|
|
bool add_idle_handler(t_handler t_callback, uint64_t timeout_ms)
|
|
{
|
|
boost::shared_ptr<idle_callback_conext_base> ptr(new idle_callback_conext<t_handler>(io_service_, t_callback, timeout_ms));
|
|
//needed call handler here ?...
|
|
ptr->m_timer.expires_from_now(boost::posix_time::milliseconds(ptr->m_period));
|
|
ptr->m_timer.async_wait(boost::bind(&boosted_tcp_server<t_protocol_handler>::global_timer_handler, this, ptr));
|
|
return true;
|
|
}
|
|
|
|
bool global_timer_handler(/*const boost::system::error_code& err, */boost::shared_ptr<idle_callback_conext_base> ptr)
|
|
{
|
|
//if handler return false - he don't want to be called anymore
|
|
if(!ptr->call_handler())
|
|
return true;
|
|
ptr->m_timer.expires_from_now(boost::posix_time::milliseconds(ptr->m_period));
|
|
ptr->m_timer.async_wait(boost::bind(&boosted_tcp_server<t_protocol_handler>::global_timer_handler, this, ptr));
|
|
return true;
|
|
}
|
|
|
|
template<class t_handler>
|
|
bool async_call(t_handler t_callback)
|
|
{
|
|
io_service_.post(t_callback);
|
|
return true;
|
|
}
|
|
|
|
protected:
|
|
typename t_protocol_handler::config_type m_config;
|
|
|
|
private:
|
|
/// Run the server's io_service loop.
|
|
bool worker_thread();
|
|
/// Handle completion of an asynchronous accept operation.
|
|
void handle_accept(const boost::system::error_code& e);
|
|
|
|
bool is_thread_worker();
|
|
|
|
bool cleanup_connections();
|
|
|
|
/// The io_service used to perform asynchronous operations.
|
|
std::unique_ptr<boost::asio::io_service> m_io_service_local_instance;
|
|
boost::asio::io_service& io_service_;
|
|
|
|
/// Acceptor used to listen for incoming connections.
|
|
boost::asio::ip::tcp::acceptor acceptor_;
|
|
|
|
std::atomic<bool> m_stop_signal_sent;
|
|
uint32_t m_port;
|
|
std::atomic<long> m_sock_count;
|
|
std::atomic<long> m_sock_number;
|
|
std::string m_address;
|
|
std::string m_thread_name_prefix; //TODO: change to enum server_type, now used
|
|
size_t m_threads_count;
|
|
i_connection_filter* m_pfilter;
|
|
std::vector<boost::shared_ptr<boost::thread> > m_threads;
|
|
boost::thread::id m_main_thread_id;
|
|
critical_section m_threads_lock;
|
|
volatile uint32_t m_thread_index; // TODO change to std::atomic
|
|
void detach_threads();
|
|
|
|
t_connection_type m_connection_type;
|
|
|
|
/// The next connection to be accepted
|
|
connection_ptr new_connection_;
|
|
|
|
std::mutex connections_mutex;
|
|
std::deque<std::pair<boost::system_time, connection_ptr>> connections_;
|
|
|
|
}; // class <>boosted_tcp_server
|
|
|
|
|
|
} // namespace
|
|
} // namespace
|
|
|
|
#include "abstract_tcp_server2.inl"
|
|
|
|
#endif
|