2019-03-05 21:05:34 +00:00
|
|
|
// Copyright (c) 2014-2019, The Monero Project
|
2014-07-23 13:03:52 +00:00
|
|
|
//
|
|
|
|
// All rights reserved.
|
|
|
|
//
|
|
|
|
// Redistribution and use in source and binary forms, with or without modification, are
|
|
|
|
// permitted provided that the following conditions are met:
|
|
|
|
//
|
|
|
|
// 1. Redistributions of source code must retain the above copyright notice, this list of
|
|
|
|
// conditions and the following disclaimer.
|
|
|
|
//
|
|
|
|
// 2. Redistributions in binary form must reproduce the above copyright notice, this list
|
|
|
|
// of conditions and the following disclaimer in the documentation and/or other
|
|
|
|
// materials provided with the distribution.
|
|
|
|
//
|
|
|
|
// 3. Neither the name of the copyright holder nor the names of its contributors may be
|
|
|
|
// used to endorse or promote products derived from this software without specific
|
|
|
|
// prior written permission.
|
|
|
|
//
|
|
|
|
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY
|
|
|
|
// EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF
|
|
|
|
// MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL
|
|
|
|
// THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
|
|
// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
|
|
|
|
// PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
|
|
// INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
|
|
|
|
// STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF
|
|
|
|
// THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
//
|
|
|
|
// Parts of this file are originally copyright (c) 2012-2013 The Cryptonote developers
|
2014-03-03 22:07:58 +00:00
|
|
|
|
|
|
|
#pragma once
|
|
|
|
|
2018-12-16 17:57:44 +00:00
|
|
|
#include <iosfwd>
|
2014-03-03 22:07:58 +00:00
|
|
|
#include <list>
|
2018-12-16 17:57:44 +00:00
|
|
|
#include <string>
|
|
|
|
#include <vector>
|
2014-03-03 22:07:58 +00:00
|
|
|
|
|
|
|
#include <boost/multi_index_container.hpp>
|
|
|
|
#include <boost/multi_index/ordered_index.hpp>
|
|
|
|
#include <boost/multi_index/identity.hpp>
|
|
|
|
#include <boost/multi_index/member.hpp>
|
2018-12-16 17:57:44 +00:00
|
|
|
#include <boost/optional/optional.hpp>
|
2017-01-22 20:47:39 +00:00
|
|
|
#include <boost/range/adaptor/reversed.hpp>
|
2014-03-03 22:07:58 +00:00
|
|
|
|
|
|
|
|
2018-12-16 17:57:44 +00:00
|
|
|
#include "cryptonote_config.h"
|
|
|
|
#include "net/enums.h"
|
2014-03-03 22:07:58 +00:00
|
|
|
#include "net/local_ip.h"
|
|
|
|
#include "p2p_protocol_defs.h"
|
2018-12-16 17:57:44 +00:00
|
|
|
#include "syncobj.h"
|
2014-03-03 22:07:58 +00:00
|
|
|
|
|
|
|
namespace nodetool
|
|
|
|
{
|
2018-12-16 17:57:44 +00:00
|
|
|
struct peerlist_types
|
|
|
|
{
|
|
|
|
std::vector<peerlist_entry> white;
|
|
|
|
std::vector<peerlist_entry> gray;
|
|
|
|
std::vector<anchor_peerlist_entry> anchor;
|
|
|
|
};
|
|
|
|
|
|
|
|
class peerlist_storage
|
|
|
|
{
|
|
|
|
public:
|
|
|
|
peerlist_storage()
|
|
|
|
: m_types{}
|
|
|
|
{}
|
|
|
|
|
|
|
|
//! \return Peers stored in stream `src` in `new_format` (portable archive or older non-portable).
|
|
|
|
static boost::optional<peerlist_storage> open(std::istream& src, const bool new_format);
|
|
|
|
|
|
|
|
//! \return Peers stored in file at `path`
|
|
|
|
static boost::optional<peerlist_storage> open(const std::string& path);
|
|
|
|
|
|
|
|
peerlist_storage(peerlist_storage&&) = default;
|
|
|
|
peerlist_storage(const peerlist_storage&) = delete;
|
|
|
|
|
|
|
|
~peerlist_storage() noexcept;
|
2014-03-03 22:07:58 +00:00
|
|
|
|
2018-12-16 17:57:44 +00:00
|
|
|
peerlist_storage& operator=(peerlist_storage&&) = default;
|
|
|
|
peerlist_storage& operator=(const peerlist_storage&) = delete;
|
|
|
|
|
|
|
|
//! Save peers from `this` and `other` in stream `dest`.
|
|
|
|
bool store(std::ostream& dest, const peerlist_types& other) const;
|
|
|
|
|
|
|
|
//! Save peers from `this` and `other` in one file at `path`.
|
|
|
|
bool store(const std::string& path, const peerlist_types& other) const;
|
|
|
|
|
|
|
|
//! \return Peers in `zone` and from remove from `this`.
|
|
|
|
peerlist_types take_zone(epee::net_utils::zone zone);
|
|
|
|
|
|
|
|
private:
|
|
|
|
peerlist_types m_types;
|
|
|
|
};
|
2014-03-03 22:07:58 +00:00
|
|
|
|
|
|
|
/************************************************************************/
|
|
|
|
/* */
|
|
|
|
/************************************************************************/
|
|
|
|
class peerlist_manager
|
|
|
|
{
|
|
|
|
public:
|
2018-12-16 17:57:44 +00:00
|
|
|
bool init(peerlist_types&& peers, bool allow_local_ip);
|
2014-03-03 22:07:58 +00:00
|
|
|
size_t get_white_peers_count(){CRITICAL_REGION_LOCAL(m_peerlist_lock); return m_peers_white.size();}
|
|
|
|
size_t get_gray_peers_count(){CRITICAL_REGION_LOCAL(m_peerlist_lock); return m_peers_gray.size();}
|
2018-12-05 22:25:27 +00:00
|
|
|
bool merge_peerlist(const std::vector<peerlist_entry>& outer_bs);
|
2019-04-23 11:07:44 +00:00
|
|
|
bool get_peerlist_head(std::vector<peerlist_entry>& bs_head, bool anonymize, uint32_t depth = P2P_DEFAULT_PEERS_IN_HANDSHAKE);
|
2018-12-16 17:57:44 +00:00
|
|
|
void get_peerlist(std::vector<peerlist_entry>& pl_gray, std::vector<peerlist_entry>& pl_white);
|
|
|
|
void get_peerlist(peerlist_types& peers);
|
2014-03-03 22:07:58 +00:00
|
|
|
bool get_white_peer_by_index(peerlist_entry& p, size_t i);
|
|
|
|
bool get_gray_peer_by_index(peerlist_entry& p, size_t i);
|
Pruning
The blockchain prunes seven eighths of prunable tx data.
This saves about two thirds of the blockchain size, while
keeping the node useful as a sync source for an eighth
of the blockchain.
No other data is currently pruned.
There are three ways to prune a blockchain:
- run monerod with --prune-blockchain
- run "prune_blockchain" in the monerod console
- run the monero-blockchain-prune utility
The first two will prune in place. Due to how LMDB works, this
will not reduce the blockchain size on disk. Instead, it will
mark parts of the file as free, so that future data will use
that free space, causing the file to not grow until free space
grows scarce.
The third way will create a second database, a pruned copy of
the original one. Since this is a new file, this one will be
smaller than the original one.
Once the database is pruned, it will stay pruned as it syncs.
That is, there is no need to use --prune-blockchain again, etc.
2018-04-29 22:30:51 +00:00
|
|
|
template<typename F> bool foreach(bool white, const F &f);
|
2014-03-03 22:07:58 +00:00
|
|
|
bool append_with_peer_white(const peerlist_entry& pr);
|
|
|
|
bool append_with_peer_gray(const peerlist_entry& pr);
|
2017-02-09 00:11:58 +00:00
|
|
|
bool append_with_peer_anchor(const anchor_peerlist_entry& ple);
|
2019-02-24 08:47:49 +00:00
|
|
|
bool set_peer_just_seen(peerid_type peer, const epee::net_utils::network_address& addr, uint32_t pruning_seed, uint16_t rpc_port);
|
2014-03-03 22:07:58 +00:00
|
|
|
bool set_peer_unreachable(const peerlist_entry& pr);
|
2017-05-27 10:35:54 +00:00
|
|
|
bool is_host_allowed(const epee::net_utils::network_address &address);
|
2017-01-21 12:04:49 +00:00
|
|
|
bool get_random_gray_peer(peerlist_entry& pe);
|
2017-01-20 23:59:04 +00:00
|
|
|
bool remove_from_peer_gray(const peerlist_entry& pe);
|
2017-02-09 00:11:58 +00:00
|
|
|
bool get_and_empty_anchor_peerlist(std::vector<anchor_peerlist_entry>& apl);
|
2017-05-27 10:35:54 +00:00
|
|
|
bool remove_from_peer_anchor(const epee::net_utils::network_address& addr);
|
Pruning
The blockchain prunes seven eighths of prunable tx data.
This saves about two thirds of the blockchain size, while
keeping the node useful as a sync source for an eighth
of the blockchain.
No other data is currently pruned.
There are three ways to prune a blockchain:
- run monerod with --prune-blockchain
- run "prune_blockchain" in the monerod console
- run the monero-blockchain-prune utility
The first two will prune in place. Due to how LMDB works, this
will not reduce the blockchain size on disk. Instead, it will
mark parts of the file as free, so that future data will use
that free space, causing the file to not grow until free space
grows scarce.
The third way will create a second database, a pruned copy of
the original one. Since this is a new file, this one will be
smaller than the original one.
Once the database is pruned, it will stay pruned as it syncs.
That is, there is no need to use --prune-blockchain again, etc.
2018-04-29 22:30:51 +00:00
|
|
|
bool remove_from_peer_white(const peerlist_entry& pe);
|
2014-03-03 22:07:58 +00:00
|
|
|
|
|
|
|
private:
|
|
|
|
struct by_time{};
|
|
|
|
struct by_id{};
|
|
|
|
struct by_addr{};
|
|
|
|
|
|
|
|
struct modify_all_but_id
|
|
|
|
{
|
|
|
|
modify_all_but_id(const peerlist_entry& ple):m_ple(ple){}
|
|
|
|
void operator()(peerlist_entry& e)
|
|
|
|
{
|
|
|
|
e.id = m_ple.id;
|
|
|
|
}
|
|
|
|
private:
|
|
|
|
const peerlist_entry& m_ple;
|
|
|
|
};
|
|
|
|
|
|
|
|
struct modify_all
|
|
|
|
{
|
|
|
|
modify_all(const peerlist_entry& ple):m_ple(ple){}
|
|
|
|
void operator()(peerlist_entry& e)
|
|
|
|
{
|
|
|
|
e = m_ple;
|
|
|
|
}
|
|
|
|
private:
|
|
|
|
const peerlist_entry& m_ple;
|
|
|
|
};
|
|
|
|
|
|
|
|
struct modify_last_seen
|
|
|
|
{
|
|
|
|
modify_last_seen(time_t last_seen):m_last_seen(last_seen){}
|
|
|
|
void operator()(peerlist_entry& e)
|
|
|
|
{
|
|
|
|
e.last_seen = m_last_seen;
|
|
|
|
}
|
|
|
|
private:
|
|
|
|
time_t m_last_seen;
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
|
|
typedef boost::multi_index_container<
|
|
|
|
peerlist_entry,
|
|
|
|
boost::multi_index::indexed_by<
|
|
|
|
// access by peerlist_entry::net_adress
|
2017-05-27 10:35:54 +00:00
|
|
|
boost::multi_index::ordered_unique<boost::multi_index::tag<by_addr>, boost::multi_index::member<peerlist_entry,epee::net_utils::network_address,&peerlist_entry::adr> >,
|
2014-03-03 22:07:58 +00:00
|
|
|
// sort by peerlist_entry::last_seen<
|
2014-08-20 15:57:29 +00:00
|
|
|
boost::multi_index::ordered_non_unique<boost::multi_index::tag<by_time>, boost::multi_index::member<peerlist_entry,int64_t,&peerlist_entry::last_seen> >
|
2014-03-03 22:07:58 +00:00
|
|
|
>
|
|
|
|
> peers_indexed;
|
|
|
|
|
2017-02-09 00:11:58 +00:00
|
|
|
typedef boost::multi_index_container<
|
|
|
|
anchor_peerlist_entry,
|
|
|
|
boost::multi_index::indexed_by<
|
|
|
|
// access by anchor_peerlist_entry::net_adress
|
2017-05-27 10:35:54 +00:00
|
|
|
boost::multi_index::ordered_unique<boost::multi_index::tag<by_addr>, boost::multi_index::member<anchor_peerlist_entry,epee::net_utils::network_address,&anchor_peerlist_entry::adr> >,
|
2017-02-09 00:11:58 +00:00
|
|
|
// sort by anchor_peerlist_entry::first_seen
|
|
|
|
boost::multi_index::ordered_non_unique<boost::multi_index::tag<by_time>, boost::multi_index::member<anchor_peerlist_entry,int64_t,&anchor_peerlist_entry::first_seen> >
|
|
|
|
>
|
|
|
|
> anchor_peers_indexed;
|
2014-03-03 22:07:58 +00:00
|
|
|
|
|
|
|
private:
|
2015-12-30 10:36:51 +00:00
|
|
|
void trim_white_peerlist();
|
|
|
|
void trim_gray_peerlist();
|
2014-03-03 22:07:58 +00:00
|
|
|
|
|
|
|
friend class boost::serialization::access;
|
|
|
|
epee::critical_section m_peerlist_lock;
|
|
|
|
std::string m_config_folder;
|
|
|
|
bool m_allow_local_ip;
|
|
|
|
|
|
|
|
|
|
|
|
peers_indexed m_peers_gray;
|
|
|
|
peers_indexed m_peers_white;
|
2017-02-09 00:11:58 +00:00
|
|
|
anchor_peers_indexed m_peers_anchor;
|
2014-03-03 22:07:58 +00:00
|
|
|
};
|
|
|
|
//--------------------------------------------------------------------------------------------------
|
2015-05-28 13:06:19 +00:00
|
|
|
inline void peerlist_manager::trim_gray_peerlist()
|
2014-03-03 22:07:58 +00:00
|
|
|
{
|
|
|
|
while(m_peers_gray.size() > P2P_LOCAL_GRAY_PEERLIST_LIMIT)
|
|
|
|
{
|
|
|
|
peers_indexed::index<by_time>::type& sorted_index=m_peers_gray.get<by_time>();
|
|
|
|
sorted_index.erase(sorted_index.begin());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
//--------------------------------------------------------------------------------------------------
|
2015-05-28 13:06:19 +00:00
|
|
|
inline void peerlist_manager::trim_white_peerlist()
|
2014-03-03 22:07:58 +00:00
|
|
|
{
|
|
|
|
while(m_peers_white.size() > P2P_LOCAL_WHITE_PEERLIST_LIMIT)
|
|
|
|
{
|
|
|
|
peers_indexed::index<by_time>::type& sorted_index=m_peers_white.get<by_time>();
|
|
|
|
sorted_index.erase(sorted_index.begin());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
//--------------------------------------------------------------------------------------------------
|
|
|
|
inline
|
2018-12-05 22:25:27 +00:00
|
|
|
bool peerlist_manager::merge_peerlist(const std::vector<peerlist_entry>& outer_bs)
|
2014-03-03 22:07:58 +00:00
|
|
|
{
|
|
|
|
CRITICAL_REGION_LOCAL(m_peerlist_lock);
|
2017-01-22 20:38:10 +00:00
|
|
|
for(const peerlist_entry& be: outer_bs)
|
2014-03-03 22:07:58 +00:00
|
|
|
{
|
|
|
|
append_with_peer_gray(be);
|
|
|
|
}
|
|
|
|
// delete extra elements
|
|
|
|
trim_gray_peerlist();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
//--------------------------------------------------------------------------------------------------
|
|
|
|
inline
|
|
|
|
bool peerlist_manager::get_white_peer_by_index(peerlist_entry& p, size_t i)
|
|
|
|
{
|
|
|
|
CRITICAL_REGION_LOCAL(m_peerlist_lock);
|
|
|
|
if(i >= m_peers_white.size())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
peers_indexed::index<by_time>::type& by_time_index = m_peers_white.get<by_time>();
|
|
|
|
p = *epee::misc_utils::move_it_backward(--by_time_index.end(), i);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
//--------------------------------------------------------------------------------------------------
|
|
|
|
inline
|
|
|
|
bool peerlist_manager::get_gray_peer_by_index(peerlist_entry& p, size_t i)
|
|
|
|
{
|
|
|
|
CRITICAL_REGION_LOCAL(m_peerlist_lock);
|
|
|
|
if(i >= m_peers_gray.size())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
peers_indexed::index<by_time>::type& by_time_index = m_peers_gray.get<by_time>();
|
|
|
|
p = *epee::misc_utils::move_it_backward(--by_time_index.end(), i);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
//--------------------------------------------------------------------------------------------------
|
|
|
|
inline
|
2017-05-27 10:35:54 +00:00
|
|
|
bool peerlist_manager::is_host_allowed(const epee::net_utils::network_address &address)
|
2014-03-03 22:07:58 +00:00
|
|
|
{
|
|
|
|
//never allow loopback ip
|
2017-05-27 10:35:54 +00:00
|
|
|
if(address.is_loopback())
|
2014-03-03 22:07:58 +00:00
|
|
|
return false;
|
|
|
|
|
2017-05-27 10:35:54 +00:00
|
|
|
if(!m_allow_local_ip && address.is_local())
|
2014-03-03 22:07:58 +00:00
|
|
|
return false;
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
//--------------------------------------------------------------------------------------------------
|
|
|
|
inline
|
2019-04-23 11:07:44 +00:00
|
|
|
bool peerlist_manager::get_peerlist_head(std::vector<peerlist_entry>& bs_head, bool anonymize, uint32_t depth)
|
2014-03-03 22:07:58 +00:00
|
|
|
{
|
|
|
|
CRITICAL_REGION_LOCAL(m_peerlist_lock);
|
|
|
|
peers_indexed::index<by_time>::type& by_time_index=m_peers_white.get<by_time>();
|
|
|
|
uint32_t cnt = 0;
|
2019-04-23 11:07:44 +00:00
|
|
|
|
|
|
|
// picks a random set of peers within the first 120%, rather than a set of the first 100%.
|
|
|
|
// The intent is that if someone asks twice, they can't easily tell:
|
|
|
|
// - this address was not in the first list, but is in the second, so the only way this can be
|
|
|
|
// is if its last_seen was recently reset, so this means the target node recently had a new
|
|
|
|
// connection to that address
|
|
|
|
// - this address was in the first list, and not in the second, which means either the address
|
|
|
|
// was moved to the gray list (if it's not accessibe, which the attacker can check if
|
|
|
|
// the address accepts incoming connections) or it was the oldest to still fit in the 250 items,
|
|
|
|
// so its last_seen is old.
|
2019-06-21 10:34:00 +00:00
|
|
|
//
|
|
|
|
// See Cao, Tong et al. "Exploring the Monero Peer-to-Peer Network". https://eprint.iacr.org/2019/411
|
|
|
|
//
|
2019-04-23 11:07:44 +00:00
|
|
|
const uint32_t pick_depth = anonymize ? depth + depth / 5 : depth;
|
|
|
|
bs_head.reserve(pick_depth);
|
2017-01-22 20:47:39 +00:00
|
|
|
for(const peers_indexed::value_type& vl: boost::adaptors::reverse(by_time_index))
|
2014-03-03 22:07:58 +00:00
|
|
|
{
|
2019-04-23 11:07:44 +00:00
|
|
|
if(cnt++ >= pick_depth)
|
2014-03-03 22:07:58 +00:00
|
|
|
break;
|
2017-01-18 17:53:11 +00:00
|
|
|
|
|
|
|
bs_head.push_back(vl);
|
2014-03-03 22:07:58 +00:00
|
|
|
}
|
2019-04-23 11:07:44 +00:00
|
|
|
|
|
|
|
if (anonymize)
|
|
|
|
{
|
2019-07-03 09:05:01 +00:00
|
|
|
std::shuffle(bs_head.begin(), bs_head.end(), crypto::random_device{});
|
2019-04-23 11:07:44 +00:00
|
|
|
if (bs_head.size() > depth)
|
|
|
|
bs_head.resize(depth);
|
|
|
|
for (auto &e: bs_head)
|
|
|
|
e.last_seen = 0;
|
|
|
|
}
|
|
|
|
|
2014-03-03 22:07:58 +00:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
//--------------------------------------------------------------------------------------------------
|
Pruning
The blockchain prunes seven eighths of prunable tx data.
This saves about two thirds of the blockchain size, while
keeping the node useful as a sync source for an eighth
of the blockchain.
No other data is currently pruned.
There are three ways to prune a blockchain:
- run monerod with --prune-blockchain
- run "prune_blockchain" in the monerod console
- run the monero-blockchain-prune utility
The first two will prune in place. Due to how LMDB works, this
will not reduce the blockchain size on disk. Instead, it will
mark parts of the file as free, so that future data will use
that free space, causing the file to not grow until free space
grows scarce.
The third way will create a second database, a pruned copy of
the original one. Since this is a new file, this one will be
smaller than the original one.
Once the database is pruned, it will stay pruned as it syncs.
That is, there is no need to use --prune-blockchain again, etc.
2018-04-29 22:30:51 +00:00
|
|
|
template<typename F> inline
|
|
|
|
bool peerlist_manager::foreach(bool white, const F &f)
|
|
|
|
{
|
|
|
|
CRITICAL_REGION_LOCAL(m_peerlist_lock);
|
|
|
|
peers_indexed::index<by_time>::type& by_time_index = white ? m_peers_white.get<by_time>() : m_peers_gray.get<by_time>();
|
|
|
|
for(const peers_indexed::value_type& vl: boost::adaptors::reverse(by_time_index))
|
|
|
|
if (!f(vl))
|
|
|
|
return false;
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
//--------------------------------------------------------------------------------------------------
|
2014-03-03 22:07:58 +00:00
|
|
|
inline
|
2019-02-24 08:47:49 +00:00
|
|
|
bool peerlist_manager::set_peer_just_seen(peerid_type peer, const epee::net_utils::network_address& addr, uint32_t pruning_seed, uint16_t rpc_port)
|
2014-03-03 22:07:58 +00:00
|
|
|
{
|
|
|
|
TRY_ENTRY();
|
|
|
|
CRITICAL_REGION_LOCAL(m_peerlist_lock);
|
|
|
|
//find in white list
|
|
|
|
peerlist_entry ple;
|
|
|
|
ple.adr = addr;
|
|
|
|
ple.id = peer;
|
|
|
|
ple.last_seen = time(NULL);
|
Pruning
The blockchain prunes seven eighths of prunable tx data.
This saves about two thirds of the blockchain size, while
keeping the node useful as a sync source for an eighth
of the blockchain.
No other data is currently pruned.
There are three ways to prune a blockchain:
- run monerod with --prune-blockchain
- run "prune_blockchain" in the monerod console
- run the monero-blockchain-prune utility
The first two will prune in place. Due to how LMDB works, this
will not reduce the blockchain size on disk. Instead, it will
mark parts of the file as free, so that future data will use
that free space, causing the file to not grow until free space
grows scarce.
The third way will create a second database, a pruned copy of
the original one. Since this is a new file, this one will be
smaller than the original one.
Once the database is pruned, it will stay pruned as it syncs.
That is, there is no need to use --prune-blockchain again, etc.
2018-04-29 22:30:51 +00:00
|
|
|
ple.pruning_seed = pruning_seed;
|
2019-02-24 08:47:49 +00:00
|
|
|
ple.rpc_port = rpc_port;
|
2014-03-03 22:07:58 +00:00
|
|
|
return append_with_peer_white(ple);
|
|
|
|
CATCH_ENTRY_L0("peerlist_manager::set_peer_just_seen()", false);
|
|
|
|
}
|
|
|
|
//--------------------------------------------------------------------------------------------------
|
|
|
|
inline
|
|
|
|
bool peerlist_manager::append_with_peer_white(const peerlist_entry& ple)
|
|
|
|
{
|
|
|
|
TRY_ENTRY();
|
2017-05-27 10:35:54 +00:00
|
|
|
if(!is_host_allowed(ple.adr))
|
2014-03-03 22:07:58 +00:00
|
|
|
return true;
|
|
|
|
|
|
|
|
CRITICAL_REGION_LOCAL(m_peerlist_lock);
|
|
|
|
//find in white list
|
|
|
|
auto by_addr_it_wt = m_peers_white.get<by_addr>().find(ple.adr);
|
|
|
|
if(by_addr_it_wt == m_peers_white.get<by_addr>().end())
|
|
|
|
{
|
|
|
|
//put new record into white list
|
|
|
|
m_peers_white.insert(ple);
|
|
|
|
trim_white_peerlist();
|
|
|
|
}else
|
|
|
|
{
|
2019-06-18 14:38:57 +00:00
|
|
|
//update record in white list
|
|
|
|
peerlist_entry new_ple = ple;
|
|
|
|
if (by_addr_it_wt->pruning_seed && ple.pruning_seed == 0) // guard against older nodes not passing pruning info around
|
|
|
|
new_ple.pruning_seed = by_addr_it_wt->pruning_seed;
|
|
|
|
if (by_addr_it_wt->rpc_port && ple.rpc_port == 0) // guard against older nodes not passing RPC port around
|
|
|
|
new_ple.rpc_port = by_addr_it_wt->rpc_port;
|
|
|
|
new_ple.last_seen = by_addr_it_wt->last_seen; // do not overwrite the last seen timestamp, incoming peer list are untrusted
|
|
|
|
m_peers_white.replace(by_addr_it_wt, new_ple);
|
2014-03-03 22:07:58 +00:00
|
|
|
}
|
|
|
|
//remove from gray list, if need
|
|
|
|
auto by_addr_it_gr = m_peers_gray.get<by_addr>().find(ple.adr);
|
|
|
|
if(by_addr_it_gr != m_peers_gray.get<by_addr>().end())
|
|
|
|
{
|
|
|
|
m_peers_gray.erase(by_addr_it_gr);
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
CATCH_ENTRY_L0("peerlist_manager::append_with_peer_white()", false);
|
|
|
|
}
|
|
|
|
//--------------------------------------------------------------------------------------------------
|
|
|
|
inline
|
|
|
|
bool peerlist_manager::append_with_peer_gray(const peerlist_entry& ple)
|
|
|
|
{
|
|
|
|
TRY_ENTRY();
|
2017-05-27 10:35:54 +00:00
|
|
|
if(!is_host_allowed(ple.adr))
|
2014-03-03 22:07:58 +00:00
|
|
|
return true;
|
|
|
|
|
|
|
|
CRITICAL_REGION_LOCAL(m_peerlist_lock);
|
|
|
|
//find in white list
|
|
|
|
auto by_addr_it_wt = m_peers_white.get<by_addr>().find(ple.adr);
|
|
|
|
if(by_addr_it_wt != m_peers_white.get<by_addr>().end())
|
|
|
|
return true;
|
|
|
|
|
|
|
|
//update gray list
|
|
|
|
auto by_addr_it_gr = m_peers_gray.get<by_addr>().find(ple.adr);
|
|
|
|
if(by_addr_it_gr == m_peers_gray.get<by_addr>().end())
|
|
|
|
{
|
|
|
|
//put new record into white list
|
|
|
|
m_peers_gray.insert(ple);
|
|
|
|
trim_gray_peerlist();
|
|
|
|
}else
|
|
|
|
{
|
2019-06-18 14:38:57 +00:00
|
|
|
//update record in gray list
|
|
|
|
peerlist_entry new_ple = ple;
|
|
|
|
if (by_addr_it_gr->pruning_seed && ple.pruning_seed == 0) // guard against older nodes not passing pruning info around
|
|
|
|
new_ple.pruning_seed = by_addr_it_gr->pruning_seed;
|
|
|
|
if (by_addr_it_gr->rpc_port && ple.rpc_port == 0) // guard against older nodes not passing RPC port around
|
|
|
|
new_ple.rpc_port = by_addr_it_gr->rpc_port;
|
|
|
|
new_ple.last_seen = by_addr_it_gr->last_seen; // do not overwrite the last seen timestamp, incoming peer list are untrusted
|
|
|
|
m_peers_gray.replace(by_addr_it_gr, new_ple);
|
2014-03-03 22:07:58 +00:00
|
|
|
}
|
|
|
|
return true;
|
|
|
|
CATCH_ENTRY_L0("peerlist_manager::append_with_peer_gray()", false);
|
|
|
|
}
|
|
|
|
//--------------------------------------------------------------------------------------------------
|
2017-01-20 23:59:04 +00:00
|
|
|
inline
|
2017-02-09 00:11:58 +00:00
|
|
|
bool peerlist_manager::append_with_peer_anchor(const anchor_peerlist_entry& ple)
|
|
|
|
{
|
|
|
|
TRY_ENTRY();
|
|
|
|
|
|
|
|
CRITICAL_REGION_LOCAL(m_peerlist_lock);
|
|
|
|
|
|
|
|
auto by_addr_it_anchor = m_peers_anchor.get<by_addr>().find(ple.adr);
|
|
|
|
|
|
|
|
if(by_addr_it_anchor == m_peers_anchor.get<by_addr>().end()) {
|
|
|
|
m_peers_anchor.insert(ple);
|
|
|
|
}
|
|
|
|
|
|
|
|
return true;
|
|
|
|
|
|
|
|
CATCH_ENTRY_L0("peerlist_manager::append_with_peer_anchor()", false);
|
|
|
|
}
|
|
|
|
//--------------------------------------------------------------------------------------------------
|
|
|
|
inline
|
2017-01-21 12:04:49 +00:00
|
|
|
bool peerlist_manager::get_random_gray_peer(peerlist_entry& pe)
|
2017-01-20 23:59:04 +00:00
|
|
|
{
|
|
|
|
TRY_ENTRY();
|
|
|
|
|
|
|
|
CRITICAL_REGION_LOCAL(m_peerlist_lock);
|
|
|
|
|
|
|
|
if (m_peers_gray.empty()) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2019-04-03 05:10:24 +00:00
|
|
|
size_t random_index = crypto::rand_idx(m_peers_gray.size());
|
2017-01-20 23:59:04 +00:00
|
|
|
|
|
|
|
peers_indexed::index<by_time>::type& by_time_index = m_peers_gray.get<by_time>();
|
2017-02-05 22:46:47 +00:00
|
|
|
pe = *epee::misc_utils::move_it_backward(--by_time_index.end(), random_index);
|
2017-01-20 23:59:04 +00:00
|
|
|
|
|
|
|
return true;
|
|
|
|
|
2017-01-21 12:04:49 +00:00
|
|
|
CATCH_ENTRY_L0("peerlist_manager::get_random_gray_peer()", false);
|
2017-01-20 23:59:04 +00:00
|
|
|
}
|
|
|
|
//--------------------------------------------------------------------------------------------------
|
|
|
|
inline
|
Pruning
The blockchain prunes seven eighths of prunable tx data.
This saves about two thirds of the blockchain size, while
keeping the node useful as a sync source for an eighth
of the blockchain.
No other data is currently pruned.
There are three ways to prune a blockchain:
- run monerod with --prune-blockchain
- run "prune_blockchain" in the monerod console
- run the monero-blockchain-prune utility
The first two will prune in place. Due to how LMDB works, this
will not reduce the blockchain size on disk. Instead, it will
mark parts of the file as free, so that future data will use
that free space, causing the file to not grow until free space
grows scarce.
The third way will create a second database, a pruned copy of
the original one. Since this is a new file, this one will be
smaller than the original one.
Once the database is pruned, it will stay pruned as it syncs.
That is, there is no need to use --prune-blockchain again, etc.
2018-04-29 22:30:51 +00:00
|
|
|
bool peerlist_manager::remove_from_peer_white(const peerlist_entry& pe)
|
|
|
|
{
|
|
|
|
TRY_ENTRY();
|
|
|
|
|
|
|
|
CRITICAL_REGION_LOCAL(m_peerlist_lock);
|
|
|
|
|
|
|
|
peers_indexed::index_iterator<by_addr>::type iterator = m_peers_white.get<by_addr>().find(pe.adr);
|
|
|
|
|
|
|
|
if (iterator != m_peers_white.get<by_addr>().end()) {
|
|
|
|
m_peers_white.erase(iterator);
|
|
|
|
}
|
|
|
|
|
|
|
|
return true;
|
|
|
|
|
|
|
|
CATCH_ENTRY_L0("peerlist_manager::remove_from_peer_white()", false);
|
|
|
|
}
|
|
|
|
//--------------------------------------------------------------------------------------------------
|
|
|
|
inline
|
2017-01-20 23:59:04 +00:00
|
|
|
bool peerlist_manager::remove_from_peer_gray(const peerlist_entry& pe)
|
|
|
|
{
|
|
|
|
TRY_ENTRY();
|
|
|
|
|
|
|
|
CRITICAL_REGION_LOCAL(m_peerlist_lock);
|
|
|
|
|
|
|
|
peers_indexed::index_iterator<by_addr>::type iterator = m_peers_gray.get<by_addr>().find(pe.adr);
|
|
|
|
|
|
|
|
if (iterator != m_peers_gray.get<by_addr>().end()) {
|
|
|
|
m_peers_gray.erase(iterator);
|
|
|
|
}
|
|
|
|
|
|
|
|
return true;
|
|
|
|
|
|
|
|
CATCH_ENTRY_L0("peerlist_manager::remove_from_peer_gray()", false);
|
|
|
|
}
|
2017-02-09 00:11:58 +00:00
|
|
|
//--------------------------------------------------------------------------------------------------
|
|
|
|
inline
|
|
|
|
bool peerlist_manager::get_and_empty_anchor_peerlist(std::vector<anchor_peerlist_entry>& apl)
|
|
|
|
{
|
|
|
|
TRY_ENTRY();
|
|
|
|
|
|
|
|
CRITICAL_REGION_LOCAL(m_peerlist_lock);
|
|
|
|
|
|
|
|
auto begin = m_peers_anchor.get<by_time>().begin();
|
|
|
|
auto end = m_peers_anchor.get<by_time>().end();
|
|
|
|
|
|
|
|
std::for_each(begin, end, [&apl](const anchor_peerlist_entry &a) {
|
|
|
|
apl.push_back(a);
|
|
|
|
});
|
|
|
|
|
|
|
|
m_peers_anchor.get<by_time>().clear();
|
|
|
|
|
|
|
|
return true;
|
|
|
|
|
|
|
|
CATCH_ENTRY_L0("peerlist_manager::get_and_empty_anchor_peerlist()", false);
|
|
|
|
}
|
|
|
|
//--------------------------------------------------------------------------------------------------
|
|
|
|
inline
|
2017-05-27 10:35:54 +00:00
|
|
|
bool peerlist_manager::remove_from_peer_anchor(const epee::net_utils::network_address& addr)
|
2017-02-09 00:11:58 +00:00
|
|
|
{
|
|
|
|
TRY_ENTRY();
|
|
|
|
|
|
|
|
CRITICAL_REGION_LOCAL(m_peerlist_lock);
|
|
|
|
|
|
|
|
anchor_peers_indexed::index_iterator<by_addr>::type iterator = m_peers_anchor.get<by_addr>().find(addr);
|
|
|
|
|
|
|
|
if (iterator != m_peers_anchor.get<by_addr>().end()) {
|
|
|
|
m_peers_anchor.erase(iterator);
|
|
|
|
}
|
|
|
|
|
|
|
|
return true;
|
|
|
|
|
|
|
|
CATCH_ENTRY_L0("peerlist_manager::remove_from_peer_anchor()", false);
|
|
|
|
}
|
|
|
|
//--------------------------------------------------------------------------------------------------
|
2014-03-03 22:07:58 +00:00
|
|
|
}
|
|
|
|
|