2021-11-22 16:28:30 +00:00
|
|
|
use super::*;
|
|
|
|
|
|
|
|
use crate::dht::*;
|
|
|
|
use crate::xx::*;
|
|
|
|
use crate::*;
|
|
|
|
|
2022-08-18 20:21:13 +00:00
|
|
|
pub type LowLevelProtocolPorts = BTreeSet<(LowLevelProtocolType, AddressType, u16)>;
|
|
|
|
pub type ProtocolToPortMapping = BTreeMap<(ProtocolType, AddressType), (LowLevelProtocolType, u16)>;
|
|
|
|
#[derive(Clone, Debug)]
|
|
|
|
pub struct MappedPortInfo {
|
|
|
|
pub low_level_protocol_ports: LowLevelProtocolPorts,
|
|
|
|
pub protocol_to_port: ProtocolToPortMapping,
|
|
|
|
}
|
|
|
|
|
2021-11-22 16:28:30 +00:00
|
|
|
impl RoutingTable {
|
2022-08-05 18:48:02 +00:00
|
|
|
// Makes a filter that finds nodes with a matching inbound dialinfo
|
|
|
|
pub fn make_inbound_dial_info_entry_filter(
|
|
|
|
dial_info_filter: DialInfoFilter,
|
|
|
|
) -> impl FnMut(&BucketEntryInner) -> bool {
|
|
|
|
// does it have matching public dial info?
|
|
|
|
move |e| {
|
|
|
|
e.node_info()
|
|
|
|
.map(|n| {
|
|
|
|
n.first_filtered_dial_info_detail(|did| did.matches_filter(&dial_info_filter))
|
|
|
|
.is_some()
|
|
|
|
})
|
|
|
|
.unwrap_or(false)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Makes a filter that finds nodes capable of dialing a particular outbound dialinfo
|
|
|
|
pub fn make_outbound_dial_info_entry_filter(
|
|
|
|
dial_info: DialInfo,
|
|
|
|
) -> impl FnMut(&BucketEntryInner) -> bool {
|
|
|
|
// does the node's outbound capabilities match the dialinfo?
|
|
|
|
move |e| {
|
|
|
|
e.node_info()
|
|
|
|
.map(|n| {
|
|
|
|
let mut dif = DialInfoFilter::all();
|
|
|
|
dif = dif.with_protocol_type_set(n.outbound_protocols);
|
|
|
|
dif = dif.with_address_type_set(n.address_types);
|
|
|
|
dial_info.matches_filter(&dif)
|
|
|
|
})
|
|
|
|
.unwrap_or(false)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Make a filter that wraps another filter
|
|
|
|
pub fn combine_filters<F, G>(mut f1: F, mut f2: G) -> impl FnMut(&BucketEntryInner) -> bool
|
|
|
|
where
|
|
|
|
F: FnMut(&BucketEntryInner) -> bool,
|
|
|
|
G: FnMut(&BucketEntryInner) -> bool,
|
|
|
|
{
|
|
|
|
move |e| {
|
|
|
|
if !f1(e) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
if !f2(e) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
true
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Retrieve the fastest nodes in the routing table matching an entry filter
|
|
|
|
pub fn find_fast_public_nodes_filtered<F>(
|
2022-04-17 17:28:39 +00:00
|
|
|
&self,
|
2022-06-25 14:57:33 +00:00
|
|
|
node_count: usize,
|
2022-08-05 18:48:02 +00:00
|
|
|
mut entry_filter: F,
|
|
|
|
) -> Vec<NodeRef>
|
|
|
|
where
|
|
|
|
F: FnMut(&BucketEntryInner) -> bool,
|
|
|
|
{
|
2021-11-22 16:28:30 +00:00
|
|
|
self.find_fastest_nodes(
|
2022-06-25 14:57:33 +00:00
|
|
|
// count
|
|
|
|
node_count,
|
2021-11-22 16:28:30 +00:00
|
|
|
// filter
|
2022-06-25 14:57:33 +00:00
|
|
|
Some(move |_k: DHTKey, v: Option<Arc<BucketEntry>>| {
|
|
|
|
let entry = v.unwrap();
|
|
|
|
entry.with(|e| {
|
2022-04-17 17:28:39 +00:00
|
|
|
// skip nodes on our local network here
|
2022-06-25 14:57:33 +00:00
|
|
|
if e.local_node_info().is_some() {
|
2022-04-17 17:28:39 +00:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2022-08-05 18:48:02 +00:00
|
|
|
// skip nodes that dont match entry filter
|
|
|
|
entry_filter(e)
|
2022-06-25 14:57:33 +00:00
|
|
|
})
|
|
|
|
}),
|
2021-11-22 16:28:30 +00:00
|
|
|
// transform
|
2022-06-25 14:57:33 +00:00
|
|
|
|k: DHTKey, v: Option<Arc<BucketEntry>>| {
|
2022-08-05 18:48:02 +00:00
|
|
|
NodeRef::new(self.clone(), k, v.unwrap().clone(), None)
|
2022-04-19 15:23:44 +00:00
|
|
|
},
|
2021-11-22 16:28:30 +00:00
|
|
|
)
|
|
|
|
}
|
|
|
|
|
2022-06-25 14:57:33 +00:00
|
|
|
// Retrieve up to N of each type of protocol capable nodes
|
|
|
|
pub fn find_bootstrap_nodes_filtered(&self, max_per_type: usize) -> Vec<NodeRef> {
|
|
|
|
let protocol_types = vec![
|
|
|
|
ProtocolType::UDP,
|
|
|
|
ProtocolType::TCP,
|
|
|
|
ProtocolType::WS,
|
|
|
|
ProtocolType::WSS,
|
|
|
|
];
|
|
|
|
let mut nodes_proto_v4 = vec![0usize, 0usize, 0usize, 0usize];
|
|
|
|
let mut nodes_proto_v6 = vec![0usize, 0usize, 0usize, 0usize];
|
|
|
|
|
|
|
|
self.find_fastest_nodes(
|
|
|
|
// count
|
|
|
|
protocol_types.len() * 2 * max_per_type,
|
|
|
|
// filter
|
|
|
|
Some(move |_k: DHTKey, v: Option<Arc<BucketEntry>>| {
|
|
|
|
let entry = v.unwrap();
|
|
|
|
entry.with(|e| {
|
|
|
|
// skip nodes on our local network here
|
|
|
|
if e.local_node_info().is_some() {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
// does it have some dial info we need?
|
|
|
|
let filter = |n: NodeInfo| {
|
|
|
|
let mut keep = false;
|
|
|
|
for did in n.dial_info_detail_list {
|
|
|
|
if did.dial_info.is_global() {
|
|
|
|
if matches!(did.dial_info.address_type(), AddressType::IPV4) {
|
|
|
|
for (n, protocol_type) in protocol_types.iter().enumerate() {
|
|
|
|
if nodes_proto_v4[n] < max_per_type
|
|
|
|
&& did.dial_info.protocol_type() == *protocol_type
|
|
|
|
{
|
|
|
|
nodes_proto_v4[n] += 1;
|
|
|
|
keep = true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
} else if matches!(did.dial_info.address_type(), AddressType::IPV6)
|
|
|
|
{
|
|
|
|
for (n, protocol_type) in protocol_types.iter().enumerate() {
|
|
|
|
if nodes_proto_v6[n] < max_per_type
|
|
|
|
&& did.dial_info.protocol_type() == *protocol_type
|
|
|
|
{
|
|
|
|
nodes_proto_v6[n] += 1;
|
|
|
|
keep = true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
keep
|
|
|
|
};
|
|
|
|
|
|
|
|
e.node_info().map(filter).unwrap_or(false)
|
|
|
|
})
|
|
|
|
}),
|
|
|
|
// transform
|
|
|
|
|k: DHTKey, v: Option<Arc<BucketEntry>>| {
|
|
|
|
NodeRef::new(self.clone(), k, v.unwrap().clone(), None)
|
|
|
|
},
|
|
|
|
)
|
|
|
|
}
|
|
|
|
|
2022-04-17 17:28:39 +00:00
|
|
|
// Get our own node's peer info (public node info) so we can share it with other nodes
|
2022-04-16 15:18:54 +00:00
|
|
|
pub fn get_own_peer_info(&self) -> PeerInfo {
|
2022-05-11 01:49:42 +00:00
|
|
|
PeerInfo::new(NodeId::new(self.node_id()), self.get_own_signed_node_info())
|
|
|
|
}
|
|
|
|
|
|
|
|
pub fn get_own_signed_node_info(&self) -> SignedNodeInfo {
|
|
|
|
let node_id = NodeId::new(self.node_id());
|
|
|
|
let secret = self.node_id_secret();
|
|
|
|
SignedNodeInfo::with_secret(self.get_own_node_info(), node_id, &secret).unwrap()
|
|
|
|
}
|
|
|
|
|
|
|
|
pub fn get_own_node_info(&self) -> NodeInfo {
|
2022-04-08 14:17:09 +00:00
|
|
|
let netman = self.network_manager();
|
|
|
|
let relay_node = netman.relay_node();
|
2022-08-02 01:06:31 +00:00
|
|
|
let pc = netman.get_protocol_config();
|
2022-05-11 01:49:42 +00:00
|
|
|
NodeInfo {
|
|
|
|
network_class: netman.get_network_class().unwrap_or(NetworkClass::Invalid),
|
2022-08-02 01:06:31 +00:00
|
|
|
outbound_protocols: pc.outbound,
|
|
|
|
address_types: pc.family_global,
|
2022-05-28 14:07:57 +00:00
|
|
|
min_version: MIN_VERSION,
|
|
|
|
max_version: MAX_VERSION,
|
2022-05-11 01:49:42 +00:00
|
|
|
dial_info_detail_list: self.dial_info_details(RoutingDomain::PublicInternet),
|
|
|
|
relay_peer_info: relay_node.and_then(|rn| rn.peer_info().map(Box::new)),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-05-24 21:13:52 +00:00
|
|
|
pub fn filter_has_valid_signed_node_info(
|
2022-06-25 14:57:33 +00:00
|
|
|
v: Option<Arc<BucketEntry>>,
|
2022-05-24 21:13:52 +00:00
|
|
|
own_peer_info_is_valid: bool,
|
|
|
|
) -> bool {
|
2022-06-25 14:57:33 +00:00
|
|
|
match v {
|
2022-05-24 21:13:52 +00:00
|
|
|
None => own_peer_info_is_valid,
|
2022-06-25 14:57:33 +00:00
|
|
|
Some(entry) => entry.with(|e| e.has_valid_signed_node_info()),
|
2021-11-22 16:28:30 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
pub fn transform_to_peer_info(
|
2022-06-25 14:57:33 +00:00
|
|
|
k: DHTKey,
|
|
|
|
v: Option<Arc<BucketEntry>>,
|
2021-11-22 16:28:30 +00:00
|
|
|
own_peer_info: &PeerInfo,
|
|
|
|
) -> PeerInfo {
|
2022-06-25 14:57:33 +00:00
|
|
|
match v {
|
2021-11-22 16:28:30 +00:00
|
|
|
None => own_peer_info.clone(),
|
2022-06-25 14:57:33 +00:00
|
|
|
Some(entry) => entry.with(|e| e.peer_info(k).unwrap()),
|
2021-11-22 16:28:30 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
pub fn find_peers_with_sort_and_filter<F, C, T, O>(
|
|
|
|
&self,
|
|
|
|
node_count: usize,
|
|
|
|
cur_ts: u64,
|
2022-06-25 14:57:33 +00:00
|
|
|
mut filter: F,
|
2021-11-22 16:28:30 +00:00
|
|
|
compare: C,
|
2022-06-25 14:57:33 +00:00
|
|
|
mut transform: T,
|
2021-11-22 16:28:30 +00:00
|
|
|
) -> Vec<O>
|
|
|
|
where
|
2022-06-25 14:57:33 +00:00
|
|
|
F: FnMut(DHTKey, Option<Arc<BucketEntry>>) -> bool,
|
|
|
|
C: FnMut(
|
|
|
|
&(DHTKey, Option<Arc<BucketEntry>>),
|
|
|
|
&(DHTKey, Option<Arc<BucketEntry>>),
|
2021-11-22 16:28:30 +00:00
|
|
|
) -> core::cmp::Ordering,
|
2022-06-25 14:57:33 +00:00
|
|
|
T: FnMut(DHTKey, Option<Arc<BucketEntry>>) -> O,
|
2021-11-22 16:28:30 +00:00
|
|
|
{
|
2022-06-25 14:57:33 +00:00
|
|
|
let inner = self.inner.read();
|
|
|
|
let self_node_id = inner.node_id;
|
2021-11-22 16:28:30 +00:00
|
|
|
|
|
|
|
// collect all the nodes for sorting
|
|
|
|
let mut nodes =
|
2022-06-25 14:57:33 +00:00
|
|
|
Vec::<(DHTKey, Option<Arc<BucketEntry>>)>::with_capacity(inner.bucket_entry_count + 1);
|
|
|
|
|
2021-11-22 16:28:30 +00:00
|
|
|
// add our own node (only one of there with the None entry)
|
2022-06-25 14:57:33 +00:00
|
|
|
if filter(self_node_id, None) {
|
|
|
|
nodes.push((self_node_id, None));
|
2021-11-22 16:28:30 +00:00
|
|
|
}
|
2022-06-25 14:57:33 +00:00
|
|
|
|
2021-11-22 16:28:30 +00:00
|
|
|
// add all nodes from buckets
|
2022-06-25 19:28:27 +00:00
|
|
|
Self::with_entries(&*inner, cur_ts, BucketEntryState::Unreliable, |k, v| {
|
2022-06-25 14:57:33 +00:00
|
|
|
// Apply filter
|
|
|
|
if filter(k, Some(v.clone())) {
|
|
|
|
nodes.push((k, Some(v.clone())));
|
2021-11-22 16:28:30 +00:00
|
|
|
}
|
2022-06-25 14:57:33 +00:00
|
|
|
Option::<()>::None
|
|
|
|
});
|
2021-11-22 16:28:30 +00:00
|
|
|
|
|
|
|
// sort by preference for returning nodes
|
|
|
|
nodes.sort_by(compare);
|
|
|
|
|
|
|
|
// return transformed vector for filtered+sorted nodes
|
|
|
|
let cnt = usize::min(node_count, nodes.len());
|
|
|
|
let mut out = Vec::<O>::with_capacity(cnt);
|
2022-06-25 14:57:33 +00:00
|
|
|
for node in nodes {
|
|
|
|
let val = transform(node.0, node.1);
|
2021-11-22 16:28:30 +00:00
|
|
|
out.push(val);
|
|
|
|
}
|
|
|
|
|
|
|
|
out
|
|
|
|
}
|
|
|
|
|
2022-06-25 14:57:33 +00:00
|
|
|
pub fn find_fastest_nodes<T, F, O>(
|
|
|
|
&self,
|
|
|
|
node_count: usize,
|
|
|
|
mut filter: Option<F>,
|
|
|
|
transform: T,
|
|
|
|
) -> Vec<O>
|
2021-11-22 16:28:30 +00:00
|
|
|
where
|
2022-06-25 14:57:33 +00:00
|
|
|
F: FnMut(DHTKey, Option<Arc<BucketEntry>>) -> bool,
|
|
|
|
T: FnMut(DHTKey, Option<Arc<BucketEntry>>) -> O,
|
2021-11-22 16:28:30 +00:00
|
|
|
{
|
2022-06-28 03:46:29 +00:00
|
|
|
let cur_ts = intf::get_timestamp();
|
2021-11-22 16:28:30 +00:00
|
|
|
let out = self.find_peers_with_sort_and_filter(
|
|
|
|
node_count,
|
|
|
|
cur_ts,
|
|
|
|
// filter
|
2022-06-25 14:57:33 +00:00
|
|
|
|k, v| {
|
|
|
|
if let Some(entry) = &v {
|
|
|
|
// always filter out dead nodes
|
|
|
|
if entry.with(|e| e.state(cur_ts) == BucketEntryState::Dead) {
|
|
|
|
false
|
|
|
|
} else {
|
|
|
|
filter.as_mut().map(|f| f(k, v)).unwrap_or(true)
|
|
|
|
}
|
|
|
|
} else {
|
2022-05-24 21:13:52 +00:00
|
|
|
// always filter out self peer, as it is irrelevant to the 'fastest nodes' search
|
|
|
|
false
|
2021-11-26 15:39:43 +00:00
|
|
|
}
|
2021-11-22 16:28:30 +00:00
|
|
|
},
|
|
|
|
// sort
|
|
|
|
|(a_key, a_entry), (b_key, b_entry)| {
|
|
|
|
// same nodes are always the same
|
|
|
|
if a_key == b_key {
|
|
|
|
return core::cmp::Ordering::Equal;
|
|
|
|
}
|
|
|
|
// our own node always comes last (should not happen, here for completeness)
|
|
|
|
if a_entry.is_none() {
|
|
|
|
return core::cmp::Ordering::Greater;
|
|
|
|
}
|
|
|
|
if b_entry.is_none() {
|
|
|
|
return core::cmp::Ordering::Less;
|
|
|
|
}
|
|
|
|
// reliable nodes come first
|
|
|
|
let ae = a_entry.as_ref().unwrap();
|
|
|
|
let be = b_entry.as_ref().unwrap();
|
2022-06-25 14:57:33 +00:00
|
|
|
ae.with(|ae| {
|
|
|
|
be.with(|be| {
|
|
|
|
let ra = ae.check_reliable(cur_ts);
|
|
|
|
let rb = be.check_reliable(cur_ts);
|
|
|
|
if ra != rb {
|
|
|
|
if ra {
|
|
|
|
return core::cmp::Ordering::Less;
|
|
|
|
} else {
|
|
|
|
return core::cmp::Ordering::Greater;
|
|
|
|
}
|
|
|
|
}
|
2021-11-22 16:28:30 +00:00
|
|
|
|
2022-06-25 14:57:33 +00:00
|
|
|
// latency is the next metric, closer nodes first
|
|
|
|
let a_latency = match ae.peer_stats().latency.as_ref() {
|
|
|
|
None => {
|
|
|
|
// treat unknown latency as slow
|
|
|
|
return core::cmp::Ordering::Greater;
|
|
|
|
}
|
|
|
|
Some(l) => l,
|
|
|
|
};
|
|
|
|
let b_latency = match be.peer_stats().latency.as_ref() {
|
|
|
|
None => {
|
|
|
|
// treat unknown latency as slow
|
|
|
|
return core::cmp::Ordering::Less;
|
|
|
|
}
|
|
|
|
Some(l) => l,
|
|
|
|
};
|
|
|
|
// Sort by average latency
|
|
|
|
a_latency.average.cmp(&b_latency.average)
|
|
|
|
})
|
|
|
|
})
|
2021-11-22 16:28:30 +00:00
|
|
|
},
|
|
|
|
// transform,
|
|
|
|
transform,
|
|
|
|
);
|
|
|
|
out
|
|
|
|
}
|
|
|
|
|
2022-06-25 14:57:33 +00:00
|
|
|
pub fn find_closest_nodes<F, T, O>(
|
2021-11-22 16:28:30 +00:00
|
|
|
&self,
|
|
|
|
node_id: DHTKey,
|
2022-06-25 14:57:33 +00:00
|
|
|
mut filter: Option<F>,
|
|
|
|
mut transform: T,
|
2021-11-22 16:28:30 +00:00
|
|
|
) -> Vec<O>
|
|
|
|
where
|
2022-06-25 14:57:33 +00:00
|
|
|
T: FnMut(DHTKey, Option<Arc<BucketEntry>>) -> O,
|
|
|
|
F: FnMut(DHTKey, Option<Arc<BucketEntry>>) -> bool,
|
2021-11-22 16:28:30 +00:00
|
|
|
{
|
2022-06-28 03:46:29 +00:00
|
|
|
let cur_ts = intf::get_timestamp();
|
2021-11-22 16:28:30 +00:00
|
|
|
let node_count = {
|
|
|
|
let c = self.config.get();
|
|
|
|
c.network.dht.max_find_node_count as usize
|
|
|
|
};
|
|
|
|
let out = self.find_peers_with_sort_and_filter(
|
|
|
|
node_count,
|
|
|
|
cur_ts,
|
|
|
|
// filter
|
2022-06-25 14:57:33 +00:00
|
|
|
|k, v| filter.as_mut().map(|f| f(k, v)).unwrap_or(true),
|
2021-11-22 16:28:30 +00:00
|
|
|
// sort
|
|
|
|
|(a_key, a_entry), (b_key, b_entry)| {
|
|
|
|
// same nodes are always the same
|
|
|
|
if a_key == b_key {
|
|
|
|
return core::cmp::Ordering::Equal;
|
|
|
|
}
|
2022-06-25 14:57:33 +00:00
|
|
|
|
2021-11-22 16:28:30 +00:00
|
|
|
// reliable nodes come first, pessimistically treating our own node as unreliable
|
2022-06-25 14:57:33 +00:00
|
|
|
let ra = a_entry
|
|
|
|
.as_ref()
|
|
|
|
.map_or(false, |x| x.with(|x| x.check_reliable(cur_ts)));
|
|
|
|
let rb = b_entry
|
|
|
|
.as_ref()
|
|
|
|
.map_or(false, |x| x.with(|x| x.check_reliable(cur_ts)));
|
2021-11-22 16:28:30 +00:00
|
|
|
if ra != rb {
|
|
|
|
if ra {
|
|
|
|
return core::cmp::Ordering::Less;
|
|
|
|
} else {
|
|
|
|
return core::cmp::Ordering::Greater;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// distance is the next metric, closer nodes first
|
|
|
|
let da = distance(a_key, &node_id);
|
|
|
|
let db = distance(b_key, &node_id);
|
|
|
|
da.cmp(&db)
|
|
|
|
},
|
|
|
|
// transform,
|
2022-06-25 14:57:33 +00:00
|
|
|
&mut transform,
|
2021-11-22 16:28:30 +00:00
|
|
|
);
|
2021-12-18 00:18:25 +00:00
|
|
|
log_rtab!(">> find_closest_nodes: node count = {}", out.len());
|
2021-11-22 16:28:30 +00:00
|
|
|
out
|
|
|
|
}
|
2022-06-25 14:57:33 +00:00
|
|
|
|
2022-08-18 20:21:13 +00:00
|
|
|
// Build a map of protocols to low level ports
|
|
|
|
// This way we can get the set of protocols required to keep our NAT mapping alive for keepalive pings
|
|
|
|
// Only one protocol per low level protocol/port combination is required
|
|
|
|
// For example, if WS/WSS and TCP protocols are on the same low-level TCP port, only TCP keepalives will be required
|
|
|
|
// and we do not need to do WS/WSS keepalive as well. If they are on different ports, then we will need WS/WSS keepalives too.
|
|
|
|
pub fn get_mapped_port_info(&self) -> MappedPortInfo {
|
|
|
|
let mut low_level_protocol_ports =
|
|
|
|
BTreeSet::<(LowLevelProtocolType, AddressType, u16)>::new();
|
|
|
|
let mut protocol_to_port =
|
|
|
|
BTreeMap::<(ProtocolType, AddressType), (LowLevelProtocolType, u16)>::new();
|
|
|
|
let our_dids = self.all_filtered_dial_info_details(
|
|
|
|
Some(RoutingDomain::PublicInternet),
|
|
|
|
&DialInfoFilter::all(),
|
|
|
|
);
|
|
|
|
for did in our_dids {
|
|
|
|
low_level_protocol_ports.insert((
|
|
|
|
did.dial_info.protocol_type().low_level_protocol_type(),
|
|
|
|
did.dial_info.address_type(),
|
|
|
|
did.dial_info.socket_address().port(),
|
|
|
|
));
|
|
|
|
protocol_to_port.insert(
|
|
|
|
(did.dial_info.protocol_type(), did.dial_info.address_type()),
|
|
|
|
(
|
|
|
|
did.dial_info.protocol_type().low_level_protocol_type(),
|
|
|
|
did.dial_info.socket_address().port(),
|
|
|
|
),
|
|
|
|
);
|
|
|
|
}
|
|
|
|
MappedPortInfo {
|
|
|
|
low_level_protocol_ports,
|
|
|
|
protocol_to_port,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-08-09 00:42:27 +00:00
|
|
|
fn make_relay_node_filter(&self) -> impl Fn(&BucketEntryInner) -> bool {
|
|
|
|
// Get all our outbound protocol/address types
|
|
|
|
let outbound_dif = self
|
|
|
|
.network_manager()
|
|
|
|
.get_outbound_dial_info_filter(RoutingDomain::PublicInternet);
|
2022-08-18 20:21:13 +00:00
|
|
|
let mapped_port_info = self.get_mapped_port_info();
|
2022-08-09 00:42:27 +00:00
|
|
|
|
|
|
|
move |e: &BucketEntryInner| {
|
|
|
|
// Ensure this node is not on our local network
|
|
|
|
let has_local_dial_info = e
|
|
|
|
.local_node_info()
|
|
|
|
.map(|l| l.has_dial_info())
|
|
|
|
.unwrap_or(false);
|
|
|
|
if has_local_dial_info {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2022-08-18 20:21:13 +00:00
|
|
|
// Disqualify nodes that don't cover all our inbound ports for tcp and udp
|
|
|
|
// as we need to be able to use the relay for keepalives for all nat mappings
|
|
|
|
let mut low_level_protocol_ports = mapped_port_info.low_level_protocol_ports.clone();
|
|
|
|
|
2022-08-09 00:42:27 +00:00
|
|
|
let can_serve_as_relay = e
|
|
|
|
.node_info()
|
|
|
|
.map(|n| {
|
|
|
|
let dids =
|
|
|
|
n.all_filtered_dial_info_details(|did| did.matches_filter(&outbound_dif));
|
2022-08-18 20:21:13 +00:00
|
|
|
for did in &dids {
|
|
|
|
let pt = did.dial_info.protocol_type();
|
|
|
|
let at = did.dial_info.address_type();
|
|
|
|
if let Some((llpt, port)) = mapped_port_info.protocol_to_port.get(&(pt, at))
|
|
|
|
{
|
|
|
|
low_level_protocol_ports.remove(&(*llpt, at, *port));
|
2022-08-09 00:42:27 +00:00
|
|
|
}
|
|
|
|
}
|
2022-08-18 20:21:13 +00:00
|
|
|
low_level_protocol_ports.is_empty()
|
2022-08-09 00:42:27 +00:00
|
|
|
})
|
|
|
|
.unwrap_or(false);
|
|
|
|
if !can_serve_as_relay {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
true
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-06-25 14:57:33 +00:00
|
|
|
#[instrument(level = "trace", skip(self), ret)]
|
|
|
|
pub fn find_inbound_relay(&self, cur_ts: u64) -> Option<NodeRef> {
|
2022-08-09 00:42:27 +00:00
|
|
|
// Get relay filter function
|
|
|
|
let relay_node_filter = self.make_relay_node_filter();
|
|
|
|
|
|
|
|
// Go through all entries and find fastest entry that matches filter function
|
2022-06-25 14:57:33 +00:00
|
|
|
let inner = self.inner.read();
|
|
|
|
let inner = &*inner;
|
|
|
|
let mut best_inbound_relay: Option<(DHTKey, Arc<BucketEntry>)> = None;
|
|
|
|
|
|
|
|
// Iterate all known nodes for candidates
|
2022-06-25 19:28:27 +00:00
|
|
|
Self::with_entries(inner, cur_ts, BucketEntryState::Unreliable, |k, v| {
|
2022-08-09 00:42:27 +00:00
|
|
|
let v2 = v.clone();
|
|
|
|
v.with(|e| {
|
|
|
|
// Ensure we have the node's status
|
|
|
|
if let Some(node_status) = e.peer_stats().status.clone() {
|
|
|
|
// Ensure the node will relay
|
|
|
|
if node_status.will_relay {
|
|
|
|
// Compare against previous candidate
|
|
|
|
if let Some(best_inbound_relay) = best_inbound_relay.as_mut() {
|
|
|
|
// Less is faster
|
|
|
|
let better = best_inbound_relay.1.with(|best| {
|
2022-06-25 14:57:33 +00:00
|
|
|
BucketEntryInner::cmp_fastest_reliable(cur_ts, e, best)
|
|
|
|
== std::cmp::Ordering::Less
|
2022-08-09 00:42:27 +00:00
|
|
|
});
|
|
|
|
// Now apply filter function and see if this node should be included
|
|
|
|
if better && relay_node_filter(e) {
|
|
|
|
*best_inbound_relay = (k, v2);
|
|
|
|
}
|
|
|
|
} else if relay_node_filter(e) {
|
|
|
|
// Always store the first candidate
|
|
|
|
best_inbound_relay = Some((k, v2));
|
2022-06-25 14:57:33 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2022-08-09 00:42:27 +00:00
|
|
|
});
|
|
|
|
// Don't end early, iterate through all entries
|
2022-06-25 14:57:33 +00:00
|
|
|
Option::<()>::None
|
|
|
|
});
|
|
|
|
// Return the best inbound relay noderef
|
|
|
|
best_inbound_relay.map(|(k, e)| NodeRef::new(self.clone(), k, e, None))
|
|
|
|
}
|
|
|
|
|
2022-07-20 13:39:38 +00:00
|
|
|
#[instrument(level = "trace", skip(self), ret)]
|
|
|
|
pub fn register_find_node_answer(&self, peers: Vec<PeerInfo>) -> Vec<NodeRef> {
|
2022-06-25 14:57:33 +00:00
|
|
|
let node_id = self.node_id();
|
|
|
|
|
|
|
|
// register nodes we'd found
|
2022-07-04 21:58:26 +00:00
|
|
|
let mut out = Vec::<NodeRef>::with_capacity(peers.len());
|
|
|
|
for p in peers {
|
2022-06-25 14:57:33 +00:00
|
|
|
// if our own node if is in the list then ignore it, as we don't add ourselves to our own routing table
|
|
|
|
if p.node_id.key == node_id {
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
2022-07-20 13:39:38 +00:00
|
|
|
// node can not be its own relay
|
|
|
|
if let Some(rpi) = &p.signed_node_info.node_info.relay_peer_info {
|
|
|
|
if rpi.node_id == p.node_id {
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-06-25 14:57:33 +00:00
|
|
|
// register the node if it's new
|
2022-07-20 13:39:38 +00:00
|
|
|
if let Some(nr) =
|
|
|
|
self.register_node_with_signed_node_info(p.node_id.key, p.signed_node_info.clone())
|
|
|
|
{
|
|
|
|
out.push(nr);
|
|
|
|
}
|
2022-06-25 14:57:33 +00:00
|
|
|
}
|
2022-07-20 13:39:38 +00:00
|
|
|
out
|
2022-06-25 14:57:33 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#[instrument(level = "trace", skip(self), ret, err)]
|
2022-07-20 13:39:38 +00:00
|
|
|
pub async fn find_node(
|
|
|
|
&self,
|
|
|
|
node_ref: NodeRef,
|
|
|
|
node_id: DHTKey,
|
|
|
|
) -> EyreResult<NetworkResult<Vec<NodeRef>>> {
|
2022-06-25 14:57:33 +00:00
|
|
|
let rpc_processor = self.rpc_processor();
|
|
|
|
|
2022-07-20 13:39:38 +00:00
|
|
|
let res = network_result_try!(
|
|
|
|
rpc_processor
|
|
|
|
.clone()
|
|
|
|
.rpc_call_find_node(
|
|
|
|
Destination::Direct(node_ref.clone()),
|
|
|
|
node_id,
|
|
|
|
None,
|
|
|
|
rpc_processor.make_respond_to_sender(node_ref.clone()),
|
|
|
|
)
|
|
|
|
.await?
|
|
|
|
);
|
2022-06-25 14:57:33 +00:00
|
|
|
|
|
|
|
// register nodes we'd found
|
2022-07-20 13:39:38 +00:00
|
|
|
Ok(NetworkResult::value(
|
|
|
|
self.register_find_node_answer(res.answer),
|
|
|
|
))
|
2022-06-25 14:57:33 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#[instrument(level = "trace", skip(self), ret, err)]
|
2022-07-20 13:39:38 +00:00
|
|
|
pub async fn find_self(&self, node_ref: NodeRef) -> EyreResult<NetworkResult<Vec<NodeRef>>> {
|
2022-06-25 14:57:33 +00:00
|
|
|
let node_id = self.node_id();
|
|
|
|
self.find_node(node_ref, node_id).await
|
|
|
|
}
|
|
|
|
|
|
|
|
#[instrument(level = "trace", skip(self), ret, err)]
|
2022-07-20 13:39:38 +00:00
|
|
|
pub async fn find_target(&self, node_ref: NodeRef) -> EyreResult<NetworkResult<Vec<NodeRef>>> {
|
2022-06-25 14:57:33 +00:00
|
|
|
let node_id = node_ref.node_id();
|
|
|
|
self.find_node(node_ref, node_id).await
|
|
|
|
}
|
|
|
|
|
|
|
|
#[instrument(level = "trace", skip(self))]
|
|
|
|
pub async fn reverse_find_node(&self, node_ref: NodeRef, wide: bool) {
|
|
|
|
// Ask bootstrap node to 'find' our own node so we can get some more nodes near ourselves
|
|
|
|
// and then contact those nodes to inform -them- that we exist
|
|
|
|
|
|
|
|
// Ask bootstrap server for nodes closest to our own node
|
2022-07-20 13:39:38 +00:00
|
|
|
let closest_nodes = network_result_value_or_log!(debug match self.find_self(node_ref.clone()).await {
|
2022-06-25 14:57:33 +00:00
|
|
|
Err(e) => {
|
|
|
|
log_rtab!(error
|
2022-07-20 13:39:38 +00:00
|
|
|
"find_self failed for {:?}: {:?}",
|
2022-06-25 14:57:33 +00:00
|
|
|
&node_ref, e
|
|
|
|
);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
Ok(v) => v,
|
2022-07-20 13:39:38 +00:00
|
|
|
} => {
|
|
|
|
return;
|
|
|
|
});
|
2022-06-25 14:57:33 +00:00
|
|
|
|
|
|
|
// Ask each node near us to find us as well
|
|
|
|
if wide {
|
|
|
|
for closest_nr in closest_nodes {
|
2022-07-20 13:39:38 +00:00
|
|
|
network_result_value_or_log!(debug match self.find_self(closest_nr.clone()).await {
|
|
|
|
Err(e) => {
|
|
|
|
log_rtab!(error
|
|
|
|
"find_self failed for {:?}: {:?}",
|
|
|
|
&closest_nr, e
|
|
|
|
);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
Ok(v) => v,
|
|
|
|
} => {
|
|
|
|
// Do nothing with non-values
|
|
|
|
continue;
|
|
|
|
});
|
2022-06-25 14:57:33 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2021-11-22 16:28:30 +00:00
|
|
|
}
|