use crate::{
behavior::{
FuelBehaviour,
FuelBehaviourEvent,
},
codecs::{
postcard::PostcardCodec,
GossipsubCodec,
},
config::{
build_transport_function,
Config,
},
dnsaddr_resolution::DnsResolver,
gossipsub::{
messages::{
GossipTopicTag,
GossipsubBroadcastRequest,
GossipsubMessage as FuelGossipsubMessage,
},
topics::GossipsubTopics,
},
heartbeat,
peer_manager::{
PeerManager,
Punisher,
},
peer_report::PeerReportEvent,
request_response::messages::{
RequestError,
RequestMessage,
ResponseError,
ResponseMessage,
ResponseSendError,
ResponseSender,
},
TryPeerId,
};
use fuel_core_metrics::{
global_registry,
p2p_metrics::increment_unique_peers,
};
use fuel_core_types::{
fuel_types::BlockHeight,
services::p2p::peer_reputation::AppScore,
};
use futures::prelude::*;
use libp2p::{
gossipsub::{
self,
MessageAcceptance,
MessageId,
PublishError,
TopicHash,
},
identify,
metrics::{
Metrics,
Recorder,
},
multiaddr::Protocol,
request_response::{
self,
InboundRequestId,
OutboundRequestId,
ResponseChannel,
},
swarm::SwarmEvent,
tcp,
Multiaddr,
PeerId,
Swarm,
SwarmBuilder,
};
use rand::seq::IteratorRandom;
use std::{
collections::HashMap,
time::Duration,
};
use tokio::sync::broadcast;
use tracing::{
debug,
warn,
};
const MAX_IDENTIFY_ADDRESSES: usize = 10;
impl Punisher for Swarm<FuelBehaviour> {
fn ban_peer(&mut self, peer_id: PeerId) {
self.behaviour_mut().block_peer(peer_id)
}
}
pub struct FuelP2PService {
pub local_peer_id: PeerId,
local_address: std::net::IpAddr,
tcp_port: u16,
swarm: Swarm<FuelBehaviour>,
outbound_requests_table: HashMap<OutboundRequestId, ResponseSender>,
inbound_requests_table: HashMap<InboundRequestId, ResponseChannel<ResponseMessage>>,
network_codec: PostcardCodec,
network_metadata: NetworkMetadata,
metrics: bool,
libp2p_metrics_registry: Option<Metrics>,
peer_manager: PeerManager,
}
#[derive(Debug)]
struct GossipsubData {
topics: GossipsubTopics,
}
impl GossipsubData {
pub fn with_topics(topics: GossipsubTopics) -> Self {
Self { topics }
}
}
#[derive(Debug)]
struct NetworkMetadata {
gossipsub_data: GossipsubData,
}
#[derive(Debug, Clone)]
#[allow(clippy::large_enum_variant)]
pub enum FuelP2PEvent {
GossipsubMessage {
peer_id: PeerId,
message_id: MessageId,
topic_hash: TopicHash,
message: FuelGossipsubMessage,
},
NewSubscription {
peer_id: PeerId,
tag: GossipTopicTag,
},
InboundRequestMessage {
request_id: InboundRequestId,
request_message: RequestMessage,
},
PeerConnected(PeerId),
PeerDisconnected(PeerId),
PeerInfoUpdated {
peer_id: PeerId,
block_height: BlockHeight,
},
}
async fn parse_multiaddrs(multiaddrs: Vec<Multiaddr>) -> anyhow::Result<Vec<Multiaddr>> {
let dnsaddr_urls = multiaddrs
.iter()
.filter_map(|multiaddr| {
if let Protocol::Dnsaddr(dnsaddr_url) = multiaddr.iter().next()? {
Some(dnsaddr_url.clone())
} else {
None
}
})
.collect::<Vec<_>>();
let dns_resolver = DnsResolver::new().await?;
let mut dnsaddr_multiaddrs = vec![];
for dnsaddr in dnsaddr_urls {
let multiaddrs = dns_resolver.lookup_dnsaddr(dnsaddr.as_ref()).await?;
dnsaddr_multiaddrs.extend(multiaddrs);
}
let resolved_multiaddrs = multiaddrs
.into_iter()
.filter(|multiaddr| !multiaddr.iter().any(|p| matches!(p, Protocol::Dnsaddr(_))))
.chain(dnsaddr_multiaddrs.into_iter())
.collect();
Ok(resolved_multiaddrs)
}
impl FuelP2PService {
pub async fn new(
reserved_peers_updates: broadcast::Sender<usize>,
config: Config,
codec: PostcardCodec,
) -> anyhow::Result<Self> {
let metrics = config.metrics;
let gossipsub_data =
GossipsubData::with_topics(GossipsubTopics::new(&config.network_name));
let network_metadata = NetworkMetadata { gossipsub_data };
let mut config = config;
config.bootstrap_nodes = parse_multiaddrs(config.bootstrap_nodes).await?;
config.reserved_nodes = parse_multiaddrs(config.reserved_nodes).await?;
let (transport_function, connection_state) = build_transport_function(&config);
let tcp_config = tcp::Config::new().port_reuse(true);
let behaviour = FuelBehaviour::new(&config, codec.clone())?;
let swarm_builder = SwarmBuilder::with_existing_identity(config.keypair.clone())
.with_tokio()
.with_tcp(
tcp_config,
transport_function,
libp2p::yamux::Config::default,
)
.map_err(|_| anyhow::anyhow!("Failed to build Swarm"))?
.with_dns()?;
let mut libp2p_metrics_registry = None;
let mut swarm = if metrics {
let mut registry = global_registry().registry.lock();
libp2p_metrics_registry = Some(Metrics::new(&mut registry));
swarm_builder
.with_bandwidth_metrics(&mut registry)
.with_behaviour(|_| behaviour)?
.with_swarm_config(|cfg| {
if let Some(timeout) = config.connection_idle_timeout {
cfg.with_idle_connection_timeout(timeout)
} else {
cfg
}
})
.build()
} else {
swarm_builder
.with_behaviour(|_| behaviour)?
.with_swarm_config(|cfg| {
if let Some(timeout) = config.connection_idle_timeout {
cfg.with_idle_connection_timeout(timeout)
} else {
cfg
}
})
.build()
};
let local_peer_id = swarm.local_peer_id().to_owned();
if let Some(public_address) = config.public_address.clone() {
swarm.add_external_address(public_address);
}
let reserved_peers = config
.reserved_nodes
.iter()
.filter_map(|m| m.try_to_peer_id())
.collect();
Ok(Self {
local_peer_id,
local_address: config.address,
tcp_port: config.tcp_port,
swarm,
network_codec: codec,
outbound_requests_table: HashMap::default(),
inbound_requests_table: HashMap::default(),
network_metadata,
metrics,
libp2p_metrics_registry,
peer_manager: PeerManager::new(
reserved_peers_updates,
reserved_peers,
connection_state,
config.max_peers_connected as usize,
),
})
}
pub async fn start(&mut self) -> anyhow::Result<()> {
let listen_multiaddr = {
let mut m = Multiaddr::from(self.local_address);
m.push(Protocol::Tcp(self.tcp_port));
m
};
let peer_id = self.local_peer_id;
tracing::info!(
"The p2p service starts on the `{listen_multiaddr}` with `{peer_id}`"
);
self.swarm.listen_on(listen_multiaddr)?;
tokio::time::timeout(Duration::from_secs(5), self.await_listeners_address())
.await
.map_err(|_| {
anyhow::anyhow!("P2PService should get a new address within 5 seconds")
})?;
Ok(())
}
async fn await_listeners_address(&mut self) {
loop {
if let SwarmEvent::NewListenAddr { .. } = self.swarm.select_next_some().await
{
break;
}
}
}
pub fn update_metrics<T>(&self, update_fn: T)
where
T: FnOnce(),
{
if self.metrics {
update_fn();
}
}
pub fn update_libp2p_metrics<E>(&self, event: &E)
where
Metrics: Recorder<E>,
{
if let Some(registry) = self.libp2p_metrics_registry.as_ref() {
self.update_metrics(|| registry.record(event));
}
}
#[cfg(feature = "test-helpers")]
pub fn multiaddrs(&self) -> Vec<Multiaddr> {
let local_peer = self.local_peer_id;
self.swarm
.listeners()
.map(|addr| {
format!("{addr}/p2p/{local_peer}")
.parse()
.expect("The format is always valid")
})
.collect()
}
pub fn get_peers_ids_iter(&self) -> impl Iterator<Item = &PeerId> {
self.peer_manager.get_peers_ids()
}
pub fn publish_message(
&mut self,
message: GossipsubBroadcastRequest,
) -> Result<MessageId, PublishError> {
let topic = self
.network_metadata
.gossipsub_data
.topics
.get_gossipsub_topic(&message);
match self.network_codec.encode(message) {
Ok(encoded_data) => self
.swarm
.behaviour_mut()
.publish_message(topic, encoded_data),
Err(e) => Err(PublishError::TransformFailed(e)),
}
}
pub fn send_request_msg(
&mut self,
peer_id: Option<PeerId>,
message_request: RequestMessage,
on_response: ResponseSender,
) -> Result<OutboundRequestId, RequestError> {
let peer_id = match peer_id {
Some(peer_id) => peer_id,
_ => {
let peers = self.get_peers_ids_iter();
let peers_count = self.peer_manager.total_peers_connected();
if peers_count == 0 {
return Err(RequestError::NoPeersConnected);
}
let mut range = rand::thread_rng();
*peers.choose(&mut range).unwrap()
}
};
let request_id = self
.swarm
.behaviour_mut()
.send_request_msg(message_request, &peer_id);
self.outbound_requests_table.insert(request_id, on_response);
Ok(request_id)
}
pub fn send_response_msg(
&mut self,
request_id: InboundRequestId,
message: ResponseMessage,
) -> Result<(), ResponseSendError> {
let Some(channel) = self.inbound_requests_table.remove(&request_id) else {
debug!("ResponseChannel for {:?} does not exist!", request_id);
return Err(ResponseSendError::ResponseChannelDoesNotExist);
};
if self
.swarm
.behaviour_mut()
.send_response_msg(channel, message)
.is_err()
{
debug!("Failed to send ResponseMessage for {:?}", request_id);
return Err(ResponseSendError::SendingResponseFailed);
}
Ok(())
}
pub fn update_block_height(&mut self, block_height: BlockHeight) {
self.swarm.behaviour_mut().update_block_height(block_height)
}
pub fn report_message_validation_result(
&mut self,
msg_id: &MessageId,
propagation_source: PeerId,
mut acceptance: MessageAcceptance,
) {
if let MessageAcceptance::Reject = acceptance {
if self.peer_manager.is_reserved(&propagation_source) {
acceptance = MessageAcceptance::Ignore;
}
}
if let Some(gossip_score) = self
.swarm
.behaviour_mut()
.report_message_validation_result(msg_id, &propagation_source, acceptance)
{
self.peer_manager.handle_gossip_score_update(
propagation_source,
gossip_score,
&mut self.swarm,
);
}
}
#[cfg(test)]
pub fn get_peer_score(&self, peer_id: &PeerId) -> Option<f64> {
self.swarm.behaviour().get_peer_score(peer_id)
}
pub fn report_peer(
&mut self,
peer_id: PeerId,
app_score: AppScore,
reporting_service: &str,
) {
self.peer_manager.update_app_score(
peer_id,
app_score,
reporting_service,
&mut self.swarm,
);
}
#[tracing::instrument(skip_all,
level = "debug",
fields(
local_peer_id = %self.local_peer_id,
local_address = %self.local_address
),
ret
)]
pub async fn next_event(&mut self) -> Option<FuelP2PEvent> {
let event = self.swarm.select_next_some().await;
tracing::debug!(?event);
match event {
SwarmEvent::Behaviour(fuel_behaviour) => {
self.handle_behaviour_event(fuel_behaviour)
}
SwarmEvent::NewListenAddr { address, .. } => {
tracing::info!("Listening for p2p traffic on `{address}`");
None
}
SwarmEvent::ListenerClosed {
addresses, reason, ..
} => {
tracing::info!(
"p2p listener(s) `{addresses:?}` closed with `{reason:?}`"
);
None
}
_ => {
self.update_libp2p_metrics(&event);
None
}
}
}
pub fn peer_manager(&self) -> &PeerManager {
&self.peer_manager
}
fn get_topic_tag(&self, topic_hash: &TopicHash) -> Option<GossipTopicTag> {
let topic = self
.network_metadata
.gossipsub_data
.topics
.get_gossipsub_tag(topic_hash);
if topic.is_none() {
warn!(target: "fuel-p2p", "GossipTopicTag does not exist for {:?}", &topic_hash);
}
topic
}
fn handle_behaviour_event(
&mut self,
event: FuelBehaviourEvent,
) -> Option<FuelP2PEvent> {
match event {
FuelBehaviourEvent::Gossipsub(event) => {
self.update_libp2p_metrics(&event);
self.handle_gossipsub_event(event)
}
FuelBehaviourEvent::PeerReport(event) => self.handle_peer_report_event(event),
FuelBehaviourEvent::RequestResponse(event) => {
self.handle_request_response_event(event)
}
FuelBehaviourEvent::Identify(event) => {
self.update_libp2p_metrics(&event);
self.handle_identify_event(event)
}
FuelBehaviourEvent::Heartbeat(event) => self.handle_heartbeat_event(event),
FuelBehaviourEvent::Discovery(event) => {
self.update_libp2p_metrics(&event);
None
}
_ => None,
}
}
fn handle_gossipsub_event(
&mut self,
event: gossipsub::Event,
) -> Option<FuelP2PEvent> {
match event {
gossipsub::Event::Message {
propagation_source,
message,
message_id,
} => {
let correct_topic = self.get_topic_tag(&message.topic)?;
match self.network_codec.decode(&message.data, correct_topic) {
Ok(decoded_message) => Some(FuelP2PEvent::GossipsubMessage {
peer_id: propagation_source,
message_id,
topic_hash: message.topic,
message: decoded_message,
}),
Err(err) => {
warn!(target: "fuel-p2p", "Failed to decode a message. ID: {}, Message: {:?} with error: {:?}", message_id, &message.data, err);
self.report_message_validation_result(
&message_id,
propagation_source,
MessageAcceptance::Reject,
);
None
}
}
}
gossipsub::Event::Subscribed { peer_id, topic } => {
let tag = self.get_topic_tag(&topic)?;
Some(FuelP2PEvent::NewSubscription { peer_id, tag })
}
_ => None,
}
}
fn handle_peer_report_event(
&mut self,
event: PeerReportEvent,
) -> Option<FuelP2PEvent> {
match event {
PeerReportEvent::PerformDecay => {
self.peer_manager.batch_update_score_with_decay()
}
PeerReportEvent::PeerConnected { peer_id } => {
if self.peer_manager.handle_peer_connected(&peer_id) {
let _ = self.swarm.disconnect_peer_id(peer_id);
} else {
return Some(FuelP2PEvent::PeerConnected(peer_id));
}
}
PeerReportEvent::PeerDisconnected { peer_id } => {
self.peer_manager.handle_peer_disconnect(peer_id);
return Some(FuelP2PEvent::PeerDisconnected(peer_id));
}
}
None
}
fn handle_request_response_event(
&mut self,
event: request_response::Event<RequestMessage, ResponseMessage>,
) -> Option<FuelP2PEvent> {
match event {
request_response::Event::Message { peer, message } => match message {
request_response::Message::Request {
request,
channel,
request_id,
} => {
self.inbound_requests_table.insert(request_id, channel);
return Some(FuelP2PEvent::InboundRequestMessage {
request_id,
request_message: request,
});
}
request_response::Message::Response {
request_id,
response,
} => {
let Some(channel) = self.outbound_requests_table.remove(&request_id)
else {
debug!("Send channel not found for {:?}", request_id);
return None;
};
let send_ok = match channel {
ResponseSender::SealedHeaders(c) => match response {
ResponseMessage::SealedHeaders(v) => {
c.send((peer, Ok(v))).is_ok()
}
_ => {
warn!(
"Invalid response type received for request {:?}",
request_id
);
c.send((peer, Err(ResponseError::TypeMismatch))).is_ok()
}
},
ResponseSender::Transactions(c) => match response {
ResponseMessage::Transactions(v) => {
c.send((peer, Ok(v))).is_ok()
}
_ => {
warn!(
"Invalid response type received for request {:?}",
request_id
);
c.send((peer, Err(ResponseError::TypeMismatch))).is_ok()
}
},
ResponseSender::TxPoolAllTransactionsIds(c) => match response {
ResponseMessage::TxPoolAllTransactionsIds(v) => {
c.send((peer, Ok(v))).is_ok()
}
_ => {
warn!(
"Invalid response type received for request {:?}",
request_id
);
c.send((peer, Err(ResponseError::TypeMismatch))).is_ok()
}
},
ResponseSender::TxPoolFullTransactions(c) => match response {
ResponseMessage::TxPoolFullTransactions(v) => {
c.send((peer, Ok(v))).is_ok()
}
_ => {
warn!(
"Invalid response type received for request {:?}",
request_id
);
c.send((peer, Err(ResponseError::TypeMismatch))).is_ok()
}
},
};
if !send_ok {
warn!("Failed to send through the channel for {:?}", request_id);
}
}
},
request_response::Event::InboundFailure {
peer,
error,
request_id,
} => {
tracing::error!("RequestResponse inbound error for peer: {:?} with id: {:?} and error: {:?}", peer, request_id, error);
let _ = self.inbound_requests_table.remove(&request_id);
}
request_response::Event::OutboundFailure {
peer,
error,
request_id,
} => {
tracing::error!("RequestResponse outbound error for peer: {:?} with id: {:?} and error: {:?}", peer, request_id, error);
if let Some(channel) = self.outbound_requests_table.remove(&request_id) {
match channel {
ResponseSender::SealedHeaders(c) => {
let _ = c.send((peer, Err(ResponseError::P2P(error))));
}
ResponseSender::Transactions(c) => {
let _ = c.send((peer, Err(ResponseError::P2P(error))));
}
ResponseSender::TxPoolAllTransactionsIds(c) => {
let _ = c.send((peer, Err(ResponseError::P2P(error))));
}
ResponseSender::TxPoolFullTransactions(c) => {
let _ = c.send((peer, Err(ResponseError::P2P(error))));
}
};
}
}
_ => {}
}
None
}
fn handle_identify_event(&mut self, event: identify::Event) -> Option<FuelP2PEvent> {
match event {
identify::Event::Received { peer_id, info } => {
self.update_metrics(increment_unique_peers);
let mut addresses = info.listen_addrs;
let agent_version = info.agent_version;
if addresses.len() > MAX_IDENTIFY_ADDRESSES {
let protocol_version = info.protocol_version;
debug!(
target: "fuel-p2p",
"Node {:?} has reported more than {} addresses; it is identified by {:?} and {:?}",
peer_id, MAX_IDENTIFY_ADDRESSES, protocol_version, agent_version
);
addresses.truncate(MAX_IDENTIFY_ADDRESSES);
}
self.peer_manager.handle_peer_identified(
&peer_id,
addresses.clone(),
agent_version,
);
self.swarm
.behaviour_mut()
.add_addresses_to_discovery(&peer_id, addresses);
}
identify::Event::Sent { .. } => {}
identify::Event::Pushed { .. } => {}
identify::Event::Error { peer_id, error } => {
debug!(target: "fuel-p2p", "Identification with peer {:?} failed => {}", peer_id, error);
}
}
None
}
fn handle_heartbeat_event(
&mut self,
event: heartbeat::Event,
) -> Option<FuelP2PEvent> {
let heartbeat::Event {
peer_id,
latest_block_height,
} = event;
self.peer_manager
.handle_peer_info_updated(&peer_id, latest_block_height);
Some(FuelP2PEvent::PeerInfoUpdated {
peer_id,
block_height: latest_block_height,
})
}
}
#[allow(clippy::cast_possible_truncation)]
#[cfg(test)]
mod tests {
use super::{
FuelP2PService,
PublishError,
};
use crate::{
codecs::postcard::PostcardCodec,
config::Config,
gossipsub::{
messages::{
GossipsubBroadcastRequest,
GossipsubMessage,
},
topics::{
GossipTopic,
NEW_TX_GOSSIP_TOPIC,
},
},
p2p_service::FuelP2PEvent,
peer_manager::PeerInfo,
request_response::messages::{
RequestMessage,
ResponseError,
ResponseMessage,
ResponseSender,
},
service::to_message_acceptance,
};
use fuel_core_types::{
blockchain::{
consensus::{
poa::PoAConsensus,
Consensus,
},
header::BlockHeader,
SealedBlockHeader,
},
fuel_tx::{
Transaction,
TransactionBuilder,
TxId,
UniqueIdentifier,
},
fuel_types::ChainId,
services::p2p::{
GossipsubMessageAcceptance,
NetworkableTransactionPool,
Transactions,
},
};
use futures::{
future::join_all,
StreamExt,
};
use libp2p::{
gossipsub::Topic,
identity::Keypair,
swarm::{
ListenError,
SwarmEvent,
},
Multiaddr,
PeerId,
};
use rand::Rng;
use std::{
collections::HashSet,
ops::Range,
sync::Arc,
time::Duration,
};
use tokio::sync::{
broadcast,
mpsc,
oneshot,
watch,
};
use tracing_attributes::instrument;
type P2PService = FuelP2PService;
async fn build_service_from_config(mut p2p_config: Config) -> P2PService {
p2p_config.keypair = Keypair::generate_secp256k1(); let max_block_size = p2p_config.max_block_size;
let (sender, _) =
broadcast::channel(p2p_config.reserved_nodes.len().saturating_add(1));
let mut service =
FuelP2PService::new(sender, p2p_config, PostcardCodec::new(max_block_size))
.await
.unwrap();
service.start().await.unwrap();
service
}
async fn setup_bootstrap_nodes(
p2p_config: &Config,
bootstrap_nodes_count: usize,
) -> (Vec<P2PService>, Vec<Multiaddr>) {
let nodes = join_all(
(0..bootstrap_nodes_count)
.map(|_| build_service_from_config(p2p_config.clone())),
)
.await;
let bootstrap_multiaddrs = nodes
.iter()
.flat_map(|b| b.multiaddrs())
.collect::<Vec<_>>();
(nodes, bootstrap_multiaddrs)
}
fn spawn(stop: &watch::Sender<()>, mut node: P2PService) {
let mut stop = stop.subscribe();
tokio::spawn(async move {
loop {
tokio::select! {
_ = node.next_event() => {}
_ = stop.changed() => {
break;
}
}
}
});
}
#[tokio::test]
#[instrument]
async fn p2p_service_works() {
build_service_from_config(Config::default_initialized("p2p_service_works")).await;
}
#[tokio::test(flavor = "multi_thread")]
#[instrument]
async fn reserved_nodes_reconnect_works() {
let p2p_config = Config::default_initialized("reserved_nodes_reconnect_works");
let max_peers_allowed: usize = 3;
let (bootstrap_nodes, bootstrap_multiaddrs) =
setup_bootstrap_nodes(&p2p_config, max_peers_allowed.saturating_mul(5)).await;
let (mut reserved_nodes, reserved_multiaddrs) =
setup_bootstrap_nodes(&p2p_config, max_peers_allowed).await;
let mut sentry_node = {
let mut p2p_config = p2p_config.clone();
p2p_config.max_peers_connected = max_peers_allowed as u32;
p2p_config.bootstrap_nodes = bootstrap_multiaddrs;
p2p_config.reserved_nodes = reserved_multiaddrs;
build_service_from_config(p2p_config).await
};
let mut reserved_node = reserved_nodes.pop();
let reserved_node_peer_id = reserved_node.as_ref().unwrap().local_peer_id;
let all_node_services: Vec<_> = bootstrap_nodes
.into_iter()
.chain(reserved_nodes.into_iter())
.collect();
let mut all_nodes_ids: Vec<PeerId> = all_node_services
.iter()
.map(|service| service.local_peer_id)
.collect();
let (stop_sender, _) = watch::channel(());
all_node_services.into_iter().for_each(|node| {
spawn(&stop_sender, node);
});
loop {
tokio::select! {
sentry_node_event = sentry_node.next_event() => {
if sentry_node.peer_manager.total_peers_connected() > max_peers_allowed {
if !all_nodes_ids
.iter()
.any(|local_peer_id| local_peer_id == &reserved_node_peer_id) {
if let Some(node) = reserved_node {
all_nodes_ids.push(node.local_peer_id);
spawn(&stop_sender, node);
reserved_node = None;
}
}
}
if let Some(FuelP2PEvent::PeerConnected(peer_id)) = sentry_node_event {
if peer_id == reserved_node_peer_id {
break
}
}
},
}
}
stop_sender.send(()).unwrap();
}
#[tokio::test]
#[instrument]
async fn max_peers_connected_works() {
let p2p_config = Config::default_initialized("max_peers_connected_works");
let bootstrap_nodes_count = 20;
let node_a_max_peers_allowed: usize = 3;
let node_b_max_peers_allowed: usize = 5;
let (mut nodes, nodes_multiaddrs) =
setup_bootstrap_nodes(&p2p_config, bootstrap_nodes_count).await;
let mut node_a = {
let mut p2p_config = p2p_config.clone();
p2p_config.max_peers_connected = node_a_max_peers_allowed as u32;
p2p_config.bootstrap_nodes.clone_from(&nodes_multiaddrs);
build_service_from_config(p2p_config).await
};
let mut node_b = {
let mut p2p_config = p2p_config.clone();
p2p_config.max_peers_connected = node_b_max_peers_allowed as u32;
p2p_config.bootstrap_nodes.clone_from(&nodes_multiaddrs);
build_service_from_config(p2p_config).await
};
let (tx, mut rx) = tokio::sync::oneshot::channel::<()>();
let jh = tokio::spawn(async move {
while rx.try_recv().is_err() {
futures::stream::iter(nodes.iter_mut())
.for_each_concurrent(4, |node| async move {
node.next_event().await;
})
.await;
}
});
let mut node_a_hit_limit = false;
let mut node_b_hit_limit = false;
let mut instance = tokio::time::Instant::now();
while instance.elapsed().as_secs() < 5 {
tokio::select! {
event_from_node_a = node_a.next_event() => {
if let Some(FuelP2PEvent::PeerConnected(_)) = event_from_node_a {
if node_a.peer_manager().total_peers_connected() > node_a_max_peers_allowed {
panic!("The node should only connect to max {node_a_max_peers_allowed} peers");
}
node_a_hit_limit |= node_a.peer_manager().total_peers_connected() == node_a_max_peers_allowed;
}
tracing::info!("Event from the node_a: {:?}", event_from_node_a);
},
event_from_node_b = node_b.next_event() => {
if let Some(FuelP2PEvent::PeerConnected(_)) = event_from_node_b {
if node_b.peer_manager().total_peers_connected() > node_b_max_peers_allowed {
panic!("The node should only connect to max {node_b_max_peers_allowed} peers");
}
node_b_hit_limit |= node_b.peer_manager().total_peers_connected() == node_b_max_peers_allowed;
}
tracing::info!("Event from the node_b: {:?}", event_from_node_b);
},
}
if !(node_a_hit_limit && node_b_hit_limit) {
instance = tokio::time::Instant::now();
}
}
tx.send(()).unwrap();
jh.await.unwrap()
}
#[tokio::test(flavor = "multi_thread")]
#[instrument]
async fn sentry_nodes_working() {
const RESERVED_NODE_SIZE: usize = 4;
let mut p2p_config = Config::default_initialized("sentry_nodes_working");
async fn build_sentry_nodes(p2p_config: Config) -> (P2PService, Vec<P2PService>) {
let (reserved_nodes, reserved_multiaddrs) =
setup_bootstrap_nodes(&p2p_config, RESERVED_NODE_SIZE).await;
let guarded_node_service = {
let mut p2p_config = p2p_config.clone();
p2p_config.reserved_nodes = reserved_multiaddrs;
p2p_config.reserved_nodes_only_mode = true;
build_service_from_config(p2p_config).await
};
let sentry_nodes = reserved_nodes;
(guarded_node_service, sentry_nodes)
}
let (mut first_guarded_node, mut first_sentry_nodes) =
build_sentry_nodes(p2p_config.clone()).await;
p2p_config.bootstrap_nodes = first_sentry_nodes
.iter()
.flat_map(|n| n.multiaddrs())
.collect();
let (mut second_guarded_node, second_sentry_nodes) =
build_sentry_nodes(p2p_config).await;
let first_sentry_set: HashSet<_> = first_sentry_nodes
.iter()
.map(|node| node.local_peer_id)
.collect();
let second_sentry_set: HashSet<_> = second_sentry_nodes
.iter()
.map(|node| node.local_peer_id)
.collect();
let mut single_sentry_node = first_sentry_nodes.pop().unwrap();
let mut sentry_node_connections = HashSet::new();
let (stop_sender, _) = watch::channel(());
first_sentry_nodes
.into_iter()
.chain(second_sentry_nodes.into_iter())
.for_each(|node| {
spawn(&stop_sender, node);
});
let mut instance = tokio::time::Instant::now();
while instance.elapsed().as_secs() < 5 {
tokio::select! {
event_from_first_guarded = first_guarded_node.next_event() => {
if let Some(FuelP2PEvent::PeerConnected(peer_id)) = event_from_first_guarded {
if !first_sentry_set.contains(&peer_id) {
panic!("The node should only connect to the specified reserved nodes!");
}
}
tracing::info!("Event from the first guarded node: {:?}", event_from_first_guarded);
},
event_from_second_guarded = second_guarded_node.next_event() => {
if let Some(FuelP2PEvent::PeerConnected(peer_id)) = event_from_second_guarded {
if !second_sentry_set.contains(&peer_id) {
panic!("The node should only connect to the specified reserved nodes!");
}
}
tracing::info!("Event from the second guarded node: {:?}", event_from_second_guarded);
},
sentry_node_event = single_sentry_node.next_event() => {
if let Some(FuelP2PEvent::PeerConnected(peer_id)) = sentry_node_event {
sentry_node_connections.insert(peer_id);
}
}
};
if sentry_node_connections.len() < 2 * RESERVED_NODE_SIZE {
instance = tokio::time::Instant::now();
}
}
stop_sender.send(()).unwrap();
}
#[tokio::test]
#[instrument]
async fn nodes_connected_via_mdns() {
let mut p2p_config = Config::default_initialized("nodes_connected_via_mdns");
p2p_config.enable_mdns = true;
let mut node_a = build_service_from_config(p2p_config.clone()).await;
let mut node_b = build_service_from_config(p2p_config).await;
loop {
tokio::select! {
node_b_event = node_b.next_event() => {
if let Some(FuelP2PEvent::PeerConnected(_)) = node_b_event {
break
}
tracing::info!("Node B Event: {:?}", node_b_event);
},
_ = node_a.swarm.select_next_some() => {},
};
}
}
#[tokio::test]
#[instrument]
async fn nodes_cannot_connect_due_to_different_checksum() {
use libp2p::TransportError;
let mut p2p_config =
Config::default_initialized("nodes_cannot_connect_due_to_different_checksum");
let mut node_a = build_service_from_config(p2p_config.clone()).await;
p2p_config.checksum = [1u8; 32].into();
p2p_config.bootstrap_nodes = node_a.multiaddrs();
let mut node_b = build_service_from_config(p2p_config).await;
loop {
tokio::select! {
node_a_event = node_a.swarm.select_next_some() => {
tracing::info!("Node A Event: {:?}", node_a_event);
if let SwarmEvent::IncomingConnectionError { error: ListenError::Transport(TransportError::Other(_)), .. } = node_a_event {
break
}
},
node_b_event = node_b.next_event() => {
if let Some(FuelP2PEvent::PeerConnected(_)) = node_b_event {
panic!("Node B should not connect to Node A!")
}
tracing::info!("Node B Event: {:?}", node_b_event);
},
};
}
}
#[tokio::test]
#[instrument]
async fn nodes_connected_via_identify() {
let mut p2p_config = Config::default_initialized("nodes_connected_via_identify");
let mut node_a = build_service_from_config(p2p_config.clone()).await;
p2p_config.bootstrap_nodes = node_a.multiaddrs();
let mut node_b = build_service_from_config(p2p_config.clone()).await;
let mut node_c = build_service_from_config(p2p_config).await;
loop {
tokio::select! {
node_a_event = node_a.next_event() => {
tracing::info!("Node A Event: {:?}", node_a_event);
},
node_b_event = node_b.next_event() => {
tracing::info!("Node B Event: {:?}", node_b_event);
},
node_c_event = node_c.next_event() => {
if let Some(FuelP2PEvent::PeerConnected(peer_id)) = node_c_event {
if peer_id == node_b.local_peer_id {
break
}
}
tracing::info!("Node C Event: {:?}", node_c_event);
}
};
}
}
#[tokio::test]
#[instrument]
async fn peer_info_updates_work() {
let mut p2p_config = Config::default_initialized("peer_info_updates_work");
let mut node_a = build_service_from_config(p2p_config.clone()).await;
p2p_config.bootstrap_nodes = node_a.multiaddrs();
let mut node_b = build_service_from_config(p2p_config).await;
let latest_block_height = 40_u32.into();
loop {
tokio::select! {
node_a_event = node_a.next_event() => {
if let Some(FuelP2PEvent::PeerInfoUpdated { peer_id, block_height: _ }) = node_a_event {
if let Some(PeerInfo { heartbeat_data, client_version, .. }) = node_a.peer_manager.get_peer_info(&peer_id) {
if client_version.is_some() && heartbeat_data.block_height == Some(latest_block_height) {
break;
}
}
}
tracing::info!("Node A Event: {:?}", node_a_event);
},
node_b_event = node_b.next_event() => {
if let Some(FuelP2PEvent::PeerConnected(_)) = node_b_event {
node_b.update_block_height(latest_block_height);
}
tracing::info!("Node B Event: {:?}", node_b_event);
}
}
}
}
#[tokio::test]
#[instrument]
async fn gossipsub_broadcast_tx_with_accept() {
for _ in 0..100 {
tokio::time::timeout(
Duration::from_secs(5),
gossipsub_broadcast(
GossipsubBroadcastRequest::NewTx(Arc::new(
Transaction::default_test_tx(),
)),
GossipsubMessageAcceptance::Accept,
),
)
.await
.unwrap();
}
}
#[tokio::test]
#[instrument]
async fn gossipsub_broadcast_tx_with_reject() {
for _ in 0..100 {
tokio::time::timeout(
Duration::from_secs(5),
gossipsub_broadcast(
GossipsubBroadcastRequest::NewTx(Arc::new(
Transaction::default_test_tx(),
)),
GossipsubMessageAcceptance::Reject,
),
)
.await
.unwrap();
}
}
#[tokio::test]
#[instrument]
#[ignore]
async fn gossipsub_scoring_with_accepted_messages() {
gossipsub_scoring_tester(
"gossipsub_scoring_with_accepted_messages",
100,
GossipsubMessageAcceptance::Accept,
)
.await;
}
#[tokio::test]
#[instrument]
#[ignore]
async fn gossipsub_scoring_with_rejected_messages() {
gossipsub_scoring_tester(
"gossipsub_scoring_with_rejected_messages",
100,
GossipsubMessageAcceptance::Reject,
)
.await;
}
async fn gossipsub_scoring_tester(
test_name: &str,
amount_of_msgs_per_second: usize,
acceptance: GossipsubMessageAcceptance,
) {
let mut p2p_config = Config::default_initialized(test_name);
let mut node_a = build_service_from_config(p2p_config.clone()).await;
p2p_config.bootstrap_nodes = node_a.multiaddrs();
let mut node_b = build_service_from_config(p2p_config.clone()).await;
p2p_config.bootstrap_nodes = node_b.multiaddrs();
let mut node_c = build_service_from_config(p2p_config.clone()).await;
let mut interval = tokio::time::interval(Duration::from_secs(1));
loop {
tokio::select! {
node_a_event = node_a.next_event() => {
if let Some(FuelP2PEvent::GossipsubMessage { message_id, peer_id, .. }) = node_a_event {
let msg_acceptance = to_message_acceptance(&acceptance);
node_a.report_message_validation_result(&message_id, peer_id, msg_acceptance);
}
}
node_b_event = node_b.next_event() => {
if let Some(FuelP2PEvent::GossipsubMessage { message_id, peer_id, .. }) = node_b_event {
let msg_acceptance = to_message_acceptance(&acceptance);
node_b.report_message_validation_result(&message_id, peer_id, msg_acceptance);
}
},
node_c_event = node_c.next_event() => {
if let Some(FuelP2PEvent::GossipsubMessage { message_id, peer_id, .. }) = node_c_event {
let msg_acceptance = to_message_acceptance(&acceptance);
node_c.report_message_validation_result(&message_id, peer_id, msg_acceptance);
}
},
_ = interval.tick() => {
let mut transactions = vec![];
for _ in 0..amount_of_msgs_per_second {
let random_tx =
TransactionBuilder::script(rand::thread_rng().gen::<[u8; 32]>().to_vec(), rand::thread_rng().gen::<[u8; 32]>().to_vec()).finalize_as_transaction();
transactions.push(random_tx.clone());
let random_tx = GossipsubBroadcastRequest::NewTx(Arc::new(random_tx));
match rand::thread_rng().gen_range(1..=3) {
1 => {
let _ = node_a.publish_message(random_tx);
},
2 => {
let _ = node_b.publish_message(random_tx);
},
3 => {
let _ = node_c.publish_message(random_tx);
},
_ => unreachable!("Random number generator is broken")
}
}
eprintln!("Node A WORLD VIEW");
eprintln!("B score: {:?}", node_a.get_peer_score(&node_b.local_peer_id).unwrap());
eprintln!("C score: {:?}", node_a.get_peer_score(&node_c.local_peer_id).unwrap());
eprintln!();
eprintln!("Node B WORLD VIEW");
eprintln!("A score: {:?}", node_b.get_peer_score(&node_a.local_peer_id).unwrap());
eprintln!("C score: {:?}", node_b.get_peer_score(&node_c.local_peer_id).unwrap());
eprintln!();
eprintln!("Node C WORLD VIEW");
eprintln!("A score: {:?}", node_c.get_peer_score(&node_a.local_peer_id).unwrap());
eprintln!("B score: {:?}", node_c.get_peer_score(&node_b.local_peer_id).unwrap());
eprintln!();
}
}
}
}
async fn gossipsub_broadcast(
broadcast_request: GossipsubBroadcastRequest,
acceptance: GossipsubMessageAcceptance,
) {
let mut p2p_config = Config::default_initialized("gossipsub_exchanges_messages");
let selected_topic: GossipTopic = {
let topic = match broadcast_request {
GossipsubBroadcastRequest::NewTx(_) => NEW_TX_GOSSIP_TOPIC,
};
Topic::new(format!("{}/{}", topic, p2p_config.network_name))
};
let mut message_sent = false;
let mut node_a = build_service_from_config(p2p_config.clone()).await;
p2p_config.bootstrap_nodes = node_a.multiaddrs();
let mut node_b = build_service_from_config(p2p_config.clone()).await;
p2p_config.bootstrap_nodes = node_b.multiaddrs();
let mut node_c = build_service_from_config(p2p_config.clone()).await;
node_c
.swarm
.behaviour_mut()
.block_peer(node_a.local_peer_id);
let mut a_connected_to_b = false;
let mut b_connected_to_c = false;
loop {
if a_connected_to_b && b_connected_to_c && !message_sent {
message_sent = true;
let broadcast_request = broadcast_request.clone();
node_a.publish_message(broadcast_request).unwrap();
}
tokio::select! {
node_a_event = node_a.next_event() => {
if let Some(FuelP2PEvent::NewSubscription { peer_id, .. }) = &node_a_event {
if peer_id == &node_b.local_peer_id {
a_connected_to_b = true;
}
}
tracing::info!("Node A Event: {:?}", node_a_event);
},
node_b_event = node_b.next_event() => {
if let Some(FuelP2PEvent::NewSubscription { peer_id, .. }) = &node_b_event {
if peer_id == &node_c.local_peer_id {
b_connected_to_c = true;
}
}
if let Some(FuelP2PEvent::GossipsubMessage { topic_hash, message, message_id, peer_id }) = node_b_event.clone() {
let msg_acceptance = to_message_acceptance(&acceptance);
node_b.report_message_validation_result(&message_id, peer_id, msg_acceptance);
if topic_hash != selected_topic.hash() {
tracing::error!("Wrong topic hash, expected: {} - actual: {}", selected_topic.hash(), topic_hash);
panic!("Wrong Topic");
}
match &message {
GossipsubMessage::NewTx(tx) => {
if tx != &Transaction::default_test_tx() {
tracing::error!("Wrong p2p message {:?}", message);
panic!("Wrong GossipsubMessage")
}
}
}
let broadcast_request = broadcast_request.clone();
matches!(node_b.publish_message(broadcast_request), Err(PublishError::Duplicate));
match acceptance {
GossipsubMessageAcceptance::Reject | GossipsubMessageAcceptance::Ignore => {
break
},
_ => {
}
}
}
tracing::info!("Node B Event: {:?}", node_b_event);
}
node_c_event = node_c.next_event() => {
if let Some(FuelP2PEvent::GossipsubMessage { peer_id, .. }) = node_c_event.clone() {
assert!(peer_id == node_b.local_peer_id);
match acceptance {
GossipsubMessageAcceptance::Reject | GossipsubMessageAcceptance::Ignore => {
panic!("Node C should not receive Rejected or Ignored messages")
},
GossipsubMessageAcceptance::Accept => {
break
}
}
}
}
};
}
}
fn arbitrary_headers_for_range(range: Range<u32>) -> Vec<SealedBlockHeader> {
let mut blocks = Vec::new();
for i in range {
let mut header: BlockHeader = Default::default();
header.set_block_height(i.into());
let sealed_block = SealedBlockHeader {
entity: header,
consensus: Consensus::PoA(PoAConsensus::new(Default::default())),
};
blocks.push(sealed_block);
}
blocks
}
fn eq_except_metadata(a: &SealedBlockHeader, b: &SealedBlockHeader) -> bool {
a.entity.application() == b.entity.application()
&& a.entity.consensus() == b.entity.consensus()
}
async fn request_response_works_with(request_msg: RequestMessage) {
let mut p2p_config = Config::default_initialized("request_response_works_with");
let mut node_a = build_service_from_config(p2p_config.clone()).await;
p2p_config.bootstrap_nodes = node_a.multiaddrs();
let mut node_b = build_service_from_config(p2p_config.clone()).await;
let (tx_test_end, mut rx_test_end) = mpsc::channel::<bool>(1);
let mut request_sent = false;
loop {
tokio::select! {
message_sent = rx_test_end.recv() => {
assert!(message_sent.unwrap(), "Received incorrect or missing message");
break;
}
node_a_event = node_a.next_event() => {
if let Some(FuelP2PEvent::PeerInfoUpdated { peer_id, block_height: _ }) = node_a_event {
if node_a.peer_manager.get_peer_info(&peer_id).is_some() {
if !request_sent {
request_sent = true;
match request_msg.clone() {
RequestMessage::SealedHeaders(range) => {
let (tx_orchestrator, rx_orchestrator) = oneshot::channel();
assert!(node_a.send_request_msg(None, request_msg.clone(), ResponseSender::SealedHeaders(tx_orchestrator)).is_ok());
let tx_test_end = tx_test_end.clone();
tokio::spawn(async move {
let response_message = rx_orchestrator.await;
let expected = arbitrary_headers_for_range(range.clone());
if let Ok((_, Ok(sealed_headers))) = response_message {
let check = expected.iter().zip(sealed_headers.unwrap().iter()).all(|(a, b)| eq_except_metadata(a, b));
let _ = tx_test_end.send(check).await;
} else {
tracing::error!("Orchestrator failed to receive a message: {:?}", response_message);
let _ = tx_test_end.send(false).await;
}
});
}
RequestMessage::Transactions(_range) => {
let (tx_orchestrator, rx_orchestrator) = oneshot::channel();
assert!(node_a.send_request_msg(None, request_msg.clone(), ResponseSender::Transactions(tx_orchestrator)).is_ok());
let tx_test_end = tx_test_end.clone();
tokio::spawn(async move {
let response_message = rx_orchestrator.await;
if let Ok((_, Ok(Some(transactions)))) = response_message {
let check = transactions.len() == 1 && transactions[0].0.len() == 5;
let _ = tx_test_end.send(check).await;
} else {
tracing::error!("Orchestrator failed to receive a message: {:?}", response_message);
let _ = tx_test_end.send(false).await;
}
});
}
RequestMessage::TxPoolAllTransactionsIds => {
let (tx_orchestrator, rx_orchestrator) = oneshot::channel();
assert!(node_a.send_request_msg(None, request_msg.clone(), ResponseSender::TxPoolAllTransactionsIds(tx_orchestrator)).is_ok());
let tx_test_end = tx_test_end.clone();
tokio::spawn(async move {
let response_message = rx_orchestrator.await;
if let Ok((_, Ok(Some(transaction_ids)))) = response_message {
let tx_ids: Vec<TxId> = (0..5).map(|_| Transaction::default_test_tx().id(&ChainId::new(1))).collect();
let check = transaction_ids.len() == 5 && transaction_ids.iter().zip(tx_ids.iter()).all(|(a, b)| a == b);
let _ = tx_test_end.send(check).await;
} else {
tracing::error!("Orchestrator failed to receive a message: {:?}", response_message);
let _ = tx_test_end.send(false).await;
}
});
}
RequestMessage::TxPoolFullTransactions(tx_ids) => {
let (tx_orchestrator, rx_orchestrator) = oneshot::channel();
assert!(node_a.send_request_msg(None, request_msg.clone(), ResponseSender::TxPoolFullTransactions(tx_orchestrator)).is_ok());
let tx_test_end = tx_test_end.clone();
tokio::spawn(async move {
let response_message = rx_orchestrator.await;
if let Ok((_, Ok(Some(transactions)))) = response_message {
let txs: Vec<Option<NetworkableTransactionPool>> = tx_ids.iter().enumerate().map(|(i, _)| {
if i == 0 {
None
} else {
Some(NetworkableTransactionPool::Transaction(Transaction::default_test_tx()))
}
}).collect();
let check = transactions.len() == tx_ids.len() && transactions.iter().zip(txs.iter()).all(|(a, b)| a == b);
let _ = tx_test_end.send(check).await;
} else {
tracing::error!("Orchestrator failed to receive a message: {:?}", response_message);
let _ = tx_test_end.send(false).await;
}
});
}
}
}
}
}
tracing::info!("Node A Event: {:?}", node_a_event);
},
node_b_event = node_b.next_event() => {
if let Some(FuelP2PEvent::InboundRequestMessage{ request_id, request_message: received_request_message }) = &node_b_event {
match received_request_message {
RequestMessage::SealedHeaders(range) => {
let sealed_headers: Vec<_> = arbitrary_headers_for_range(range.clone());
let _ = node_b.send_response_msg(*request_id, ResponseMessage::SealedHeaders(Some(sealed_headers)));
}
RequestMessage::Transactions(_) => {
let txs = (0..5).map(|_| Transaction::default_test_tx()).collect();
let transactions = vec![Transactions(txs)];
let _ = node_b.send_response_msg(*request_id, ResponseMessage::Transactions(Some(transactions)));
}
RequestMessage::TxPoolAllTransactionsIds => {
let tx_ids = (0..5).map(|_| Transaction::default_test_tx().id(&ChainId::new(1))).collect();
let _ = node_b.send_response_msg(*request_id, ResponseMessage::TxPoolAllTransactionsIds(Some(tx_ids)));
}
RequestMessage::TxPoolFullTransactions(tx_ids) => {
let txs = tx_ids.iter().enumerate().map(|(i, _)| {
if i == 0 {
None
} else {
Some(NetworkableTransactionPool::Transaction(Transaction::default_test_tx()))
}
}).collect();
let _ = node_b.send_response_msg(*request_id, ResponseMessage::TxPoolFullTransactions(Some(txs)));
}
}
}
tracing::info!("Node B Event: {:?}", node_b_event);
}
};
}
}
#[tokio::test]
#[instrument]
async fn request_response_works_with_transactions() {
let arbitrary_range = 2..6;
request_response_works_with(RequestMessage::Transactions(arbitrary_range)).await
}
#[tokio::test]
#[instrument]
async fn request_response_works_with_sealed_headers_range_inclusive() {
let arbitrary_range = 2..6;
request_response_works_with(RequestMessage::SealedHeaders(arbitrary_range)).await
}
#[tokio::test]
#[instrument]
async fn request_response_works_with_transactions_ids() {
request_response_works_with(RequestMessage::TxPoolAllTransactionsIds).await
}
#[tokio::test]
#[instrument]
async fn request_response_works_with_full_transactions() {
let tx_ids = (0..10)
.map(|_| Transaction::default_test_tx().id(&ChainId::new(1)))
.collect();
request_response_works_with(RequestMessage::TxPoolFullTransactions(tx_ids)).await
}
#[tokio::test]
#[instrument]
async fn invalid_response_type_is_detected() {
let mut p2p_config =
Config::default_initialized("invalid_response_type_is_detected");
let mut node_a = build_service_from_config(p2p_config.clone()).await;
p2p_config.bootstrap_nodes = node_a.multiaddrs();
let mut node_b = build_service_from_config(p2p_config.clone()).await;
let (tx_test_end, mut rx_test_end) = mpsc::channel::<bool>(1);
let mut request_sent = false;
loop {
tokio::select! {
message_sent = rx_test_end.recv() => {
assert!(message_sent.unwrap(), "Received incorrect or missing message");
break;
}
node_a_event = node_a.next_event() => {
if let Some(FuelP2PEvent::PeerInfoUpdated { peer_id, block_height: _ }) = node_a_event {
if node_a.peer_manager.get_peer_info(&peer_id).is_some() {
if !request_sent {
request_sent = true;
let (tx_orchestrator, rx_orchestrator) = oneshot::channel();
assert!(node_a.send_request_msg(None, RequestMessage::Transactions(0..2), ResponseSender::Transactions(tx_orchestrator)).is_ok());
let tx_test_end = tx_test_end.clone();
tokio::spawn(async move {
let response_message = rx_orchestrator.await;
match response_message {
Ok((_, Ok(_))) => {
let _ = tx_test_end.send(false).await;
panic!("Request succeeded unexpectedly");
},
Ok((_, Err(ResponseError::TypeMismatch))) => {
let _ = tx_test_end.send(true).await;
},
Ok((_, Err(err))) => {
let _ = tx_test_end.send(false).await;
panic!("Unexpected error: {:?}", err);
},
Err(_) => {
let _ = tx_test_end.send(false).await;
panic!("Channel closed unexpectedly");
},
}
});
}
}
}
tracing::info!("Node A Event: {:?}", node_a_event);
},
node_b_event = node_b.next_event() => {
if let Some(FuelP2PEvent::InboundRequestMessage{ request_id, request_message: _ }) = &node_b_event {
let sealed_headers: Vec<_> = arbitrary_headers_for_range(1..3);
let _ = node_b.send_response_msg(*request_id, ResponseMessage::SealedHeaders(Some(sealed_headers)));
}
tracing::info!("Node B Event: {:?}", node_b_event);
}
};
}
}
#[tokio::test]
#[instrument]
async fn req_res_outbound_timeout_works() {
let mut p2p_config =
Config::default_initialized("req_res_outbound_timeout_works");
p2p_config.set_request_timeout = Duration::from_millis(1);
let mut node_a = build_service_from_config(p2p_config.clone()).await;
p2p_config.bootstrap_nodes = node_a.multiaddrs();
p2p_config.set_request_timeout = Duration::from_secs(20);
let mut node_b = build_service_from_config(p2p_config.clone()).await;
let (tx_test_end, mut rx_test_end) = tokio::sync::mpsc::channel(1);
let mut request_sent = false;
loop {
tokio::select! {
node_a_event = node_a.next_event() => {
if let Some(FuelP2PEvent::PeerInfoUpdated { peer_id, block_height: _ }) = node_a_event {
if node_a.peer_manager.get_peer_info(&peer_id).is_some() {
if !request_sent {
request_sent = true;
let (tx_orchestrator, rx_orchestrator) = oneshot::channel();
assert_eq!(node_a.outbound_requests_table.len(), 0);
let requested_block_height = RequestMessage::SealedHeaders(0..0);
assert!(node_a.send_request_msg(None, requested_block_height, ResponseSender::SealedHeaders(tx_orchestrator)).is_ok());
assert_eq!(node_a.outbound_requests_table.len(), 1);
let tx_test_end = tx_test_end.clone();
tokio::spawn(async move {
match rx_orchestrator.await {
Ok((_, Ok(_))) => {
let _ = tx_test_end.send(false).await;
panic!("Request succeeded unexpectedly")},
Ok((_, Err(ResponseError::P2P(_)))) => {
let _ = tx_test_end.send(true).await;
},
Ok((_, Err(err))) => {
let _ = tx_test_end.send(false).await;
panic!("Unexpected error: {:?}", err);
},
Err(e) => {
let _ = tx_test_end.send(false).await;
panic!("Channel closed unexpectedly: {:?}", e)},
}
});
}
}
}
tracing::info!("Node A Event: {:?}", node_a_event);
},
recv = rx_test_end.recv() => {
assert_eq!(recv, Some(true), "Test failed");
assert_eq!(node_a.outbound_requests_table.len(), 0);
break;
},
node_b_event = node_b.next_event() => {
tracing::info!("Node B Event: {:?}", node_b_event);
}
};
}
}
}