mirror of
https://github.com/pezkuwichain/pezkuwi-subxt.git
synced 2026-04-27 06:57:58 +00:00
Make sure we don't send messages multiple times in the availability distribution subsystem (#2015)
This commit is contained in:
@@ -52,6 +52,9 @@ use std::collections::{HashMap, HashSet};
|
||||
use std::iter;
|
||||
use thiserror::Error;
|
||||
|
||||
#[cfg(test)]
|
||||
mod tests;
|
||||
|
||||
const LOG_TARGET: &'static str = "availability_distribution";
|
||||
|
||||
#[derive(Debug, Error)]
|
||||
@@ -166,6 +169,14 @@ struct PerCandidate {
|
||||
validator_index: Option<ValidatorIndex>,
|
||||
}
|
||||
|
||||
impl PerCandidate {
|
||||
/// Returns `true` iff the given `message` is required by the given `peer`.
|
||||
fn message_required_by_peer(&self, peer: &PeerId, message: &(CandidateHash, ValidatorIndex)) -> bool {
|
||||
self.received_messages.get(peer).map(|v| !v.contains(message)).unwrap_or(true)
|
||||
&& self.sent_messages.get(peer).map(|v| !v.contains(message)).unwrap_or(true)
|
||||
}
|
||||
}
|
||||
|
||||
#[derive(Debug, Clone, Default)]
|
||||
struct PerRelayParent {
|
||||
/// Set of `K` ancestors for this relay parent.
|
||||
@@ -364,24 +375,22 @@ where
|
||||
{
|
||||
let _timer = metrics.time_handle_our_view_change();
|
||||
|
||||
let old_view = std::mem::replace(&mut (state.view), view);
|
||||
let old_view = std::mem::replace(&mut state.view, view);
|
||||
|
||||
// needed due to borrow rules
|
||||
let view = state.view.clone();
|
||||
let added = view.difference(&old_view).collect::<Vec<&'_ Hash>>();
|
||||
|
||||
// add all the relay parents and fill the cache
|
||||
for added in added.iter() {
|
||||
let added = **added;
|
||||
let validators = query_validators(ctx, added).await?;
|
||||
for added in view.difference(&old_view) {
|
||||
let validators = query_validators(ctx, *added).await?;
|
||||
let validator_index = obtain_our_validator_index(&validators, keystore.clone()).await;
|
||||
state
|
||||
.add_relay_parent(ctx, added, validators, validator_index)
|
||||
.add_relay_parent(ctx, *added, validators, validator_index)
|
||||
.await?;
|
||||
}
|
||||
|
||||
// handle all candidates
|
||||
for (candidate_hash, _receipt) in state.cached_live_candidates_unioned(added) {
|
||||
for (candidate_hash, _receipt) in state.cached_live_candidates_unioned(view.difference(&old_view)) {
|
||||
let per_candidate = state.per_candidate.entry(candidate_hash).or_default();
|
||||
|
||||
// assure the node has the validator role
|
||||
@@ -418,12 +427,9 @@ where
|
||||
// obtain the chunks from the cache, if not fallback
|
||||
// and query the availability store
|
||||
let message_id = (candidate_hash, chunk_index);
|
||||
let erasure_chunk = if let Some(message) = per_candidate.message_vault.get(&chunk_index)
|
||||
{
|
||||
let erasure_chunk = if let Some(message) = per_candidate.message_vault.get(&chunk_index) {
|
||||
message.erasure_chunk.clone()
|
||||
} else if let Some(erasure_chunk) =
|
||||
query_chunk(ctx, candidate_hash, chunk_index as ValidatorIndex).await?
|
||||
{
|
||||
} else if let Some(erasure_chunk) = query_chunk(ctx, candidate_hash, chunk_index as ValidatorIndex).await? {
|
||||
erasure_chunk
|
||||
} else {
|
||||
continue;
|
||||
@@ -433,23 +439,15 @@ where
|
||||
|
||||
let peers = peers
|
||||
.iter()
|
||||
.filter(|peer| {
|
||||
// only pick those which were not sent before
|
||||
!per_candidate
|
||||
.sent_messages
|
||||
.get(*peer)
|
||||
.filter(|set| set.contains(&message_id))
|
||||
.is_some()
|
||||
})
|
||||
.map(|peer| peer.clone())
|
||||
.filter(|peer| per_candidate.message_required_by_peer(peer, &message_id))
|
||||
.cloned()
|
||||
.collect::<Vec<_>>();
|
||||
let message = AvailabilityGossipMessage {
|
||||
candidate_hash,
|
||||
erasure_chunk,
|
||||
};
|
||||
|
||||
send_tracked_gossip_message_to_peers(ctx, per_candidate, metrics, peers, message)
|
||||
.await;
|
||||
send_tracked_gossip_message_to_peers(ctx, per_candidate, metrics, peers, message).await;
|
||||
}
|
||||
}
|
||||
|
||||
@@ -472,8 +470,7 @@ async fn send_tracked_gossip_message_to_peers<Context>(
|
||||
where
|
||||
Context: SubsystemContext<Message = AvailabilityDistributionMessage>,
|
||||
{
|
||||
send_tracked_gossip_messages_to_peers(ctx, per_candidate, metrics, peers, iter::once(message))
|
||||
.await
|
||||
send_tracked_gossip_messages_to_peers(ctx, per_candidate, metrics, peers, iter::once(message)).await
|
||||
}
|
||||
|
||||
#[inline(always)]
|
||||
@@ -487,8 +484,7 @@ async fn send_tracked_gossip_messages_to_peer<Context>(
|
||||
where
|
||||
Context: SubsystemContext<Message = AvailabilityDistributionMessage>,
|
||||
{
|
||||
send_tracked_gossip_messages_to_peers(ctx, per_candidate, metrics, vec![peer], message_iter)
|
||||
.await
|
||||
send_tracked_gossip_messages_to_peers(ctx, per_candidate, metrics, vec![peer], message_iter).await
|
||||
}
|
||||
|
||||
#[tracing::instrument(level = "trace", skip(ctx, metrics, message_iter), fields(subsystem = LOG_TARGET))]
|
||||
@@ -576,21 +572,12 @@ where
|
||||
per_candidate
|
||||
.message_vault
|
||||
.get(&erasure_chunk_index)
|
||||
.filter(|_| {
|
||||
// check if that erasure chunk was already sent before
|
||||
if let Some(sent_set) = per_candidate.sent_messages.get(&origin) {
|
||||
if sent_set.contains(&message_id) {
|
||||
return false;
|
||||
}
|
||||
}
|
||||
true
|
||||
})
|
||||
.filter(|_| per_candidate.message_required_by_peer(&origin, &message_id))
|
||||
})
|
||||
.cloned()
|
||||
.collect::<HashSet<_>>();
|
||||
|
||||
send_tracked_gossip_messages_to_peer(ctx, per_candidate, metrics, origin.clone(), messages)
|
||||
.await;
|
||||
send_tracked_gossip_messages_to_peer(ctx, per_candidate, metrics, origin.clone(), messages).await;
|
||||
}
|
||||
}
|
||||
|
||||
@@ -727,15 +714,7 @@ where
|
||||
|
||||
let peers = peers
|
||||
.into_iter()
|
||||
.filter(|peer| {
|
||||
let peer: PeerId = peer.clone();
|
||||
// avoid sending duplicate messages
|
||||
per_candidate
|
||||
.sent_messages
|
||||
.entry(peer)
|
||||
.or_default()
|
||||
.contains(&message_id)
|
||||
})
|
||||
.filter(|peer| per_candidate.message_required_by_peer(peer, &message_id))
|
||||
.collect::<Vec<_>>();
|
||||
|
||||
// gossip that message to interested peers
|
||||
@@ -1210,6 +1189,3 @@ impl metrics::Metrics for Metrics {
|
||||
Ok(Metrics(Some(metrics)))
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod tests;
|
||||
|
||||
@@ -314,7 +314,7 @@ fn helper_integrity() {
|
||||
let candidate = TestCandidateBuilder {
|
||||
para_id: test_state.chain_ids[0],
|
||||
relay_parent: test_state.relay_parent,
|
||||
pov_hash: pov_hash,
|
||||
pov_hash,
|
||||
erasure_root: make_erasure_root(&test_state, pov_block.clone()),
|
||||
..Default::default()
|
||||
}
|
||||
@@ -395,7 +395,7 @@ fn reputation_verification() {
|
||||
}
|
||||
.build(),
|
||||
TestCandidateBuilder {
|
||||
para_id: test_state.chain_ids[0],
|
||||
para_id: test_state.chain_ids[1],
|
||||
relay_parent: test_state.relay_parent,
|
||||
pov_hash: pov_hash_b,
|
||||
erasure_root: make_erasure_root(&test_state, pov_block_b.clone()),
|
||||
@@ -643,10 +643,8 @@ fn reputation_verification() {
|
||||
)
|
||||
) => {
|
||||
let index = candidates2.iter().enumerate().find(|x| { x.1.hash() == candidate_hash }).map(|x| x.0).unwrap();
|
||||
expected = dbg!(candidates2.swap_remove(index).hash());
|
||||
tx.send(
|
||||
i == 0
|
||||
).unwrap();
|
||||
expected = candidates2.swap_remove(index).hash();
|
||||
tx.send(i == 0).unwrap();
|
||||
}
|
||||
);
|
||||
|
||||
@@ -762,6 +760,23 @@ fn reputation_verification() {
|
||||
)
|
||||
.await;
|
||||
|
||||
assert_matches!(
|
||||
overseer_recv(&mut virtual_overseer).await,
|
||||
AllMessages::NetworkBridge(
|
||||
NetworkBridgeMessage::SendValidationMessage(
|
||||
peers,
|
||||
protocol_v1::ValidationProtocol::AvailabilityDistribution(
|
||||
protocol_v1::AvailabilityDistributionMessage::Chunk(hash, chunk),
|
||||
),
|
||||
)
|
||||
) => {
|
||||
assert_eq!(1, peers.len());
|
||||
assert_eq!(peers[0], peer_a);
|
||||
assert_eq!(candidates[0].hash(), hash);
|
||||
assert_eq!(valid.erasure_chunk, chunk);
|
||||
}
|
||||
);
|
||||
|
||||
assert_matches!(
|
||||
overseer_recv(&mut virtual_overseer).await,
|
||||
AllMessages::NetworkBridge(
|
||||
@@ -861,7 +876,7 @@ fn reputation_verification() {
|
||||
|
||||
{
|
||||
// send another message
|
||||
let valid2: AvailabilityGossipMessage = make_valid_availability_gossip(
|
||||
let valid2 = make_valid_availability_gossip(
|
||||
&test_state,
|
||||
candidates[2].hash(),
|
||||
1,
|
||||
@@ -890,6 +905,102 @@ fn reputation_verification() {
|
||||
}
|
||||
);
|
||||
}
|
||||
|
||||
{
|
||||
// send another message
|
||||
let valid = make_valid_availability_gossip(
|
||||
&test_state,
|
||||
candidates[1].hash(),
|
||||
2,
|
||||
pov_block_b.clone(),
|
||||
);
|
||||
|
||||
// Make peer a and b listen on `current`
|
||||
overseer_send(
|
||||
&mut virtual_overseer,
|
||||
AvailabilityDistributionMessage::NetworkBridgeUpdateV1(
|
||||
NetworkBridgeEvent::PeerViewChange(peer_a.clone(), view![current]),
|
||||
),
|
||||
)
|
||||
.await;
|
||||
|
||||
overseer_send(
|
||||
&mut virtual_overseer,
|
||||
AvailabilityDistributionMessage::NetworkBridgeUpdateV1(
|
||||
NetworkBridgeEvent::PeerViewChange(peer_b.clone(), view![current]),
|
||||
),
|
||||
)
|
||||
.await;
|
||||
|
||||
overseer_send(
|
||||
&mut virtual_overseer,
|
||||
AvailabilityDistributionMessage::NetworkBridgeUpdateV1(
|
||||
NetworkBridgeEvent::PeerMessage(
|
||||
peer_a.clone(),
|
||||
chunk_protocol_message(valid.clone()),
|
||||
),
|
||||
),
|
||||
)
|
||||
.await;
|
||||
|
||||
assert_matches!(
|
||||
overseer_recv(&mut virtual_overseer).await,
|
||||
AllMessages::NetworkBridge(
|
||||
NetworkBridgeMessage::ReportPeer(
|
||||
peer,
|
||||
rep
|
||||
)
|
||||
) => {
|
||||
assert_eq!(peer, peer_a);
|
||||
assert_eq!(rep, BENEFIT_VALID_MESSAGE_FIRST);
|
||||
}
|
||||
);
|
||||
|
||||
assert_matches!(
|
||||
overseer_recv(&mut virtual_overseer).await,
|
||||
AllMessages::NetworkBridge(
|
||||
NetworkBridgeMessage::SendValidationMessage(
|
||||
peers,
|
||||
protocol_v1::ValidationProtocol::AvailabilityDistribution(
|
||||
protocol_v1::AvailabilityDistributionMessage::Chunk(hash, chunk),
|
||||
),
|
||||
)
|
||||
) => {
|
||||
assert_eq!(1, peers.len());
|
||||
assert_eq!(peers[0], peer_b);
|
||||
assert_eq!(candidates[1].hash(), hash);
|
||||
assert_eq!(valid.erasure_chunk, chunk);
|
||||
}
|
||||
);
|
||||
|
||||
// Let B send the same message
|
||||
overseer_send(
|
||||
&mut virtual_overseer,
|
||||
AvailabilityDistributionMessage::NetworkBridgeUpdateV1(
|
||||
NetworkBridgeEvent::PeerMessage(
|
||||
peer_b.clone(),
|
||||
chunk_protocol_message(valid.clone()),
|
||||
),
|
||||
),
|
||||
)
|
||||
.await;
|
||||
|
||||
assert_matches!(
|
||||
overseer_recv(&mut virtual_overseer).await,
|
||||
AllMessages::NetworkBridge(
|
||||
NetworkBridgeMessage::ReportPeer(
|
||||
peer,
|
||||
rep
|
||||
)
|
||||
) => {
|
||||
assert_eq!(peer, peer_b);
|
||||
assert_eq!(rep, BENEFIT_VALID_MESSAGE);
|
||||
}
|
||||
);
|
||||
|
||||
// There shouldn't be any other message.
|
||||
assert!(virtual_overseer.recv().timeout(TIMEOUT).await.is_none());
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
|
||||
Reference in New Issue
Block a user