Route the coordinator, fix race conditions in the signers library

This commit is contained in:
Luke Parker 2024-09-08 22:13:42 -04:00
parent 7484eadbbb
commit f07ec7bee0
15 changed files with 356 additions and 172 deletions

2
Cargo.lock generated
View file

@ -8630,7 +8630,6 @@ name = "serai-processor-frost-attempt-manager"
version = "0.1.0" version = "0.1.0"
dependencies = [ dependencies = [
"borsh", "borsh",
"hex",
"log", "log",
"modular-frost", "modular-frost",
"parity-scale-codec", "parity-scale-codec",
@ -8666,6 +8665,7 @@ version = "0.1.0"
dependencies = [ dependencies = [
"borsh", "borsh",
"dkg", "dkg",
"hex",
"parity-scale-codec", "parity-scale-codec",
"serai-coins-primitives", "serai-coins-primitives",
"serai-in-instructions-primitives", "serai-in-instructions-primitives",

View file

@ -26,7 +26,6 @@ frost = { package = "modular-frost", path = "../../crypto/frost", version = "^0.
serai-validator-sets-primitives = { path = "../../substrate/validator-sets/primitives", default-features = false, features = ["std"] } serai-validator-sets-primitives = { path = "../../substrate/validator-sets/primitives", default-features = false, features = ["std"] }
hex = { version = "0.4", default-features = false, features = ["std"] }
log = { version = "0.4", default-features = false, features = ["std"] } log = { version = "0.4", default-features = false, features = ["std"] }
scale = { package = "parity-scale-codec", version = "3", default-features = false, features = ["std"] } scale = { package = "parity-scale-codec", version = "3", default-features = false, features = ["std"] }

View file

@ -10,11 +10,11 @@ use frost::{
use serai_validator_sets_primitives::Session; use serai_validator_sets_primitives::Session;
use serai_db::{Get, DbTxn, Db, create_db}; use serai_db::{Get, DbTxn, Db, create_db};
use messages::sign::{SignId, ProcessorMessage}; use messages::sign::{VariantSignId, SignId, ProcessorMessage};
create_db!( create_db!(
FrostAttemptManager { FrostAttemptManager {
Attempted: (id: [u8; 32]) -> u32, Attempted: (id: VariantSignId) -> u32,
} }
); );
@ -28,7 +28,7 @@ pub(crate) struct SigningProtocol<D: Db, M: Clone + PreprocessMachine> {
// The key shares we sign with are expected to be continguous from this position. // The key shares we sign with are expected to be continguous from this position.
start_i: Participant, start_i: Participant,
// The ID of this signing protocol. // The ID of this signing protocol.
id: [u8; 32], id: VariantSignId,
// This accepts a vector of `root` machines in order to support signing with multiple key shares. // This accepts a vector of `root` machines in order to support signing with multiple key shares.
root: Vec<M>, root: Vec<M>,
preprocessed: HashMap<u32, (Vec<M::SignMachine>, HashMap<Participant, Vec<u8>>)>, preprocessed: HashMap<u32, (Vec<M::SignMachine>, HashMap<Participant, Vec<u8>>)>,
@ -48,10 +48,10 @@ impl<D: Db, M: Clone + PreprocessMachine> SigningProtocol<D, M> {
db: D, db: D,
session: Session, session: Session,
start_i: Participant, start_i: Participant,
id: [u8; 32], id: VariantSignId,
root: Vec<M>, root: Vec<M>,
) -> Self { ) -> Self {
log::info!("starting signing protocol {}", hex::encode(id)); log::info!("starting signing protocol {id:?}");
Self { Self {
db, db,
@ -100,7 +100,7 @@ impl<D: Db, M: Clone + PreprocessMachine> SigningProtocol<D, M> {
txn.commit(); txn.commit();
} }
log::debug!("attemting a new instance of signing protocol {}", hex::encode(self.id)); log::debug!("attemting a new instance of signing protocol {:?}", self.id);
let mut our_preprocesses = HashMap::with_capacity(self.root.len()); let mut our_preprocesses = HashMap::with_capacity(self.root.len());
let mut preprocessed = Vec::with_capacity(self.root.len()); let mut preprocessed = Vec::with_capacity(self.root.len());
@ -137,7 +137,7 @@ impl<D: Db, M: Clone + PreprocessMachine> SigningProtocol<D, M> {
attempt: u32, attempt: u32,
serialized_preprocesses: HashMap<Participant, Vec<u8>>, serialized_preprocesses: HashMap<Participant, Vec<u8>>,
) -> Vec<ProcessorMessage> { ) -> Vec<ProcessorMessage> {
log::debug!("handling preprocesses for signing protocol {}", hex::encode(self.id)); log::debug!("handling preprocesses for signing protocol {:?}", self.id);
let Some((machines, our_serialized_preprocesses)) = self.preprocessed.remove(&attempt) else { let Some((machines, our_serialized_preprocesses)) = self.preprocessed.remove(&attempt) else {
return vec![]; return vec![];
@ -211,8 +211,8 @@ impl<D: Db, M: Clone + PreprocessMachine> SigningProtocol<D, M> {
assert!(self.shared.insert(attempt, (shared.swap_remove(0), our_shares)).is_none()); assert!(self.shared.insert(attempt, (shared.swap_remove(0), our_shares)).is_none());
log::debug!( log::debug!(
"successfully handled preprocesses for signing protocol {}, sending shares", "successfully handled preprocesses for signing protocol {:?}, sending shares",
hex::encode(self.id) self.id,
); );
msgs.push(ProcessorMessage::Shares { msgs.push(ProcessorMessage::Shares {
id: SignId { session: self.session, id: self.id, attempt }, id: SignId { session: self.session, id: self.id, attempt },
@ -229,7 +229,7 @@ impl<D: Db, M: Clone + PreprocessMachine> SigningProtocol<D, M> {
attempt: u32, attempt: u32,
serialized_shares: HashMap<Participant, Vec<u8>>, serialized_shares: HashMap<Participant, Vec<u8>>,
) -> Result<M::Signature, Vec<ProcessorMessage>> { ) -> Result<M::Signature, Vec<ProcessorMessage>> {
log::debug!("handling shares for signing protocol {}", hex::encode(self.id)); log::debug!("handling shares for signing protocol {:?}", self.id);
let Some((machine, our_serialized_shares)) = self.shared.remove(&attempt) else { Err(vec![])? }; let Some((machine, our_serialized_shares)) = self.shared.remove(&attempt) else { Err(vec![])? };
@ -272,13 +272,13 @@ impl<D: Db, M: Clone + PreprocessMachine> SigningProtocol<D, M> {
}, },
}; };
log::info!("finished signing for protocol {}", hex::encode(self.id)); log::info!("finished signing for protocol {:?}", self.id);
Ok(signature) Ok(signature)
} }
/// Cleanup the database entries for a specified signing protocol. /// Cleanup the database entries for a specified signing protocol.
pub(crate) fn cleanup(txn: &mut impl DbTxn, id: [u8; 32]) { pub(crate) fn cleanup(txn: &mut impl DbTxn, id: VariantSignId) {
Attempted::del(txn, id); Attempted::del(txn, id);
} }
} }

View file

@ -9,7 +9,7 @@ use frost::{Participant, sign::PreprocessMachine};
use serai_validator_sets_primitives::Session; use serai_validator_sets_primitives::Session;
use serai_db::{DbTxn, Db}; use serai_db::{DbTxn, Db};
use messages::sign::{ProcessorMessage, CoordinatorMessage}; use messages::sign::{VariantSignId, ProcessorMessage, CoordinatorMessage};
mod individual; mod individual;
use individual::SigningProtocol; use individual::SigningProtocol;
@ -21,7 +21,7 @@ pub enum Response<M: PreprocessMachine> {
/// A produced signature. /// A produced signature.
Signature { Signature {
/// The ID of the protocol this is for. /// The ID of the protocol this is for.
id: [u8; 32], id: VariantSignId,
/// The signature. /// The signature.
signature: M::Signature, signature: M::Signature,
}, },
@ -32,7 +32,7 @@ pub struct AttemptManager<D: Db, M: Clone + PreprocessMachine> {
db: D, db: D,
session: Session, session: Session,
start_i: Participant, start_i: Participant,
active: HashMap<[u8; 32], SigningProtocol<D, M>>, active: HashMap<VariantSignId, SigningProtocol<D, M>>,
} }
impl<D: Db, M: Clone + PreprocessMachine> AttemptManager<D, M> { impl<D: Db, M: Clone + PreprocessMachine> AttemptManager<D, M> {
@ -46,7 +46,7 @@ impl<D: Db, M: Clone + PreprocessMachine> AttemptManager<D, M> {
/// Register a signing protocol to attempt. /// Register a signing protocol to attempt.
/// ///
/// This ID must be unique across all sessions, attempt managers, protocols, etc. /// This ID must be unique across all sessions, attempt managers, protocols, etc.
pub fn register(&mut self, id: [u8; 32], machines: Vec<M>) -> Vec<ProcessorMessage> { pub fn register(&mut self, id: VariantSignId, machines: Vec<M>) -> Vec<ProcessorMessage> {
let mut protocol = let mut protocol =
SigningProtocol::new(self.db.clone(), self.session, self.start_i, id, machines); SigningProtocol::new(self.db.clone(), self.session, self.start_i, id, machines);
let messages = protocol.attempt(0); let messages = protocol.attempt(0);
@ -60,11 +60,11 @@ impl<D: Db, M: Clone + PreprocessMachine> AttemptManager<D, M> {
/// This does not stop the protocol from being re-registered and further worked on (with /// This does not stop the protocol from being re-registered and further worked on (with
/// undefined behavior) then. The higher-level context must never call `register` again with this /// undefined behavior) then. The higher-level context must never call `register` again with this
/// ID accordingly. /// ID accordingly.
pub fn retire(&mut self, txn: &mut impl DbTxn, id: [u8; 32]) { pub fn retire(&mut self, txn: &mut impl DbTxn, id: VariantSignId) {
if self.active.remove(&id).is_none() { if self.active.remove(&id).is_none() {
log::info!("retiring protocol {}, which we didn't register/already retired", hex::encode(id)); log::info!("retiring protocol {id:?}, which we didn't register/already retired");
} else { } else {
log::info!("retired signing protocol {}", hex::encode(id)); log::info!("retired signing protocol {id:?}");
} }
SigningProtocol::<D, M>::cleanup(txn, id); SigningProtocol::<D, M>::cleanup(txn, id);
} }
@ -79,8 +79,8 @@ impl<D: Db, M: Clone + PreprocessMachine> AttemptManager<D, M> {
CoordinatorMessage::Preprocesses { id, preprocesses } => { CoordinatorMessage::Preprocesses { id, preprocesses } => {
let Some(protocol) = self.active.get_mut(&id.id) else { let Some(protocol) = self.active.get_mut(&id.id) else {
log::trace!( log::trace!(
"handling preprocesses for signing protocol {}, which we're not actively running", "handling preprocesses for signing protocol {:?}, which we're not actively running",
hex::encode(id.id) id.id,
); );
return Response::Messages(vec![]); return Response::Messages(vec![]);
}; };
@ -89,8 +89,8 @@ impl<D: Db, M: Clone + PreprocessMachine> AttemptManager<D, M> {
CoordinatorMessage::Shares { id, shares } => { CoordinatorMessage::Shares { id, shares } => {
let Some(protocol) = self.active.get_mut(&id.id) else { let Some(protocol) = self.active.get_mut(&id.id) else {
log::trace!( log::trace!(
"handling shares for signing protocol {}, which we're not actively running", "handling shares for signing protocol {:?}, which we're not actively running",
hex::encode(id.id) id.id,
); );
return Response::Messages(vec![]); return Response::Messages(vec![]);
}; };
@ -102,8 +102,8 @@ impl<D: Db, M: Clone + PreprocessMachine> AttemptManager<D, M> {
CoordinatorMessage::Reattempt { id } => { CoordinatorMessage::Reattempt { id } => {
let Some(protocol) = self.active.get_mut(&id.id) else { let Some(protocol) = self.active.get_mut(&id.id) else {
log::trace!( log::trace!(
"reattempting signing protocol {}, which we're not actively running", "reattempting signing protocol {:?}, which we're not actively running",
hex::encode(id.id) id.id,
); );
return Response::Messages(vec![]); return Response::Messages(vec![]);
}; };

View file

@ -17,6 +17,8 @@ rustdoc-args = ["--cfg", "docsrs"]
workspace = true workspace = true
[dependencies] [dependencies]
hex = { version = "0.4", default-features = false, features = ["std"] }
scale = { package = "parity-scale-codec", version = "3", default-features = false, features = ["std"] } scale = { package = "parity-scale-codec", version = "3", default-features = false, features = ["std"] }
borsh = { version = "1", default-features = false, features = ["std", "derive", "de_strict_order"] } borsh = { version = "1", default-features = false, features = ["std", "derive", "de_strict_order"] }

View file

@ -1,3 +1,4 @@
use core::fmt;
use std::collections::HashMap; use std::collections::HashMap;
use scale::{Encode, Decode}; use scale::{Encode, Decode};
@ -85,10 +86,37 @@ pub mod key_gen {
pub mod sign { pub mod sign {
use super::*; use super::*;
#[derive(Clone, PartialEq, Eq, Hash, Debug, Encode, Decode, BorshSerialize, BorshDeserialize)] #[derive(Clone, Copy, PartialEq, Eq, Hash, Encode, Decode, BorshSerialize, BorshDeserialize)]
pub enum VariantSignId {
Cosign([u8; 32]),
Batch(u32),
SlashReport([u8; 32]),
Transaction([u8; 32]),
}
impl fmt::Debug for VariantSignId {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> Result<(), fmt::Error> {
match self {
Self::Cosign(cosign) => {
f.debug_struct("VariantSignId::Cosign").field("0", &hex::encode(cosign)).finish()
}
Self::Batch(batch) => f.debug_struct("VariantSignId::Batch").field("0", &batch).finish(),
Self::SlashReport(slash_report) => f
.debug_struct("VariantSignId::SlashReport")
.field("0", &hex::encode(slash_report))
.finish(),
Self::Transaction(tx) => {
f.debug_struct("VariantSignId::Transaction").field("0", &hex::encode(tx)).finish()
}
}
}
}
#[derive(
Clone, Copy, PartialEq, Eq, Hash, Debug, Encode, Decode, BorshSerialize, BorshDeserialize,
)]
pub struct SignId { pub struct SignId {
pub session: Session, pub session: Session,
pub id: [u8; 32], pub id: VariantSignId,
pub attempt: u32, pub attempt: u32,
} }
@ -109,11 +137,11 @@ pub mod sign {
None None
} }
pub fn session(&self) -> Session { pub fn sign_id(&self) -> &SignId {
match self { match self {
CoordinatorMessage::Preprocesses { id, .. } | CoordinatorMessage::Preprocesses { id, .. } |
CoordinatorMessage::Shares { id, .. } | CoordinatorMessage::Shares { id, .. } |
CoordinatorMessage::Reattempt { id, .. } => id.session, CoordinatorMessage::Reattempt { id, .. } => id,
} }
} }
} }

View file

@ -60,6 +60,8 @@ pub trait Block: Send + Sync + Sized + Clone + Debug {
/// Check if this block resolved any Eventualities. /// Check if this block resolved any Eventualities.
/// ///
/// This MUST mutate `eventualities` to no longer contain the resolved Eventualities.
///
/// Returns tbe resolved Eventualities, indexed by the ID of the transactions which resolved /// Returns tbe resolved Eventualities, indexed by the ID of the transactions which resolved
/// them. /// them.
fn check_for_eventuality_resolutions( fn check_for_eventuality_resolutions(

View file

@ -1,6 +1,7 @@
use core::marker::PhantomData; use core::marker::PhantomData;
use std::io::{self, Read, Write}; use std::io::{self, Read, Write};
use group::GroupEncoding;
use scale::{Encode, Decode, IoReader}; use scale::{Encode, Decode, IoReader};
use borsh::{BorshSerialize, BorshDeserialize}; use borsh::{BorshSerialize, BorshDeserialize};
use serai_db::{Get, DbTxn, create_db, db_channel}; use serai_db::{Get, DbTxn, create_db, db_channel};
@ -526,20 +527,20 @@ mod _completed_eventualities {
db_channel! { db_channel! {
ScannerPublic { ScannerPublic {
CompletedEventualities: (empty_key: ()) -> [u8; 32], CompletedEventualities: (key: &[u8]) -> [u8; 32],
} }
} }
} }
/// The IDs of completed Eventualities found on-chain, within a finalized block. /// The IDs of completed Eventualities found on-chain, within a finalized block.
pub struct CompletedEventualities<S: ScannerFeed>(PhantomData<S>); pub struct CompletedEventualities<K: GroupEncoding>(PhantomData<K>);
impl<S: ScannerFeed> CompletedEventualities<S> { impl<K: GroupEncoding> CompletedEventualities<K> {
pub(crate) fn send(txn: &mut impl DbTxn, id: [u8; 32]) { pub(crate) fn send(txn: &mut impl DbTxn, key: &K, id: [u8; 32]) {
_completed_eventualities::CompletedEventualities::send(txn, (), &id); _completed_eventualities::CompletedEventualities::send(txn, key.to_bytes().as_ref(), &id);
} }
/// Receive the ID of a completed Eventuality. /// Receive the ID of a completed Eventuality.
pub fn try_recv(txn: &mut impl DbTxn) -> Option<[u8; 32]> { pub fn try_recv(txn: &mut impl DbTxn, key: &K) -> Option<[u8; 32]> {
_completed_eventualities::CompletedEventualities::try_recv(txn, ()) _completed_eventualities::CompletedEventualities::try_recv(txn, key.to_bytes().as_ref())
} }
} }

View file

@ -298,7 +298,7 @@ impl<D: Db, S: ScannerFeed, Sch: Scheduler<S>> ContinuallyRan for EventualityTas
hex::encode(eventuality.id()), hex::encode(eventuality.id()),
hex::encode(tx.as_ref()) hex::encode(tx.as_ref())
); );
CompletedEventualities::<S>::send(&mut txn, eventuality.id()); CompletedEventualities::send(&mut txn, &key.key, eventuality.id());
} }
// Fetch all non-External outputs // Fetch all non-External outputs

View file

@ -362,24 +362,24 @@ impl<S: ScannerFeed> Scanner<S> {
let substrate_task = substrate::SubstrateTask::<_, S>::new(db.clone()); let substrate_task = substrate::SubstrateTask::<_, S>::new(db.clone());
let eventuality_task = eventuality::EventualityTask::<_, _, Sch>::new(db, feed, start_block); let eventuality_task = eventuality::EventualityTask::<_, _, Sch>::new(db, feed, start_block);
let (index_run, _index_handle) = Task::new(); let (index_task_def, _index_handle) = Task::new();
let (scan_run, scan_handle) = Task::new(); let (scan_task_def, scan_handle) = Task::new();
let (report_run, report_handle) = Task::new(); let (report_task_def, report_handle) = Task::new();
let (substrate_run, substrate_handle) = Task::new(); let (substrate_task_def, substrate_handle) = Task::new();
let (eventuality_run, eventuality_handle) = Task::new(); let (eventuality_task_def, eventuality_handle) = Task::new();
// Upon indexing a new block, scan it // Upon indexing a new block, scan it
tokio::spawn(index_task.continually_run(index_run, vec![scan_handle.clone()])); tokio::spawn(index_task.continually_run(index_task_def, vec![scan_handle.clone()]));
// Upon scanning a block, report it // Upon scanning a block, report it
tokio::spawn(scan_task.continually_run(scan_run, vec![report_handle])); tokio::spawn(scan_task.continually_run(scan_task_def, vec![report_handle]));
// Upon reporting a block, we do nothing (as the burden is on Substrate which won't be // Upon reporting a block, we do nothing (as the burden is on Substrate which won't be
// immediately ready) // immediately ready)
tokio::spawn(report_task.continually_run(report_run, vec![])); tokio::spawn(report_task.continually_run(report_task_def, vec![]));
// Upon handling an event from Substrate, we run the Eventuality task (as it's what's affected) // Upon handling an event from Substrate, we run the Eventuality task (as it's what's affected)
tokio::spawn(substrate_task.continually_run(substrate_run, vec![eventuality_handle])); tokio::spawn(substrate_task.continually_run(substrate_task_def, vec![eventuality_handle]));
// Upon handling the Eventualities in a block, we run the scan task as we've advanced the // Upon handling the Eventualities in a block, we run the scan task as we've advanced the
// window its allowed to scan // window its allowed to scan
tokio::spawn(eventuality_task.continually_run(eventuality_run, vec![scan_handle])); tokio::spawn(eventuality_task.continually_run(eventuality_task_def, vec![scan_handle]));
Self { substrate_handle, _S: PhantomData } Self { substrate_handle, _S: PhantomData }
} }

View file

@ -41,6 +41,10 @@ pub trait SignableTransaction: 'static + Sized + Send + Sync + Clone {
fn sign(self, keys: ThresholdKeys<Self::Ciphersuite>) -> Self::PreprocessMachine; fn sign(self, keys: ThresholdKeys<Self::Ciphersuite>) -> Self::PreprocessMachine;
} }
/// The transaction type for a SignableTransaction.
pub type TransactionFor<ST> =
<<ST as SignableTransaction>::PreprocessMachine as PreprocessMachine>::Signature;
mod db { mod db {
use serai_db::{Get, DbTxn, create_db, db_channel}; use serai_db::{Get, DbTxn, create_db, db_channel};

View file

@ -0,0 +1,98 @@
use serai_db::{DbTxn, Db};
use primitives::task::ContinuallyRan;
use crate::{
db::{
RegisteredKeys, CosignerToCoordinatorMessages, BatchSignerToCoordinatorMessages,
SlashReportSignerToCoordinatorMessages, TransactionSignerToCoordinatorMessages,
},
Coordinator,
};
// Fetches messages to send the coordinator and sends them.
pub(crate) struct CoordinatorTask<D: Db, C: Coordinator> {
db: D,
coordinator: C,
}
impl<D: Db, C: Coordinator> CoordinatorTask<D, C> {
pub(crate) fn new(db: D, coordinator: C) -> Self {
Self { db, coordinator }
}
}
#[async_trait::async_trait]
impl<D: Db, C: Coordinator> ContinuallyRan for CoordinatorTask<D, C> {
async fn run_iteration(&mut self) -> Result<bool, String> {
let mut iterated = false;
for session in RegisteredKeys::get(&self.db).unwrap_or(vec![]) {
loop {
let mut txn = self.db.txn();
let Some(msg) = CosignerToCoordinatorMessages::try_recv(&mut txn, session) else {
break;
};
iterated = true;
self
.coordinator
.send(msg)
.await
.map_err(|e| format!("couldn't send sign message to the coordinator: {e:?}"))?;
txn.commit();
}
loop {
let mut txn = self.db.txn();
let Some(msg) = BatchSignerToCoordinatorMessages::try_recv(&mut txn, session) else {
break;
};
iterated = true;
self
.coordinator
.send(msg)
.await
.map_err(|e| format!("couldn't send sign message to the coordinator: {e:?}"))?;
txn.commit();
}
loop {
let mut txn = self.db.txn();
let Some(msg) = SlashReportSignerToCoordinatorMessages::try_recv(&mut txn, session) else {
break;
};
iterated = true;
self
.coordinator
.send(msg)
.await
.map_err(|e| format!("couldn't send sign message to the coordinator: {e:?}"))?;
txn.commit();
}
loop {
let mut txn = self.db.txn();
let Some(msg) = TransactionSignerToCoordinatorMessages::try_recv(&mut txn, session) else {
break;
};
iterated = true;
self
.coordinator
.send(msg)
.await
.map_err(|e| format!("couldn't send sign message to the coordinator: {e:?}"))?;
txn.commit();
}
}
Ok(iterated)
}
}

View file

@ -15,14 +15,8 @@ create_db! {
db_channel! { db_channel! {
SignersGlobal { SignersGlobal {
// CompletedEventualities needs to be handled by each signer, meaning we need to turn its CoordinatorToCosignerMessages: (session: Session) -> CoordinatorMessage,
// effective spsc into a spmc. We do this by duplicating its message for all keys we're CosignerToCoordinatorMessages: (session: Session) -> ProcessorMessage,
// signing for.
// TODO: Populate from CompletedEventualities
CompletedEventualitiesForEachKey: (session: Session) -> [u8; 32],
CoordinatorToTransactionSignerMessages: (session: Session) -> CoordinatorMessage,
TransactionSignerToCoordinatorMessages: (session: Session) -> ProcessorMessage,
CoordinatorToBatchSignerMessages: (session: Session) -> CoordinatorMessage, CoordinatorToBatchSignerMessages: (session: Session) -> CoordinatorMessage,
BatchSignerToCoordinatorMessages: (session: Session) -> ProcessorMessage, BatchSignerToCoordinatorMessages: (session: Session) -> ProcessorMessage,
@ -30,7 +24,7 @@ db_channel! {
CoordinatorToSlashReportSignerMessages: (session: Session) -> CoordinatorMessage, CoordinatorToSlashReportSignerMessages: (session: Session) -> CoordinatorMessage,
SlashReportSignerToCoordinatorMessages: (session: Session) -> ProcessorMessage, SlashReportSignerToCoordinatorMessages: (session: Session) -> ProcessorMessage,
CoordinatorToCosignerMessages: (session: Session) -> CoordinatorMessage, CoordinatorToTransactionSignerMessages: (session: Session) -> CoordinatorMessage,
CosignerToCoordinatorMessages: (session: Session) -> ProcessorMessage, TransactionSignerToCoordinatorMessages: (session: Session) -> ProcessorMessage,
} }
} }

View file

@ -7,23 +7,42 @@ use std::collections::HashMap;
use zeroize::Zeroizing; use zeroize::Zeroizing;
use serai_validator_sets_primitives::Session; use ciphersuite::{group::GroupEncoding, Ciphersuite, Ristretto};
use ciphersuite::{group::GroupEncoding, Ristretto};
use frost::dkg::{ThresholdCore, ThresholdKeys}; use frost::dkg::{ThresholdCore, ThresholdKeys};
use serai_validator_sets_primitives::Session;
use serai_db::{DbTxn, Db}; use serai_db::{DbTxn, Db};
use primitives::task::TaskHandle; use messages::sign::{VariantSignId, ProcessorMessage, CoordinatorMessage};
use scheduler::{Transaction, SignableTransaction, TransactionsToSign};
use primitives::task::{Task, TaskHandle, ContinuallyRan};
use scheduler::{Transaction, SignableTransaction, TransactionFor};
pub(crate) mod db; pub(crate) mod db;
mod coordinator;
use coordinator::CoordinatorTask;
mod transaction; mod transaction;
use transaction::TransactionTask;
/// A connection to the Coordinator which messages can be published with.
#[async_trait::async_trait]
pub trait Coordinator: 'static + Send + Sync {
/// An error encountered when sending a message.
///
/// This MUST be an ephemeral error. Retrying sending a message MUST eventually resolve without
/// manual intervention/changing the arguments.
type EphemeralError: Debug;
/// Send a `messages::sign::ProcessorMessage`.
async fn send(&mut self, message: ProcessorMessage) -> Result<(), Self::EphemeralError>;
}
/// An object capable of publishing a transaction. /// An object capable of publishing a transaction.
#[async_trait::async_trait] #[async_trait::async_trait]
pub trait TransactionPublisher<T: Transaction>: 'static + Send + Sync { pub trait TransactionPublisher<T: Transaction>: 'static + Send + Sync + Clone {
/// An error encountered when publishing a transaction. /// An error encountered when publishing a transaction.
/// ///
/// This MUST be an ephemeral error. Retrying publication MUST eventually resolve without manual /// This MUST be an ephemeral error. Retrying publication MUST eventually resolve without manual
@ -40,9 +59,18 @@ pub trait TransactionPublisher<T: Transaction>: 'static + Send + Sync {
async fn publish(&self, tx: T) -> Result<(), Self::EphemeralError>; async fn publish(&self, tx: T) -> Result<(), Self::EphemeralError>;
} }
struct Tasks {
cosigner: TaskHandle,
batch: TaskHandle,
slash_report: TaskHandle,
transaction: TaskHandle,
}
/// The signers used by a processor. /// The signers used by a processor.
#[allow(non_snake_case)]
pub struct Signers<ST: SignableTransaction> { pub struct Signers<ST: SignableTransaction> {
tasks: HashMap<Session, Vec<TaskHandle>>, coordinator_handle: TaskHandle,
tasks: HashMap<Session, Tasks>,
_ST: PhantomData<ST>, _ST: PhantomData<ST>,
} }
@ -62,9 +90,57 @@ impl<ST: SignableTransaction> Signers<ST> {
/// Initialize the signers. /// Initialize the signers.
/// ///
/// This will spawn tasks for any historically registered keys. /// This will spawn tasks for any historically registered keys.
pub fn new(db: impl Db) -> Self { pub fn new(
mut db: impl Db,
coordinator: impl Coordinator,
publisher: &impl TransactionPublisher<TransactionFor<ST>>,
) -> Self {
/*
On boot, perform any database cleanup which was queued.
We don't do this cleanup at time of dropping the task as we'd need to wait an unbounded
amount of time for the task to stop (requiring an async task), then we'd have to drain the
channels (which would be on a distinct DB transaction and risk not occurring if we rebooted
while waiting for the task to stop). This is the easiest way to handle this.
*/
{
let mut txn = db.txn();
for (session, external_key_bytes) in db::ToCleanup::get(&txn).unwrap_or(vec![]) {
let mut external_key_bytes = external_key_bytes.as_slice();
let external_key =
<ST::Ciphersuite as Ciphersuite>::read_G(&mut external_key_bytes).unwrap();
assert!(external_key_bytes.is_empty());
// Drain the transactions to sign
// TransactionsToSign will be fully populated by the scheduler before retiry occurs, making
// this perfect in not leaving any pending blobs behind
while scheduler::TransactionsToSign::<ST>::try_recv(&mut txn, &external_key).is_some() {}
// Drain the completed Eventualities
// This will be fully populated by the scanner before retiry
while scanner::CompletedEventualities::try_recv(&mut txn, &external_key).is_some() {}
// Drain our DB channels
while db::CoordinatorToCosignerMessages::try_recv(&mut txn, session).is_some() {}
while db::CosignerToCoordinatorMessages::try_recv(&mut txn, session).is_some() {}
while db::CoordinatorToBatchSignerMessages::try_recv(&mut txn, session).is_some() {}
while db::BatchSignerToCoordinatorMessages::try_recv(&mut txn, session).is_some() {}
while db::CoordinatorToSlashReportSignerMessages::try_recv(&mut txn, session).is_some() {}
while db::SlashReportSignerToCoordinatorMessages::try_recv(&mut txn, session).is_some() {}
while db::CoordinatorToTransactionSignerMessages::try_recv(&mut txn, session).is_some() {}
while db::TransactionSignerToCoordinatorMessages::try_recv(&mut txn, session).is_some() {}
}
db::ToCleanup::del(&mut txn);
txn.commit();
}
let mut tasks = HashMap::new(); let mut tasks = HashMap::new();
let (coordinator_task, coordinator_handle) = Task::new();
tokio::spawn(
CoordinatorTask::new(db.clone(), coordinator).continually_run(coordinator_task, vec![]),
);
for session in db::RegisteredKeys::get(&db).unwrap_or(vec![]) { for session in db::RegisteredKeys::get(&db).unwrap_or(vec![]) {
let buf = db::SerializedKeys::get(&db, session).unwrap(); let buf = db::SerializedKeys::get(&db, session).unwrap();
let mut buf = buf.as_slice(); let mut buf = buf.as_slice();
@ -78,10 +154,23 @@ impl<ST: SignableTransaction> Signers<ST> {
.push(ThresholdKeys::from(ThresholdCore::<ST::Ciphersuite>::read(&mut buf).unwrap())); .push(ThresholdKeys::from(ThresholdCore::<ST::Ciphersuite>::read(&mut buf).unwrap()));
} }
todo!("TODO") // TODO: Batch signer, cosigner, slash report signers
let (transaction_task, transaction_handle) = Task::new();
tokio::spawn(
TransactionTask::<_, ST, _>::new(db.clone(), publisher.clone(), session, external_keys)
.continually_run(transaction_task, vec![coordinator_handle.clone()]),
);
tasks.insert(session, Tasks {
cosigner: todo!("TODO"),
batch: todo!("TODO"),
slash_report: todo!("TODO"),
transaction: transaction_handle,
});
} }
Self { tasks, _ST: PhantomData } Self { coordinator_handle, tasks, _ST: PhantomData }
} }
/// Register a set of keys to sign with. /// Register a set of keys to sign with.
@ -146,82 +235,31 @@ impl<ST: SignableTransaction> Signers<ST> {
let mut to_cleanup = db::ToCleanup::get(txn).unwrap_or(vec![]); let mut to_cleanup = db::ToCleanup::get(txn).unwrap_or(vec![]);
to_cleanup.push((session, external_key.to_bytes().as_ref().to_vec())); to_cleanup.push((session, external_key.to_bytes().as_ref().to_vec()));
db::ToCleanup::set(txn, &to_cleanup); db::ToCleanup::set(txn, &to_cleanup);
}
// TODO: Handle all of the following cleanup on a task /// Queue handling a message.
/* ///
// Kill the tasks /// This is a cheap call and able to be done inline with a higher-level loop.
if let Some(tasks) = self.tasks.remove(&session) { pub fn queue_message(&mut self, txn: &mut impl DbTxn, message: &CoordinatorMessage) {
for task in tasks { let sign_id = message.sign_id();
task.close().await; let tasks = self.tasks.get(&sign_id.session);
match sign_id.id {
VariantSignId::Cosign(_) => {
db::CoordinatorToCosignerMessages::send(txn, sign_id.session, message);
if let Some(tasks) = tasks { tasks.cosigner.run_now(); }
}
VariantSignId::Batch(_) => {
db::CoordinatorToBatchSignerMessages::send(txn, sign_id.session, message);
if let Some(tasks) = tasks { tasks.batch.run_now(); }
}
VariantSignId::SlashReport(_) => {
db::CoordinatorToSlashReportSignerMessages::send(txn, sign_id.session, message);
if let Some(tasks) = tasks { tasks.slash_report.run_now(); }
}
VariantSignId::Transaction(_) => {
db::CoordinatorToTransactionSignerMessages::send(txn, sign_id.session, message);
if let Some(tasks) = tasks { tasks.transaction.run_now(); }
} }
} }
// Drain the transactions to sign
// Presumably, TransactionsToSign will be fully populated before retiry occurs, making this
// perfect in not leaving any pending blobs behind
while TransactionsToSign::<ST>::try_recv(txn, external_key).is_some() {}
// Drain our DB channels
while db::CompletedEventualitiesForEachKey::try_recv(txn, session).is_some() {}
while db::CoordinatorToTransactionSignerMessages::try_recv(txn, session).is_some() {}
while db::TransactionSignerToCoordinatorMessages::try_recv(txn, session).is_some() {}
while db::CoordinatorToBatchSignerMessages::try_recv(txn, session).is_some() {}
while db::BatchSignerToCoordinatorMessages::try_recv(txn, session).is_some() {}
while db::CoordinatorToSlashReportSignerMessages::try_recv(txn, session).is_some() {}
while db::SlashReportSignerToCoordinatorMessages::try_recv(txn, session).is_some() {}
while db::CoordinatorToCosignerMessages::try_recv(txn, session).is_some() {}
while db::CosignerToCoordinatorMessages::try_recv(txn, session).is_some() {}
*/
} }
} }
/*
// The signers used by a Processor, key-scoped.
struct KeySigners<D: Db, T: Clone + PreprocessMachine> {
transaction: AttemptManager<D, T>,
substrate: AttemptManager<D, AlgorithmMachine<Ristretto, Schnorrkel>>,
cosigner: AttemptManager<D, AlgorithmMachine<Ristretto, Schnorrkel>>,
}
/// The signers used by a protocol.
pub struct Signers<D: Db, T: Clone + PreprocessMachine>(HashMap<Vec<u8>, KeySigners<D, T>>);
impl<D: Db, T: Clone + PreprocessMachine> Signers<D, T> {
/// Create a new set of signers.
pub fn new(db: D) -> Self {
// TODO: Load the registered keys
// TODO: Load the transactions being signed
// TODO: Load the batches being signed
todo!("TODO")
}
/// Register a transaction to sign.
pub fn sign_transaction(&mut self) -> Vec<ProcessorMessage> {
todo!("TODO")
}
/// Mark a transaction as signed.
pub fn signed_transaction(&mut self) { todo!("TODO") }
/// Register a batch to sign.
pub fn sign_batch(&mut self, key: KeyFor<S>, batch: Batch) -> Vec<ProcessorMessage> {
todo!("TODO")
}
/// Mark a batch as signed.
pub fn signed_batch(&mut self, batch: u32) { todo!("TODO") }
/// Register a slash report to sign.
pub fn sign_slash_report(&mut self) -> Vec<ProcessorMessage> {
todo!("TODO")
}
/// Mark a slash report as signed.
pub fn signed_slash_report(&mut self) { todo!("TODO") }
/// Start a cosigning protocol.
pub fn cosign(&mut self) { todo!("TODO") }
/// Handle a message for a signing protocol.
pub fn handle(&mut self, msg: CoordinatorMessage) -> Vec<ProcessorMessage> {
todo!("TODO")
}
}
*/

View file

@ -3,31 +3,28 @@ use std::{
time::{Duration, Instant}, time::{Duration, Instant},
}; };
use frost::{dkg::ThresholdKeys, sign::PreprocessMachine}; use frost::dkg::ThresholdKeys;
use serai_validator_sets_primitives::Session; use serai_validator_sets_primitives::Session;
use serai_db::{DbTxn, Db}; use serai_db::{DbTxn, Db};
use messages::sign::VariantSignId;
use primitives::task::ContinuallyRan; use primitives::task::ContinuallyRan;
use scheduler::{Transaction, SignableTransaction, TransactionsToSign}; use scheduler::{Transaction, SignableTransaction, TransactionFor, TransactionsToSign};
use scanner::CompletedEventualities;
use frost_attempt_manager::*; use frost_attempt_manager::*;
use crate::{ use crate::{
db::{ db::{CoordinatorToTransactionSignerMessages, TransactionSignerToCoordinatorMessages},
CoordinatorToTransactionSignerMessages, TransactionSignerToCoordinatorMessages,
CompletedEventualitiesForEachKey,
},
TransactionPublisher, TransactionPublisher,
}; };
mod db; mod db;
use db::*; use db::*;
type TransactionFor<ST> =
<<ST as SignableTransaction>::PreprocessMachine as PreprocessMachine>::Signature;
// Fetches transactions to sign and signs them. // Fetches transactions to sign and signs them.
pub(crate) struct TransactionTask< pub(crate) struct TransactionTask<
D: Db, D: Db,
@ -76,7 +73,7 @@ impl<D: Db, ST: SignableTransaction, P: TransactionPublisher<TransactionFor<ST>>
for keys in &keys { for keys in &keys {
machines.push(signable_transaction.clone().sign(keys.clone())); machines.push(signable_transaction.clone().sign(keys.clone()));
} }
attempt_manager.register(tx, machines); attempt_manager.register(VariantSignId::Transaction(tx), machines);
} }
Self { Self {
@ -123,7 +120,7 @@ impl<D: Db, ST: SignableTransaction, P: TransactionPublisher<TransactionFor<ST>>
for keys in &self.keys { for keys in &self.keys {
machines.push(tx.clone().sign(keys.clone())); machines.push(tx.clone().sign(keys.clone()));
} }
for msg in self.attempt_manager.register(tx.id(), machines) { for msg in self.attempt_manager.register(VariantSignId::Transaction(tx.id()), machines) {
TransactionSignerToCoordinatorMessages::send(&mut txn, self.session, &msg); TransactionSignerToCoordinatorMessages::send(&mut txn, self.session, &msg);
} }
@ -133,28 +130,42 @@ impl<D: Db, ST: SignableTransaction, P: TransactionPublisher<TransactionFor<ST>>
// Check for completed Eventualities (meaning we should no longer sign for these transactions) // Check for completed Eventualities (meaning we should no longer sign for these transactions)
loop { loop {
let mut txn = self.db.txn(); let mut txn = self.db.txn();
let Some(id) = CompletedEventualitiesForEachKey::try_recv(&mut txn, self.session) else { let Some(id) = CompletedEventualities::try_recv(&mut txn, &self.keys[0].group_key()) else {
break; break;
}; };
/*
We may have yet to register this signing protocol.
While `TransactionsToSign` is populated before `CompletedEventualities`, we could
theoretically have `TransactionsToSign` populated with a new transaction _while iterating
over `CompletedEventualities`_, and then have `CompletedEventualities` populated. In that
edge case, we will see the completion notification before we see the transaction.
In such a case, we break (dropping the txn, re-queueing the completion notification). On
the task's next iteration, we'll process the transaction from `TransactionsToSign` and be
able to make progress.
*/
if !self.active_signing_protocols.remove(&id) {
break;
}
iterated = true; iterated = true;
// This may or may not be an ID this key was responsible for // Since it was, remove this as an active signing protocol
if self.active_signing_protocols.remove(&id) { ActiveSigningProtocols::set(
// Since it was, remove this as an active signing protocol &mut txn,
ActiveSigningProtocols::set( self.session,
&mut txn, &self.active_signing_protocols.iter().copied().collect(),
self.session, );
&self.active_signing_protocols.iter().copied().collect(), // Clean up the database
); SerializedSignableTransactions::del(&mut txn, id);
// Clean up the database SerializedTransactions::del(&mut txn, id);
SerializedSignableTransactions::del(&mut txn, id);
SerializedTransactions::del(&mut txn, id); // We retire with a txn so we either successfully flag this Eventuality as completed, and
// won't re-register it (making this retire safe), or we don't flag it, meaning we will
// re-register it, yet that's safe as we have yet to retire it
self.attempt_manager.retire(&mut txn, VariantSignId::Transaction(id));
// We retire with a txn so we either successfully flag this Eventuality as completed, and
// won't re-register it (making this retire safe), or we don't flag it, meaning we will
// re-register it, yet that's safe as we have yet to retire it
self.attempt_manager.retire(&mut txn, id);
}
txn.commit(); txn.commit();
} }
@ -178,7 +189,14 @@ impl<D: Db, ST: SignableTransaction, P: TransactionPublisher<TransactionFor<ST>>
{ {
let mut buf = Vec::with_capacity(256); let mut buf = Vec::with_capacity(256);
signed_tx.write(&mut buf).unwrap(); signed_tx.write(&mut buf).unwrap();
SerializedTransactions::set(&mut txn, id, &buf); SerializedTransactions::set(
&mut txn,
match id {
VariantSignId::Transaction(id) => id,
_ => panic!("TransactionTask signed a non-transaction"),
},
&buf,
);
} }
self self