mirror of
https://github.com/serai-dex/serai.git
synced 2025-02-02 19:26:26 +00:00
Further work on transaction signing
This commit is contained in:
parent
b62fc3a1fa
commit
a353f9e2da
13 changed files with 299 additions and 74 deletions
135
Cargo.lock
generated
135
Cargo.lock
generated
|
@ -8120,6 +8120,29 @@ dependencies = [
|
||||||
"sp-runtime",
|
"sp-runtime",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "serai-bitcoin-processor"
|
||||||
|
version = "0.1.0"
|
||||||
|
dependencies = [
|
||||||
|
"async-trait",
|
||||||
|
"bitcoin-serai",
|
||||||
|
"borsh",
|
||||||
|
"const-hex",
|
||||||
|
"env_logger",
|
||||||
|
"hex",
|
||||||
|
"k256",
|
||||||
|
"log",
|
||||||
|
"parity-scale-codec",
|
||||||
|
"secp256k1",
|
||||||
|
"serai-db",
|
||||||
|
"serai-env",
|
||||||
|
"serai-message-queue",
|
||||||
|
"serai-processor-messages",
|
||||||
|
"serde_json",
|
||||||
|
"tokio",
|
||||||
|
"zalloc",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "serai-client"
|
name = "serai-client"
|
||||||
version = "0.1.0"
|
version = "0.1.0"
|
||||||
|
@ -8315,6 +8338,28 @@ dependencies = [
|
||||||
name = "serai-env"
|
name = "serai-env"
|
||||||
version = "0.1.0"
|
version = "0.1.0"
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "serai-ethereum-processor"
|
||||||
|
version = "0.1.0"
|
||||||
|
dependencies = [
|
||||||
|
"async-trait",
|
||||||
|
"borsh",
|
||||||
|
"const-hex",
|
||||||
|
"env_logger",
|
||||||
|
"ethereum-serai",
|
||||||
|
"hex",
|
||||||
|
"k256",
|
||||||
|
"log",
|
||||||
|
"parity-scale-codec",
|
||||||
|
"serai-db",
|
||||||
|
"serai-env",
|
||||||
|
"serai-message-queue",
|
||||||
|
"serai-processor-messages",
|
||||||
|
"serde_json",
|
||||||
|
"tokio",
|
||||||
|
"zalloc",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "serai-ethereum-relayer"
|
name = "serai-ethereum-relayer"
|
||||||
version = "0.1.0"
|
version = "0.1.0"
|
||||||
|
@ -8343,7 +8388,6 @@ dependencies = [
|
||||||
"serai-coordinator-tests",
|
"serai-coordinator-tests",
|
||||||
"serai-docker-tests",
|
"serai-docker-tests",
|
||||||
"serai-message-queue-tests",
|
"serai-message-queue-tests",
|
||||||
"serai-processor",
|
|
||||||
"serai-processor-tests",
|
"serai-processor-tests",
|
||||||
"serde",
|
"serde",
|
||||||
"serde_json",
|
"serde_json",
|
||||||
|
@ -8459,6 +8503,29 @@ dependencies = [
|
||||||
"zeroize",
|
"zeroize",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "serai-monero-processor"
|
||||||
|
version = "0.1.0"
|
||||||
|
dependencies = [
|
||||||
|
"async-trait",
|
||||||
|
"borsh",
|
||||||
|
"const-hex",
|
||||||
|
"dalek-ff-group",
|
||||||
|
"env_logger",
|
||||||
|
"hex",
|
||||||
|
"log",
|
||||||
|
"monero-simple-request-rpc",
|
||||||
|
"monero-wallet",
|
||||||
|
"parity-scale-codec",
|
||||||
|
"serai-db",
|
||||||
|
"serai-env",
|
||||||
|
"serai-message-queue",
|
||||||
|
"serai-processor-messages",
|
||||||
|
"serde_json",
|
||||||
|
"tokio",
|
||||||
|
"zalloc",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "serai-no-std-tests"
|
name = "serai-no-std-tests"
|
||||||
version = "0.1.0"
|
version = "0.1.0"
|
||||||
|
@ -8558,47 +8625,6 @@ dependencies = [
|
||||||
"zeroize",
|
"zeroize",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "serai-processor"
|
|
||||||
version = "0.1.0"
|
|
||||||
dependencies = [
|
|
||||||
"async-trait",
|
|
||||||
"bitcoin-serai",
|
|
||||||
"borsh",
|
|
||||||
"ciphersuite",
|
|
||||||
"const-hex",
|
|
||||||
"dalek-ff-group",
|
|
||||||
"dkg",
|
|
||||||
"dockertest",
|
|
||||||
"ec-divisors",
|
|
||||||
"env_logger",
|
|
||||||
"ethereum-serai",
|
|
||||||
"flexible-transcript",
|
|
||||||
"frost-schnorrkel",
|
|
||||||
"hex",
|
|
||||||
"k256",
|
|
||||||
"log",
|
|
||||||
"modular-frost",
|
|
||||||
"monero-simple-request-rpc",
|
|
||||||
"monero-wallet",
|
|
||||||
"parity-scale-codec",
|
|
||||||
"rand_chacha",
|
|
||||||
"rand_core",
|
|
||||||
"secp256k1",
|
|
||||||
"serai-client",
|
|
||||||
"serai-db",
|
|
||||||
"serai-docker-tests",
|
|
||||||
"serai-env",
|
|
||||||
"serai-message-queue",
|
|
||||||
"serai-processor-messages",
|
|
||||||
"serde_json",
|
|
||||||
"sp-application-crypto",
|
|
||||||
"thiserror",
|
|
||||||
"tokio",
|
|
||||||
"zalloc",
|
|
||||||
"zeroize",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "serai-processor-frost-attempt-manager"
|
name = "serai-processor-frost-attempt-manager"
|
||||||
version = "0.1.0"
|
version = "0.1.0"
|
||||||
|
@ -8676,6 +8702,7 @@ dependencies = [
|
||||||
"serai-in-instructions-primitives",
|
"serai-in-instructions-primitives",
|
||||||
"serai-primitives",
|
"serai-primitives",
|
||||||
"serai-processor-primitives",
|
"serai-processor-primitives",
|
||||||
|
"serai-processor-scheduler-primitives",
|
||||||
"tokio",
|
"tokio",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
@ -8684,11 +8711,32 @@ name = "serai-processor-scheduler-primitives"
|
||||||
version = "0.1.0"
|
version = "0.1.0"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"borsh",
|
"borsh",
|
||||||
"group",
|
"ciphersuite",
|
||||||
|
"modular-frost",
|
||||||
"parity-scale-codec",
|
"parity-scale-codec",
|
||||||
"serai-db",
|
"serai-db",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "serai-processor-signers"
|
||||||
|
version = "0.1.0"
|
||||||
|
dependencies = [
|
||||||
|
"async-trait",
|
||||||
|
"borsh",
|
||||||
|
"ciphersuite",
|
||||||
|
"log",
|
||||||
|
"modular-frost",
|
||||||
|
"parity-scale-codec",
|
||||||
|
"serai-db",
|
||||||
|
"serai-processor-frost-attempt-manager",
|
||||||
|
"serai-processor-messages",
|
||||||
|
"serai-processor-primitives",
|
||||||
|
"serai-processor-scanner",
|
||||||
|
"serai-processor-scheduler-primitives",
|
||||||
|
"serai-validator-sets-primitives",
|
||||||
|
"tokio",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "serai-processor-tests"
|
name = "serai-processor-tests"
|
||||||
version = "0.1.0"
|
version = "0.1.0"
|
||||||
|
@ -8711,7 +8759,6 @@ dependencies = [
|
||||||
"serai-docker-tests",
|
"serai-docker-tests",
|
||||||
"serai-message-queue",
|
"serai-message-queue",
|
||||||
"serai-message-queue-tests",
|
"serai-message-queue-tests",
|
||||||
"serai-processor",
|
|
||||||
"serai-processor-messages",
|
"serai-processor-messages",
|
||||||
"serde_json",
|
"serde_json",
|
||||||
"tokio",
|
"tokio",
|
||||||
|
|
|
@ -80,10 +80,15 @@ impl<D: Db, M: Clone + PreprocessMachine> SigningProtocol<D, M> {
|
||||||
|
|
||||||
We avoid this by saving to the DB we preprocessed before sending our preprocessed, and only
|
We avoid this by saving to the DB we preprocessed before sending our preprocessed, and only
|
||||||
keeping our preprocesses for this instance of the processor. Accordingly, on reboot, we will
|
keeping our preprocesses for this instance of the processor. Accordingly, on reboot, we will
|
||||||
flag the prior preprocess and not send new preprocesses.
|
flag the prior preprocess and not send new preprocesses. This does require our own DB
|
||||||
|
transaction (to ensure we save to the DB we preprocessed before yielding the preprocess
|
||||||
|
messages).
|
||||||
|
|
||||||
We also won't send the share we were supposed to, unfortunately, yet caching/reloading the
|
We also won't send the share we were supposed to, unfortunately, yet caching/reloading the
|
||||||
preprocess has enough safety issues it isn't worth the headache.
|
preprocess has enough safety issues it isn't worth the headache.
|
||||||
|
|
||||||
|
Since we bind a signing attempt to the lifetime of the application, we're also safe against
|
||||||
|
nonce reuse (as the state machines enforce single-use and we never reuse a preprocess).
|
||||||
*/
|
*/
|
||||||
{
|
{
|
||||||
let mut txn = self.db.txn();
|
let mut txn = self.db.txn();
|
||||||
|
|
|
@ -65,6 +65,10 @@ impl<D: Db, M: Clone + PreprocessMachine> AttemptManager<D, M> {
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Handle a message for a signing protocol.
|
/// Handle a message for a signing protocol.
|
||||||
|
///
|
||||||
|
/// Handling a message multiple times is safe and will cause subsequent calls to return
|
||||||
|
/// `Response::Messages(vec![])`. Handling a message for a signing protocol which isn't being
|
||||||
|
/// worked on (potentially due to rebooting) will also return `Response::Messages(vec![])`.
|
||||||
pub fn handle(&mut self, msg: CoordinatorMessage) -> Response<M> {
|
pub fn handle(&mut self, msg: CoordinatorMessage) -> Response<M> {
|
||||||
match msg {
|
match msg {
|
||||||
CoordinatorMessage::Preprocesses { id, preprocesses } => {
|
CoordinatorMessage::Preprocesses { id, preprocesses } => {
|
||||||
|
|
|
@ -7,7 +7,7 @@ pub trait Eventuality: Sized + Send + Sync {
|
||||||
/// The type used to identify a received output.
|
/// The type used to identify a received output.
|
||||||
type OutputId: Id;
|
type OutputId: Id;
|
||||||
|
|
||||||
/// The ID of the transaction this Eventuality is for.
|
/// The ID of the SignableTransaction this Eventuality is for.
|
||||||
///
|
///
|
||||||
/// This is an internal ID arbitrarily definable so long as it's unique.
|
/// This is an internal ID arbitrarily definable so long as it's unique.
|
||||||
fn id(&self) -> [u8; 32];
|
fn id(&self) -> [u8; 32];
|
||||||
|
|
|
@ -520,3 +520,26 @@ impl SubstrateToEventualityDb {
|
||||||
Burns::try_recv(txn, acknowledged_block)
|
Burns::try_recv(txn, acknowledged_block)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
mod _completed_eventualities {
|
||||||
|
use serai_db::{Get, DbTxn, create_db, db_channel};
|
||||||
|
|
||||||
|
db_channel! {
|
||||||
|
ScannerPublic {
|
||||||
|
CompletedEventualities: (empty_key: ()) -> [u8; 32],
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// The IDs of completed Eventualities found on-chain, within a finalized block.
|
||||||
|
pub struct CompletedEventualities<S: ScannerFeed>(PhantomData<S>);
|
||||||
|
impl<S: ScannerFeed> CompletedEventualities<S> {
|
||||||
|
pub(crate) fn send(txn: &mut impl DbTxn, id: [u8; 32]) {
|
||||||
|
_completed_eventualities::CompletedEventualities::send(txn, (), &id);
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Receive the ID of a completed Eventuality.
|
||||||
|
pub fn try_recv(txn: &mut impl DbTxn) -> Option<[u8; 32]> {
|
||||||
|
_completed_eventualities::CompletedEventualities::try_recv(txn, ())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
|
@ -14,7 +14,7 @@ use crate::{
|
||||||
ScanToEventualityDb,
|
ScanToEventualityDb,
|
||||||
},
|
},
|
||||||
BlockExt, ScannerFeed, KeyFor, AddressFor, OutputFor, EventualityFor, SchedulerUpdate, Scheduler,
|
BlockExt, ScannerFeed, KeyFor, AddressFor, OutputFor, EventualityFor, SchedulerUpdate, Scheduler,
|
||||||
sort_outputs,
|
CompletedEventualities, sort_outputs,
|
||||||
scan::{next_to_scan_for_outputs_block, queue_output_until_block},
|
scan::{next_to_scan_for_outputs_block, queue_output_until_block},
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -292,8 +292,13 @@ impl<D: Db, S: ScannerFeed, Sch: Scheduler<S>> ContinuallyRan for EventualityTas
|
||||||
completed_eventualities
|
completed_eventualities
|
||||||
};
|
};
|
||||||
|
|
||||||
for tx in completed_eventualities.keys() {
|
for (tx, eventuality) in &completed_eventualities {
|
||||||
log::info!("eventuality resolved by {}", hex::encode(tx.as_ref()));
|
log::info!(
|
||||||
|
"eventuality {} resolved by {}",
|
||||||
|
hex::encode(eventuality.id()),
|
||||||
|
hex::encode(tx.as_ref())
|
||||||
|
);
|
||||||
|
CompletedEventualities::<S>::send(&mut txn, eventuality.id());
|
||||||
}
|
}
|
||||||
|
|
||||||
// Fetch all non-External outputs
|
// Fetch all non-External outputs
|
||||||
|
|
|
@ -21,6 +21,7 @@ pub use lifetime::LifetimeStage;
|
||||||
|
|
||||||
// Database schema definition and associated functions.
|
// Database schema definition and associated functions.
|
||||||
mod db;
|
mod db;
|
||||||
|
pub use db::CompletedEventualities;
|
||||||
// Task to index the blockchain, ensuring we don't reorganize finalized blocks.
|
// Task to index the blockchain, ensuring we don't reorganize finalized blocks.
|
||||||
mod index;
|
mod index;
|
||||||
// Scans blocks for received coins.
|
// Scans blocks for received coins.
|
||||||
|
@ -170,6 +171,10 @@ pub type EventualityFor<S> = <<S as ScannerFeed>::Block as Block>::Eventuality;
|
||||||
/// The block type for this ScannerFeed.
|
/// The block type for this ScannerFeed.
|
||||||
pub type BlockFor<S> = <S as ScannerFeed>::Block;
|
pub type BlockFor<S> = <S as ScannerFeed>::Block;
|
||||||
|
|
||||||
|
/// An object usable to publish a Batch.
|
||||||
|
// This will presumably be the Batch signer defined in `serai-processor-signers` or a test shim.
|
||||||
|
// It could also be some app-layer database for the purpose of verifying the Batches published to
|
||||||
|
// Serai.
|
||||||
#[async_trait::async_trait]
|
#[async_trait::async_trait]
|
||||||
pub trait BatchPublisher: 'static + Send + Sync {
|
pub trait BatchPublisher: 'static + Send + Sync {
|
||||||
/// An error encountered when publishing the Batch.
|
/// An error encountered when publishing the Batch.
|
||||||
|
|
|
@ -20,7 +20,8 @@ ignored = ["scale", "borsh"]
|
||||||
workspace = true
|
workspace = true
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
group = { version = "0.13", default-features = false }
|
ciphersuite = { path = "../../../crypto/ciphersuite", default-features = false, features = ["std"] }
|
||||||
|
frost = { package = "modular-frost", path = "../../../crypto/frost", default-features = false }
|
||||||
|
|
||||||
scale = { package = "parity-scale-codec", version = "3", default-features = false, features = ["std"] }
|
scale = { package = "parity-scale-codec", version = "3", default-features = false, features = ["std"] }
|
||||||
borsh = { version = "1", default-features = false, features = ["std", "derive", "de_strict_order"] }
|
borsh = { version = "1", default-features = false, features = ["std", "derive", "de_strict_order"] }
|
||||||
|
|
|
@ -5,16 +5,25 @@
|
||||||
use core::marker::PhantomData;
|
use core::marker::PhantomData;
|
||||||
use std::io;
|
use std::io;
|
||||||
|
|
||||||
use group::GroupEncoding;
|
use ciphersuite::{group::GroupEncoding, Ciphersuite};
|
||||||
|
use frost::{dkg::ThresholdKeys, sign::PreprocessMachine};
|
||||||
|
|
||||||
use serai_db::DbTxn;
|
use serai_db::DbTxn;
|
||||||
|
|
||||||
|
/// A transaction.
|
||||||
|
pub trait Transaction: Sized {
|
||||||
|
/// Read a `Transaction`.
|
||||||
|
fn read(reader: &mut impl io::Read) -> io::Result<Self>;
|
||||||
|
/// Write a `Transaction`.
|
||||||
|
fn write(&self, writer: &mut impl io::Write) -> io::Result<()>;
|
||||||
|
}
|
||||||
|
|
||||||
/// A signable transaction.
|
/// A signable transaction.
|
||||||
pub trait SignableTransaction: 'static + Sized + Send + Sync + Clone {
|
pub trait SignableTransaction: 'static + Sized + Send + Sync + Clone {
|
||||||
/// The ciphersuite used to sign this transaction.
|
/// The ciphersuite used to sign this transaction.
|
||||||
type Ciphersuite: Cuphersuite;
|
type Ciphersuite: Ciphersuite;
|
||||||
/// The preprocess machine for the signing protocol for this transaction.
|
/// The preprocess machine for the signing protocol for this transaction.
|
||||||
type PreprocessMachine: PreprocessMachine;
|
type PreprocessMachine: Clone + PreprocessMachine<Signature: Send>;
|
||||||
|
|
||||||
/// Read a `SignableTransaction`.
|
/// Read a `SignableTransaction`.
|
||||||
fn read(reader: &mut impl io::Read) -> io::Result<Self>;
|
fn read(reader: &mut impl io::Read) -> io::Result<Self>;
|
||||||
|
|
|
@ -20,13 +20,23 @@ ignored = ["borsh", "scale"]
|
||||||
workspace = true
|
workspace = true
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
group = { version = "0.13", default-features = false }
|
async-trait = { version = "0.1", default-features = false }
|
||||||
|
|
||||||
|
ciphersuite = { path = "../../crypto/ciphersuite", default-features = false, features = ["std"] }
|
||||||
|
frost = { package = "modular-frost", path = "../../crypto/frost", default-features = false }
|
||||||
|
|
||||||
|
scale = { package = "parity-scale-codec", version = "3", default-features = false, features = ["std"] }
|
||||||
|
borsh = { version = "1", default-features = false, features = ["std", "derive", "de_strict_order"] }
|
||||||
|
|
||||||
|
serai-validator-sets-primitives = { path = "../../substrate/validator-sets/primitives", default-features = false, features = ["std"] }
|
||||||
|
|
||||||
|
serai-db = { path = "../../common/db" }
|
||||||
log = { version = "0.4", default-features = false, features = ["std"] }
|
log = { version = "0.4", default-features = false, features = ["std"] }
|
||||||
tokio = { version = "1", default-features = false, features = ["rt-multi-thread", "sync", "time", "macros"] }
|
tokio = { version = "1", default-features = false, features = ["rt-multi-thread", "sync", "time", "macros"] }
|
||||||
|
|
||||||
|
messages = { package = "serai-processor-messages", path = "../messages" }
|
||||||
primitives = { package = "serai-processor-primitives", path = "../primitives" }
|
primitives = { package = "serai-processor-primitives", path = "../primitives" }
|
||||||
scanner = { package = "serai-processor-scanner", path = "../scanner" }
|
scanner = { package = "serai-processor-scanner", path = "../scanner" }
|
||||||
scheduler = { package = "serai-scheduler-primitives", path = "../scheduler/primitives" }
|
scheduler = { package = "serai-processor-scheduler-primitives", path = "../scheduler/primitives" }
|
||||||
|
|
||||||
frost-attempt-manager = { package = "serai-processor-frost-attempt-manager", path = "../frost-attempt-manager" }
|
frost-attempt-manager = { package = "serai-processor-frost-attempt-manager", path = "../frost-attempt-manager" }
|
||||||
|
|
27
processor/signers/src/db.rs
Normal file
27
processor/signers/src/db.rs
Normal file
|
@ -0,0 +1,27 @@
|
||||||
|
use serai_validator_sets_primitives::Session;
|
||||||
|
|
||||||
|
use serai_db::{Get, DbTxn, create_db, db_channel};
|
||||||
|
|
||||||
|
use messages::sign::{ProcessorMessage, CoordinatorMessage};
|
||||||
|
|
||||||
|
db_channel! {
|
||||||
|
SignersGlobal {
|
||||||
|
// CompletedEventualities needs to be handled by each signer, meaning we need to turn its
|
||||||
|
// effective spsc into a spmc. We do this by duplicating its message for all keys we're
|
||||||
|
// signing for.
|
||||||
|
// TODO: Populate from CompletedEventualities
|
||||||
|
CompletedEventualitiesForEachKey: (session: Session) -> [u8; 32],
|
||||||
|
|
||||||
|
CoordinatorToTransactionSignerMessages: (session: Session) -> CoordinatorMessage,
|
||||||
|
TransactionSignerToCoordinatorMessages: (session: Session) -> ProcessorMessage,
|
||||||
|
|
||||||
|
CoordinatorToBatchSignerMessages: (session: Session) -> CoordinatorMessage,
|
||||||
|
BatchSignerToCoordinatorMessages: (session: Session) -> ProcessorMessage,
|
||||||
|
|
||||||
|
CoordinatorToSlashReportSignerMessages: (session: Session) -> CoordinatorMessage,
|
||||||
|
SlashReportSignerToCoordinatorMessages: (session: Session) -> ProcessorMessage,
|
||||||
|
|
||||||
|
CoordinatorToCosignerMessages: (session: Session) -> CoordinatorMessage,
|
||||||
|
CosignerToCoordinatorMessages: (session: Session) -> ProcessorMessage,
|
||||||
|
}
|
||||||
|
}
|
|
@ -2,8 +2,38 @@
|
||||||
#![doc = include_str!("../README.md")]
|
#![doc = include_str!("../README.md")]
|
||||||
#![deny(missing_docs)]
|
#![deny(missing_docs)]
|
||||||
|
|
||||||
|
use core::fmt::Debug;
|
||||||
|
|
||||||
|
use frost::sign::PreprocessMachine;
|
||||||
|
|
||||||
|
use scheduler::SignableTransaction;
|
||||||
|
|
||||||
|
pub(crate) mod db;
|
||||||
|
|
||||||
mod transaction;
|
mod transaction;
|
||||||
|
|
||||||
|
/// An object capable of publishing a transaction.
|
||||||
|
#[async_trait::async_trait]
|
||||||
|
pub trait TransactionPublisher<S: SignableTransaction>: 'static + Send + Sync {
|
||||||
|
/// An error encountered when publishing a transaction.
|
||||||
|
///
|
||||||
|
/// This MUST be an ephemeral error. Retrying publication MUST eventually resolve without manual
|
||||||
|
/// intervention/changing the arguments.
|
||||||
|
///
|
||||||
|
/// The transaction already being present in the mempool/on-chain SHOULD NOT be considered an
|
||||||
|
/// error.
|
||||||
|
type EphemeralError: Debug;
|
||||||
|
|
||||||
|
/// Publish a transaction.
|
||||||
|
///
|
||||||
|
/// This will be called multiple times, with the same transaction, until the transaction is
|
||||||
|
/// confirmed on-chain.
|
||||||
|
async fn publish(
|
||||||
|
&self,
|
||||||
|
tx: <S::PreprocessMachine as PreprocessMachine>::Signature,
|
||||||
|
) -> Result<(), Self::EphemeralError>;
|
||||||
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
// The signers used by a Processor, key-scoped.
|
// The signers used by a Processor, key-scoped.
|
||||||
struct KeySigners<D: Db, T: Clone + PreprocessMachine> {
|
struct KeySigners<D: Db, T: Clone + PreprocessMachine> {
|
||||||
|
|
|
@ -1,69 +1,128 @@
|
||||||
use serai_db::{Get, DbTxn, Db};
|
use frost::dkg::ThresholdKeys;
|
||||||
|
|
||||||
|
use serai_validator_sets_primitives::Session;
|
||||||
|
|
||||||
|
use serai_db::{DbTxn, Db};
|
||||||
|
|
||||||
use primitives::task::ContinuallyRan;
|
use primitives::task::ContinuallyRan;
|
||||||
use scanner::ScannerFeed;
|
use scheduler::{SignableTransaction, TransactionsToSign};
|
||||||
use scheduler::TransactionsToSign;
|
use scanner::{ScannerFeed, Scheduler};
|
||||||
|
|
||||||
|
use frost_attempt_manager::*;
|
||||||
|
|
||||||
|
use crate::{
|
||||||
|
db::{
|
||||||
|
CoordinatorToTransactionSignerMessages, TransactionSignerToCoordinatorMessages,
|
||||||
|
CompletedEventualitiesForEachKey,
|
||||||
|
},
|
||||||
|
TransactionPublisher,
|
||||||
|
};
|
||||||
|
|
||||||
mod db;
|
mod db;
|
||||||
use db::IndexDb;
|
|
||||||
|
|
||||||
// Fetches transactions to sign and signs them.
|
// Fetches transactions to sign and signs them.
|
||||||
pub(crate) struct TransactionTask<D: Db, S: ScannerFeed, Sch: Scheduler> {
|
pub(crate) struct TransactionTask<
|
||||||
|
D: Db,
|
||||||
|
S: ScannerFeed,
|
||||||
|
Sch: Scheduler<S>,
|
||||||
|
P: TransactionPublisher<Sch::SignableTransaction>,
|
||||||
|
> {
|
||||||
db: D,
|
db: D,
|
||||||
keys: ThresholdKeys<<Sch::SignableTransaction as SignableTransaction>::Ciphersuite>,
|
session: Session,
|
||||||
|
keys: Vec<ThresholdKeys<<Sch::SignableTransaction as SignableTransaction>::Ciphersuite>>,
|
||||||
attempt_manager:
|
attempt_manager:
|
||||||
AttemptManager<D, <Sch::SignableTransaction as SignableTransaction>::PreprocessMachine>,
|
AttemptManager<D, <Sch::SignableTransaction as SignableTransaction>::PreprocessMachine>,
|
||||||
|
publisher: P,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<D: Db, S: ScannerFeed> TransactionTask<D, S> {
|
impl<D: Db, S: ScannerFeed, Sch: Scheduler<S>, P: TransactionPublisher<Sch::SignableTransaction>>
|
||||||
pub(crate) async fn new(
|
TransactionTask<D, S, Sch, P>
|
||||||
|
{
|
||||||
|
pub(crate) fn new(
|
||||||
db: D,
|
db: D,
|
||||||
keys: ThresholdKeys<<Sch::SignableTransaction as SignableTransaction>::Ciphersuite>,
|
session: Session,
|
||||||
|
keys: Vec<ThresholdKeys<<Sch::SignableTransaction as SignableTransaction>::Ciphersuite>>,
|
||||||
|
publisher: P,
|
||||||
) -> Self {
|
) -> Self {
|
||||||
Self { db, keys, attempt_manager: AttemptManager::new() }
|
let attempt_manager = AttemptManager::new(
|
||||||
|
db.clone(),
|
||||||
|
session,
|
||||||
|
keys.first().expect("creating a transaction signer with 0 keys").params().i(),
|
||||||
|
);
|
||||||
|
Self { db, session, keys, attempt_manager, publisher }
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[async_trait::async_trait]
|
#[async_trait::async_trait]
|
||||||
impl<D: Db, S: ScannerFeed> ContinuallyRan for TransactionTask<D, S> {
|
impl<D: Db, S: ScannerFeed, Sch: Scheduler<S>, P: TransactionPublisher<Sch::SignableTransaction>>
|
||||||
|
ContinuallyRan for TransactionTask<D, S, Sch, P>
|
||||||
|
{
|
||||||
async fn run_iteration(&mut self) -> Result<bool, String> {
|
async fn run_iteration(&mut self) -> Result<bool, String> {
|
||||||
let mut iterated = false;
|
let mut iterated = false;
|
||||||
|
|
||||||
// Check for new transactions to sign
|
// Check for new transactions to sign
|
||||||
loop {
|
loop {
|
||||||
let mut txn = self.db.txn();
|
let mut txn = self.db.txn();
|
||||||
let Some(tx) = TransactionsToSign::try_recv(&mut txn, self.key) else { break };
|
let Some(tx) = TransactionsToSign::<Sch::SignableTransaction>::try_recv(
|
||||||
|
&mut txn,
|
||||||
|
&self.keys[0].group_key(),
|
||||||
|
) else {
|
||||||
|
break;
|
||||||
|
};
|
||||||
iterated = true;
|
iterated = true;
|
||||||
|
|
||||||
let mut machines = Vec::with_capacity(self.keys.len());
|
let mut machines = Vec::with_capacity(self.keys.len());
|
||||||
for keys in &self.keys {
|
for keys in &self.keys {
|
||||||
machines.push(tx.clone().sign(keys.clone()));
|
machines.push(tx.clone().sign(keys.clone()));
|
||||||
}
|
}
|
||||||
let messages = self.attempt_manager.register(tx.id(), machines);
|
for msg in self.attempt_manager.register(tx.id(), machines) {
|
||||||
todo!("TODO");
|
TransactionSignerToCoordinatorMessages::send(&mut txn, self.session, &msg);
|
||||||
|
}
|
||||||
txn.commit();
|
txn.commit();
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check for completed Eventualities (meaning we should no longer sign for these transactions)
|
// Check for completed Eventualities (meaning we should no longer sign for these transactions)
|
||||||
loop {
|
loop {
|
||||||
let mut txn = self.db.txn();
|
let mut txn = self.db.txn();
|
||||||
let Some(tx) = CompletedEventualities::try_recv(&mut txn, self.key) else { break };
|
let Some(id) = CompletedEventualitiesForEachKey::try_recv(&mut txn, self.session) else {
|
||||||
|
break;
|
||||||
|
};
|
||||||
iterated = true;
|
iterated = true;
|
||||||
|
|
||||||
self.attempt_manager.retire(tx);
|
self.attempt_manager.retire(id);
|
||||||
|
// TODO: Stop rebroadcasting this transaction
|
||||||
txn.commit();
|
txn.commit();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Handle any messages sent to us
|
||||||
loop {
|
loop {
|
||||||
let mut txn = self.db.txn();
|
let mut txn = self.db.txn();
|
||||||
let Some(msg) = TransactionSignMessages::try_recv(&mut txn, self.key) else { break };
|
let Some(msg) = CoordinatorToTransactionSignerMessages::try_recv(&mut txn, self.session)
|
||||||
|
else {
|
||||||
|
break;
|
||||||
|
};
|
||||||
iterated = true;
|
iterated = true;
|
||||||
|
|
||||||
match self.attempt_manager.handle(msg) {
|
match self.attempt_manager.handle(msg) {
|
||||||
Response::Messages(messages) => todo!("TODO"),
|
Response::Messages(msgs) => {
|
||||||
Response::Signature(signature) => todo!("TODO"),
|
for msg in msgs {
|
||||||
|
TransactionSignerToCoordinatorMessages::send(&mut txn, self.session, &msg);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
Response::Signature(signed_tx) => {
|
||||||
|
// TODO: Save this TX to the DB
|
||||||
|
// TODO: Attempt publication every minute
|
||||||
|
// TODO: On boot, reload all TXs to rebroadcast
|
||||||
|
self
|
||||||
|
.publisher
|
||||||
|
.publish(signed_tx)
|
||||||
|
.await
|
||||||
|
.map_err(|e| format!("couldn't publish transaction: {e:?}"))?;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
txn.commit();
|
||||||
|
}
|
||||||
|
|
||||||
Ok(iterated)
|
Ok(iterated)
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in a new issue