mirror of
https://github.com/serai-dex/serai.git
synced 2025-12-08 12:19:24 +00:00
Fill out code for the rest of the Substrate events
This commit is contained in:
@@ -11,6 +11,20 @@ impl<D: Db> MainDb<D> {
|
|||||||
D::key(b"MAIN", dst, key)
|
D::key(b"MAIN", dst, key)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn substrate_block_key() -> Vec<u8> {
|
||||||
|
Self::main_key(b"substrate_block", [])
|
||||||
|
}
|
||||||
|
pub fn set_last_substrate_block(&mut self, block: u64) {
|
||||||
|
let mut txn = self.0.txn();
|
||||||
|
txn.put(Self::substrate_block_key(), block.to_le_bytes());
|
||||||
|
txn.commit();
|
||||||
|
}
|
||||||
|
pub fn last_substrate_block(&self) -> u64 {
|
||||||
|
u64::from_le_bytes(
|
||||||
|
self.0.get(Self::substrate_block_key()).unwrap_or(vec![0; 8]).try_into().unwrap(),
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
fn event_key(id: &[u8], index: u32) -> Vec<u8> {
|
fn event_key(id: &[u8], index: u32) -> Vec<u8> {
|
||||||
Self::main_key(b"event", [id, index.to_le_bytes().as_ref()].concat())
|
Self::main_key(b"event", [id, index.to_le_bytes().as_ref()].concat())
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,7 +1,12 @@
|
|||||||
#![allow(unused_variables)]
|
#![allow(unused_variables)]
|
||||||
|
#![allow(unreachable_code)]
|
||||||
|
#![allow(clippy::diverging_sub_expression)]
|
||||||
|
|
||||||
|
use zeroize::Zeroizing;
|
||||||
|
|
||||||
|
use ciphersuite::{group::ff::Field, Ciphersuite, Ristretto};
|
||||||
|
|
||||||
use serai_db::{Db, MemDb};
|
use serai_db::{Db, MemDb};
|
||||||
|
|
||||||
use serai_client::Serai;
|
use serai_client::Serai;
|
||||||
|
|
||||||
mod db;
|
mod db;
|
||||||
@@ -18,13 +23,19 @@ mod substrate;
|
|||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod tests;
|
mod tests;
|
||||||
|
|
||||||
async fn run<D: Db, P: P2p>(db: D, p2p: P, serai: Serai) {
|
async fn run<D: Db, P: P2p>(
|
||||||
|
db: D,
|
||||||
|
key: Zeroizing<<Ristretto as Ciphersuite>::F>,
|
||||||
|
p2p: P,
|
||||||
|
serai: Serai,
|
||||||
|
) {
|
||||||
let mut db = MainDb::new(db);
|
let mut db = MainDb::new(db);
|
||||||
|
|
||||||
let mut last_substrate_block = 0; // TODO: Load from DB
|
let mut last_substrate_block = db.last_substrate_block();
|
||||||
|
|
||||||
loop {
|
loop {
|
||||||
match substrate::handle_new_blocks(&mut db, &p2p, &serai, &mut last_substrate_block).await {
|
match substrate::handle_new_blocks(&mut db, &key, &p2p, &serai, &mut last_substrate_block).await
|
||||||
|
{
|
||||||
Ok(()) => {}
|
Ok(()) => {}
|
||||||
Err(e) => log::error!("couldn't communicate with serai node: {e}"),
|
Err(e) => log::error!("couldn't communicate with serai node: {e}"),
|
||||||
}
|
}
|
||||||
@@ -38,6 +49,7 @@ async fn run<D: Db, P: P2p>(db: D, p2p: P, serai: Serai) {
|
|||||||
#[tokio::main]
|
#[tokio::main]
|
||||||
async fn main() {
|
async fn main() {
|
||||||
let db = MemDb::new(); // TODO
|
let db = MemDb::new(); // TODO
|
||||||
|
let key = Zeroizing::new(<Ristretto as Ciphersuite>::F::ZERO); // TODO
|
||||||
let p2p = LocalP2p {}; // TODO
|
let p2p = LocalP2p {}; // TODO
|
||||||
let serai = || async {
|
let serai = || async {
|
||||||
loop {
|
loop {
|
||||||
@@ -48,5 +60,5 @@ async fn main() {
|
|||||||
return serai;
|
return serai;
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
run(db, p2p, serai().await).await
|
run(db, key, p2p, serai().await).await
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,18 +1,68 @@
|
|||||||
|
use core::{time::Duration, ops::Deref};
|
||||||
use std::collections::HashMap;
|
use std::collections::HashMap;
|
||||||
|
|
||||||
use zeroize::Zeroizing;
|
use zeroize::Zeroizing;
|
||||||
|
|
||||||
use transcript::{Transcript, RecommendedTranscript};
|
use transcript::{Transcript, RecommendedTranscript};
|
||||||
use ciphersuite::{group::ff::Field, Ciphersuite, Ristretto};
|
use ciphersuite::{group::GroupEncoding, Ciphersuite, Ristretto};
|
||||||
|
use frost::{Participant, ThresholdParams};
|
||||||
|
|
||||||
use serai_client::{SeraiError, Block, Serai, validator_sets::ValidatorSetsEvent};
|
use tokio::time::sleep;
|
||||||
|
|
||||||
|
use serai_client::{
|
||||||
|
SeraiError, Block, Serai,
|
||||||
|
primitives::BlockHash,
|
||||||
|
validator_sets::{
|
||||||
|
primitives::{Session, ValidatorSet, ValidatorSetData},
|
||||||
|
ValidatorSetsEvent,
|
||||||
|
},
|
||||||
|
in_instructions::InInstructionsEvent,
|
||||||
|
tokens::{primitives::OutInstructionWithBalance, TokensEvent},
|
||||||
|
};
|
||||||
|
|
||||||
use tributary::Tributary;
|
use tributary::Tributary;
|
||||||
|
|
||||||
|
use processor_messages::{SubstrateContext, key_gen::KeyGenId};
|
||||||
|
|
||||||
use crate::{Db, MainDb, TributaryTransaction, P2p};
|
use crate::{Db, MainDb, TributaryTransaction, P2p};
|
||||||
|
|
||||||
|
async fn get_set(serai: &Serai, set: ValidatorSet) -> ValidatorSetData {
|
||||||
|
loop {
|
||||||
|
match serai.get_validator_set(set).await {
|
||||||
|
Ok(data) => return data.unwrap(),
|
||||||
|
Err(e) => {
|
||||||
|
log::error!("couldn't get validator set data: {e}");
|
||||||
|
sleep(Duration::from_secs(5)).await;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn get_coin_keys(serai: &Serai, set: ValidatorSet) -> Vec<u8> {
|
||||||
|
loop {
|
||||||
|
match serai.get_keys(set).await {
|
||||||
|
Ok(data) => return data.unwrap().1.into_inner(),
|
||||||
|
Err(e) => {
|
||||||
|
log::error!("couldn't get validator set's keys: {e}");
|
||||||
|
sleep(Duration::from_secs(5)).await;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn in_set(
|
||||||
|
key: &Zeroizing<<Ristretto as Ciphersuite>::F>,
|
||||||
|
serai: &Serai,
|
||||||
|
set: ValidatorSet,
|
||||||
|
) -> bool {
|
||||||
|
let data = get_set(serai, set).await;
|
||||||
|
let key = Ristretto::generator() * key.deref();
|
||||||
|
data.participants.iter().any(|(participant, _)| participant.0 == key.to_bytes())
|
||||||
|
}
|
||||||
|
|
||||||
async fn handle_block<D: Db, P: P2p>(
|
async fn handle_block<D: Db, P: P2p>(
|
||||||
db: &mut MainDb<D>,
|
db: &mut MainDb<D>,
|
||||||
|
key: &Zeroizing<<Ristretto as Ciphersuite>::F>,
|
||||||
p2p: &P,
|
p2p: &P,
|
||||||
serai: &Serai,
|
serai: &Serai,
|
||||||
block: Block,
|
block: Block,
|
||||||
@@ -27,36 +77,52 @@ async fn handle_block<D: Db, P: P2p>(
|
|||||||
if let ValidatorSetsEvent::NewSet { set } = new_set {
|
if let ValidatorSetsEvent::NewSet { set } = new_set {
|
||||||
let set_data = serai.get_validator_set(set).await?.unwrap();
|
let set_data = serai.get_validator_set(set).await?.unwrap();
|
||||||
|
|
||||||
let mut genesis = RecommendedTranscript::new(b"Serai Tributary Genesis");
|
let mut i = None;
|
||||||
genesis.append_message(b"serai_block", hash);
|
|
||||||
genesis.append_message(b"session", set.session.0.to_le_bytes());
|
|
||||||
genesis.append_message(b"network", set.network.0.to_le_bytes());
|
|
||||||
let genesis = genesis.challenge(b"genesis");
|
|
||||||
let genesis_ref: &[u8] = genesis.as_ref();
|
|
||||||
let genesis = genesis_ref[.. 32].try_into().unwrap();
|
|
||||||
|
|
||||||
let mut validators = HashMap::new();
|
let mut validators = HashMap::new();
|
||||||
for (participant, amount) in &set_data.participants {
|
for (l, (participant, amount)) in set_data.participants.iter().enumerate() {
|
||||||
validators.insert(
|
// TODO2: Ensure an invalid public key can't be a validator
|
||||||
// TODO2: Ensure an invalid public key can't be a validator
|
let participant =
|
||||||
<Ristretto as Ciphersuite>::read_G::<&[u8]>(&mut participant.0.as_ref()).unwrap(),
|
<Ristretto as Ciphersuite>::read_G::<&[u8]>(&mut participant.0.as_ref()).unwrap();
|
||||||
// Give one weight on Tributary per bond instance
|
if participant == (Ristretto::generator() * key.deref()) {
|
||||||
amount.0 / set_data.bond.0,
|
i = Some(Participant::new((l + 1).try_into().unwrap()).unwrap());
|
||||||
);
|
}
|
||||||
|
|
||||||
|
// Give one weight on Tributary per bond instance
|
||||||
|
validators.insert(participant, amount.0 / set_data.bond.0);
|
||||||
}
|
}
|
||||||
|
|
||||||
// TODO: Do something with this
|
if let Some(i) = i {
|
||||||
let tributary = Tributary::<_, TributaryTransaction, _>::new(
|
let n = u16::try_from(set_data.participants.len()).unwrap();
|
||||||
// TODO2: Use a DB on a dedicated volume
|
let t = (2 * (n / 3)) + 1;
|
||||||
db.0.clone(),
|
|
||||||
genesis,
|
let mut genesis = RecommendedTranscript::new(b"Serai Tributary Genesis");
|
||||||
block.time().unwrap(),
|
genesis.append_message(b"serai_block", hash);
|
||||||
Zeroizing::new(<Ristretto as Ciphersuite>::F::ZERO), // TODO
|
genesis.append_message(b"session", set.session.0.to_le_bytes());
|
||||||
validators,
|
genesis.append_message(b"network", set.network.0.to_le_bytes());
|
||||||
p2p.clone(),
|
let genesis = genesis.challenge(b"genesis");
|
||||||
)
|
let genesis_ref: &[u8] = genesis.as_ref();
|
||||||
.await
|
let genesis = genesis_ref[.. 32].try_into().unwrap();
|
||||||
.unwrap();
|
|
||||||
|
// TODO: Do something with this
|
||||||
|
let tributary = Tributary::<_, TributaryTransaction, _>::new(
|
||||||
|
// TODO2: Use a DB on a dedicated volume
|
||||||
|
db.0.clone(),
|
||||||
|
genesis,
|
||||||
|
block.time().unwrap(),
|
||||||
|
key.clone(),
|
||||||
|
validators,
|
||||||
|
p2p.clone(),
|
||||||
|
)
|
||||||
|
.await
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
// Trigger a DKG
|
||||||
|
// TODO: Send this to processor. Check how it handles it being fired multiple times
|
||||||
|
let msg = processor_messages::key_gen::CoordinatorMessage::GenerateKey {
|
||||||
|
id: KeyGenId { set, attempt: 0 },
|
||||||
|
params: ThresholdParams::new(t, n, i).unwrap(),
|
||||||
|
};
|
||||||
|
}
|
||||||
} else {
|
} else {
|
||||||
panic!("NewSet event wasn't NewSet: {new_set:?}");
|
panic!("NewSet event wasn't NewSet: {new_set:?}");
|
||||||
}
|
}
|
||||||
@@ -68,26 +134,106 @@ async fn handle_block<D: Db, P: P2p>(
|
|||||||
// If a key pair was confirmed, inform the processor
|
// If a key pair was confirmed, inform the processor
|
||||||
for key_gen in serai.get_key_gen_events(hash).await? {
|
for key_gen in serai.get_key_gen_events(hash).await? {
|
||||||
if !db.handled_event(hash, event_id) {
|
if !db.handled_event(hash, event_id) {
|
||||||
// TODO: Handle key_gen
|
if let ValidatorSetsEvent::KeyGen { set, key_pair } = key_gen {
|
||||||
|
if in_set(key, serai, set).await {
|
||||||
|
// TODO: Send this to processor. Check how it handles it being fired multiple times
|
||||||
|
let msg = processor_messages::key_gen::CoordinatorMessage::ConfirmKeyPair {
|
||||||
|
context: SubstrateContext {
|
||||||
|
coin_latest_finalized_block: serai
|
||||||
|
.get_latest_block_for_network(hash, set.network)
|
||||||
|
.await?
|
||||||
|
.unwrap_or(BlockHash([0; 32])), // TODO: Have the processor override this
|
||||||
|
},
|
||||||
|
id: KeyGenId { set, attempt: todo!() },
|
||||||
|
};
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
panic!("KeyGen event wasn't KeyGen: {key_gen:?}");
|
||||||
|
}
|
||||||
db.handle_event(hash, event_id);
|
db.handle_event(hash, event_id);
|
||||||
}
|
}
|
||||||
event_id += 1;
|
event_id += 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
// If batch, tell processor of block acknowledged/burns
|
if !db.handled_event(hash, event_id) {
|
||||||
for batch in serai.get_batch_events(hash).await? {
|
// Finally, tell the processor of acknowledged blocks/burns
|
||||||
if !db.handled_event(hash, event_id) {
|
let mut coins_with_event = vec![];
|
||||||
// TODO: Handle batch
|
let mut batch_block = HashMap::new();
|
||||||
db.handle_event(hash, event_id);
|
let mut burns = HashMap::new();
|
||||||
|
|
||||||
|
for batch in serai.get_batch_events(hash).await? {
|
||||||
|
if let InInstructionsEvent::Batch { network, id: _, block: coin_block } = batch {
|
||||||
|
// Don't insert this multiple times, yet use a Vec to maintain the insertion order
|
||||||
|
if !coins_with_event.contains(&network) {
|
||||||
|
coins_with_event.push(network);
|
||||||
|
burns.insert(network, vec![]);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Use the last specified block
|
||||||
|
batch_block.insert(network, coin_block);
|
||||||
|
|
||||||
|
// TODO: Send this to processor. Check how it handles it being fired multiple times
|
||||||
|
let msg = processor_messages::coordinator::CoordinatorMessage::BatchSigned {
|
||||||
|
key: get_coin_keys(serai, ValidatorSet { network, session: Session(0) }).await, // TODO2
|
||||||
|
block: coin_block,
|
||||||
|
};
|
||||||
|
} else {
|
||||||
|
panic!("Batch event wasn't Batch: {batch:?}");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
for burn in serai.get_burn_events(hash).await? {
|
||||||
|
if let TokensEvent::Burn { address: _, balance, instruction } = burn {
|
||||||
|
let network = {
|
||||||
|
use serai_client::primitives::*;
|
||||||
|
match balance.coin {
|
||||||
|
BITCOIN => BITCOIN_NET_ID,
|
||||||
|
ETHER => ETHEREUM_NET_ID,
|
||||||
|
DAI => ETHEREUM_NET_ID,
|
||||||
|
MONERO => MONERO_NET_ID,
|
||||||
|
invalid => panic!("burn from unrecognized coin: {invalid:?}"),
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
if !coins_with_event.contains(&network) {
|
||||||
|
coins_with_event.push(network);
|
||||||
|
burns.insert(network, vec![]);
|
||||||
|
}
|
||||||
|
|
||||||
|
let mut burns_so_far = burns.remove(&network).unwrap_or(vec![]);
|
||||||
|
burns_so_far.push(OutInstructionWithBalance { balance, instruction });
|
||||||
|
burns.insert(network, burns_so_far);
|
||||||
|
} else {
|
||||||
|
panic!("Burn event wasn't Burn: {burn:?}");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
for network in coins_with_event {
|
||||||
|
let coin_latest_finalized_block = if let Some(block) = batch_block.remove(&network) {
|
||||||
|
block
|
||||||
|
} else {
|
||||||
|
// If it's had a batch or a burn, it must have had a block acknowledged
|
||||||
|
serai.get_latest_block_for_network(hash, network).await?.unwrap()
|
||||||
|
};
|
||||||
|
|
||||||
|
// TODO: Send this to processor. Check how it handles it being fired multiple times
|
||||||
|
let msg = processor_messages::substrate::CoordinatorMessage::SubstrateBlock {
|
||||||
|
context: SubstrateContext { coin_latest_finalized_block },
|
||||||
|
key: get_coin_keys(serai, ValidatorSet { network, session: Session(0) }).await, // TODO2
|
||||||
|
// Use remove not only to avoid a clone, yet so if network is present twice somehow, this
|
||||||
|
// isn't fired multiple times
|
||||||
|
burns: burns.remove(&network).unwrap(),
|
||||||
|
};
|
||||||
}
|
}
|
||||||
event_id += 1;
|
|
||||||
}
|
}
|
||||||
|
db.handle_event(hash, event_id);
|
||||||
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn handle_new_blocks<D: Db, P: P2p>(
|
pub async fn handle_new_blocks<D: Db, P: P2p>(
|
||||||
db: &mut MainDb<D>,
|
db: &mut MainDb<D>,
|
||||||
|
key: &Zeroizing<<Ristretto as Ciphersuite>::F>,
|
||||||
p2p: &P,
|
p2p: &P,
|
||||||
serai: &Serai,
|
serai: &Serai,
|
||||||
last_substrate_block: &mut u64,
|
last_substrate_block: &mut u64,
|
||||||
@@ -103,6 +249,7 @@ pub async fn handle_new_blocks<D: Db, P: P2p>(
|
|||||||
for b in (*last_substrate_block + 1) ..= latest_number {
|
for b in (*last_substrate_block + 1) ..= latest_number {
|
||||||
handle_block(
|
handle_block(
|
||||||
db,
|
db,
|
||||||
|
key,
|
||||||
p2p,
|
p2p,
|
||||||
serai,
|
serai,
|
||||||
if b == latest_number {
|
if b == latest_number {
|
||||||
@@ -112,8 +259,8 @@ pub async fn handle_new_blocks<D: Db, P: P2p>(
|
|||||||
},
|
},
|
||||||
)
|
)
|
||||||
.await?;
|
.await?;
|
||||||
// TODO: Update the DB
|
|
||||||
*last_substrate_block += 1;
|
*last_substrate_block += 1;
|
||||||
|
db.set_last_substrate_block(*last_substrate_block);
|
||||||
}
|
}
|
||||||
|
|
||||||
Ok(())
|
Ok(())
|
||||||
|
|||||||
Reference in New Issue
Block a user