mirror of
https://github.com/serai-dex/serai.git
synced 2025-12-14 06:59:24 +00:00
Compare commits
69 Commits
fe41b09fd4
...
next
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
ca93c82156 | ||
|
|
5b1875dae6 | ||
|
|
bcd68441be | ||
|
|
4ebf9ad9c7 | ||
|
|
807572199c | ||
|
|
3cdc1536c5 | ||
|
|
9e13e5ebff | ||
|
|
9b2c254eee | ||
|
|
0883479068 | ||
|
|
c5480c63be | ||
|
|
4280ee6987 | ||
|
|
91673d7ae3 | ||
|
|
927f07b62b | ||
|
|
7e774d6d2d | ||
|
|
fccd06b376 | ||
|
|
e3edc0a7fc | ||
|
|
9c47ef2658 | ||
|
|
e1b6b638c6 | ||
|
|
c24768f922 | ||
|
|
87ee879dea | ||
|
|
b5603560e8 | ||
|
|
5818f1a41c | ||
|
|
1b781b4b57 | ||
|
|
94faf098b6 | ||
|
|
03e45f73cd | ||
|
|
63f7e220c0 | ||
|
|
7d49366373 | ||
|
|
55ed33d2d1 | ||
|
|
138a0e9b40 | ||
|
|
4fc7263ac3 | ||
|
|
f27fd59fa6 | ||
|
|
437f0e9a93 | ||
|
|
cc5d38f1ce | ||
|
|
0ce025e0c2 | ||
|
|
224cf4ea21 | ||
|
|
a9b1e5293c | ||
|
|
80009ab67f | ||
|
|
df9fda2971 | ||
|
|
ca8afb83a1 | ||
|
|
18a9cf2535 | ||
|
|
10c126ad92 | ||
|
|
19305aebc9 | ||
|
|
be68e27551 | ||
|
|
d6d96fe8ff | ||
|
|
95909d83a4 | ||
|
|
3bd48974f3 | ||
|
|
29093715e3 | ||
|
|
87b4dfc8f3 | ||
|
|
4db78b1787 | ||
|
|
02a5f15535 | ||
|
|
865e351f96 | ||
|
|
ea275df26c | ||
|
|
2216ade8c4 | ||
|
|
5265cc69de | ||
|
|
a141deaf36 | ||
|
|
215e41fdb6 | ||
|
|
41c34d7f11 | ||
|
|
974bc82387 | ||
|
|
47ef24a7cc | ||
|
|
c0e48867e1 | ||
|
|
0066b94d38 | ||
|
|
7d54c02ec6 | ||
|
|
568324f631 | ||
|
|
2a02a8dc59 | ||
|
|
eaa9a0e5a6 | ||
|
|
251996c1b0 | ||
|
|
98b9cc82a7 | ||
|
|
f8adfb56ad | ||
|
|
7a790f3a20 |
2
.github/actions/bitcoin/action.yml
vendored
2
.github/actions/bitcoin/action.yml
vendored
@@ -5,7 +5,7 @@ inputs:
|
|||||||
version:
|
version:
|
||||||
description: "Version to download and run"
|
description: "Version to download and run"
|
||||||
required: false
|
required: false
|
||||||
default: "27.0"
|
default: "30.0"
|
||||||
|
|
||||||
runs:
|
runs:
|
||||||
using: "composite"
|
using: "composite"
|
||||||
|
|||||||
21
.github/actions/build-dependencies/action.yml
vendored
21
.github/actions/build-dependencies/action.yml
vendored
@@ -7,6 +7,10 @@ runs:
|
|||||||
- name: Remove unused packages
|
- name: Remove unused packages
|
||||||
shell: bash
|
shell: bash
|
||||||
run: |
|
run: |
|
||||||
|
# Ensure the repositories are synced
|
||||||
|
sudo apt update -y
|
||||||
|
|
||||||
|
# Actually perform the removals
|
||||||
sudo apt remove -y "*powershell*" "*nuget*" "*bazel*" "*ansible*" "*terraform*" "*heroku*" "*aws*" azure-cli
|
sudo apt remove -y "*powershell*" "*nuget*" "*bazel*" "*ansible*" "*terraform*" "*heroku*" "*aws*" azure-cli
|
||||||
sudo apt remove -y "*nodejs*" "*npm*" "*yarn*" "*java*" "*kotlin*" "*golang*" "*swift*" "*julia*" "*fortran*" "*android*"
|
sudo apt remove -y "*nodejs*" "*npm*" "*yarn*" "*java*" "*kotlin*" "*golang*" "*swift*" "*julia*" "*fortran*" "*android*"
|
||||||
sudo apt remove -y "*apache2*" "*nginx*" "*firefox*" "*chromium*" "*chrome*" "*edge*"
|
sudo apt remove -y "*apache2*" "*nginx*" "*firefox*" "*chromium*" "*chrome*" "*edge*"
|
||||||
@@ -14,8 +18,9 @@ runs:
|
|||||||
sudo apt remove -y --allow-remove-essential -f shim-signed *python3*
|
sudo apt remove -y --allow-remove-essential -f shim-signed *python3*
|
||||||
# This removal command requires the prior removals due to unmet dependencies otherwise
|
# This removal command requires the prior removals due to unmet dependencies otherwise
|
||||||
sudo apt remove -y "*qemu*" "*sql*" "*texinfo*" "*imagemagick*"
|
sudo apt remove -y "*qemu*" "*sql*" "*texinfo*" "*imagemagick*"
|
||||||
|
|
||||||
# Reinstall python3 as a general dependency of a functional operating system
|
# Reinstall python3 as a general dependency of a functional operating system
|
||||||
sudo apt install python3
|
sudo apt install -y python3 --fix-missing
|
||||||
if: runner.os == 'Linux'
|
if: runner.os == 'Linux'
|
||||||
|
|
||||||
- name: Remove unused packages
|
- name: Remove unused packages
|
||||||
@@ -33,19 +38,23 @@ runs:
|
|||||||
shell: bash
|
shell: bash
|
||||||
run: |
|
run: |
|
||||||
if [ "$RUNNER_OS" == "Linux" ]; then
|
if [ "$RUNNER_OS" == "Linux" ]; then
|
||||||
sudo apt install -y ca-certificates protobuf-compiler
|
sudo apt install -y ca-certificates protobuf-compiler libclang-dev
|
||||||
elif [ "$RUNNER_OS" == "Windows" ]; then
|
elif [ "$RUNNER_OS" == "Windows" ]; then
|
||||||
choco install protoc
|
choco install protoc
|
||||||
elif [ "$RUNNER_OS" == "macOS" ]; then
|
elif [ "$RUNNER_OS" == "macOS" ]; then
|
||||||
brew install protobuf
|
brew install protobuf llvm
|
||||||
|
HOMEBREW_ROOT_PATH=/opt/homebrew # Apple Silicon
|
||||||
|
if [ $(uname -m) = "x86_64" ]; then HOMEBREW_ROOT_PATH=/usr/local; fi # Intel
|
||||||
|
ls $HOMEBREW_ROOT_PATH/opt/llvm/lib | grep "libclang.dylib" # Make sure this installed `libclang`
|
||||||
|
echo "DYLD_LIBRARY_PATH=$HOMEBREW_ROOT_PATH/opt/llvm/lib:$DYLD_LIBRARY_PATH" >> "$GITHUB_ENV"
|
||||||
fi
|
fi
|
||||||
|
|
||||||
- name: Install solc
|
- name: Install solc
|
||||||
shell: bash
|
shell: bash
|
||||||
run: |
|
run: |
|
||||||
cargo +1.89 install svm-rs --version =0.5.18
|
cargo +1.91 install svm-rs --version =0.5.19
|
||||||
svm install 0.8.26
|
svm install 0.8.29
|
||||||
svm use 0.8.26
|
svm use 0.8.29
|
||||||
|
|
||||||
- name: Remove preinstalled Docker
|
- name: Remove preinstalled Docker
|
||||||
shell: bash
|
shell: bash
|
||||||
|
|||||||
2
.github/actions/monero-wallet-rpc/action.yml
vendored
2
.github/actions/monero-wallet-rpc/action.yml
vendored
@@ -5,7 +5,7 @@ inputs:
|
|||||||
version:
|
version:
|
||||||
description: "Version to download and run"
|
description: "Version to download and run"
|
||||||
required: false
|
required: false
|
||||||
default: v0.18.3.4
|
default: v0.18.4.3
|
||||||
|
|
||||||
runs:
|
runs:
|
||||||
using: "composite"
|
using: "composite"
|
||||||
|
|||||||
2
.github/actions/monero/action.yml
vendored
2
.github/actions/monero/action.yml
vendored
@@ -5,7 +5,7 @@ inputs:
|
|||||||
version:
|
version:
|
||||||
description: "Version to download and run"
|
description: "Version to download and run"
|
||||||
required: false
|
required: false
|
||||||
default: v0.18.3.4
|
default: v0.18.4.3
|
||||||
|
|
||||||
runs:
|
runs:
|
||||||
using: "composite"
|
using: "composite"
|
||||||
|
|||||||
4
.github/actions/test-dependencies/action.yml
vendored
4
.github/actions/test-dependencies/action.yml
vendored
@@ -5,12 +5,12 @@ inputs:
|
|||||||
monero-version:
|
monero-version:
|
||||||
description: "Monero version to download and run as a regtest node"
|
description: "Monero version to download and run as a regtest node"
|
||||||
required: false
|
required: false
|
||||||
default: v0.18.3.4
|
default: v0.18.4.3
|
||||||
|
|
||||||
bitcoin-version:
|
bitcoin-version:
|
||||||
description: "Bitcoin version to download and run as a regtest node"
|
description: "Bitcoin version to download and run as a regtest node"
|
||||||
required: false
|
required: false
|
||||||
default: "27.1"
|
default: "30.0"
|
||||||
|
|
||||||
runs:
|
runs:
|
||||||
using: "composite"
|
using: "composite"
|
||||||
|
|||||||
2
.github/nightly-version
vendored
2
.github/nightly-version
vendored
@@ -1 +1 @@
|
|||||||
nightly-2025-08-01
|
nightly-2025-11-11
|
||||||
|
|||||||
2
.github/workflows/daily-deny.yml
vendored
2
.github/workflows/daily-deny.yml
vendored
@@ -18,7 +18,7 @@ jobs:
|
|||||||
key: rust-advisory-db
|
key: rust-advisory-db
|
||||||
|
|
||||||
- name: Install cargo deny
|
- name: Install cargo deny
|
||||||
run: cargo +1.89 install cargo-deny --version =0.18.3
|
run: cargo +1.91 install cargo-deny --version =0.18.5
|
||||||
|
|
||||||
- name: Run cargo deny
|
- name: Run cargo deny
|
||||||
run: cargo deny -L error --all-features check --hide-inclusion-graph
|
run: cargo deny -L error --all-features check --hide-inclusion-graph
|
||||||
|
|||||||
38
.github/workflows/lint.yml
vendored
38
.github/workflows/lint.yml
vendored
@@ -11,7 +11,7 @@ jobs:
|
|||||||
clippy:
|
clippy:
|
||||||
strategy:
|
strategy:
|
||||||
matrix:
|
matrix:
|
||||||
os: [ubuntu-latest, macos-13, macos-14, windows-latest]
|
os: [ubuntu-latest, macos-15-intel, macos-latest, windows-latest]
|
||||||
runs-on: ${{ matrix.os }}
|
runs-on: ${{ matrix.os }}
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
@@ -26,7 +26,7 @@ jobs:
|
|||||||
uses: ./.github/actions/build-dependencies
|
uses: ./.github/actions/build-dependencies
|
||||||
|
|
||||||
- name: Install nightly rust
|
- name: Install nightly rust
|
||||||
run: rustup toolchain install ${{ steps.nightly.outputs.version }} --profile minimal -t wasm32v1-none -c rust-src -c clippy
|
run: rustup toolchain install ${{ steps.nightly.outputs.version }} --profile minimal -t wasm32v1-none -c clippy
|
||||||
|
|
||||||
- name: Run Clippy
|
- name: Run Clippy
|
||||||
run: cargo +${{ steps.nightly.outputs.version }} clippy --all-features --all-targets -- -D warnings -A clippy::items_after_test_module
|
run: cargo +${{ steps.nightly.outputs.version }} clippy --all-features --all-targets -- -D warnings -A clippy::items_after_test_module
|
||||||
@@ -52,7 +52,7 @@ jobs:
|
|||||||
key: rust-advisory-db
|
key: rust-advisory-db
|
||||||
|
|
||||||
- name: Install cargo deny
|
- name: Install cargo deny
|
||||||
run: cargo +1.89 install cargo-deny --version =0.18.3
|
run: cargo +1.91 install cargo-deny --version =0.18.5
|
||||||
|
|
||||||
- name: Run cargo deny
|
- name: Run cargo deny
|
||||||
run: cargo deny -L error --all-features check --hide-inclusion-graph
|
run: cargo deny -L error --all-features check --hide-inclusion-graph
|
||||||
@@ -88,8 +88,8 @@ jobs:
|
|||||||
- uses: actions/checkout@3df4ab11eba7bda6032a0b82a6bb43b11571feac
|
- uses: actions/checkout@3df4ab11eba7bda6032a0b82a6bb43b11571feac
|
||||||
- name: Verify all dependencies are in use
|
- name: Verify all dependencies are in use
|
||||||
run: |
|
run: |
|
||||||
cargo +1.89 install cargo-machete --version =0.8.0
|
cargo +1.91 install cargo-machete --version =0.9.1
|
||||||
cargo +1.89 machete
|
cargo +1.91 machete
|
||||||
|
|
||||||
msrv:
|
msrv:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
@@ -98,7 +98,7 @@ jobs:
|
|||||||
- name: Verify claimed `rust-version`
|
- name: Verify claimed `rust-version`
|
||||||
shell: bash
|
shell: bash
|
||||||
run: |
|
run: |
|
||||||
cargo +1.89 install cargo-msrv --version =0.18.4
|
cargo +1.91 install cargo-msrv --version =0.18.4
|
||||||
|
|
||||||
function check_msrv {
|
function check_msrv {
|
||||||
# We `cd` into the directory passed as the first argument, but will return to the
|
# We `cd` into the directory passed as the first argument, but will return to the
|
||||||
@@ -144,18 +144,17 @@ jobs:
|
|||||||
function check_workspace {
|
function check_workspace {
|
||||||
# Get the members array from the workspace's `Cargo.toml`
|
# Get the members array from the workspace's `Cargo.toml`
|
||||||
cargo_toml_lines=$(cat ./Cargo.toml | wc -l)
|
cargo_toml_lines=$(cat ./Cargo.toml | wc -l)
|
||||||
|
# Keep all lines after the start of the array, then keep all lines before the next "]"
|
||||||
members=$(cat Cargo.toml | grep "members\ \=\ \[" -m1 -A$cargo_toml_lines | grep "]" -m1 -B$cargo_toml_lines)
|
members=$(cat Cargo.toml | grep "members\ \=\ \[" -m1 -A$cargo_toml_lines | grep "]" -m1 -B$cargo_toml_lines)
|
||||||
# Parse out any comments, including comments post-fixed on the same line as an entry
|
|
||||||
members=$(echo "$members" | grep -Ev "^[[:space:]]+#" | grep -Ev "^[[:space:]]?$" | awk -F',' '{print $1","}')
|
# Parse out any comments, whitespace, including comments post-fixed on the same line as an entry
|
||||||
# Prune `members = [` to `[` by replacing the first line with just `[`
|
# We accomplish the latter by pruning all characters after the entry's ","
|
||||||
|
members=$(echo "$members" | grep -Ev "^[[:space:]]*(#|$)" | awk -F',' '{print $1","}')
|
||||||
|
# Replace the first line, which was "members = [" and is now "members = [,", with "["
|
||||||
members=$(echo "$members" | sed "1s/.*/\[/")
|
members=$(echo "$members" | sed "1s/.*/\[/")
|
||||||
# Remove the trailing comma by replacing the last line's "," with ""
|
# Correct the last line, which was malleated to "],"
|
||||||
members=$(echo "$members" | sed "$(($(echo "$members" | wc -l) - 1))s/\,//")
|
|
||||||
# Correct the last line, which was malleated to "]," when pruning comments
|
|
||||||
members=$(echo "$members" | sed "$(echo "$members" | wc -l)s/\]\,/\]/")
|
members=$(echo "$members" | sed "$(echo "$members" | wc -l)s/\]\,/\]/")
|
||||||
|
|
||||||
# Don't check the patches
|
|
||||||
members=$(echo "$members" | grep -v "patches")
|
|
||||||
# Don't check the following
|
# Don't check the following
|
||||||
# Most of these are binaries, with the exception of the Substrate runtime which has a
|
# Most of these are binaries, with the exception of the Substrate runtime which has a
|
||||||
# bespoke build pipeline
|
# bespoke build pipeline
|
||||||
@@ -174,6 +173,9 @@ jobs:
|
|||||||
members=$(echo "$members" | grep -v "mini\"")
|
members=$(echo "$members" | grep -v "mini\"")
|
||||||
members=$(echo "$members" | grep -v "tests/")
|
members=$(echo "$members" | grep -v "tests/")
|
||||||
|
|
||||||
|
# Remove the trailing comma by replacing the last line's "," with ""
|
||||||
|
members=$(echo "$members" | sed "$(($(echo "$members" | wc -l) - 1))s/\,//")
|
||||||
|
|
||||||
echo $members | jq -r ".[]" | while read -r member; do
|
echo $members | jq -r ".[]" | while read -r member; do
|
||||||
check_msrv $member
|
check_msrv $member
|
||||||
correct=$?
|
correct=$?
|
||||||
@@ -188,12 +190,12 @@ jobs:
|
|||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@3df4ab11eba7bda6032a0b82a6bb43b11571feac
|
- uses: actions/checkout@3df4ab11eba7bda6032a0b82a6bb43b11571feac
|
||||||
|
|
||||||
|
- name: Build Dependencies
|
||||||
|
uses: ./.github/actions/build-dependencies
|
||||||
|
|
||||||
- name: Slither
|
- name: Slither
|
||||||
run: |
|
run: |
|
||||||
python3 -m pip install solc-select
|
|
||||||
solc-select install 0.8.26
|
|
||||||
solc-select use 0.8.26
|
|
||||||
|
|
||||||
python3 -m pip install slither-analyzer
|
python3 -m pip install slither-analyzer
|
||||||
|
|
||||||
slither --include-paths ./networks/ethereum/schnorr/contracts/Schnorr.sol
|
slither --include-paths ./networks/ethereum/schnorr/contracts/Schnorr.sol
|
||||||
|
|||||||
4
.github/workflows/pages.yml
vendored
4
.github/workflows/pages.yml
vendored
@@ -69,8 +69,8 @@ jobs:
|
|||||||
uses: ./.github/actions/build-dependencies
|
uses: ./.github/actions/build-dependencies
|
||||||
- name: Buld Rust docs
|
- name: Buld Rust docs
|
||||||
run: |
|
run: |
|
||||||
rustup toolchain install ${{ steps.nightly.outputs.version }} --profile minimal -t wasm32v1-none -c rust-docs -c rust-src
|
rustup toolchain install ${{ steps.nightly.outputs.version }} --profile minimal -t wasm32v1-none -c rust-docs
|
||||||
RUSTDOCFLAGS="--cfg docsrs" cargo +${{ steps.nightly.outputs.version }} doc --workspace --all-features
|
RUSTDOCFLAGS="--cfg docsrs" cargo +${{ steps.nightly.outputs.version }} doc --workspace --no-deps --all-features
|
||||||
mv target/doc docs/_site/rust
|
mv target/doc docs/_site/rust
|
||||||
|
|
||||||
- name: Upload artifact
|
- name: Upload artifact
|
||||||
|
|||||||
1
.github/workflows/tests.yml
vendored
1
.github/workflows/tests.yml
vendored
@@ -61,7 +61,6 @@ jobs:
|
|||||||
-p serai-monero-processor \
|
-p serai-monero-processor \
|
||||||
-p tendermint-machine \
|
-p tendermint-machine \
|
||||||
-p tributary-sdk \
|
-p tributary-sdk \
|
||||||
-p serai-cosign-types \
|
|
||||||
-p serai-cosign \
|
-p serai-cosign \
|
||||||
-p serai-coordinator-substrate \
|
-p serai-coordinator-substrate \
|
||||||
-p serai-coordinator-tributary \
|
-p serai-coordinator-tributary \
|
||||||
|
|||||||
7
.gitignore
vendored
7
.gitignore
vendored
@@ -1,7 +1,14 @@
|
|||||||
target
|
target
|
||||||
|
|
||||||
|
# Don't commit any `Cargo.lock` which aren't the workspace's
|
||||||
|
Cargo.lock
|
||||||
|
!./Cargo.lock
|
||||||
|
|
||||||
|
# Don't commit any `Dockerfile`, as they're auto-generated, except the only one which isn't
|
||||||
Dockerfile
|
Dockerfile
|
||||||
Dockerfile.fast-epoch
|
Dockerfile.fast-epoch
|
||||||
!orchestration/runtime/Dockerfile
|
!orchestration/runtime/Dockerfile
|
||||||
|
|
||||||
.test-logs
|
.test-logs
|
||||||
|
|
||||||
.vscode
|
.vscode
|
||||||
|
|||||||
4997
Cargo.lock
generated
4997
Cargo.lock
generated
File diff suppressed because it is too large
Load Diff
87
Cargo.toml
87
Cargo.toml
@@ -1,19 +1,6 @@
|
|||||||
[workspace]
|
[workspace]
|
||||||
resolver = "2"
|
resolver = "2"
|
||||||
members = [
|
members = [
|
||||||
# Version patches
|
|
||||||
"patches/parking_lot",
|
|
||||||
"patches/rocksdb",
|
|
||||||
|
|
||||||
# Rewrites/redirects
|
|
||||||
"patches/option-ext",
|
|
||||||
"patches/directories-next",
|
|
||||||
|
|
||||||
# monero-oxide expects ciphersuite, yet the ciphersuite in-tree here has breaking changes
|
|
||||||
# This re-exports the in-tree ciphersuite _without_ changes breaking to monero-oxide
|
|
||||||
# Not included in workspace to prevent having two crates with the same name (an error)
|
|
||||||
# "patches/ciphersuite",
|
|
||||||
|
|
||||||
"common/std-shims",
|
"common/std-shims",
|
||||||
"common/zalloc",
|
"common/zalloc",
|
||||||
"common/patchable-async-sleep",
|
"common/patchable-async-sleep",
|
||||||
@@ -82,7 +69,6 @@ members = [
|
|||||||
|
|
||||||
"coordinator/tributary-sdk/tendermint",
|
"coordinator/tributary-sdk/tendermint",
|
||||||
"coordinator/tributary-sdk",
|
"coordinator/tributary-sdk",
|
||||||
"coordinator/cosign/types",
|
|
||||||
"coordinator/cosign",
|
"coordinator/cosign",
|
||||||
"coordinator/substrate",
|
"coordinator/substrate",
|
||||||
"coordinator/tributary",
|
"coordinator/tributary",
|
||||||
@@ -91,16 +77,30 @@ members = [
|
|||||||
"coordinator",
|
"coordinator",
|
||||||
|
|
||||||
"substrate/primitives",
|
"substrate/primitives",
|
||||||
"substrate/abi",
|
|
||||||
|
|
||||||
"substrate/coins",
|
"substrate/coins/primitives",
|
||||||
"substrate/validator-sets",
|
"substrate/coins/pallet",
|
||||||
"substrate/signals",
|
|
||||||
"substrate/dex",
|
"substrate/dex/pallet",
|
||||||
"substrate/genesis-liquidity",
|
|
||||||
"substrate/economic-security",
|
"substrate/validator-sets/primitives",
|
||||||
"substrate/emissions",
|
"substrate/validator-sets/pallet",
|
||||||
"substrate/in-instructions",
|
|
||||||
|
"substrate/genesis-liquidity/primitives",
|
||||||
|
"substrate/genesis-liquidity/pallet",
|
||||||
|
|
||||||
|
"substrate/emissions/primitives",
|
||||||
|
"substrate/emissions/pallet",
|
||||||
|
|
||||||
|
"substrate/economic-security/pallet",
|
||||||
|
|
||||||
|
"substrate/in-instructions/primitives",
|
||||||
|
"substrate/in-instructions/pallet",
|
||||||
|
|
||||||
|
"substrate/signals/primitives",
|
||||||
|
"substrate/signals/pallet",
|
||||||
|
|
||||||
|
"substrate/abi",
|
||||||
|
|
||||||
"substrate/runtime",
|
"substrate/runtime",
|
||||||
"substrate/node",
|
"substrate/node",
|
||||||
@@ -172,23 +172,32 @@ panic = "unwind"
|
|||||||
overflow-checks = true
|
overflow-checks = true
|
||||||
|
|
||||||
[patch.crates-io]
|
[patch.crates-io]
|
||||||
|
# Point to empty crates for unused crates in our tree
|
||||||
|
ark-ff-3 = { package = "ark-ff", path = "patches/ethereum/ark-ff-0.3" }
|
||||||
|
ark-ff-4 = { package = "ark-ff", path = "patches/ethereum/ark-ff-0.4" }
|
||||||
|
c-kzg = { path = "patches/ethereum/c-kzg" }
|
||||||
|
secp256k1-30 = { package = "secp256k1", path = "patches/ethereum/secp256k1-30" }
|
||||||
|
|
||||||
# Dependencies from monero-oxide which originate from within our own tree
|
# Dependencies from monero-oxide which originate from within our own tree
|
||||||
std-shims = { path = "common/std-shims" }
|
std-shims = { path = "patches/std-shims" }
|
||||||
simple-request = { path = "common/request" }
|
simple-request = { path = "patches/simple-request" }
|
||||||
multiexp = { path = "crypto/multiexp" }
|
multiexp = { path = "crypto/multiexp" }
|
||||||
flexible-transcript = { path = "crypto/transcript" }
|
flexible-transcript = { path = "crypto/transcript" }
|
||||||
ciphersuite = { path = "patches/ciphersuite" }
|
ciphersuite = { path = "patches/ciphersuite" }
|
||||||
dalek-ff-group = { path = "crypto/dalek-ff-group" }
|
dalek-ff-group = { path = "patches/dalek-ff-group" }
|
||||||
minimal-ed448 = { path = "crypto/ed448" }
|
minimal-ed448 = { path = "crypto/ed448" }
|
||||||
modular-frost = { path = "crypto/frost" }
|
modular-frost = { path = "crypto/frost" }
|
||||||
|
|
||||||
|
# This has a non-deprecated `std` alternative since Rust's 2024 edition
|
||||||
|
home = { path = "patches/home" }
|
||||||
|
|
||||||
|
# Updates to the latest version
|
||||||
|
darling = { path = "patches/darling" }
|
||||||
|
thiserror = { path = "patches/thiserror" }
|
||||||
|
|
||||||
# https://github.com/rust-lang-nursery/lazy-static.rs/issues/201
|
# https://github.com/rust-lang-nursery/lazy-static.rs/issues/201
|
||||||
lazy_static = { git = "https://github.com/rust-lang-nursery/lazy-static.rs", rev = "5735630d46572f1e5377c8f2ba0f79d18f53b10c" }
|
lazy_static = { git = "https://github.com/rust-lang-nursery/lazy-static.rs", rev = "5735630d46572f1e5377c8f2ba0f79d18f53b10c" }
|
||||||
|
|
||||||
parking_lot = { path = "patches/parking_lot" }
|
|
||||||
# Needed for WAL compression
|
|
||||||
rocksdb = { path = "patches/rocksdb" }
|
|
||||||
|
|
||||||
# directories-next was created because directories was unmaintained
|
# directories-next was created because directories was unmaintained
|
||||||
# directories-next is now unmaintained while directories is maintained
|
# directories-next is now unmaintained while directories is maintained
|
||||||
# The directories author pulls in ridiculously pointless crates and prefers
|
# The directories author pulls in ridiculously pointless crates and prefers
|
||||||
@@ -197,19 +206,22 @@ rocksdb = { path = "patches/rocksdb" }
|
|||||||
option-ext = { path = "patches/option-ext" }
|
option-ext = { path = "patches/option-ext" }
|
||||||
directories-next = { path = "patches/directories-next" }
|
directories-next = { path = "patches/directories-next" }
|
||||||
|
|
||||||
# Patch to include `FromUniformBytes<64>` over Scalar
|
# Patch from a fork back to upstream
|
||||||
|
parity-bip39 = { path = "patches/parity-bip39" }
|
||||||
|
|
||||||
|
# Patch to include `FromUniformBytes<64>` over `Scalar`
|
||||||
k256 = { git = "https://github.com/kayabaNerve/elliptic-curves", rev = "4994c9ab163781a88cd4a49beae812a89a44e8c3" }
|
k256 = { git = "https://github.com/kayabaNerve/elliptic-curves", rev = "4994c9ab163781a88cd4a49beae812a89a44e8c3" }
|
||||||
p256 = { git = "https://github.com/kayabaNerve/elliptic-curves", rev = "4994c9ab163781a88cd4a49beae812a89a44e8c3" }
|
p256 = { git = "https://github.com/kayabaNerve/elliptic-curves", rev = "4994c9ab163781a88cd4a49beae812a89a44e8c3" }
|
||||||
|
|
||||||
# https://github.com/RustCrypto/hybrid-array/issues/131
|
# `jemalloc` conflicts with `mimalloc`, so patch to a `rocksdb` which never uses `jemalloc`
|
||||||
hybrid-array = { git = "https://github.com/kayabaNerve/hybrid-array", rev = "8caa508976c93696a67f40734537c91be7cecd96" }
|
librocksdb-sys = { path = "patches/librocksdb-sys" }
|
||||||
|
|
||||||
[workspace.lints.clippy]
|
[workspace.lints.clippy]
|
||||||
incompatible_msrv = "allow" # Manually verified with a GitHub workflow
|
|
||||||
manual_is_multiple_of = "allow"
|
|
||||||
unwrap_or_default = "allow"
|
unwrap_or_default = "allow"
|
||||||
map_unwrap_or = "allow"
|
map_unwrap_or = "allow"
|
||||||
needless_continue = "allow"
|
needless_continue = "allow"
|
||||||
|
manual_is_multiple_of = "allow"
|
||||||
|
incompatible_msrv = "allow" # Manually verified with a GitHub workflow
|
||||||
borrow_as_ptr = "deny"
|
borrow_as_ptr = "deny"
|
||||||
cast_lossless = "deny"
|
cast_lossless = "deny"
|
||||||
cast_possible_truncation = "deny"
|
cast_possible_truncation = "deny"
|
||||||
@@ -248,7 +260,7 @@ redundant_closure_for_method_calls = "deny"
|
|||||||
redundant_else = "deny"
|
redundant_else = "deny"
|
||||||
string_add_assign = "deny"
|
string_add_assign = "deny"
|
||||||
string_slice = "deny"
|
string_slice = "deny"
|
||||||
unchecked_duration_subtraction = "deny"
|
unchecked_time_subtraction = "deny"
|
||||||
uninlined_format_args = "deny"
|
uninlined_format_args = "deny"
|
||||||
unnecessary_box_returns = "deny"
|
unnecessary_box_returns = "deny"
|
||||||
unnecessary_join = "deny"
|
unnecessary_join = "deny"
|
||||||
@@ -257,3 +269,6 @@ unnested_or_patterns = "deny"
|
|||||||
unused_async = "deny"
|
unused_async = "deny"
|
||||||
unused_self = "deny"
|
unused_self = "deny"
|
||||||
zero_sized_map_values = "deny"
|
zero_sized_map_values = "deny"
|
||||||
|
|
||||||
|
[workspace.lints.rust]
|
||||||
|
unused = "allow" # TODO: https://github.com/rust-lang/rust/issues/147648
|
||||||
|
|||||||
50
audits/crypto/dkg/evrf/README.md
Normal file
50
audits/crypto/dkg/evrf/README.md
Normal file
@@ -0,0 +1,50 @@
|
|||||||
|
# eVRF DKG
|
||||||
|
|
||||||
|
In 2024, the [eVRF paper](https://eprint.iacr.org/2024/397) was published to
|
||||||
|
the IACR preprint server. Within it was a one-round unbiased DKG and a
|
||||||
|
one-round unbiased threshold DKG. Unfortunately, both simply describe
|
||||||
|
communication of the secret shares as 'Alice sends $s_b$ to Bob'. This causes,
|
||||||
|
in practice, the need for an additional round of communication to occur where
|
||||||
|
all participants confirm they received their secret shares.
|
||||||
|
|
||||||
|
Within Serai, it was posited to use the same premises as the DDH eVRF itself to
|
||||||
|
achieve a verifiable encryption scheme. This allows the secret shares to be
|
||||||
|
posted to any 'bulletin board' (such as a blockchain) and for all observers to
|
||||||
|
confirm:
|
||||||
|
|
||||||
|
- A participant participated
|
||||||
|
- The secret shares sent can be received by the intended recipient so long as
|
||||||
|
they can access the bulletin board
|
||||||
|
|
||||||
|
Additionally, Serai desired a robust scheme (albeit with an biased key as the
|
||||||
|
output, which is fine for our purposes). Accordingly, our implementation
|
||||||
|
instantiates the threshold eVRF DKG from the eVRF paper, with our own proposal
|
||||||
|
for verifiable encryption, with the caller allowed to decide the set of
|
||||||
|
participants. They may:
|
||||||
|
|
||||||
|
- Select everyone, collapsing to the non-threshold unbiased DKG from the eVRF
|
||||||
|
paper
|
||||||
|
- Select a pre-determined set, collapsing to the threshold unbaised DKG from
|
||||||
|
the eVRF paper
|
||||||
|
- Select a post-determined set (with any solution for the Common Subset
|
||||||
|
problem), allowing achieving a robust threshold biased DKG
|
||||||
|
|
||||||
|
Note that the eVRF paper proposes using the eVRF to sample coefficients yet
|
||||||
|
this is unnecessary when the resulting key will be biased. Any proof of
|
||||||
|
knowledge for the coefficients, as necessary for their extraction within the
|
||||||
|
security proofs, would be sufficient.
|
||||||
|
|
||||||
|
MAGIC Grants contracted HashCloak to formalize Serai's proposal for a DKG and
|
||||||
|
provide proofs for its security. This resulted in
|
||||||
|
[this paper](<./Security Proofs.pdf>).
|
||||||
|
|
||||||
|
Our implementation itself is then built on top of the audited
|
||||||
|
[`generalized-bulletproofs`](https://github.com/kayabaNerve/monero-oxide/tree/generalized-bulletproofs/audits/crypto/generalized-bulletproofs)
|
||||||
|
and
|
||||||
|
[`generalized-bulletproofs-ec-gadgets`](https://github.com/monero-oxide/monero-oxide/tree/fcmp%2B%2B/audits/fcmps).
|
||||||
|
|
||||||
|
Note we do not use the originally premised DDH eVRF yet the one premised on
|
||||||
|
elliptic curve divisors, the methodology of which is commented on
|
||||||
|
[here](https://github.com/monero-oxide/monero-oxide/tree/fcmp%2B%2B/audits/divisors).
|
||||||
|
|
||||||
|
Our implementation itself is unaudited at this time however.
|
||||||
BIN
audits/crypto/dkg/evrf/Security Proofs.pdf
Normal file
BIN
audits/crypto/dkg/evrf/Security Proofs.pdf
Normal file
Binary file not shown.
@@ -17,7 +17,7 @@ rustdoc-args = ["--cfg", "docsrs"]
|
|||||||
workspace = true
|
workspace = true
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
parity-db = { version = "0.4", default-features = false, optional = true }
|
parity-db = { version = "0.5", default-features = false, optional = true }
|
||||||
rocksdb = { version = "0.24", default-features = false, features = ["zstd"], optional = true }
|
rocksdb = { version = "0.24", default-features = false, features = ["zstd"], optional = true }
|
||||||
|
|
||||||
[features]
|
[features]
|
||||||
|
|||||||
@@ -15,7 +15,7 @@ pub fn serai_db_key(
|
|||||||
///
|
///
|
||||||
/// Creates a unit struct and a default implementation for the `key`, `get`, and `set`. The macro
|
/// Creates a unit struct and a default implementation for the `key`, `get`, and `set`. The macro
|
||||||
/// uses a syntax similar to defining a function. Parameters are concatenated to produce a key,
|
/// uses a syntax similar to defining a function. Parameters are concatenated to produce a key,
|
||||||
/// they must be `borsh` serializable. The return type is used to auto (de)serialize the database
|
/// they must be `scale` encodable. The return type is used to auto encode and decode the database
|
||||||
/// value bytes using `borsh`.
|
/// value bytes using `borsh`.
|
||||||
///
|
///
|
||||||
/// # Arguments
|
/// # Arguments
|
||||||
@@ -54,10 +54,11 @@ macro_rules! create_db {
|
|||||||
)?;
|
)?;
|
||||||
impl$(<$($generic_name: $generic_type),+>)? $field_name$(<$($generic_name),+>)? {
|
impl$(<$($generic_name: $generic_type),+>)? $field_name$(<$($generic_name),+>)? {
|
||||||
pub(crate) fn key($($arg: $arg_type),*) -> Vec<u8> {
|
pub(crate) fn key($($arg: $arg_type),*) -> Vec<u8> {
|
||||||
|
use scale::Encode;
|
||||||
$crate::serai_db_key(
|
$crate::serai_db_key(
|
||||||
stringify!($db_name).as_bytes(),
|
stringify!($db_name).as_bytes(),
|
||||||
stringify!($field_name).as_bytes(),
|
stringify!($field_name).as_bytes(),
|
||||||
&borsh::to_vec(&($($arg),*)).unwrap(),
|
($($arg),*).encode()
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
pub(crate) fn set(
|
pub(crate) fn set(
|
||||||
|
|||||||
2
common/env/src/lib.rs
vendored
2
common/env/src/lib.rs
vendored
@@ -1,5 +1,5 @@
|
|||||||
#![cfg_attr(docsrs, feature(doc_cfg))]
|
#![cfg_attr(docsrs, feature(doc_cfg))]
|
||||||
#![cfg_attr(docsrs, feature(doc_auto_cfg))]
|
#![cfg_attr(docsrs, feature(doc_cfg))]
|
||||||
|
|
||||||
// Obtain a variable from the Serai environment/secret store.
|
// Obtain a variable from the Serai environment/secret store.
|
||||||
pub fn var(variable: &str) -> Option<String> {
|
pub fn var(variable: &str) -> Option<String> {
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#![cfg_attr(docsrs, feature(doc_auto_cfg))]
|
#![cfg_attr(docsrs, feature(doc_cfg))]
|
||||||
#![doc = include_str!("../README.md")]
|
#![doc = include_str!("../README.md")]
|
||||||
#![deny(missing_docs)]
|
#![deny(missing_docs)]
|
||||||
|
|
||||||
|
|||||||
@@ -1,9 +1,9 @@
|
|||||||
[package]
|
[package]
|
||||||
name = "simple-request"
|
name = "simple-request"
|
||||||
version = "0.1.0"
|
version = "0.3.0"
|
||||||
description = "A simple HTTP(S) request library"
|
description = "A simple HTTP(S) request library"
|
||||||
license = "MIT"
|
license = "MIT"
|
||||||
repository = "https://github.com/serai-dex/serai/tree/develop/common/simple-request"
|
repository = "https://github.com/serai-dex/serai/tree/develop/common/request"
|
||||||
authors = ["Luke Parker <lukeparker5132@gmail.com>"]
|
authors = ["Luke Parker <lukeparker5132@gmail.com>"]
|
||||||
keywords = ["http", "https", "async", "request", "ssl"]
|
keywords = ["http", "https", "async", "request", "ssl"]
|
||||||
edition = "2021"
|
edition = "2021"
|
||||||
@@ -19,9 +19,10 @@ workspace = true
|
|||||||
[dependencies]
|
[dependencies]
|
||||||
tower-service = { version = "0.3", default-features = false }
|
tower-service = { version = "0.3", default-features = false }
|
||||||
hyper = { version = "1", default-features = false, features = ["http1", "client"] }
|
hyper = { version = "1", default-features = false, features = ["http1", "client"] }
|
||||||
hyper-util = { version = "0.1", default-features = false, features = ["http1", "client-legacy", "tokio"] }
|
hyper-util = { version = "0.1", default-features = false, features = ["http1", "client-legacy"] }
|
||||||
http-body-util = { version = "0.1", default-features = false }
|
http-body-util = { version = "0.1", default-features = false }
|
||||||
tokio = { version = "1", default-features = false }
|
futures-util = { version = "0.3", default-features = false, features = ["std"] }
|
||||||
|
tokio = { version = "1", default-features = false, features = ["sync"] }
|
||||||
|
|
||||||
hyper-rustls = { version = "0.27", default-features = false, features = ["http1", "ring", "rustls-native-certs", "native-tokio"], optional = true }
|
hyper-rustls = { version = "0.27", default-features = false, features = ["http1", "ring", "rustls-native-certs", "native-tokio"], optional = true }
|
||||||
|
|
||||||
@@ -29,6 +30,8 @@ zeroize = { version = "1", optional = true }
|
|||||||
base64ct = { version = "1", features = ["alloc"], optional = true }
|
base64ct = { version = "1", features = ["alloc"], optional = true }
|
||||||
|
|
||||||
[features]
|
[features]
|
||||||
tls = ["hyper-rustls"]
|
tokio = ["hyper-util/tokio"]
|
||||||
|
tls = ["tokio", "hyper-rustls"]
|
||||||
|
webpki-roots = ["tls", "hyper-rustls/webpki-roots"]
|
||||||
basic-auth = ["zeroize", "base64ct"]
|
basic-auth = ["zeroize", "base64ct"]
|
||||||
default = ["tls"]
|
default = ["tls"]
|
||||||
|
|||||||
@@ -1,19 +1,20 @@
|
|||||||
#![cfg_attr(docsrs, feature(doc_auto_cfg))]
|
#![cfg_attr(docsrs, feature(doc_cfg))]
|
||||||
#![doc = include_str!("../README.md")]
|
#![doc = include_str!("../README.md")]
|
||||||
|
|
||||||
|
use core::{pin::Pin, future::Future};
|
||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
|
|
||||||
use tokio::sync::Mutex;
|
use futures_util::FutureExt;
|
||||||
|
use ::tokio::sync::Mutex;
|
||||||
|
|
||||||
use tower_service::Service as TowerService;
|
use tower_service::Service as TowerService;
|
||||||
|
use hyper::{Uri, header::HeaderValue, body::Bytes, client::conn::http1::SendRequest, rt::Executor};
|
||||||
|
pub use hyper;
|
||||||
|
|
||||||
|
use hyper_util::client::legacy::{Client as HyperClient, connect::HttpConnector};
|
||||||
|
|
||||||
#[cfg(feature = "tls")]
|
#[cfg(feature = "tls")]
|
||||||
use hyper_rustls::{HttpsConnectorBuilder, HttpsConnector};
|
use hyper_rustls::{HttpsConnectorBuilder, HttpsConnector};
|
||||||
use hyper::{Uri, header::HeaderValue, body::Bytes, client::conn::http1::SendRequest};
|
|
||||||
use hyper_util::{
|
|
||||||
rt::tokio::TokioExecutor,
|
|
||||||
client::legacy::{Client as HyperClient, connect::HttpConnector},
|
|
||||||
};
|
|
||||||
pub use hyper;
|
|
||||||
|
|
||||||
mod request;
|
mod request;
|
||||||
pub use request::*;
|
pub use request::*;
|
||||||
@@ -37,52 +38,86 @@ type Connector = HttpConnector;
|
|||||||
type Connector = HttpsConnector<HttpConnector>;
|
type Connector = HttpsConnector<HttpConnector>;
|
||||||
|
|
||||||
#[derive(Clone, Debug)]
|
#[derive(Clone, Debug)]
|
||||||
enum Connection {
|
enum Connection<
|
||||||
|
E: 'static + Send + Sync + Clone + Executor<Pin<Box<dyn Send + Future<Output = ()>>>>,
|
||||||
|
> {
|
||||||
ConnectionPool(HyperClient<Connector, Full<Bytes>>),
|
ConnectionPool(HyperClient<Connector, Full<Bytes>>),
|
||||||
Connection {
|
Connection {
|
||||||
|
executor: E,
|
||||||
connector: Connector,
|
connector: Connector,
|
||||||
host: Uri,
|
host: Uri,
|
||||||
connection: Arc<Mutex<Option<SendRequest<Full<Bytes>>>>>,
|
connection: Arc<Mutex<Option<SendRequest<Full<Bytes>>>>>,
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// An HTTP client.
|
||||||
|
///
|
||||||
|
/// `tls` is only guaranteed to work when using the `tokio` executor. Instantiating a client when
|
||||||
|
/// the `tls` feature is active without using the `tokio` executor will cause errors.
|
||||||
#[derive(Clone, Debug)]
|
#[derive(Clone, Debug)]
|
||||||
pub struct Client {
|
pub struct Client<
|
||||||
connection: Connection,
|
E: 'static + Send + Sync + Clone + Executor<Pin<Box<dyn Send + Future<Output = ()>>>>,
|
||||||
|
> {
|
||||||
|
connection: Connection<E>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Client {
|
impl<E: 'static + Send + Sync + Clone + Executor<Pin<Box<dyn Send + Future<Output = ()>>>>>
|
||||||
fn connector() -> Connector {
|
Client<E>
|
||||||
|
{
|
||||||
|
#[allow(clippy::unnecessary_wraps)]
|
||||||
|
fn connector() -> Result<Connector, Error> {
|
||||||
let mut res = HttpConnector::new();
|
let mut res = HttpConnector::new();
|
||||||
res.set_keepalive(Some(core::time::Duration::from_secs(60)));
|
res.set_keepalive(Some(core::time::Duration::from_secs(60)));
|
||||||
res.set_nodelay(true);
|
res.set_nodelay(true);
|
||||||
res.set_reuse_address(true);
|
res.set_reuse_address(true);
|
||||||
|
|
||||||
|
#[cfg(feature = "tls")]
|
||||||
|
if core::any::TypeId::of::<E>() !=
|
||||||
|
core::any::TypeId::of::<hyper_util::rt::tokio::TokioExecutor>()
|
||||||
|
{
|
||||||
|
Err(Error::ConnectionError(
|
||||||
|
"`tls` feature enabled but not using the `tokio` executor".into(),
|
||||||
|
))?;
|
||||||
|
}
|
||||||
|
|
||||||
#[cfg(feature = "tls")]
|
#[cfg(feature = "tls")]
|
||||||
res.enforce_http(false);
|
res.enforce_http(false);
|
||||||
#[cfg(feature = "tls")]
|
#[cfg(feature = "tls")]
|
||||||
let res = HttpsConnectorBuilder::new()
|
let https = HttpsConnectorBuilder::new().with_native_roots();
|
||||||
.with_native_roots()
|
#[cfg(all(feature = "tls", not(feature = "webpki-roots")))]
|
||||||
.expect("couldn't fetch system's SSL roots")
|
let https = https.map_err(|e| {
|
||||||
.https_or_http()
|
Error::ConnectionError(
|
||||||
.enable_http1()
|
format!("couldn't load system's SSL root certificates and webpki-roots unavilable: {e:?}")
|
||||||
.wrap_connector(res);
|
.into(),
|
||||||
res
|
)
|
||||||
|
})?;
|
||||||
|
// Fallback to `webpki-roots` if present
|
||||||
|
#[cfg(all(feature = "tls", feature = "webpki-roots"))]
|
||||||
|
let https = https.unwrap_or(HttpsConnectorBuilder::new().with_webpki_roots());
|
||||||
|
#[cfg(feature = "tls")]
|
||||||
|
let res = https.https_or_http().enable_http1().wrap_connector(res);
|
||||||
|
|
||||||
|
Ok(res)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn with_connection_pool() -> Client {
|
pub fn with_executor_and_connection_pool(executor: E) -> Result<Client<E>, Error> {
|
||||||
Client {
|
Ok(Client {
|
||||||
connection: Connection::ConnectionPool(
|
connection: Connection::ConnectionPool(
|
||||||
HyperClient::builder(TokioExecutor::new())
|
HyperClient::builder(executor)
|
||||||
.pool_idle_timeout(core::time::Duration::from_secs(60))
|
.pool_idle_timeout(core::time::Duration::from_secs(60))
|
||||||
.build(Self::connector()),
|
.build(Self::connector()?),
|
||||||
),
|
),
|
||||||
}
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn without_connection_pool(host: &str) -> Result<Client, Error> {
|
pub fn with_executor_and_without_connection_pool(
|
||||||
|
executor: E,
|
||||||
|
host: &str,
|
||||||
|
) -> Result<Client<E>, Error> {
|
||||||
Ok(Client {
|
Ok(Client {
|
||||||
connection: Connection::Connection {
|
connection: Connection::Connection {
|
||||||
connector: Self::connector(),
|
executor,
|
||||||
|
connector: Self::connector()?,
|
||||||
host: {
|
host: {
|
||||||
let uri: Uri = host.parse().map_err(|_| Error::InvalidUri)?;
|
let uri: Uri = host.parse().map_err(|_| Error::InvalidUri)?;
|
||||||
if uri.host().is_none() {
|
if uri.host().is_none() {
|
||||||
@@ -95,9 +130,9 @@ impl Client {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn request<R: Into<Request>>(&self, request: R) -> Result<Response<'_>, Error> {
|
pub async fn request<R: Into<Request>>(&self, request: R) -> Result<Response<'_, E>, Error> {
|
||||||
let request: Request = request.into();
|
let request: Request = request.into();
|
||||||
let mut request = request.0;
|
let Request { mut request, response_size_limit } = request;
|
||||||
if let Some(header_host) = request.headers().get(hyper::header::HOST) {
|
if let Some(header_host) = request.headers().get(hyper::header::HOST) {
|
||||||
match &self.connection {
|
match &self.connection {
|
||||||
Connection::ConnectionPool(_) => {}
|
Connection::ConnectionPool(_) => {}
|
||||||
@@ -131,7 +166,7 @@ impl Client {
|
|||||||
Connection::ConnectionPool(client) => {
|
Connection::ConnectionPool(client) => {
|
||||||
client.request(request).await.map_err(Error::HyperUtil)?
|
client.request(request).await.map_err(Error::HyperUtil)?
|
||||||
}
|
}
|
||||||
Connection::Connection { connector, host, connection } => {
|
Connection::Connection { executor, connector, host, connection } => {
|
||||||
let mut connection_lock = connection.lock().await;
|
let mut connection_lock = connection.lock().await;
|
||||||
|
|
||||||
// If there's not a connection...
|
// If there's not a connection...
|
||||||
@@ -143,28 +178,46 @@ impl Client {
|
|||||||
let call_res = call_res.map_err(Error::ConnectionError);
|
let call_res = call_res.map_err(Error::ConnectionError);
|
||||||
let (requester, connection) =
|
let (requester, connection) =
|
||||||
hyper::client::conn::http1::handshake(call_res?).await.map_err(Error::Hyper)?;
|
hyper::client::conn::http1::handshake(call_res?).await.map_err(Error::Hyper)?;
|
||||||
// This will die when we drop the requester, so we don't need to track an AbortHandle
|
// This task will die when we drop the requester
|
||||||
// for it
|
executor.execute(Box::pin(connection.map(|_| ())));
|
||||||
tokio::spawn(connection);
|
|
||||||
*connection_lock = Some(requester);
|
*connection_lock = Some(requester);
|
||||||
}
|
}
|
||||||
|
|
||||||
let connection = connection_lock.as_mut().unwrap();
|
let connection = connection_lock.as_mut().expect("lock over the connection was poisoned");
|
||||||
let mut err = connection.ready().await.err();
|
let mut err = connection.ready().await.err();
|
||||||
if err.is_none() {
|
if err.is_none() {
|
||||||
// Send the request
|
// Send the request
|
||||||
let res = connection.send_request(request).await;
|
let response = connection.send_request(request).await;
|
||||||
if let Ok(res) = res {
|
if let Ok(response) = response {
|
||||||
return Ok(Response(res, self));
|
return Ok(Response { response, size_limit: response_size_limit, client: self });
|
||||||
}
|
}
|
||||||
err = res.err();
|
err = response.err();
|
||||||
}
|
}
|
||||||
// Since this connection has been put into an error state, drop it
|
// Since this connection has been put into an error state, drop it
|
||||||
*connection_lock = None;
|
*connection_lock = None;
|
||||||
Err(Error::Hyper(err.unwrap()))?
|
Err(Error::Hyper(err.expect("only here if `err` is some yet no error")))?
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
Ok(Response(response, self))
|
Ok(Response { response, size_limit: response_size_limit, client: self })
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[cfg(feature = "tokio")]
|
||||||
|
mod tokio {
|
||||||
|
use hyper_util::rt::tokio::TokioExecutor;
|
||||||
|
use super::*;
|
||||||
|
|
||||||
|
pub type TokioClient = Client<TokioExecutor>;
|
||||||
|
impl Client<TokioExecutor> {
|
||||||
|
pub fn with_connection_pool() -> Result<Self, Error> {
|
||||||
|
Self::with_executor_and_connection_pool(TokioExecutor::new())
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn without_connection_pool(host: &str) -> Result<Self, Error> {
|
||||||
|
Self::with_executor_and_without_connection_pool(TokioExecutor::new(), host)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
#[cfg(feature = "tokio")]
|
||||||
|
pub use tokio::TokioClient;
|
||||||
|
|||||||
@@ -7,11 +7,15 @@ pub use http_body_util::Full;
|
|||||||
use crate::Error;
|
use crate::Error;
|
||||||
|
|
||||||
#[derive(Debug)]
|
#[derive(Debug)]
|
||||||
pub struct Request(pub(crate) hyper::Request<Full<Bytes>>);
|
pub struct Request {
|
||||||
|
pub(crate) request: hyper::Request<Full<Bytes>>,
|
||||||
|
pub(crate) response_size_limit: Option<usize>,
|
||||||
|
}
|
||||||
|
|
||||||
impl Request {
|
impl Request {
|
||||||
#[cfg(feature = "basic-auth")]
|
#[cfg(feature = "basic-auth")]
|
||||||
fn username_password_from_uri(&self) -> Result<(String, String), Error> {
|
fn username_password_from_uri(&self) -> Result<(String, String), Error> {
|
||||||
if let Some(authority) = self.0.uri().authority() {
|
if let Some(authority) = self.request.uri().authority() {
|
||||||
let authority = authority.as_str();
|
let authority = authority.as_str();
|
||||||
if authority.contains('@') {
|
if authority.contains('@') {
|
||||||
// Decode the username and password from the URI
|
// Decode the username and password from the URI
|
||||||
@@ -36,9 +40,10 @@ impl Request {
|
|||||||
let mut formatted = format!("{username}:{password}");
|
let mut formatted = format!("{username}:{password}");
|
||||||
let mut encoded = Base64::encode_string(formatted.as_bytes());
|
let mut encoded = Base64::encode_string(formatted.as_bytes());
|
||||||
formatted.zeroize();
|
formatted.zeroize();
|
||||||
self.0.headers_mut().insert(
|
self.request.headers_mut().insert(
|
||||||
hyper::header::AUTHORIZATION,
|
hyper::header::AUTHORIZATION,
|
||||||
HeaderValue::from_str(&format!("Basic {encoded}")).unwrap(),
|
HeaderValue::from_str(&format!("Basic {encoded}"))
|
||||||
|
.expect("couldn't form header from base64-encoded string"),
|
||||||
);
|
);
|
||||||
encoded.zeroize();
|
encoded.zeroize();
|
||||||
}
|
}
|
||||||
@@ -59,9 +64,17 @@ impl Request {
|
|||||||
pub fn with_basic_auth(&mut self) {
|
pub fn with_basic_auth(&mut self) {
|
||||||
let _ = self.basic_auth_from_uri();
|
let _ = self.basic_auth_from_uri();
|
||||||
}
|
}
|
||||||
}
|
|
||||||
impl From<hyper::Request<Full<Bytes>>> for Request {
|
/// Set a size limit for the response.
|
||||||
fn from(request: hyper::Request<Full<Bytes>>) -> Request {
|
///
|
||||||
Request(request)
|
/// This may be exceeded by a single HTTP frame and accordingly isn't perfect.
|
||||||
|
pub fn set_response_size_limit(&mut self, response_size_limit: Option<usize>) {
|
||||||
|
self.response_size_limit = response_size_limit;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl From<hyper::Request<Full<Bytes>>> for Request {
|
||||||
|
fn from(request: hyper::Request<Full<Bytes>>) -> Request {
|
||||||
|
Request { request, response_size_limit: None }
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,24 +1,54 @@
|
|||||||
|
use core::{pin::Pin, future::Future};
|
||||||
|
use std::io;
|
||||||
|
|
||||||
use hyper::{
|
use hyper::{
|
||||||
StatusCode,
|
StatusCode,
|
||||||
header::{HeaderValue, HeaderMap},
|
header::{HeaderValue, HeaderMap},
|
||||||
body::{Buf, Incoming},
|
body::Incoming,
|
||||||
|
rt::Executor,
|
||||||
};
|
};
|
||||||
use http_body_util::BodyExt;
|
use http_body_util::BodyExt;
|
||||||
|
|
||||||
|
use futures_util::{Stream, StreamExt};
|
||||||
|
|
||||||
use crate::{Client, Error};
|
use crate::{Client, Error};
|
||||||
|
|
||||||
// Borrows the client so its async task lives as long as this response exists.
|
// Borrows the client so its async task lives as long as this response exists.
|
||||||
#[allow(dead_code)]
|
#[allow(dead_code)]
|
||||||
#[derive(Debug)]
|
#[derive(Debug)]
|
||||||
pub struct Response<'a>(pub(crate) hyper::Response<Incoming>, pub(crate) &'a Client);
|
pub struct Response<
|
||||||
impl Response<'_> {
|
'a,
|
||||||
|
E: 'static + Send + Sync + Clone + Executor<Pin<Box<dyn Send + Future<Output = ()>>>>,
|
||||||
|
> {
|
||||||
|
pub(crate) response: hyper::Response<Incoming>,
|
||||||
|
pub(crate) size_limit: Option<usize>,
|
||||||
|
pub(crate) client: &'a Client<E>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<E: 'static + Send + Sync + Clone + Executor<Pin<Box<dyn Send + Future<Output = ()>>>>>
|
||||||
|
Response<'_, E>
|
||||||
|
{
|
||||||
pub fn status(&self) -> StatusCode {
|
pub fn status(&self) -> StatusCode {
|
||||||
self.0.status()
|
self.response.status()
|
||||||
}
|
}
|
||||||
pub fn headers(&self) -> &HeaderMap<HeaderValue> {
|
pub fn headers(&self) -> &HeaderMap<HeaderValue> {
|
||||||
self.0.headers()
|
self.response.headers()
|
||||||
}
|
}
|
||||||
pub async fn body(self) -> Result<impl std::io::Read, Error> {
|
pub async fn body(self) -> Result<impl std::io::Read, Error> {
|
||||||
Ok(self.0.into_body().collect().await.map_err(Error::Hyper)?.aggregate().reader())
|
let mut body = self.response.into_body().into_data_stream();
|
||||||
|
let mut res: Vec<u8> = vec![];
|
||||||
|
loop {
|
||||||
|
if let Some(size_limit) = self.size_limit {
|
||||||
|
let (lower, upper) = body.size_hint();
|
||||||
|
if res.len().wrapping_add(upper.unwrap_or(lower)) > size_limit.min(usize::MAX - 1) {
|
||||||
|
Err(Error::ConnectionError("response exceeded size limit".into()))?;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
let Some(part) = body.next().await else { break };
|
||||||
|
let part = part.map_err(Error::Hyper)?;
|
||||||
|
res.extend(part.as_ref());
|
||||||
|
}
|
||||||
|
Ok(io::Cursor::new(res))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
[package]
|
[package]
|
||||||
name = "std-shims"
|
name = "std-shims"
|
||||||
version = "0.1.4"
|
version = "0.1.5"
|
||||||
description = "A series of std shims to make alloc more feasible"
|
description = "A series of std shims to make alloc more feasible"
|
||||||
license = "MIT"
|
license = "MIT"
|
||||||
repository = "https://github.com/serai-dex/serai/tree/develop/common/std-shims"
|
repository = "https://github.com/serai-dex/serai/tree/develop/common/std-shims"
|
||||||
@@ -18,9 +18,10 @@ workspace = true
|
|||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
rustversion = { version = "1", default-features = false }
|
rustversion = { version = "1", default-features = false }
|
||||||
spin = { version = "0.10", default-features = false, features = ["use_ticket_mutex", "once", "lazy"] }
|
spin = { version = "0.10", default-features = false, features = ["use_ticket_mutex", "fair_mutex", "once", "lazy"] }
|
||||||
hashbrown = { version = "0.16", default-features = false, features = ["default-hasher", "inline-more"] }
|
hashbrown = { version = "0.16", default-features = false, features = ["default-hasher", "inline-more"], optional = true }
|
||||||
|
|
||||||
[features]
|
[features]
|
||||||
std = []
|
alloc = ["hashbrown"]
|
||||||
|
std = ["alloc", "spin/std"]
|
||||||
default = ["std"]
|
default = ["std"]
|
||||||
|
|||||||
@@ -1,11 +1,28 @@
|
|||||||
# std shims
|
# `std` shims
|
||||||
|
|
||||||
A crate which passes through to std when the default `std` feature is enabled,
|
`std-shims` is a Rust crate with two purposes:
|
||||||
yet provides a series of shims when it isn't.
|
- Expand the functionality of `core` and `alloc`
|
||||||
|
- Polyfill functionality only available on newer version of Rust
|
||||||
|
|
||||||
No guarantee of one-to-one parity is provided. The shims provided aim to be sufficient for the
|
The goal is to make supporting no-`std` environments, and older versions of
|
||||||
average case.
|
Rust, as simple as possible. For most use cases, replacing `std::` with
|
||||||
|
`std_shims::` and adding `use std_shims::prelude::*` is sufficient to take full
|
||||||
|
advantage of `std-shims`.
|
||||||
|
|
||||||
`HashSet` and `HashMap` are provided via `hashbrown`. Synchronization primitives are provided via
|
# API Surface
|
||||||
`spin` (avoiding a requirement on `critical-section`).
|
|
||||||
types are not guaranteed to be
|
`std-shims` only aims to have items _mutually available_ between `alloc` (with
|
||||||
|
extra dependencies) and `std` publicly exposed. Items exclusive to `std`, with
|
||||||
|
no shims available, will not be exported by `std-shims`.
|
||||||
|
|
||||||
|
# Dependencies
|
||||||
|
|
||||||
|
`HashSet` and `HashMap` are provided via `hashbrown`. Synchronization
|
||||||
|
primitives are provided via `spin` (avoiding a requirement on
|
||||||
|
`critical-section`). Sections of `std::io` are independently matched as
|
||||||
|
possible. `rustversion` is used to detect when to provide polyfills.
|
||||||
|
|
||||||
|
# Disclaimer
|
||||||
|
|
||||||
|
No guarantee of one-to-one parity is provided. The shims provided aim to be
|
||||||
|
sufficient for the average case. Pull requests are _welcome_.
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
|
#[cfg(all(feature = "alloc", not(feature = "std")))]
|
||||||
|
pub use extern_alloc::collections::*;
|
||||||
|
#[cfg(all(feature = "alloc", not(feature = "std")))]
|
||||||
|
pub use hashbrown::{HashSet, HashMap};
|
||||||
|
|
||||||
#[cfg(feature = "std")]
|
#[cfg(feature = "std")]
|
||||||
pub use std::collections::*;
|
pub use std::collections::*;
|
||||||
|
|
||||||
#[cfg(not(feature = "std"))]
|
|
||||||
pub use alloc::collections::*;
|
|
||||||
#[cfg(not(feature = "std"))]
|
|
||||||
pub use hashbrown::{HashSet, HashMap};
|
|
||||||
|
|||||||
@@ -1,42 +1,74 @@
|
|||||||
#[cfg(feature = "std")]
|
|
||||||
pub use std::io::*;
|
|
||||||
|
|
||||||
#[cfg(not(feature = "std"))]
|
#[cfg(not(feature = "std"))]
|
||||||
mod shims {
|
mod shims {
|
||||||
use core::fmt::{Debug, Formatter};
|
use core::fmt::{self, Debug, Display, Formatter};
|
||||||
use alloc::{boxed::Box, vec::Vec};
|
#[cfg(feature = "alloc")]
|
||||||
|
use extern_alloc::{boxed::Box, vec::Vec};
|
||||||
|
use crate::error::Error as CoreError;
|
||||||
|
|
||||||
|
/// The kind of error.
|
||||||
#[derive(Clone, Copy, PartialEq, Eq, Debug)]
|
#[derive(Clone, Copy, PartialEq, Eq, Debug)]
|
||||||
pub enum ErrorKind {
|
pub enum ErrorKind {
|
||||||
UnexpectedEof,
|
UnexpectedEof,
|
||||||
Other,
|
Other,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// An error.
|
||||||
|
#[derive(Debug)]
|
||||||
pub struct Error {
|
pub struct Error {
|
||||||
kind: ErrorKind,
|
kind: ErrorKind,
|
||||||
error: Box<dyn Send + Sync>,
|
#[cfg(feature = "alloc")]
|
||||||
|
error: Box<dyn Send + Sync + CoreError>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Debug for Error {
|
impl Display for Error {
|
||||||
fn fmt(&self, fmt: &mut Formatter<'_>) -> core::result::Result<(), core::fmt::Error> {
|
fn fmt(&self, f: &mut Formatter<'_>) -> fmt::Result {
|
||||||
fmt.debug_struct("Error").field("kind", &self.kind).finish_non_exhaustive()
|
<Self as Debug>::fmt(self, f)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
impl CoreError for Error {}
|
||||||
|
|
||||||
|
#[cfg(not(feature = "alloc"))]
|
||||||
|
pub trait IntoBoxSendSyncError {}
|
||||||
|
#[cfg(not(feature = "alloc"))]
|
||||||
|
impl<I> IntoBoxSendSyncError for I {}
|
||||||
|
#[cfg(feature = "alloc")]
|
||||||
|
pub trait IntoBoxSendSyncError: Into<Box<dyn Send + Sync + CoreError>> {}
|
||||||
|
#[cfg(feature = "alloc")]
|
||||||
|
impl<I: Into<Box<dyn Send + Sync + CoreError>>> IntoBoxSendSyncError for I {}
|
||||||
|
|
||||||
impl Error {
|
impl Error {
|
||||||
pub fn new<E: 'static + Send + Sync>(kind: ErrorKind, error: E) -> Error {
|
/// Create a new error.
|
||||||
Error { kind, error: Box::new(error) }
|
///
|
||||||
|
/// The error object itself is silently dropped when `alloc` is not enabled.
|
||||||
|
#[allow(unused)]
|
||||||
|
pub fn new<E: 'static + IntoBoxSendSyncError>(kind: ErrorKind, error: E) -> Error {
|
||||||
|
#[cfg(not(feature = "alloc"))]
|
||||||
|
let res = Error { kind };
|
||||||
|
#[cfg(feature = "alloc")]
|
||||||
|
let res = Error { kind, error: error.into() };
|
||||||
|
res
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn other<E: 'static + Send + Sync>(error: E) -> Error {
|
/// Create a new error with `io::ErrorKind::Other` as its kind.
|
||||||
Error { kind: ErrorKind::Other, error: Box::new(error) }
|
///
|
||||||
|
/// The error object itself is silently dropped when `alloc` is not enabled.
|
||||||
|
#[allow(unused)]
|
||||||
|
pub fn other<E: 'static + IntoBoxSendSyncError>(error: E) -> Error {
|
||||||
|
#[cfg(not(feature = "alloc"))]
|
||||||
|
let res = Error { kind: ErrorKind::Other };
|
||||||
|
#[cfg(feature = "alloc")]
|
||||||
|
let res = Error { kind: ErrorKind::Other, error: error.into() };
|
||||||
|
res
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// The kind of error.
|
||||||
pub fn kind(&self) -> ErrorKind {
|
pub fn kind(&self) -> ErrorKind {
|
||||||
self.kind
|
self.kind
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn into_inner(self) -> Option<Box<dyn Send + Sync>> {
|
/// Retrieve the inner error.
|
||||||
|
#[cfg(feature = "alloc")]
|
||||||
|
pub fn into_inner(self) -> Option<Box<dyn Send + Sync + CoreError>> {
|
||||||
Some(self.error)
|
Some(self.error)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -64,6 +96,12 @@ mod shims {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
impl<R: Read> Read for &mut R {
|
||||||
|
fn read(&mut self, buf: &mut [u8]) -> Result<usize> {
|
||||||
|
R::read(*self, buf)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
pub trait BufRead: Read {
|
pub trait BufRead: Read {
|
||||||
fn fill_buf(&mut self) -> Result<&[u8]>;
|
fn fill_buf(&mut self) -> Result<&[u8]>;
|
||||||
fn consume(&mut self, amt: usize);
|
fn consume(&mut self, amt: usize);
|
||||||
@@ -88,6 +126,7 @@ mod shims {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[cfg(feature = "alloc")]
|
||||||
impl Write for Vec<u8> {
|
impl Write for Vec<u8> {
|
||||||
fn write(&mut self, buf: &[u8]) -> Result<usize> {
|
fn write(&mut self, buf: &[u8]) -> Result<usize> {
|
||||||
self.extend(buf);
|
self.extend(buf);
|
||||||
@@ -95,6 +134,8 @@ mod shims {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(not(feature = "std"))]
|
#[cfg(not(feature = "std"))]
|
||||||
pub use shims::*;
|
pub use shims::*;
|
||||||
|
|
||||||
|
#[cfg(feature = "std")]
|
||||||
|
pub use std::io::{ErrorKind, Error, Result, Read, BufRead, Write};
|
||||||
|
|||||||
@@ -1,18 +1,45 @@
|
|||||||
#![cfg_attr(docsrs, feature(doc_auto_cfg))]
|
#![cfg_attr(docsrs, feature(doc_cfg))]
|
||||||
#![doc = include_str!("../README.md")]
|
#![doc = include_str!("../README.md")]
|
||||||
#![cfg_attr(not(feature = "std"), no_std)]
|
#![cfg_attr(not(feature = "std"), no_std)]
|
||||||
|
|
||||||
pub extern crate alloc;
|
#[cfg(not(feature = "alloc"))]
|
||||||
|
pub use core::*;
|
||||||
|
#[cfg(not(feature = "alloc"))]
|
||||||
|
pub use core::{alloc, borrow, ffi, fmt, slice, str, task};
|
||||||
|
|
||||||
|
#[cfg(not(feature = "std"))]
|
||||||
|
#[rustversion::before(1.81)]
|
||||||
|
pub mod error {
|
||||||
|
use core::fmt::Debug::Display;
|
||||||
|
pub trait Error: Debug + Display {}
|
||||||
|
}
|
||||||
|
#[cfg(not(feature = "std"))]
|
||||||
|
#[rustversion::since(1.81)]
|
||||||
|
pub use core::error;
|
||||||
|
|
||||||
|
#[cfg(feature = "alloc")]
|
||||||
|
extern crate alloc as extern_alloc;
|
||||||
|
#[cfg(all(feature = "alloc", not(feature = "std")))]
|
||||||
|
pub use extern_alloc::{alloc, borrow, boxed, ffi, fmt, rc, slice, str, string, task, vec, format};
|
||||||
|
#[cfg(feature = "std")]
|
||||||
|
pub use std::{alloc, borrow, boxed, error, ffi, fmt, rc, slice, str, string, task, vec, format};
|
||||||
|
|
||||||
pub mod sync;
|
|
||||||
pub mod collections;
|
pub mod collections;
|
||||||
pub mod io;
|
pub mod io;
|
||||||
|
pub mod sync;
|
||||||
pub use alloc::vec;
|
|
||||||
pub use alloc::str;
|
|
||||||
pub use alloc::string;
|
|
||||||
|
|
||||||
pub mod prelude {
|
pub mod prelude {
|
||||||
|
// Shim the `std` prelude
|
||||||
|
#[cfg(feature = "alloc")]
|
||||||
|
pub use extern_alloc::{
|
||||||
|
format, vec,
|
||||||
|
borrow::ToOwned,
|
||||||
|
boxed::Box,
|
||||||
|
vec::Vec,
|
||||||
|
string::{String, ToString},
|
||||||
|
};
|
||||||
|
|
||||||
|
// Shim `div_ceil`
|
||||||
#[rustversion::before(1.73)]
|
#[rustversion::before(1.73)]
|
||||||
#[doc(hidden)]
|
#[doc(hidden)]
|
||||||
pub trait StdShimsDivCeil {
|
pub trait StdShimsDivCeil {
|
||||||
@@ -53,6 +80,7 @@ pub mod prelude {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Shim `io::Error::other`
|
||||||
#[cfg(feature = "std")]
|
#[cfg(feature = "std")]
|
||||||
#[rustversion::before(1.74)]
|
#[rustversion::before(1.74)]
|
||||||
#[doc(hidden)]
|
#[doc(hidden)]
|
||||||
|
|||||||
@@ -1,19 +1,28 @@
|
|||||||
pub use core::sync::*;
|
pub use core::sync::atomic;
|
||||||
pub use alloc::sync::*;
|
#[cfg(all(feature = "alloc", not(feature = "std")))]
|
||||||
|
pub use extern_alloc::sync::{Arc, Weak};
|
||||||
|
#[cfg(feature = "std")]
|
||||||
|
pub use std::sync::{Arc, Weak};
|
||||||
|
|
||||||
mod mutex_shim {
|
mod mutex_shim {
|
||||||
#[cfg(feature = "std")]
|
|
||||||
pub use std::sync::*;
|
|
||||||
#[cfg(not(feature = "std"))]
|
#[cfg(not(feature = "std"))]
|
||||||
pub use spin::*;
|
pub use spin::{Mutex, MutexGuard};
|
||||||
|
#[cfg(feature = "std")]
|
||||||
|
pub use std::sync::{Mutex, MutexGuard};
|
||||||
|
|
||||||
|
/// A shimmed `Mutex` with an API mutual to `spin` and `std`.
|
||||||
#[derive(Default, Debug)]
|
#[derive(Default, Debug)]
|
||||||
pub struct ShimMutex<T>(Mutex<T>);
|
pub struct ShimMutex<T>(Mutex<T>);
|
||||||
impl<T> ShimMutex<T> {
|
impl<T> ShimMutex<T> {
|
||||||
|
/// Construct a new `Mutex`.
|
||||||
pub const fn new(value: T) -> Self {
|
pub const fn new(value: T) -> Self {
|
||||||
Self(Mutex::new(value))
|
Self(Mutex::new(value))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Acquire a lock on the contents of the `Mutex`.
|
||||||
|
///
|
||||||
|
/// On no-`std` environments, this may spin until the lock is acquired. On `std` environments,
|
||||||
|
/// this may panic if the `Mutex` was poisoned.
|
||||||
pub fn lock(&self) -> MutexGuard<'_, T> {
|
pub fn lock(&self) -> MutexGuard<'_, T> {
|
||||||
#[cfg(feature = "std")]
|
#[cfg(feature = "std")]
|
||||||
let res = self.0.lock().unwrap();
|
let res = self.0.lock().unwrap();
|
||||||
@@ -25,10 +34,11 @@ mod mutex_shim {
|
|||||||
}
|
}
|
||||||
pub use mutex_shim::{ShimMutex as Mutex, MutexGuard};
|
pub use mutex_shim::{ShimMutex as Mutex, MutexGuard};
|
||||||
|
|
||||||
#[cfg(not(feature = "std"))]
|
|
||||||
pub use spin::Lazy as LazyLock;
|
|
||||||
#[rustversion::before(1.80)]
|
#[rustversion::before(1.80)]
|
||||||
#[cfg(feature = "std")]
|
pub use spin::Lazy as LazyLock;
|
||||||
|
|
||||||
|
#[rustversion::since(1.80)]
|
||||||
|
#[cfg(not(feature = "std"))]
|
||||||
pub use spin::Lazy as LazyLock;
|
pub use spin::Lazy as LazyLock;
|
||||||
#[rustversion::since(1.80)]
|
#[rustversion::since(1.80)]
|
||||||
#[cfg(feature = "std")]
|
#[cfg(feature = "std")]
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#![cfg_attr(docsrs, feature(doc_auto_cfg))]
|
#![cfg_attr(docsrs, feature(doc_cfg))]
|
||||||
#![doc = include_str!("../README.md")]
|
#![doc = include_str!("../README.md")]
|
||||||
#![deny(missing_docs)]
|
#![deny(missing_docs)]
|
||||||
|
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
#![cfg_attr(docsrs, feature(doc_cfg))]
|
#![cfg_attr(docsrs, feature(doc_cfg))]
|
||||||
#![cfg_attr(docsrs, feature(doc_auto_cfg))]
|
#![cfg_attr(docsrs, feature(doc_cfg))]
|
||||||
#![cfg_attr(all(zalloc_rustc_nightly, feature = "allocator"), feature(allocator_api))]
|
#![cfg_attr(all(zalloc_rustc_nightly, feature = "allocator"), feature(allocator_api))]
|
||||||
|
|
||||||
//! Implementation of a Zeroizing Allocator, enabling zeroizing memory on deallocation.
|
//! Implementation of a Zeroizing Allocator, enabling zeroizing memory on deallocation.
|
||||||
|
|||||||
@@ -24,15 +24,14 @@ rand_core = { version = "0.6", default-features = false, features = ["std"] }
|
|||||||
blake2 = { version = "0.11.0-rc.0", default-features = false, features = ["alloc"] }
|
blake2 = { version = "0.11.0-rc.0", default-features = false, features = ["alloc"] }
|
||||||
schnorrkel = { version = "0.11", default-features = false, features = ["std"] }
|
schnorrkel = { version = "0.11", default-features = false, features = ["std"] }
|
||||||
|
|
||||||
transcript = { package = "flexible-transcript", path = "../crypto/transcript", default-features = false, features = ["std", "recommended"] }
|
|
||||||
dalek-ff-group = { path = "../crypto/dalek-ff-group", default-features = false, features = ["std"] }
|
dalek-ff-group = { path = "../crypto/dalek-ff-group", default-features = false, features = ["std"] }
|
||||||
ciphersuite = { path = "../crypto/ciphersuite", default-features = false, features = ["std"] }
|
ciphersuite = { path = "../crypto/ciphersuite", default-features = false, features = ["std"] }
|
||||||
schnorr = { package = "schnorr-signatures", path = "../crypto/schnorr", default-features = false, features = ["std", "aggregate"] }
|
|
||||||
dkg = { package = "dkg-musig", path = "../crypto/dkg/musig", default-features = false, features = ["std"] }
|
dkg = { package = "dkg-musig", path = "../crypto/dkg/musig", default-features = false, features = ["std"] }
|
||||||
frost = { package = "modular-frost", path = "../crypto/frost" }
|
frost = { package = "modular-frost", path = "../crypto/frost" }
|
||||||
frost-schnorrkel = { path = "../crypto/schnorrkel" }
|
frost-schnorrkel = { path = "../crypto/schnorrkel" }
|
||||||
|
|
||||||
hex = { version = "0.4", default-features = false, features = ["std"] }
|
hex = { version = "0.4", default-features = false, features = ["std"] }
|
||||||
|
scale = { package = "parity-scale-codec", version = "3", default-features = false, features = ["std", "derive", "bit-vec"] }
|
||||||
borsh = { version = "1", default-features = false, features = ["std", "derive", "de_strict_order"] }
|
borsh = { version = "1", default-features = false, features = ["std", "derive", "de_strict_order"] }
|
||||||
|
|
||||||
zalloc = { path = "../common/zalloc" }
|
zalloc = { path = "../common/zalloc" }
|
||||||
@@ -44,7 +43,7 @@ messages = { package = "serai-processor-messages", path = "../processor/messages
|
|||||||
message-queue = { package = "serai-message-queue", path = "../message-queue" }
|
message-queue = { package = "serai-message-queue", path = "../message-queue" }
|
||||||
tributary-sdk = { path = "./tributary-sdk" }
|
tributary-sdk = { path = "./tributary-sdk" }
|
||||||
|
|
||||||
serai-client = { path = "../substrate/client", default-features = false, features = ["serai"] }
|
serai-client = { path = "../substrate/client", default-features = false, features = ["serai", "borsh"] }
|
||||||
|
|
||||||
log = { version = "0.4", default-features = false, features = ["std"] }
|
log = { version = "0.4", default-features = false, features = ["std"] }
|
||||||
env_logger = { version = "0.10", default-features = false, features = ["humantime"] }
|
env_logger = { version = "0.10", default-features = false, features = ["humantime"] }
|
||||||
|
|||||||
@@ -21,8 +21,9 @@ workspace = true
|
|||||||
blake2 = { version = "0.11.0-rc.0", default-features = false, features = ["alloc"] }
|
blake2 = { version = "0.11.0-rc.0", default-features = false, features = ["alloc"] }
|
||||||
schnorrkel = { version = "0.11", default-features = false, features = ["std"] }
|
schnorrkel = { version = "0.11", default-features = false, features = ["std"] }
|
||||||
|
|
||||||
|
scale = { package = "parity-scale-codec", version = "3", default-features = false, features = ["std", "derive"] }
|
||||||
borsh = { version = "1", default-features = false, features = ["std", "derive", "de_strict_order"] }
|
borsh = { version = "1", default-features = false, features = ["std", "derive", "de_strict_order"] }
|
||||||
serai-client = { path = "../../substrate/client", default-features = false, features = ["serai"] }
|
serai-client = { path = "../../substrate/client", default-features = false, features = ["serai", "borsh"] }
|
||||||
|
|
||||||
log = { version = "0.4", default-features = false, features = ["std"] }
|
log = { version = "0.4", default-features = false, features = ["std"] }
|
||||||
|
|
||||||
@@ -30,5 +31,3 @@ tokio = { version = "1", default-features = false }
|
|||||||
|
|
||||||
serai-db = { path = "../../common/db", version = "0.1.1" }
|
serai-db = { path = "../../common/db", version = "0.1.1" }
|
||||||
serai-task = { path = "../../common/task", version = "0.1" }
|
serai-task = { path = "../../common/task", version = "0.1" }
|
||||||
|
|
||||||
serai-cosign-types = { path = "./types" }
|
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#![cfg_attr(docsrs, feature(doc_auto_cfg))]
|
#![cfg_attr(docsrs, feature(doc_cfg))]
|
||||||
#![doc = include_str!("../README.md")]
|
#![doc = include_str!("../README.md")]
|
||||||
#![deny(missing_docs)]
|
#![deny(missing_docs)]
|
||||||
|
|
||||||
@@ -7,6 +7,7 @@ use std::{sync::Arc, collections::HashMap, time::Instant};
|
|||||||
|
|
||||||
use blake2::{Digest, Blake2s256};
|
use blake2::{Digest, Blake2s256};
|
||||||
|
|
||||||
|
use scale::{Encode, Decode};
|
||||||
use borsh::{BorshSerialize, BorshDeserialize};
|
use borsh::{BorshSerialize, BorshDeserialize};
|
||||||
|
|
||||||
use serai_client::{
|
use serai_client::{
|
||||||
@@ -18,8 +19,6 @@ use serai_client::{
|
|||||||
use serai_db::*;
|
use serai_db::*;
|
||||||
use serai_task::*;
|
use serai_task::*;
|
||||||
|
|
||||||
use serai_cosign_types::*;
|
|
||||||
|
|
||||||
/// The cosigns which are intended to be performed.
|
/// The cosigns which are intended to be performed.
|
||||||
mod intend;
|
mod intend;
|
||||||
/// The evaluator of the cosigns.
|
/// The evaluator of the cosigns.
|
||||||
@@ -79,6 +78,68 @@ enum HasEvents {
|
|||||||
No,
|
No,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// An intended cosign.
|
||||||
|
#[derive(Clone, Copy, PartialEq, Eq, Debug, BorshSerialize, BorshDeserialize)]
|
||||||
|
pub struct CosignIntent {
|
||||||
|
/// The global session this cosign is being performed under.
|
||||||
|
pub global_session: [u8; 32],
|
||||||
|
/// The number of the block to cosign.
|
||||||
|
pub block_number: u64,
|
||||||
|
/// The hash of the block to cosign.
|
||||||
|
pub block_hash: [u8; 32],
|
||||||
|
/// If this cosign must be handled before further cosigns are.
|
||||||
|
pub notable: bool,
|
||||||
|
}
|
||||||
|
|
||||||
|
/// A cosign.
|
||||||
|
#[derive(Clone, PartialEq, Eq, Debug, Encode, Decode, BorshSerialize, BorshDeserialize)]
|
||||||
|
pub struct Cosign {
|
||||||
|
/// The global session this cosign is being performed under.
|
||||||
|
pub global_session: [u8; 32],
|
||||||
|
/// The number of the block to cosign.
|
||||||
|
pub block_number: u64,
|
||||||
|
/// The hash of the block to cosign.
|
||||||
|
pub block_hash: [u8; 32],
|
||||||
|
/// The actual cosigner.
|
||||||
|
pub cosigner: ExternalNetworkId,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl CosignIntent {
|
||||||
|
/// Convert this into a `Cosign`.
|
||||||
|
pub fn into_cosign(self, cosigner: ExternalNetworkId) -> Cosign {
|
||||||
|
let CosignIntent { global_session, block_number, block_hash, notable: _ } = self;
|
||||||
|
Cosign { global_session, block_number, block_hash, cosigner }
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Cosign {
|
||||||
|
/// The message to sign to sign this cosign.
|
||||||
|
///
|
||||||
|
/// This must be signed with schnorrkel, the context set to `COSIGN_CONTEXT`.
|
||||||
|
pub fn signature_message(&self) -> Vec<u8> {
|
||||||
|
// We use a schnorrkel context to domain-separate this
|
||||||
|
self.encode()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// A signed cosign.
|
||||||
|
#[derive(Clone, Debug, BorshSerialize, BorshDeserialize)]
|
||||||
|
pub struct SignedCosign {
|
||||||
|
/// The cosign.
|
||||||
|
pub cosign: Cosign,
|
||||||
|
/// The signature for the cosign.
|
||||||
|
pub signature: [u8; 64],
|
||||||
|
}
|
||||||
|
|
||||||
|
impl SignedCosign {
|
||||||
|
fn verify_signature(&self, signer: serai_client::Public) -> bool {
|
||||||
|
let Ok(signer) = schnorrkel::PublicKey::from_bytes(&signer.0) else { return false };
|
||||||
|
let Ok(signature) = schnorrkel::Signature::from_bytes(&self.signature) else { return false };
|
||||||
|
|
||||||
|
signer.verify_simple(COSIGN_CONTEXT, &self.cosign.signature_message(), &signature).is_ok()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
create_db! {
|
create_db! {
|
||||||
Cosign {
|
Cosign {
|
||||||
// The following are populated by the intend task and used throughout the library
|
// The following are populated by the intend task and used throughout the library
|
||||||
|
|||||||
@@ -1,25 +0,0 @@
|
|||||||
[package]
|
|
||||||
name = "serai-cosign-types"
|
|
||||||
version = "0.1.0"
|
|
||||||
description = "Evaluator of cosigns for the Serai network"
|
|
||||||
license = "AGPL-3.0-only"
|
|
||||||
repository = "https://github.com/serai-dex/serai/tree/develop/coordinator/cosign"
|
|
||||||
authors = ["Luke Parker <lukeparker5132@gmail.com>"]
|
|
||||||
keywords = []
|
|
||||||
edition = "2021"
|
|
||||||
publish = false
|
|
||||||
rust-version = "1.85"
|
|
||||||
|
|
||||||
[package.metadata.docs.rs]
|
|
||||||
all-features = true
|
|
||||||
rustdoc-args = ["--cfg", "docsrs"]
|
|
||||||
|
|
||||||
[lints]
|
|
||||||
workspace = true
|
|
||||||
|
|
||||||
[dependencies]
|
|
||||||
schnorrkel = { version = "0.11", default-features = false, features = ["std"] }
|
|
||||||
|
|
||||||
borsh = { version = "1", default-features = false, features = ["std", "derive", "de_strict_order"] }
|
|
||||||
|
|
||||||
serai-primitives = { path = "../../../substrate/primitives", default-features = false, features = ["std"] }
|
|
||||||
@@ -1,72 +0,0 @@
|
|||||||
#![cfg_attr(docsrs, feature(doc_auto_cfg))]
|
|
||||||
#![deny(missing_docs)]
|
|
||||||
//! Types used when cosigning Serai. For more info, please see `serai-cosign`.
|
|
||||||
use borsh::{BorshSerialize, BorshDeserialize};
|
|
||||||
|
|
||||||
use serai_primitives::{crypto::Public, network_id::ExternalNetworkId};
|
|
||||||
|
|
||||||
/// The schnorrkel context to used when signing a cosign.
|
|
||||||
pub const COSIGN_CONTEXT: &[u8] = b"/serai/coordinator/cosign";
|
|
||||||
|
|
||||||
/// An intended cosign.
|
|
||||||
#[derive(Clone, Copy, PartialEq, Eq, Debug, BorshSerialize, BorshDeserialize)]
|
|
||||||
pub struct CosignIntent {
|
|
||||||
/// The global session this cosign is being performed under.
|
|
||||||
pub global_session: [u8; 32],
|
|
||||||
/// The number of the block to cosign.
|
|
||||||
pub block_number: u64,
|
|
||||||
/// The hash of the block to cosign.
|
|
||||||
pub block_hash: [u8; 32],
|
|
||||||
/// If this cosign must be handled before further cosigns are.
|
|
||||||
pub notable: bool,
|
|
||||||
}
|
|
||||||
|
|
||||||
/// A cosign.
|
|
||||||
#[derive(Clone, PartialEq, Eq, Debug, BorshSerialize, BorshDeserialize)]
|
|
||||||
pub struct Cosign {
|
|
||||||
/// The global session this cosign is being performed under.
|
|
||||||
pub global_session: [u8; 32],
|
|
||||||
/// The number of the block to cosign.
|
|
||||||
pub block_number: u64,
|
|
||||||
/// The hash of the block to cosign.
|
|
||||||
pub block_hash: [u8; 32],
|
|
||||||
/// The actual cosigner.
|
|
||||||
pub cosigner: ExternalNetworkId,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl CosignIntent {
|
|
||||||
/// Convert this into a `Cosign`.
|
|
||||||
pub fn into_cosign(self, cosigner: ExternalNetworkId) -> Cosign {
|
|
||||||
let CosignIntent { global_session, block_number, block_hash, notable: _ } = self;
|
|
||||||
Cosign { global_session, block_number, block_hash, cosigner }
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl Cosign {
|
|
||||||
/// The message to sign to sign this cosign.
|
|
||||||
///
|
|
||||||
/// This must be signed with schnorrkel, the context set to `COSIGN_CONTEXT`.
|
|
||||||
pub fn signature_message(&self) -> Vec<u8> {
|
|
||||||
// We use a schnorrkel context to domain-separate this
|
|
||||||
borsh::to_vec(self).unwrap()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// A signed cosign.
|
|
||||||
#[derive(Clone, Debug, BorshSerialize, BorshDeserialize)]
|
|
||||||
pub struct SignedCosign {
|
|
||||||
/// The cosign.
|
|
||||||
pub cosign: Cosign,
|
|
||||||
/// The signature for the cosign.
|
|
||||||
pub signature: [u8; 64],
|
|
||||||
}
|
|
||||||
|
|
||||||
impl SignedCosign {
|
|
||||||
/// Verify a cosign's signature.
|
|
||||||
pub fn verify_signature(&self, signer: Public) -> bool {
|
|
||||||
let Ok(signer) = schnorrkel::PublicKey::from_bytes(&signer.0) else { return false };
|
|
||||||
let Ok(signature) = schnorrkel::Signature::from_bytes(&self.signature) else { return false };
|
|
||||||
|
|
||||||
signer.verify_simple(COSIGN_CONTEXT, &self.cosign.signature_message(), &signature).is_ok()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -22,7 +22,7 @@ borsh = { version = "1", default-features = false, features = ["std", "derive",
|
|||||||
|
|
||||||
serai-db = { path = "../../common/db", version = "0.1" }
|
serai-db = { path = "../../common/db", version = "0.1" }
|
||||||
|
|
||||||
serai-primitives = { path = "../../substrate/primitives", default-features = false, features = ["std"] }
|
serai-client = { path = "../../substrate/client", default-features = false, features = ["serai", "borsh"] }
|
||||||
serai-cosign = { path = "../cosign" }
|
serai-cosign = { path = "../cosign" }
|
||||||
tributary-sdk = { path = "../tributary-sdk" }
|
tributary-sdk = { path = "../tributary-sdk" }
|
||||||
|
|
||||||
|
|||||||
@@ -29,13 +29,13 @@ schnorrkel = { version = "0.11", default-features = false, features = ["std"] }
|
|||||||
hex = { version = "0.4", default-features = false, features = ["std"] }
|
hex = { version = "0.4", default-features = false, features = ["std"] }
|
||||||
borsh = { version = "1", default-features = false, features = ["std", "derive", "de_strict_order"] }
|
borsh = { version = "1", default-features = false, features = ["std", "derive", "de_strict_order"] }
|
||||||
|
|
||||||
serai-client = { path = "../../../substrate/client", default-features = false, features = ["serai"] }
|
serai-client = { path = "../../../substrate/client", default-features = false, features = ["serai", "borsh"] }
|
||||||
serai-cosign = { path = "../../cosign" }
|
serai-cosign = { path = "../../cosign" }
|
||||||
tributary-sdk = { path = "../../tributary-sdk" }
|
tributary-sdk = { path = "../../tributary-sdk" }
|
||||||
|
|
||||||
futures-util = { version = "0.3", default-features = false, features = ["std"] }
|
futures-util = { version = "0.3", default-features = false, features = ["std"] }
|
||||||
tokio = { version = "1", default-features = false, features = ["sync"] }
|
tokio = { version = "1", default-features = false, features = ["sync"] }
|
||||||
libp2p = { version = "0.54", default-features = false, features = ["tokio", "tcp", "noise", "yamux", "ping", "request-response", "gossipsub", "macros"] }
|
libp2p = { version = "0.56", default-features = false, features = ["tokio", "tcp", "noise", "yamux", "ping", "request-response", "gossipsub", "macros"] }
|
||||||
|
|
||||||
log = { version = "0.4", default-features = false, features = ["std"] }
|
log = { version = "0.4", default-features = false, features = ["std"] }
|
||||||
serai-task = { path = "../../../common/task", version = "0.1" }
|
serai-task = { path = "../../../common/task", version = "0.1" }
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#![cfg_attr(docsrs, feature(doc_auto_cfg))]
|
#![cfg_attr(docsrs, feature(doc_cfg))]
|
||||||
#![doc = include_str!("../README.md")]
|
#![doc = include_str!("../README.md")]
|
||||||
#![deny(missing_docs)]
|
#![deny(missing_docs)]
|
||||||
|
|
||||||
|
|||||||
@@ -92,7 +92,8 @@ impl SwarmTask {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
gossip::Event::Subscribed { .. } | gossip::Event::Unsubscribed { .. } => {}
|
gossip::Event::Subscribed { .. } | gossip::Event::Unsubscribed { .. } => {}
|
||||||
gossip::Event::GossipsubNotSupported { peer_id } => {
|
gossip::Event::GossipsubNotSupported { peer_id } |
|
||||||
|
gossip::Event::SlowPeer { peer_id, .. } => {
|
||||||
let _: Result<_, _> = self.swarm.disconnect_peer_id(peer_id);
|
let _: Result<_, _> = self.swarm.disconnect_peer_id(peer_id);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
use core::future::Future;
|
use core::future::Future;
|
||||||
use std::time::{Duration, SystemTime};
|
use std::time::{Duration, SystemTime};
|
||||||
|
|
||||||
use serai_primitives::{MAX_KEY_SHARES_PER_SET, ExternalValidatorSet};
|
use serai_client::validator_sets::primitives::{MAX_KEY_SHARES_PER_SET, ExternalValidatorSet};
|
||||||
|
|
||||||
use futures_lite::FutureExt;
|
use futures_lite::FutureExt;
|
||||||
|
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#![cfg_attr(docsrs, feature(doc_auto_cfg))]
|
#![cfg_attr(docsrs, feature(doc_cfg))]
|
||||||
#![doc = include_str!("../README.md")]
|
#![doc = include_str!("../README.md")]
|
||||||
#![deny(missing_docs)]
|
#![deny(missing_docs)]
|
||||||
|
|
||||||
@@ -7,7 +7,7 @@ use std::collections::HashMap;
|
|||||||
|
|
||||||
use borsh::{BorshSerialize, BorshDeserialize};
|
use borsh::{BorshSerialize, BorshDeserialize};
|
||||||
|
|
||||||
use serai_primitives::{network_id::ExternalNetworkId, validator_sets::ExternalValidatorSet};
|
use serai_client::{primitives::ExternalNetworkId, validator_sets::primitives::ExternalValidatorSet};
|
||||||
|
|
||||||
use serai_db::Db;
|
use serai_db::Db;
|
||||||
use tributary_sdk::{ReadWrite, TransactionTrait, Tributary, TributaryReader};
|
use tributary_sdk::{ReadWrite, TransactionTrait, Tributary, TributaryReader};
|
||||||
|
|||||||
@@ -103,7 +103,7 @@ mod _internal_db {
|
|||||||
// Tributary transactions to publish from the DKG confirmation task
|
// Tributary transactions to publish from the DKG confirmation task
|
||||||
TributaryTransactionsFromDkgConfirmation: (set: ExternalValidatorSet) -> Transaction,
|
TributaryTransactionsFromDkgConfirmation: (set: ExternalValidatorSet) -> Transaction,
|
||||||
// Participants to remove
|
// Participants to remove
|
||||||
RemoveParticipant: (set: ExternalValidatorSet) -> Participant,
|
RemoveParticipant: (set: ExternalValidatorSet) -> u16,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -139,10 +139,11 @@ impl RemoveParticipant {
|
|||||||
pub(crate) fn send(txn: &mut impl DbTxn, set: ExternalValidatorSet, participant: Participant) {
|
pub(crate) fn send(txn: &mut impl DbTxn, set: ExternalValidatorSet, participant: Participant) {
|
||||||
// If this set has yet to be retired, send this transaction
|
// If this set has yet to be retired, send this transaction
|
||||||
if RetiredTributary::get(txn, set.network).map(|session| session.0) < Some(set.session.0) {
|
if RetiredTributary::get(txn, set.network).map(|session| session.0) < Some(set.session.0) {
|
||||||
_internal_db::RemoveParticipant::send(txn, set, &participant);
|
_internal_db::RemoveParticipant::send(txn, set, &u16::from(participant));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
pub(crate) fn try_recv(txn: &mut impl DbTxn, set: ExternalValidatorSet) -> Option<Participant> {
|
pub(crate) fn try_recv(txn: &mut impl DbTxn, set: ExternalValidatorSet) -> Option<Participant> {
|
||||||
_internal_db::RemoveParticipant::try_recv(txn, set)
|
_internal_db::RemoveParticipant::try_recv(txn, set)
|
||||||
|
.map(|i| Participant::new(i).expect("sent invalid participant index for removal"))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -3,11 +3,10 @@ use std::{boxed::Box, collections::HashMap};
|
|||||||
|
|
||||||
use zeroize::Zeroizing;
|
use zeroize::Zeroizing;
|
||||||
use rand_core::OsRng;
|
use rand_core::OsRng;
|
||||||
use ciphersuite::{group::GroupEncoding, Ciphersuite};
|
use ciphersuite::{group::GroupEncoding, *};
|
||||||
use dalek_ff_group::Ristretto;
|
|
||||||
use dkg::{Participant, musig};
|
use dkg::{Participant, musig};
|
||||||
use frost_schnorrkel::{
|
use frost_schnorrkel::{
|
||||||
frost::{FrostError, sign::*},
|
frost::{curve::Ristretto, FrostError, sign::*},
|
||||||
Schnorrkel,
|
Schnorrkel,
|
||||||
};
|
};
|
||||||
|
|
||||||
@@ -31,7 +30,7 @@ fn schnorrkel() -> Schnorrkel {
|
|||||||
|
|
||||||
fn our_i(
|
fn our_i(
|
||||||
set: &NewSetInformation,
|
set: &NewSetInformation,
|
||||||
key: &Zeroizing<<Ristretto as Ciphersuite>::F>,
|
key: &Zeroizing<<Ristretto as WrappedGroup>::F>,
|
||||||
data: &HashMap<Participant, Vec<u8>>,
|
data: &HashMap<Participant, Vec<u8>>,
|
||||||
) -> Participant {
|
) -> Participant {
|
||||||
let public = SeraiAddress((Ristretto::generator() * key.deref()).to_bytes());
|
let public = SeraiAddress((Ristretto::generator() * key.deref()).to_bytes());
|
||||||
@@ -125,7 +124,7 @@ pub(crate) struct ConfirmDkgTask<CD: DbTrait, TD: DbTrait> {
|
|||||||
set: NewSetInformation,
|
set: NewSetInformation,
|
||||||
tributary_db: TD,
|
tributary_db: TD,
|
||||||
|
|
||||||
key: Zeroizing<<Ristretto as Ciphersuite>::F>,
|
key: Zeroizing<<Ristretto as WrappedGroup>::F>,
|
||||||
signer: Option<Signer>,
|
signer: Option<Signer>,
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -134,7 +133,7 @@ impl<CD: DbTrait, TD: DbTrait> ConfirmDkgTask<CD, TD> {
|
|||||||
db: CD,
|
db: CD,
|
||||||
set: NewSetInformation,
|
set: NewSetInformation,
|
||||||
tributary_db: TD,
|
tributary_db: TD,
|
||||||
key: Zeroizing<<Ristretto as Ciphersuite>::F>,
|
key: Zeroizing<<Ristretto as WrappedGroup>::F>,
|
||||||
) -> Self {
|
) -> Self {
|
||||||
Self { db, set, tributary_db, key, signer: None }
|
Self { db, set, tributary_db, key, signer: None }
|
||||||
}
|
}
|
||||||
@@ -153,7 +152,7 @@ impl<CD: DbTrait, TD: DbTrait> ConfirmDkgTask<CD, TD> {
|
|||||||
db: &mut CD,
|
db: &mut CD,
|
||||||
set: ExternalValidatorSet,
|
set: ExternalValidatorSet,
|
||||||
attempt: u32,
|
attempt: u32,
|
||||||
key: Zeroizing<<Ristretto as Ciphersuite>::F>,
|
key: Zeroizing<<Ristretto as WrappedGroup>::F>,
|
||||||
signer: &mut Option<Signer>,
|
signer: &mut Option<Signer>,
|
||||||
) {
|
) {
|
||||||
// Perform the preprocess
|
// Perform the preprocess
|
||||||
|
|||||||
@@ -7,7 +7,7 @@ use rand_core::{RngCore, OsRng};
|
|||||||
use dalek_ff_group::Ristretto;
|
use dalek_ff_group::Ristretto;
|
||||||
use ciphersuite::{
|
use ciphersuite::{
|
||||||
group::{ff::PrimeField, GroupEncoding},
|
group::{ff::PrimeField, GroupEncoding},
|
||||||
Ciphersuite,
|
*,
|
||||||
};
|
};
|
||||||
|
|
||||||
use borsh::BorshDeserialize;
|
use borsh::BorshDeserialize;
|
||||||
@@ -284,7 +284,7 @@ async fn handle_network(
|
|||||||
&mut txn,
|
&mut txn,
|
||||||
ExternalValidatorSet { network, session },
|
ExternalValidatorSet { network, session },
|
||||||
slash_report,
|
slash_report,
|
||||||
Signature(signature),
|
Signature::from(signature),
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
@@ -352,7 +352,7 @@ async fn main() {
|
|||||||
let mut key_bytes = [0; 32];
|
let mut key_bytes = [0; 32];
|
||||||
key_bytes.copy_from_slice(&key_vec);
|
key_bytes.copy_from_slice(&key_vec);
|
||||||
key_vec.zeroize();
|
key_vec.zeroize();
|
||||||
let key = Zeroizing::new(<Ristretto as Ciphersuite>::F::from_repr(key_bytes).unwrap());
|
let key = Zeroizing::new(<Ristretto as WrappedGroup>::F::from_repr(key_bytes).unwrap());
|
||||||
key_bytes.zeroize();
|
key_bytes.zeroize();
|
||||||
key
|
key
|
||||||
};
|
};
|
||||||
@@ -439,7 +439,7 @@ async fn main() {
|
|||||||
EphemeralEventStream::new(
|
EphemeralEventStream::new(
|
||||||
db.clone(),
|
db.clone(),
|
||||||
serai.clone(),
|
serai.clone(),
|
||||||
SeraiAddress((<Ristretto as Ciphersuite>::generator() * serai_key.deref()).to_bytes()),
|
SeraiAddress((<Ristretto as WrappedGroup>::generator() * serai_key.deref()).to_bytes()),
|
||||||
)
|
)
|
||||||
.continually_run(substrate_ephemeral_task_def, vec![substrate_task]),
|
.continually_run(substrate_ephemeral_task_def, vec![substrate_task]),
|
||||||
);
|
);
|
||||||
|
|||||||
@@ -3,7 +3,7 @@ use std::sync::Arc;
|
|||||||
|
|
||||||
use zeroize::Zeroizing;
|
use zeroize::Zeroizing;
|
||||||
|
|
||||||
use ciphersuite::Ciphersuite;
|
use ciphersuite::*;
|
||||||
use dalek_ff_group::Ristretto;
|
use dalek_ff_group::Ristretto;
|
||||||
|
|
||||||
use tokio::sync::mpsc;
|
use tokio::sync::mpsc;
|
||||||
@@ -23,7 +23,7 @@ use serai_coordinator_p2p::P2p;
|
|||||||
use crate::{Db, KeySet};
|
use crate::{Db, KeySet};
|
||||||
|
|
||||||
pub(crate) struct SubstrateTask<P: P2p> {
|
pub(crate) struct SubstrateTask<P: P2p> {
|
||||||
pub(crate) serai_key: Zeroizing<<Ristretto as Ciphersuite>::F>,
|
pub(crate) serai_key: Zeroizing<<Ristretto as WrappedGroup>::F>,
|
||||||
pub(crate) db: Db,
|
pub(crate) db: Db,
|
||||||
pub(crate) message_queue: Arc<MessageQueue>,
|
pub(crate) message_queue: Arc<MessageQueue>,
|
||||||
pub(crate) p2p: P,
|
pub(crate) p2p: P,
|
||||||
|
|||||||
@@ -4,13 +4,14 @@ use std::sync::Arc;
|
|||||||
use zeroize::Zeroizing;
|
use zeroize::Zeroizing;
|
||||||
use rand_core::OsRng;
|
use rand_core::OsRng;
|
||||||
use blake2::{digest::typenum::U32, Digest, Blake2s};
|
use blake2::{digest::typenum::U32, Digest, Blake2s};
|
||||||
use ciphersuite::Ciphersuite;
|
use ciphersuite::*;
|
||||||
use dalek_ff_group::Ristretto;
|
use dalek_ff_group::Ristretto;
|
||||||
|
|
||||||
use tokio::sync::mpsc;
|
use tokio::sync::mpsc;
|
||||||
|
|
||||||
use serai_db::{Get, DbTxn, Db as DbTrait, create_db, db_channel};
|
use serai_db::{Get, DbTxn, Db as DbTrait, create_db, db_channel};
|
||||||
|
|
||||||
|
use scale::Encode;
|
||||||
use serai_client::validator_sets::primitives::ExternalValidatorSet;
|
use serai_client::validator_sets::primitives::ExternalValidatorSet;
|
||||||
|
|
||||||
use tributary_sdk::{TransactionKind, TransactionError, ProvidedError, TransactionTrait, Tributary};
|
use tributary_sdk::{TransactionKind, TransactionError, ProvidedError, TransactionTrait, Tributary};
|
||||||
@@ -159,7 +160,7 @@ impl<CD: DbTrait, TD: DbTrait, P: P2p> ContinuallyRan
|
|||||||
#[must_use]
|
#[must_use]
|
||||||
async fn add_signed_unsigned_transaction<TD: DbTrait, P: P2p>(
|
async fn add_signed_unsigned_transaction<TD: DbTrait, P: P2p>(
|
||||||
tributary: &Tributary<TD, Transaction, P>,
|
tributary: &Tributary<TD, Transaction, P>,
|
||||||
key: &Zeroizing<<Ristretto as Ciphersuite>::F>,
|
key: &Zeroizing<<Ristretto as WrappedGroup>::F>,
|
||||||
mut tx: Transaction,
|
mut tx: Transaction,
|
||||||
) -> bool {
|
) -> bool {
|
||||||
// If this is a signed transaction, sign it
|
// If this is a signed transaction, sign it
|
||||||
@@ -212,7 +213,7 @@ async fn add_with_recognition_check<TD: DbTrait, P: P2p>(
|
|||||||
set: ExternalValidatorSet,
|
set: ExternalValidatorSet,
|
||||||
tributary_db: &mut TD,
|
tributary_db: &mut TD,
|
||||||
tributary: &Tributary<TD, Transaction, P>,
|
tributary: &Tributary<TD, Transaction, P>,
|
||||||
key: &Zeroizing<<Ristretto as Ciphersuite>::F>,
|
key: &Zeroizing<<Ristretto as WrappedGroup>::F>,
|
||||||
tx: Transaction,
|
tx: Transaction,
|
||||||
) -> bool {
|
) -> bool {
|
||||||
let kind = tx.kind();
|
let kind = tx.kind();
|
||||||
@@ -251,7 +252,7 @@ pub(crate) struct AddTributaryTransactionsTask<CD: DbTrait, TD: DbTrait, P: P2p>
|
|||||||
tributary_db: TD,
|
tributary_db: TD,
|
||||||
tributary: Tributary<TD, Transaction, P>,
|
tributary: Tributary<TD, Transaction, P>,
|
||||||
set: NewSetInformation,
|
set: NewSetInformation,
|
||||||
key: Zeroizing<<Ristretto as Ciphersuite>::F>,
|
key: Zeroizing<<Ristretto as WrappedGroup>::F>,
|
||||||
}
|
}
|
||||||
impl<CD: DbTrait, TD: DbTrait, P: P2p> ContinuallyRan for AddTributaryTransactionsTask<CD, TD, P> {
|
impl<CD: DbTrait, TD: DbTrait, P: P2p> ContinuallyRan for AddTributaryTransactionsTask<CD, TD, P> {
|
||||||
type Error = DoesNotError;
|
type Error = DoesNotError;
|
||||||
@@ -381,7 +382,7 @@ pub(crate) struct SignSlashReportTask<CD: DbTrait, TD: DbTrait, P: P2p> {
|
|||||||
tributary_db: TD,
|
tributary_db: TD,
|
||||||
tributary: Tributary<TD, Transaction, P>,
|
tributary: Tributary<TD, Transaction, P>,
|
||||||
set: NewSetInformation,
|
set: NewSetInformation,
|
||||||
key: Zeroizing<<Ristretto as Ciphersuite>::F>,
|
key: Zeroizing<<Ristretto as WrappedGroup>::F>,
|
||||||
}
|
}
|
||||||
impl<CD: DbTrait, TD: DbTrait, P: P2p> ContinuallyRan for SignSlashReportTask<CD, TD, P> {
|
impl<CD: DbTrait, TD: DbTrait, P: P2p> ContinuallyRan for SignSlashReportTask<CD, TD, P> {
|
||||||
type Error = DoesNotError;
|
type Error = DoesNotError;
|
||||||
@@ -469,7 +470,7 @@ pub(crate) async fn spawn_tributary<P: P2p>(
|
|||||||
p2p: P,
|
p2p: P,
|
||||||
p2p_add_tributary: &mpsc::UnboundedSender<(ExternalValidatorSet, Tributary<Db, Transaction, P>)>,
|
p2p_add_tributary: &mpsc::UnboundedSender<(ExternalValidatorSet, Tributary<Db, Transaction, P>)>,
|
||||||
set: NewSetInformation,
|
set: NewSetInformation,
|
||||||
serai_key: Zeroizing<<Ristretto as Ciphersuite>::F>,
|
serai_key: Zeroizing<<Ristretto as WrappedGroup>::F>,
|
||||||
) {
|
) {
|
||||||
// Don't spawn retired Tributaries
|
// Don't spawn retired Tributaries
|
||||||
if crate::db::RetiredTributary::get(&db, set.set.network).map(|session| session.0) >=
|
if crate::db::RetiredTributary::get(&db, set.set.network).map(|session| session.0) >=
|
||||||
@@ -478,8 +479,7 @@ pub(crate) async fn spawn_tributary<P: P2p>(
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
let genesis =
|
let genesis = <[u8; 32]>::from(Blake2s::<U32>::digest((set.serai_block, set.set).encode()));
|
||||||
<[u8; 32]>::from(Blake2s::<U32>::digest(borsh::to_vec(&(set.serai_block, set.set)).unwrap()));
|
|
||||||
|
|
||||||
// Since the Serai block will be finalized, then cosigned, before we handle this, this time will
|
// Since the Serai block will be finalized, then cosigned, before we handle this, this time will
|
||||||
// be a couple of minutes stale. While the Tributary will still function with a start time in the
|
// be a couple of minutes stale. While the Tributary will still function with a start time in the
|
||||||
@@ -490,7 +490,7 @@ pub(crate) async fn spawn_tributary<P: P2p>(
|
|||||||
|
|
||||||
let mut tributary_validators = Vec::with_capacity(set.validators.len());
|
let mut tributary_validators = Vec::with_capacity(set.validators.len());
|
||||||
for (validator, weight) in set.validators.iter().copied() {
|
for (validator, weight) in set.validators.iter().copied() {
|
||||||
let validator_key = <Ristretto as Ciphersuite>::read_G(&mut validator.0.as_slice())
|
let validator_key = <Ristretto as GroupIo>::read_G(&mut validator.0.as_slice())
|
||||||
.expect("Serai validator had an invalid public key");
|
.expect("Serai validator had an invalid public key");
|
||||||
let weight = u64::from(weight);
|
let weight = u64::from(weight);
|
||||||
tributary_validators.push((validator_key, weight));
|
tributary_validators.push((validator_key, weight));
|
||||||
|
|||||||
@@ -20,11 +20,12 @@ workspace = true
|
|||||||
[dependencies]
|
[dependencies]
|
||||||
bitvec = { version = "1", default-features = false, features = ["std"] }
|
bitvec = { version = "1", default-features = false, features = ["std"] }
|
||||||
|
|
||||||
|
scale = { package = "parity-scale-codec", version = "3", default-features = false, features = ["std", "derive", "bit-vec"] }
|
||||||
borsh = { version = "1", default-features = false, features = ["std", "derive", "de_strict_order"] }
|
borsh = { version = "1", default-features = false, features = ["std", "derive", "de_strict_order"] }
|
||||||
|
|
||||||
dkg = { path = "../../crypto/dkg", default-features = false, features = ["std"] }
|
dkg = { path = "../../crypto/dkg", default-features = false, features = ["std"] }
|
||||||
|
|
||||||
serai-client = { path = "../../substrate/client", version = "0.1", default-features = false, features = ["serai"] }
|
serai-client = { path = "../../substrate/client", version = "0.1", default-features = false, features = ["serai", "borsh"] }
|
||||||
|
|
||||||
log = { version = "0.4", default-features = false, features = ["std"] }
|
log = { version = "0.4", default-features = false, features = ["std"] }
|
||||||
|
|
||||||
|
|||||||
@@ -1,9 +1,10 @@
|
|||||||
#![cfg_attr(docsrs, feature(doc_auto_cfg))]
|
#![cfg_attr(docsrs, feature(doc_cfg))]
|
||||||
#![doc = include_str!("../README.md")]
|
#![doc = include_str!("../README.md")]
|
||||||
#![deny(missing_docs)]
|
#![deny(missing_docs)]
|
||||||
|
|
||||||
use std::collections::HashMap;
|
use std::collections::HashMap;
|
||||||
|
|
||||||
|
use scale::{Encode, Decode};
|
||||||
use borsh::{BorshSerialize, BorshDeserialize};
|
use borsh::{BorshSerialize, BorshDeserialize};
|
||||||
|
|
||||||
use dkg::Participant;
|
use dkg::Participant;
|
||||||
@@ -177,13 +178,14 @@ impl Keys {
|
|||||||
signature_participants,
|
signature_participants,
|
||||||
signature,
|
signature,
|
||||||
);
|
);
|
||||||
_public_db::Keys::set(txn, set.network, &(set.session, tx));
|
_public_db::Keys::set(txn, set.network, &(set.session, tx.encode()));
|
||||||
}
|
}
|
||||||
pub(crate) fn take(
|
pub(crate) fn take(
|
||||||
txn: &mut impl DbTxn,
|
txn: &mut impl DbTxn,
|
||||||
network: ExternalNetworkId,
|
network: ExternalNetworkId,
|
||||||
) -> Option<(Session, Transaction)> {
|
) -> Option<(Session, Transaction)> {
|
||||||
_public_db::Keys::take(txn, network)
|
let (session, tx) = _public_db::Keys::take(txn, network)?;
|
||||||
|
Some((session, <_>::decode(&mut tx.as_slice()).unwrap()))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -224,12 +226,13 @@ impl SlashReports {
|
|||||||
slash_report,
|
slash_report,
|
||||||
signature,
|
signature,
|
||||||
);
|
);
|
||||||
_public_db::SlashReports::set(txn, set.network, &(set.session, tx));
|
_public_db::SlashReports::set(txn, set.network, &(set.session, tx.encode()));
|
||||||
}
|
}
|
||||||
pub(crate) fn take(
|
pub(crate) fn take(
|
||||||
txn: &mut impl DbTxn,
|
txn: &mut impl DbTxn,
|
||||||
network: ExternalNetworkId,
|
network: ExternalNetworkId,
|
||||||
) -> Option<(Session, Transaction)> {
|
) -> Option<(Session, Transaction)> {
|
||||||
_public_db::SlashReports::take(txn, network)
|
let (session, tx) = _public_db::SlashReports::take(txn, network)?;
|
||||||
|
Some((session, <_>::decode(&mut tx.as_slice()).unwrap()))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -36,7 +36,7 @@ log = { version = "0.4", default-features = false, features = ["std"] }
|
|||||||
|
|
||||||
serai-db = { path = "../../common/db", version = "0.1" }
|
serai-db = { path = "../../common/db", version = "0.1" }
|
||||||
|
|
||||||
borsh = { version = "1", default-features = false, features = ["std", "derive", "de_strict_order"] }
|
scale = { package = "parity-scale-codec", version = "3", default-features = false, features = ["std", "derive"] }
|
||||||
futures-util = { version = "0.3", default-features = false, features = ["std", "sink", "channel"] }
|
futures-util = { version = "0.3", default-features = false, features = ["std", "sink", "channel"] }
|
||||||
futures-channel = { version = "0.3", default-features = false, features = ["std", "sink"] }
|
futures-channel = { version = "0.3", default-features = false, features = ["std", "sink"] }
|
||||||
tendermint = { package = "tendermint-machine", path = "./tendermint", version = "0.2" }
|
tendermint = { package = "tendermint-machine", path = "./tendermint", version = "0.2" }
|
||||||
|
|||||||
@@ -1,11 +1,11 @@
|
|||||||
use std::collections::{VecDeque, HashSet};
|
use std::collections::{VecDeque, HashSet};
|
||||||
|
|
||||||
use dalek_ff_group::Ristretto;
|
use dalek_ff_group::Ristretto;
|
||||||
use ciphersuite::{group::GroupEncoding, Ciphersuite};
|
use ciphersuite::{group::GroupEncoding, *};
|
||||||
|
|
||||||
use serai_db::{Get, DbTxn, Db};
|
use serai_db::{Get, DbTxn, Db};
|
||||||
|
|
||||||
use borsh::BorshDeserialize;
|
use scale::Decode;
|
||||||
|
|
||||||
use tendermint::ext::{Network, Commit};
|
use tendermint::ext::{Network, Commit};
|
||||||
|
|
||||||
@@ -21,7 +21,7 @@ pub(crate) struct Blockchain<D: Db, T: TransactionTrait> {
|
|||||||
|
|
||||||
block_number: u64,
|
block_number: u64,
|
||||||
tip: [u8; 32],
|
tip: [u8; 32],
|
||||||
participants: HashSet<<Ristretto as Ciphersuite>::G>,
|
participants: HashSet<[u8; 32]>,
|
||||||
|
|
||||||
provided: ProvidedTransactions<D, T>,
|
provided: ProvidedTransactions<D, T>,
|
||||||
mempool: Mempool<D, T>,
|
mempool: Mempool<D, T>,
|
||||||
@@ -56,25 +56,28 @@ impl<D: Db, T: TransactionTrait> Blockchain<D, T> {
|
|||||||
}
|
}
|
||||||
fn next_nonce_key(
|
fn next_nonce_key(
|
||||||
genesis: &[u8; 32],
|
genesis: &[u8; 32],
|
||||||
signer: &<Ristretto as Ciphersuite>::G,
|
signer: &<Ristretto as WrappedGroup>::G,
|
||||||
order: &[u8],
|
order: &[u8],
|
||||||
) -> Vec<u8> {
|
) -> Vec<u8> {
|
||||||
D::key(
|
D::key(
|
||||||
b"tributary_blockchain",
|
b"tributary_blockchain",
|
||||||
b"next_nonce",
|
b"next_nonce",
|
||||||
[genesis.as_slice(), signer.to_bytes().as_slice(), order].concat(),
|
[genesis.as_ref(), signer.to_bytes().as_ref(), order].concat(),
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub(crate) fn new(
|
pub(crate) fn new(
|
||||||
db: D,
|
db: D,
|
||||||
genesis: [u8; 32],
|
genesis: [u8; 32],
|
||||||
participants: &[<Ristretto as Ciphersuite>::G],
|
participants: &[<Ristretto as WrappedGroup>::G],
|
||||||
) -> Self {
|
) -> Self {
|
||||||
let mut res = Self {
|
let mut res = Self {
|
||||||
db: Some(db.clone()),
|
db: Some(db.clone()),
|
||||||
genesis,
|
genesis,
|
||||||
participants: participants.iter().copied().collect(),
|
participants: participants
|
||||||
|
.iter()
|
||||||
|
.map(<<Ristretto as WrappedGroup>::G as GroupEncoding>::to_bytes)
|
||||||
|
.collect(),
|
||||||
|
|
||||||
block_number: 0,
|
block_number: 0,
|
||||||
tip: genesis,
|
tip: genesis,
|
||||||
@@ -106,7 +109,7 @@ impl<D: Db, T: TransactionTrait> Blockchain<D, T> {
|
|||||||
|
|
||||||
pub(crate) fn block_from_db(db: &D, genesis: [u8; 32], block: &[u8; 32]) -> Option<Block<T>> {
|
pub(crate) fn block_from_db(db: &D, genesis: [u8; 32], block: &[u8; 32]) -> Option<Block<T>> {
|
||||||
db.get(Self::block_key(&genesis, block))
|
db.get(Self::block_key(&genesis, block))
|
||||||
.map(|bytes| Block::<T>::read::<&[u8]>(&mut bytes.as_slice()).unwrap())
|
.map(|bytes| Block::<T>::read::<&[u8]>(&mut bytes.as_ref()).unwrap())
|
||||||
}
|
}
|
||||||
|
|
||||||
pub(crate) fn commit_from_db(db: &D, genesis: [u8; 32], block: &[u8; 32]) -> Option<Vec<u8>> {
|
pub(crate) fn commit_from_db(db: &D, genesis: [u8; 32], block: &[u8; 32]) -> Option<Vec<u8>> {
|
||||||
@@ -166,14 +169,14 @@ impl<D: Db, T: TransactionTrait> Blockchain<D, T> {
|
|||||||
// we must have a commit per valid hash
|
// we must have a commit per valid hash
|
||||||
let commit = Self::commit_from_db(db, genesis, &hash).unwrap();
|
let commit = Self::commit_from_db(db, genesis, &hash).unwrap();
|
||||||
// commit has to be valid if it is coming from our db
|
// commit has to be valid if it is coming from our db
|
||||||
Some(Commit::<N::SignatureScheme>::deserialize_reader(&mut commit.as_slice()).unwrap())
|
Some(Commit::<N::SignatureScheme>::decode(&mut commit.as_ref()).unwrap())
|
||||||
};
|
};
|
||||||
let unsigned_in_chain =
|
let unsigned_in_chain =
|
||||||
|hash: [u8; 32]| db.get(Self::unsigned_included_key(&self.genesis, &hash)).is_some();
|
|hash: [u8; 32]| db.get(Self::unsigned_included_key(&self.genesis, &hash)).is_some();
|
||||||
|
|
||||||
self.mempool.add::<N, _>(
|
self.mempool.add::<N, _>(
|
||||||
|signer, order| {
|
|signer, order| {
|
||||||
if self.participants.contains(&signer) {
|
if self.participants.contains(&signer.to_bytes()) {
|
||||||
Some(
|
Some(
|
||||||
db.get(Self::next_nonce_key(&self.genesis, &signer, &order))
|
db.get(Self::next_nonce_key(&self.genesis, &signer, &order))
|
||||||
.map_or(0, |bytes| u32::from_le_bytes(bytes.try_into().unwrap())),
|
.map_or(0, |bytes| u32::from_le_bytes(bytes.try_into().unwrap())),
|
||||||
@@ -196,13 +199,13 @@ impl<D: Db, T: TransactionTrait> Blockchain<D, T> {
|
|||||||
|
|
||||||
pub(crate) fn next_nonce(
|
pub(crate) fn next_nonce(
|
||||||
&self,
|
&self,
|
||||||
signer: &<Ristretto as Ciphersuite>::G,
|
signer: &<Ristretto as WrappedGroup>::G,
|
||||||
order: &[u8],
|
order: &[u8],
|
||||||
) -> Option<u32> {
|
) -> Option<u32> {
|
||||||
if let Some(next_nonce) = self.mempool.next_nonce_in_mempool(signer, order.to_vec()) {
|
if let Some(next_nonce) = self.mempool.next_nonce_in_mempool(signer, order.to_vec()) {
|
||||||
return Some(next_nonce);
|
return Some(next_nonce);
|
||||||
}
|
}
|
||||||
if self.participants.contains(signer) {
|
if self.participants.contains(&signer.to_bytes()) {
|
||||||
Some(
|
Some(
|
||||||
self
|
self
|
||||||
.db
|
.db
|
||||||
@@ -241,7 +244,7 @@ impl<D: Db, T: TransactionTrait> Blockchain<D, T> {
|
|||||||
let commit = |block: u64| -> Option<Commit<N::SignatureScheme>> {
|
let commit = |block: u64| -> Option<Commit<N::SignatureScheme>> {
|
||||||
let commit = self.commit_by_block_number(block)?;
|
let commit = self.commit_by_block_number(block)?;
|
||||||
// commit has to be valid if it is coming from our db
|
// commit has to be valid if it is coming from our db
|
||||||
Some(Commit::<N::SignatureScheme>::deserialize_reader(&mut commit.as_slice()).unwrap())
|
Some(Commit::<N::SignatureScheme>::decode(&mut commit.as_ref()).unwrap())
|
||||||
};
|
};
|
||||||
|
|
||||||
let mut txn_db = db.clone();
|
let mut txn_db = db.clone();
|
||||||
@@ -251,7 +254,7 @@ impl<D: Db, T: TransactionTrait> Blockchain<D, T> {
|
|||||||
self.tip,
|
self.tip,
|
||||||
self.provided.transactions.clone(),
|
self.provided.transactions.clone(),
|
||||||
&mut |signer, order| {
|
&mut |signer, order| {
|
||||||
if self.participants.contains(signer) {
|
if self.participants.contains(&signer.to_bytes()) {
|
||||||
let key = Self::next_nonce_key(&self.genesis, signer, order);
|
let key = Self::next_nonce_key(&self.genesis, signer, order);
|
||||||
let next = txn
|
let next = txn
|
||||||
.get(&key)
|
.get(&key)
|
||||||
|
|||||||
@@ -3,11 +3,10 @@ use std::{sync::Arc, io};
|
|||||||
|
|
||||||
use zeroize::Zeroizing;
|
use zeroize::Zeroizing;
|
||||||
|
|
||||||
use borsh::BorshDeserialize;
|
use ciphersuite::*;
|
||||||
|
|
||||||
use ciphersuite::Ciphersuite;
|
|
||||||
use dalek_ff_group::Ristretto;
|
use dalek_ff_group::Ristretto;
|
||||||
|
|
||||||
|
use scale::Decode;
|
||||||
use futures_channel::mpsc::UnboundedReceiver;
|
use futures_channel::mpsc::UnboundedReceiver;
|
||||||
use futures_util::{StreamExt, SinkExt};
|
use futures_util::{StreamExt, SinkExt};
|
||||||
use ::tendermint::{
|
use ::tendermint::{
|
||||||
@@ -163,8 +162,8 @@ impl<D: Db, T: TransactionTrait, P: P2p> Tributary<D, T, P> {
|
|||||||
db: D,
|
db: D,
|
||||||
genesis: [u8; 32],
|
genesis: [u8; 32],
|
||||||
start_time: u64,
|
start_time: u64,
|
||||||
key: Zeroizing<<Ristretto as Ciphersuite>::F>,
|
key: Zeroizing<<Ristretto as WrappedGroup>::F>,
|
||||||
validators: Vec<(<Ristretto as Ciphersuite>::G, u64)>,
|
validators: Vec<(<Ristretto as WrappedGroup>::G, u64)>,
|
||||||
p2p: P,
|
p2p: P,
|
||||||
) -> Option<Self> {
|
) -> Option<Self> {
|
||||||
log::info!("new Tributary with genesis {}", hex::encode(genesis));
|
log::info!("new Tributary with genesis {}", hex::encode(genesis));
|
||||||
@@ -178,7 +177,7 @@ impl<D: Db, T: TransactionTrait, P: P2p> Tributary<D, T, P> {
|
|||||||
let block_number = BlockNumber(blockchain.block_number());
|
let block_number = BlockNumber(blockchain.block_number());
|
||||||
|
|
||||||
let start_time = if let Some(commit) = blockchain.commit(&blockchain.tip()) {
|
let start_time = if let Some(commit) = blockchain.commit(&blockchain.tip()) {
|
||||||
Commit::<Validators>::deserialize_reader(&mut commit.as_slice()).unwrap().end_time
|
Commit::<Validators>::decode(&mut commit.as_ref()).unwrap().end_time
|
||||||
} else {
|
} else {
|
||||||
start_time
|
start_time
|
||||||
};
|
};
|
||||||
@@ -236,7 +235,7 @@ impl<D: Db, T: TransactionTrait, P: P2p> Tributary<D, T, P> {
|
|||||||
|
|
||||||
pub async fn next_nonce(
|
pub async fn next_nonce(
|
||||||
&self,
|
&self,
|
||||||
signer: &<Ristretto as Ciphersuite>::G,
|
signer: &<Ristretto as WrappedGroup>::G,
|
||||||
order: &[u8],
|
order: &[u8],
|
||||||
) -> Option<u32> {
|
) -> Option<u32> {
|
||||||
self.network.blockchain.read().await.next_nonce(signer, order)
|
self.network.blockchain.read().await.next_nonce(signer, order)
|
||||||
@@ -277,8 +276,8 @@ impl<D: Db, T: TransactionTrait, P: P2p> Tributary<D, T, P> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
let block = TendermintBlock(block.serialize());
|
let block = TendermintBlock(block.serialize());
|
||||||
let mut commit_ref = commit.as_slice();
|
let mut commit_ref = commit.as_ref();
|
||||||
let Ok(commit) = Commit::<Arc<Validators>>::deserialize_reader(&mut commit_ref) else {
|
let Ok(commit) = Commit::<Arc<Validators>>::decode(&mut commit_ref) else {
|
||||||
log::error!("sent an invalidly serialized commit");
|
log::error!("sent an invalidly serialized commit");
|
||||||
return false;
|
return false;
|
||||||
};
|
};
|
||||||
@@ -328,7 +327,7 @@ impl<D: Db, T: TransactionTrait, P: P2p> Tributary<D, T, P> {
|
|||||||
|
|
||||||
Some(&TENDERMINT_MESSAGE) => {
|
Some(&TENDERMINT_MESSAGE) => {
|
||||||
let Ok(msg) =
|
let Ok(msg) =
|
||||||
SignedMessageFor::<TendermintNetwork<D, T, P>>::deserialize_reader(&mut &msg[1 ..])
|
SignedMessageFor::<TendermintNetwork<D, T, P>>::decode::<&[u8]>(&mut &msg[1 ..])
|
||||||
else {
|
else {
|
||||||
log::error!("received invalid tendermint message");
|
log::error!("received invalid tendermint message");
|
||||||
return false;
|
return false;
|
||||||
@@ -368,17 +367,15 @@ impl<D: Db, T: TransactionTrait> TributaryReader<D, T> {
|
|||||||
Blockchain::<D, T>::commit_from_db(&self.0, self.1, hash)
|
Blockchain::<D, T>::commit_from_db(&self.0, self.1, hash)
|
||||||
}
|
}
|
||||||
pub fn parsed_commit(&self, hash: &[u8; 32]) -> Option<Commit<Validators>> {
|
pub fn parsed_commit(&self, hash: &[u8; 32]) -> Option<Commit<Validators>> {
|
||||||
self
|
self.commit(hash).map(|commit| Commit::<Validators>::decode(&mut commit.as_ref()).unwrap())
|
||||||
.commit(hash)
|
|
||||||
.map(|commit| Commit::<Validators>::deserialize_reader(&mut commit.as_slice()).unwrap())
|
|
||||||
}
|
}
|
||||||
pub fn block_after(&self, hash: &[u8; 32]) -> Option<[u8; 32]> {
|
pub fn block_after(&self, hash: &[u8; 32]) -> Option<[u8; 32]> {
|
||||||
Blockchain::<D, T>::block_after(&self.0, self.1, hash)
|
Blockchain::<D, T>::block_after(&self.0, self.1, hash)
|
||||||
}
|
}
|
||||||
pub fn time_of_block(&self, hash: &[u8; 32]) -> Option<u64> {
|
pub fn time_of_block(&self, hash: &[u8; 32]) -> Option<u64> {
|
||||||
self.commit(hash).map(|commit| {
|
self
|
||||||
Commit::<Validators>::deserialize_reader(&mut commit.as_slice()).unwrap().end_time
|
.commit(hash)
|
||||||
})
|
.map(|commit| Commit::<Validators>::decode(&mut commit.as_ref()).unwrap().end_time)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn locally_provided_txs_in_block(&self, hash: &[u8; 32], order: &str) -> bool {
|
pub fn locally_provided_txs_in_block(&self, hash: &[u8; 32], order: &str) -> bool {
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
use std::collections::HashMap;
|
use std::collections::HashMap;
|
||||||
|
|
||||||
use dalek_ff_group::Ristretto;
|
use dalek_ff_group::Ristretto;
|
||||||
use ciphersuite::Ciphersuite;
|
use ciphersuite::{group::GroupEncoding, *};
|
||||||
|
|
||||||
use serai_db::{DbTxn, Db};
|
use serai_db::{DbTxn, Db};
|
||||||
|
|
||||||
@@ -21,9 +21,9 @@ pub(crate) struct Mempool<D: Db, T: TransactionTrait> {
|
|||||||
db: D,
|
db: D,
|
||||||
genesis: [u8; 32],
|
genesis: [u8; 32],
|
||||||
|
|
||||||
last_nonce_in_mempool: HashMap<(<Ristretto as Ciphersuite>::G, Vec<u8>), u32>,
|
last_nonce_in_mempool: HashMap<([u8; 32], Vec<u8>), u32>,
|
||||||
txs: HashMap<[u8; 32], Transaction<T>>,
|
txs: HashMap<[u8; 32], Transaction<T>>,
|
||||||
txs_per_signer: HashMap<<Ristretto as Ciphersuite>::G, u32>,
|
txs_per_signer: HashMap<[u8; 32], u32>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<D: Db, T: TransactionTrait> Mempool<D, T> {
|
impl<D: Db, T: TransactionTrait> Mempool<D, T> {
|
||||||
@@ -82,6 +82,7 @@ impl<D: Db, T: TransactionTrait> Mempool<D, T> {
|
|||||||
}
|
}
|
||||||
Transaction::Application(tx) => match tx.kind() {
|
Transaction::Application(tx) => match tx.kind() {
|
||||||
TransactionKind::Signed(order, Signed { signer, nonce, .. }) => {
|
TransactionKind::Signed(order, Signed { signer, nonce, .. }) => {
|
||||||
|
let signer = signer.to_bytes();
|
||||||
let amount = *res.txs_per_signer.get(&signer).unwrap_or(&0) + 1;
|
let amount = *res.txs_per_signer.get(&signer).unwrap_or(&0) + 1;
|
||||||
res.txs_per_signer.insert(signer, amount);
|
res.txs_per_signer.insert(signer, amount);
|
||||||
|
|
||||||
@@ -107,7 +108,7 @@ impl<D: Db, T: TransactionTrait> Mempool<D, T> {
|
|||||||
// Returns Ok(true) if new, Ok(false) if an already present unsigned, or the error.
|
// Returns Ok(true) if new, Ok(false) if an already present unsigned, or the error.
|
||||||
pub(crate) fn add<
|
pub(crate) fn add<
|
||||||
N: Network,
|
N: Network,
|
||||||
F: FnOnce(<Ristretto as Ciphersuite>::G, Vec<u8>) -> Option<u32>,
|
F: FnOnce(<Ristretto as WrappedGroup>::G, Vec<u8>) -> Option<u32>,
|
||||||
>(
|
>(
|
||||||
&mut self,
|
&mut self,
|
||||||
blockchain_next_nonce: F,
|
blockchain_next_nonce: F,
|
||||||
@@ -140,6 +141,8 @@ impl<D: Db, T: TransactionTrait> Mempool<D, T> {
|
|||||||
};
|
};
|
||||||
let mut next_nonce = blockchain_next_nonce;
|
let mut next_nonce = blockchain_next_nonce;
|
||||||
|
|
||||||
|
let signer = signer.to_bytes();
|
||||||
|
|
||||||
if let Some(mempool_last_nonce) =
|
if let Some(mempool_last_nonce) =
|
||||||
self.last_nonce_in_mempool.get(&(signer, order.clone()))
|
self.last_nonce_in_mempool.get(&(signer, order.clone()))
|
||||||
{
|
{
|
||||||
@@ -179,10 +182,10 @@ impl<D: Db, T: TransactionTrait> Mempool<D, T> {
|
|||||||
// Returns None if the mempool doesn't have a nonce tracked.
|
// Returns None if the mempool doesn't have a nonce tracked.
|
||||||
pub(crate) fn next_nonce_in_mempool(
|
pub(crate) fn next_nonce_in_mempool(
|
||||||
&self,
|
&self,
|
||||||
signer: &<Ristretto as Ciphersuite>::G,
|
signer: &<Ristretto as WrappedGroup>::G,
|
||||||
order: Vec<u8>,
|
order: Vec<u8>,
|
||||||
) -> Option<u32> {
|
) -> Option<u32> {
|
||||||
self.last_nonce_in_mempool.get(&(*signer, order)).copied().map(|nonce| nonce + 1)
|
self.last_nonce_in_mempool.get(&(signer.to_bytes(), order)).copied().map(|nonce| nonce + 1)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Get transactions to include in a block.
|
/// Get transactions to include in a block.
|
||||||
@@ -243,6 +246,8 @@ impl<D: Db, T: TransactionTrait> Mempool<D, T> {
|
|||||||
|
|
||||||
if let Some(tx) = self.txs.remove(tx) {
|
if let Some(tx) = self.txs.remove(tx) {
|
||||||
if let TransactionKind::Signed(order, Signed { signer, nonce, .. }) = tx.kind() {
|
if let TransactionKind::Signed(order, Signed { signer, nonce, .. }) = tx.kind() {
|
||||||
|
let signer = signer.to_bytes();
|
||||||
|
|
||||||
let amount = *self.txs_per_signer.get(&signer).unwrap() - 1;
|
let amount = *self.txs_per_signer.get(&signer).unwrap() - 1;
|
||||||
self.txs_per_signer.insert(signer, amount);
|
self.txs_per_signer.insert(signer, amount);
|
||||||
|
|
||||||
|
|||||||
@@ -10,11 +10,8 @@ use rand_chacha::ChaCha12Rng;
|
|||||||
use transcript::{Transcript, RecommendedTranscript};
|
use transcript::{Transcript, RecommendedTranscript};
|
||||||
|
|
||||||
use ciphersuite::{
|
use ciphersuite::{
|
||||||
group::{
|
group::{ff::PrimeField, GroupEncoding},
|
||||||
GroupEncoding,
|
*,
|
||||||
ff::{Field, PrimeField},
|
|
||||||
},
|
|
||||||
Ciphersuite,
|
|
||||||
};
|
};
|
||||||
use dalek_ff_group::Ristretto;
|
use dalek_ff_group::Ristretto;
|
||||||
use schnorr::{
|
use schnorr::{
|
||||||
@@ -24,7 +21,7 @@ use schnorr::{
|
|||||||
|
|
||||||
use serai_db::Db;
|
use serai_db::Db;
|
||||||
|
|
||||||
use borsh::{BorshSerialize, BorshDeserialize};
|
use scale::{Encode, Decode};
|
||||||
use tendermint::{
|
use tendermint::{
|
||||||
SignedMessageFor,
|
SignedMessageFor,
|
||||||
ext::{
|
ext::{
|
||||||
@@ -51,24 +48,26 @@ fn challenge(
|
|||||||
key: [u8; 32],
|
key: [u8; 32],
|
||||||
nonce: &[u8],
|
nonce: &[u8],
|
||||||
msg: &[u8],
|
msg: &[u8],
|
||||||
) -> <Ristretto as Ciphersuite>::F {
|
) -> <Ristretto as WrappedGroup>::F {
|
||||||
let mut transcript = RecommendedTranscript::new(b"Tributary Chain Tendermint Message");
|
let mut transcript = RecommendedTranscript::new(b"Tributary Chain Tendermint Message");
|
||||||
transcript.append_message(b"genesis", genesis);
|
transcript.append_message(b"genesis", genesis);
|
||||||
transcript.append_message(b"key", key);
|
transcript.append_message(b"key", key);
|
||||||
transcript.append_message(b"nonce", nonce);
|
transcript.append_message(b"nonce", nonce);
|
||||||
transcript.append_message(b"message", msg);
|
transcript.append_message(b"message", msg);
|
||||||
|
|
||||||
<Ristretto as Ciphersuite>::F::from_bytes_mod_order_wide(&transcript.challenge(b"schnorr").into())
|
<Ristretto as WrappedGroup>::F::from_bytes_mod_order_wide(
|
||||||
|
&transcript.challenge(b"schnorr").into(),
|
||||||
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, PartialEq, Eq, Debug)]
|
#[derive(Clone, PartialEq, Eq, Debug)]
|
||||||
pub struct Signer {
|
pub struct Signer {
|
||||||
genesis: [u8; 32],
|
genesis: [u8; 32],
|
||||||
key: Zeroizing<<Ristretto as Ciphersuite>::F>,
|
key: Zeroizing<<Ristretto as WrappedGroup>::F>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Signer {
|
impl Signer {
|
||||||
pub(crate) fn new(genesis: [u8; 32], key: Zeroizing<<Ristretto as Ciphersuite>::F>) -> Signer {
|
pub(crate) fn new(genesis: [u8; 32], key: Zeroizing<<Ristretto as WrappedGroup>::F>) -> Signer {
|
||||||
Signer { genesis, key }
|
Signer { genesis, key }
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -101,10 +100,10 @@ impl SignerTrait for Signer {
|
|||||||
assert_eq!(nonce_ref, [0; 64].as_ref());
|
assert_eq!(nonce_ref, [0; 64].as_ref());
|
||||||
|
|
||||||
let nonce =
|
let nonce =
|
||||||
Zeroizing::new(<Ristretto as Ciphersuite>::F::from_bytes_mod_order_wide(&nonce_arr));
|
Zeroizing::new(<Ristretto as WrappedGroup>::F::from_bytes_mod_order_wide(&nonce_arr));
|
||||||
nonce_arr.zeroize();
|
nonce_arr.zeroize();
|
||||||
|
|
||||||
assert!(!bool::from(nonce.ct_eq(&<Ristretto as Ciphersuite>::F::ZERO)));
|
assert!(!bool::from(nonce.ct_eq(&<Ristretto as WrappedGroup>::F::ZERO)));
|
||||||
|
|
||||||
let challenge = challenge(
|
let challenge = challenge(
|
||||||
self.genesis,
|
self.genesis,
|
||||||
@@ -133,7 +132,7 @@ pub struct Validators {
|
|||||||
impl Validators {
|
impl Validators {
|
||||||
pub(crate) fn new(
|
pub(crate) fn new(
|
||||||
genesis: [u8; 32],
|
genesis: [u8; 32],
|
||||||
validators: Vec<(<Ristretto as Ciphersuite>::G, u64)>,
|
validators: Vec<(<Ristretto as WrappedGroup>::G, u64)>,
|
||||||
) -> Option<Validators> {
|
) -> Option<Validators> {
|
||||||
let mut total_weight = 0;
|
let mut total_weight = 0;
|
||||||
let mut weights = HashMap::new();
|
let mut weights = HashMap::new();
|
||||||
@@ -220,7 +219,7 @@ impl SignatureScheme for Validators {
|
|||||||
signers
|
signers
|
||||||
.iter()
|
.iter()
|
||||||
.zip(challenges)
|
.zip(challenges)
|
||||||
.map(|(s, c)| (<Ristretto as Ciphersuite>::read_G(&mut s.as_slice()).unwrap(), c))
|
.map(|(s, c)| (<Ristretto as GroupIo>::read_G(&mut s.as_slice()).unwrap(), c))
|
||||||
.collect::<Vec<_>>()
|
.collect::<Vec<_>>()
|
||||||
.as_slice(),
|
.as_slice(),
|
||||||
)
|
)
|
||||||
@@ -249,7 +248,7 @@ impl Weights for Validators {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, PartialEq, Eq, Debug, BorshSerialize, BorshDeserialize)]
|
#[derive(Clone, PartialEq, Eq, Debug, Encode, Decode)]
|
||||||
pub struct TendermintBlock(pub Vec<u8>);
|
pub struct TendermintBlock(pub Vec<u8>);
|
||||||
impl BlockTrait for TendermintBlock {
|
impl BlockTrait for TendermintBlock {
|
||||||
type Id = [u8; 32];
|
type Id = [u8; 32];
|
||||||
@@ -301,7 +300,7 @@ impl<D: Db, T: TransactionTrait, P: P2p> Network for TendermintNetwork<D, T, P>
|
|||||||
fn broadcast(&mut self, msg: SignedMessageFor<Self>) -> impl Send + Future<Output = ()> {
|
fn broadcast(&mut self, msg: SignedMessageFor<Self>) -> impl Send + Future<Output = ()> {
|
||||||
async move {
|
async move {
|
||||||
let mut to_broadcast = vec![TENDERMINT_MESSAGE];
|
let mut to_broadcast = vec![TENDERMINT_MESSAGE];
|
||||||
msg.serialize(&mut to_broadcast).unwrap();
|
to_broadcast.extend(msg.encode());
|
||||||
self.p2p.broadcast(self.genesis, to_broadcast).await
|
self.p2p.broadcast(self.genesis, to_broadcast).await
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -391,7 +390,7 @@ impl<D: Db, T: TransactionTrait, P: P2p> Network for TendermintNetwork<D, T, P>
|
|||||||
return invalid_block();
|
return invalid_block();
|
||||||
};
|
};
|
||||||
|
|
||||||
let encoded_commit = borsh::to_vec(&commit).unwrap();
|
let encoded_commit = commit.encode();
|
||||||
loop {
|
loop {
|
||||||
let block_res = self.blockchain.write().await.add_block::<Self>(
|
let block_res = self.blockchain.write().await.add_block::<Self>(
|
||||||
&block,
|
&block,
|
||||||
|
|||||||
@@ -1,11 +1,11 @@
|
|||||||
use std::io;
|
use std::io;
|
||||||
|
|
||||||
use borsh::BorshDeserialize;
|
use scale::{Encode, Decode, IoReader};
|
||||||
|
|
||||||
use blake2::{Digest, Blake2s256};
|
use blake2::{Digest, Blake2s256};
|
||||||
|
|
||||||
use dalek_ff_group::Ristretto;
|
use dalek_ff_group::Ristretto;
|
||||||
use ciphersuite::Ciphersuite;
|
use ciphersuite::*;
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
transaction::{Transaction, TransactionKind, TransactionError},
|
transaction::{Transaction, TransactionKind, TransactionError},
|
||||||
@@ -27,14 +27,14 @@ pub enum TendermintTx {
|
|||||||
|
|
||||||
impl ReadWrite for TendermintTx {
|
impl ReadWrite for TendermintTx {
|
||||||
fn read<R: io::Read>(reader: &mut R) -> io::Result<Self> {
|
fn read<R: io::Read>(reader: &mut R) -> io::Result<Self> {
|
||||||
Evidence::deserialize_reader(reader)
|
Evidence::decode(&mut IoReader(reader))
|
||||||
.map(TendermintTx::SlashEvidence)
|
.map(TendermintTx::SlashEvidence)
|
||||||
.map_err(|_| io::Error::new(io::ErrorKind::InvalidData, "invalid evidence format"))
|
.map_err(|_| io::Error::new(io::ErrorKind::InvalidData, "invalid evidence format"))
|
||||||
}
|
}
|
||||||
|
|
||||||
fn write<W: io::Write>(&self, writer: &mut W) -> io::Result<()> {
|
fn write<W: io::Write>(&self, writer: &mut W) -> io::Result<()> {
|
||||||
match self {
|
match self {
|
||||||
TendermintTx::SlashEvidence(ev) => writer.write_all(&borsh::to_vec(&ev).unwrap()),
|
TendermintTx::SlashEvidence(ev) => writer.write_all(&ev.encode()),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -50,7 +50,7 @@ impl Transaction for TendermintTx {
|
|||||||
Blake2s256::digest(self.serialize()).into()
|
Blake2s256::digest(self.serialize()).into()
|
||||||
}
|
}
|
||||||
|
|
||||||
fn sig_hash(&self, _genesis: [u8; 32]) -> <Ristretto as Ciphersuite>::F {
|
fn sig_hash(&self, _genesis: [u8; 32]) -> <Ristretto as WrappedGroup>::F {
|
||||||
match self {
|
match self {
|
||||||
TendermintTx::SlashEvidence(_) => panic!("sig_hash called on slash evidence transaction"),
|
TendermintTx::SlashEvidence(_) => panic!("sig_hash called on slash evidence transaction"),
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -3,10 +3,7 @@ use std::{sync::Arc, io, collections::HashMap, fmt::Debug};
|
|||||||
use blake2::{Digest, Blake2s256};
|
use blake2::{Digest, Blake2s256};
|
||||||
|
|
||||||
use dalek_ff_group::Ristretto;
|
use dalek_ff_group::Ristretto;
|
||||||
use ciphersuite::{
|
use ciphersuite::{group::Group, *};
|
||||||
group::{ff::Field, Group},
|
|
||||||
Ciphersuite,
|
|
||||||
};
|
|
||||||
use schnorr::SchnorrSignature;
|
use schnorr::SchnorrSignature;
|
||||||
|
|
||||||
use serai_db::MemDb;
|
use serai_db::MemDb;
|
||||||
@@ -32,11 +29,11 @@ impl NonceTransaction {
|
|||||||
nonce,
|
nonce,
|
||||||
distinguisher,
|
distinguisher,
|
||||||
Signed {
|
Signed {
|
||||||
signer: <Ristretto as Ciphersuite>::G::identity(),
|
signer: <Ristretto as WrappedGroup>::G::identity(),
|
||||||
nonce,
|
nonce,
|
||||||
signature: SchnorrSignature::<Ristretto> {
|
signature: SchnorrSignature::<Ristretto> {
|
||||||
R: <Ristretto as Ciphersuite>::G::identity(),
|
R: <Ristretto as WrappedGroup>::G::identity(),
|
||||||
s: <Ristretto as Ciphersuite>::F::ZERO,
|
s: <Ristretto as WrappedGroup>::F::ZERO,
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
)
|
)
|
||||||
|
|||||||
@@ -11,7 +11,7 @@ use rand::rngs::OsRng;
|
|||||||
use blake2::{Digest, Blake2s256};
|
use blake2::{Digest, Blake2s256};
|
||||||
|
|
||||||
use dalek_ff_group::Ristretto;
|
use dalek_ff_group::Ristretto;
|
||||||
use ciphersuite::{group::ff::Field, Ciphersuite};
|
use ciphersuite::*;
|
||||||
|
|
||||||
use serai_db::{DbTxn, Db, MemDb};
|
use serai_db::{DbTxn, Db, MemDb};
|
||||||
|
|
||||||
@@ -31,7 +31,7 @@ type N = TendermintNetwork<MemDb, SignedTransaction, DummyP2p>;
|
|||||||
|
|
||||||
fn new_blockchain<T: TransactionTrait>(
|
fn new_blockchain<T: TransactionTrait>(
|
||||||
genesis: [u8; 32],
|
genesis: [u8; 32],
|
||||||
participants: &[<Ristretto as Ciphersuite>::G],
|
participants: &[<Ristretto as WrappedGroup>::G],
|
||||||
) -> (MemDb, Blockchain<MemDb, T>) {
|
) -> (MemDb, Blockchain<MemDb, T>) {
|
||||||
let db = MemDb::new();
|
let db = MemDb::new();
|
||||||
let blockchain = Blockchain::new(db.clone(), genesis, participants);
|
let blockchain = Blockchain::new(db.clone(), genesis, participants);
|
||||||
@@ -82,7 +82,7 @@ fn invalid_block() {
|
|||||||
assert!(blockchain.verify_block::<N>(&block, &validators, false).is_err());
|
assert!(blockchain.verify_block::<N>(&block, &validators, false).is_err());
|
||||||
}
|
}
|
||||||
|
|
||||||
let key = Zeroizing::new(<Ristretto as Ciphersuite>::F::random(&mut OsRng));
|
let key = Zeroizing::new(<Ristretto as WrappedGroup>::F::random(&mut OsRng));
|
||||||
let tx = crate::tests::signed_transaction(&mut OsRng, genesis, &key, 0);
|
let tx = crate::tests::signed_transaction(&mut OsRng, genesis, &key, 0);
|
||||||
|
|
||||||
// Not a participant
|
// Not a participant
|
||||||
@@ -134,7 +134,7 @@ fn invalid_block() {
|
|||||||
blockchain.verify_block::<N>(&block, &validators, false).unwrap();
|
blockchain.verify_block::<N>(&block, &validators, false).unwrap();
|
||||||
match &mut block.transactions[0] {
|
match &mut block.transactions[0] {
|
||||||
Transaction::Application(tx) => {
|
Transaction::Application(tx) => {
|
||||||
tx.1.signature.s += <Ristretto as Ciphersuite>::F::ONE;
|
tx.1.signature.s += <Ristretto as WrappedGroup>::F::ONE;
|
||||||
}
|
}
|
||||||
_ => panic!("non-signed tx found"),
|
_ => panic!("non-signed tx found"),
|
||||||
}
|
}
|
||||||
@@ -150,7 +150,7 @@ fn invalid_block() {
|
|||||||
fn signed_transaction() {
|
fn signed_transaction() {
|
||||||
let genesis = new_genesis();
|
let genesis = new_genesis();
|
||||||
let validators = Arc::new(Validators::new(genesis, vec![]).unwrap());
|
let validators = Arc::new(Validators::new(genesis, vec![]).unwrap());
|
||||||
let key = Zeroizing::new(<Ristretto as Ciphersuite>::F::random(&mut OsRng));
|
let key = Zeroizing::new(<Ristretto as WrappedGroup>::F::random(&mut OsRng));
|
||||||
let tx = crate::tests::signed_transaction(&mut OsRng, genesis, &key, 0);
|
let tx = crate::tests::signed_transaction(&mut OsRng, genesis, &key, 0);
|
||||||
let signer = tx.1.signer;
|
let signer = tx.1.signer;
|
||||||
|
|
||||||
@@ -339,7 +339,7 @@ fn provided_transaction() {
|
|||||||
#[tokio::test]
|
#[tokio::test]
|
||||||
async fn tendermint_evidence_tx() {
|
async fn tendermint_evidence_tx() {
|
||||||
let genesis = new_genesis();
|
let genesis = new_genesis();
|
||||||
let key = Zeroizing::new(<Ristretto as Ciphersuite>::F::random(&mut OsRng));
|
let key = Zeroizing::new(<Ristretto as WrappedGroup>::F::random(&mut OsRng));
|
||||||
let signer = Signer::new(genesis, key.clone());
|
let signer = Signer::new(genesis, key.clone());
|
||||||
let signer_id = Ristretto::generator() * key.deref();
|
let signer_id = Ristretto::generator() * key.deref();
|
||||||
let validators = Arc::new(Validators::new(genesis, vec![(signer_id, 1)]).unwrap());
|
let validators = Arc::new(Validators::new(genesis, vec![(signer_id, 1)]).unwrap());
|
||||||
@@ -379,7 +379,7 @@ async fn tendermint_evidence_tx() {
|
|||||||
let mut mempool: Vec<Transaction<SignedTransaction>> = vec![];
|
let mut mempool: Vec<Transaction<SignedTransaction>> = vec![];
|
||||||
let mut signers = vec![];
|
let mut signers = vec![];
|
||||||
for _ in 0 .. 5 {
|
for _ in 0 .. 5 {
|
||||||
let key = Zeroizing::new(<Ristretto as Ciphersuite>::F::random(&mut OsRng));
|
let key = Zeroizing::new(<Ristretto as WrappedGroup>::F::random(&mut OsRng));
|
||||||
let signer = Signer::new(genesis, key.clone());
|
let signer = Signer::new(genesis, key.clone());
|
||||||
let signer_id = Ristretto::generator() * key.deref();
|
let signer_id = Ristretto::generator() * key.deref();
|
||||||
signers.push((signer_id, 1));
|
signers.push((signer_id, 1));
|
||||||
@@ -446,7 +446,7 @@ async fn block_tx_ordering() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
let genesis = new_genesis();
|
let genesis = new_genesis();
|
||||||
let key = Zeroizing::new(<Ristretto as Ciphersuite>::F::random(&mut OsRng));
|
let key = Zeroizing::new(<Ristretto as WrappedGroup>::F::random(&mut OsRng));
|
||||||
|
|
||||||
// signer
|
// signer
|
||||||
let signer = crate::tests::signed_transaction(&mut OsRng, genesis, &key, 0).1.signer;
|
let signer = crate::tests::signed_transaction(&mut OsRng, genesis, &key, 0).1.signer;
|
||||||
|
|||||||
@@ -4,7 +4,7 @@ use zeroize::Zeroizing;
|
|||||||
use rand::{RngCore, rngs::OsRng};
|
use rand::{RngCore, rngs::OsRng};
|
||||||
|
|
||||||
use dalek_ff_group::Ristretto;
|
use dalek_ff_group::Ristretto;
|
||||||
use ciphersuite::{group::ff::Field, Ciphersuite};
|
use ciphersuite::*;
|
||||||
|
|
||||||
use tendermint::ext::Commit;
|
use tendermint::ext::Commit;
|
||||||
|
|
||||||
@@ -33,7 +33,7 @@ async fn mempool_addition() {
|
|||||||
Some(Commit::<Arc<Validators>> { end_time: 0, validators: vec![], signature: vec![] })
|
Some(Commit::<Arc<Validators>> { end_time: 0, validators: vec![], signature: vec![] })
|
||||||
};
|
};
|
||||||
let unsigned_in_chain = |_: [u8; 32]| false;
|
let unsigned_in_chain = |_: [u8; 32]| false;
|
||||||
let key = Zeroizing::new(<Ristretto as Ciphersuite>::F::random(&mut OsRng));
|
let key = Zeroizing::new(<Ristretto as WrappedGroup>::F::random(&mut OsRng));
|
||||||
|
|
||||||
let first_tx = signed_transaction(&mut OsRng, genesis, &key, 0);
|
let first_tx = signed_transaction(&mut OsRng, genesis, &key, 0);
|
||||||
let signer = first_tx.1.signer;
|
let signer = first_tx.1.signer;
|
||||||
@@ -125,7 +125,7 @@ async fn mempool_addition() {
|
|||||||
|
|
||||||
// If the mempool doesn't have a nonce for an account, it should successfully use the
|
// If the mempool doesn't have a nonce for an account, it should successfully use the
|
||||||
// blockchain's
|
// blockchain's
|
||||||
let second_key = Zeroizing::new(<Ristretto as Ciphersuite>::F::random(&mut OsRng));
|
let second_key = Zeroizing::new(<Ristretto as WrappedGroup>::F::random(&mut OsRng));
|
||||||
let tx = signed_transaction(&mut OsRng, genesis, &second_key, 2);
|
let tx = signed_transaction(&mut OsRng, genesis, &second_key, 2);
|
||||||
let second_signer = tx.1.signer;
|
let second_signer = tx.1.signer;
|
||||||
assert_eq!(mempool.next_nonce_in_mempool(&second_signer, vec![]), None);
|
assert_eq!(mempool.next_nonce_in_mempool(&second_signer, vec![]), None);
|
||||||
@@ -165,7 +165,7 @@ fn too_many_mempool() {
|
|||||||
Some(Commit::<Arc<Validators>> { end_time: 0, validators: vec![], signature: vec![] })
|
Some(Commit::<Arc<Validators>> { end_time: 0, validators: vec![], signature: vec![] })
|
||||||
};
|
};
|
||||||
let unsigned_in_chain = |_: [u8; 32]| false;
|
let unsigned_in_chain = |_: [u8; 32]| false;
|
||||||
let key = Zeroizing::new(<Ristretto as Ciphersuite>::F::random(&mut OsRng));
|
let key = Zeroizing::new(<Ristretto as WrappedGroup>::F::random(&mut OsRng));
|
||||||
|
|
||||||
// We should be able to add transactions up to the limit
|
// We should be able to add transactions up to the limit
|
||||||
for i in 0 .. ACCOUNT_MEMPOOL_LIMIT {
|
for i in 0 .. ACCOUNT_MEMPOOL_LIMIT {
|
||||||
|
|||||||
@@ -7,12 +7,11 @@ use rand::{RngCore, CryptoRng, rngs::OsRng};
|
|||||||
use blake2::{Digest, Blake2s256};
|
use blake2::{Digest, Blake2s256};
|
||||||
|
|
||||||
use dalek_ff_group::Ristretto;
|
use dalek_ff_group::Ristretto;
|
||||||
use ciphersuite::{
|
use ciphersuite::*;
|
||||||
group::{ff::Field, Group},
|
|
||||||
Ciphersuite,
|
|
||||||
};
|
|
||||||
use schnorr::SchnorrSignature;
|
use schnorr::SchnorrSignature;
|
||||||
|
|
||||||
|
use scale::Encode;
|
||||||
|
|
||||||
use ::tendermint::{
|
use ::tendermint::{
|
||||||
ext::{Network, Signer as SignerTrait, SignatureScheme, BlockNumber, RoundNumber},
|
ext::{Network, Signer as SignerTrait, SignatureScheme, BlockNumber, RoundNumber},
|
||||||
SignedMessageFor, DataFor, Message, SignedMessage, Data, Evidence,
|
SignedMessageFor, DataFor, Message, SignedMessage, Data, Evidence,
|
||||||
@@ -32,11 +31,11 @@ mod tendermint;
|
|||||||
|
|
||||||
pub fn random_signed<R: RngCore + CryptoRng>(rng: &mut R) -> Signed {
|
pub fn random_signed<R: RngCore + CryptoRng>(rng: &mut R) -> Signed {
|
||||||
Signed {
|
Signed {
|
||||||
signer: <Ristretto as Ciphersuite>::G::random(&mut *rng),
|
signer: <Ristretto as WrappedGroup>::G::random(&mut *rng),
|
||||||
nonce: u32::try_from(rng.next_u64() >> 32 >> 1).unwrap(),
|
nonce: u32::try_from(rng.next_u64() >> 32 >> 1).unwrap(),
|
||||||
signature: SchnorrSignature::<Ristretto> {
|
signature: SchnorrSignature::<Ristretto> {
|
||||||
R: <Ristretto as Ciphersuite>::G::random(&mut *rng),
|
R: <Ristretto as WrappedGroup>::G::random(&mut *rng),
|
||||||
s: <Ristretto as Ciphersuite>::F::random(rng),
|
s: <Ristretto as WrappedGroup>::F::random(rng),
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -135,18 +134,18 @@ impl Transaction for SignedTransaction {
|
|||||||
pub fn signed_transaction<R: RngCore + CryptoRng>(
|
pub fn signed_transaction<R: RngCore + CryptoRng>(
|
||||||
rng: &mut R,
|
rng: &mut R,
|
||||||
genesis: [u8; 32],
|
genesis: [u8; 32],
|
||||||
key: &Zeroizing<<Ristretto as Ciphersuite>::F>,
|
key: &Zeroizing<<Ristretto as WrappedGroup>::F>,
|
||||||
nonce: u32,
|
nonce: u32,
|
||||||
) -> SignedTransaction {
|
) -> SignedTransaction {
|
||||||
let mut data = vec![0; 512];
|
let mut data = vec![0; 512];
|
||||||
rng.fill_bytes(&mut data);
|
rng.fill_bytes(&mut data);
|
||||||
|
|
||||||
let signer = <Ristretto as Ciphersuite>::generator() * **key;
|
let signer = <Ristretto as WrappedGroup>::generator() * **key;
|
||||||
|
|
||||||
let mut tx =
|
let mut tx =
|
||||||
SignedTransaction(data, Signed { signer, nonce, signature: random_signed(rng).signature });
|
SignedTransaction(data, Signed { signer, nonce, signature: random_signed(rng).signature });
|
||||||
|
|
||||||
let sig_nonce = Zeroizing::new(<Ristretto as Ciphersuite>::F::random(rng));
|
let sig_nonce = Zeroizing::new(<Ristretto as WrappedGroup>::F::random(rng));
|
||||||
tx.1.signature.R = Ristretto::generator() * sig_nonce.deref();
|
tx.1.signature.R = Ristretto::generator() * sig_nonce.deref();
|
||||||
tx.1.signature = SchnorrSignature::sign(key, sig_nonce, tx.sig_hash(genesis));
|
tx.1.signature = SchnorrSignature::sign(key, sig_nonce, tx.sig_hash(genesis));
|
||||||
|
|
||||||
@@ -161,7 +160,7 @@ pub fn random_signed_transaction<R: RngCore + CryptoRng>(
|
|||||||
let mut genesis = [0; 32];
|
let mut genesis = [0; 32];
|
||||||
rng.fill_bytes(&mut genesis);
|
rng.fill_bytes(&mut genesis);
|
||||||
|
|
||||||
let key = Zeroizing::new(<Ristretto as Ciphersuite>::F::random(&mut *rng));
|
let key = Zeroizing::new(<Ristretto as WrappedGroup>::F::random(&mut *rng));
|
||||||
// Shift over an additional bit to ensure it won't overflow when incremented
|
// Shift over an additional bit to ensure it won't overflow when incremented
|
||||||
let nonce = u32::try_from(rng.next_u64() >> 32 >> 1).unwrap();
|
let nonce = u32::try_from(rng.next_u64() >> 32 >> 1).unwrap();
|
||||||
|
|
||||||
@@ -178,12 +177,11 @@ pub async fn tendermint_meta() -> ([u8; 32], Signer, [u8; 32], Arc<Validators>)
|
|||||||
// signer
|
// signer
|
||||||
let genesis = new_genesis();
|
let genesis = new_genesis();
|
||||||
let signer =
|
let signer =
|
||||||
Signer::new(genesis, Zeroizing::new(<Ristretto as Ciphersuite>::F::random(&mut OsRng)));
|
Signer::new(genesis, Zeroizing::new(<Ristretto as WrappedGroup>::F::random(&mut OsRng)));
|
||||||
let validator_id = signer.validator_id().await.unwrap();
|
let validator_id = signer.validator_id().await.unwrap();
|
||||||
|
|
||||||
// schema
|
// schema
|
||||||
let signer_pub =
|
let signer_pub = <Ristretto as GroupIo>::read_G::<&[u8]>(&mut validator_id.as_slice()).unwrap();
|
||||||
<Ristretto as Ciphersuite>::read_G::<&[u8]>(&mut validator_id.as_slice()).unwrap();
|
|
||||||
let validators = Arc::new(Validators::new(genesis, vec![(signer_pub, 1)]).unwrap());
|
let validators = Arc::new(Validators::new(genesis, vec![(signer_pub, 1)]).unwrap());
|
||||||
|
|
||||||
(genesis, signer, validator_id, validators)
|
(genesis, signer, validator_id, validators)
|
||||||
@@ -202,7 +200,7 @@ pub async fn signed_from_data<N: Network>(
|
|||||||
round: RoundNumber(round_number),
|
round: RoundNumber(round_number),
|
||||||
data,
|
data,
|
||||||
};
|
};
|
||||||
let sig = signer.sign(&borsh::to_vec(&msg).unwrap()).await;
|
let sig = signer.sign(&msg.encode()).await;
|
||||||
SignedMessage { msg, sig }
|
SignedMessage { msg, sig }
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -215,5 +213,5 @@ pub async fn random_evidence_tx<N: Network>(
|
|||||||
let data = Data::Proposal(Some(RoundNumber(0)), b);
|
let data = Data::Proposal(Some(RoundNumber(0)), b);
|
||||||
let signer_id = signer.validator_id().await.unwrap();
|
let signer_id = signer.validator_id().await.unwrap();
|
||||||
let signed = signed_from_data::<N>(signer, signer_id, 0, 0, data).await;
|
let signed = signed_from_data::<N>(signer, signer_id, 0, 0, data).await;
|
||||||
TendermintTx::SlashEvidence(Evidence::InvalidValidRound(borsh::to_vec(&signed).unwrap()))
|
TendermintTx::SlashEvidence(Evidence::InvalidValidRound(signed.encode()))
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -3,7 +3,7 @@ use rand::rngs::OsRng;
|
|||||||
use blake2::{Digest, Blake2s256};
|
use blake2::{Digest, Blake2s256};
|
||||||
|
|
||||||
use dalek_ff_group::Ristretto;
|
use dalek_ff_group::Ristretto;
|
||||||
use ciphersuite::{group::ff::Field, Ciphersuite};
|
use ciphersuite::*;
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
ReadWrite,
|
ReadWrite,
|
||||||
@@ -69,7 +69,7 @@ fn signed_transaction() {
|
|||||||
}
|
}
|
||||||
{
|
{
|
||||||
let mut tx = tx.clone();
|
let mut tx = tx.clone();
|
||||||
tx.1.signature.s += <Ristretto as Ciphersuite>::F::ONE;
|
tx.1.signature.s += <Ristretto as WrappedGroup>::F::ONE;
|
||||||
assert!(verify_transaction(&tx, genesis, &mut |_, _| Some(tx.1.nonce)).is_err());
|
assert!(verify_transaction(&tx, genesis, &mut |_, _| Some(tx.1.nonce)).is_err());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -4,7 +4,9 @@ use zeroize::Zeroizing;
|
|||||||
use rand::{RngCore, rngs::OsRng};
|
use rand::{RngCore, rngs::OsRng};
|
||||||
|
|
||||||
use dalek_ff_group::Ristretto;
|
use dalek_ff_group::Ristretto;
|
||||||
use ciphersuite::{Ciphersuite, group::ff::Field};
|
use ciphersuite::*;
|
||||||
|
|
||||||
|
use scale::Encode;
|
||||||
|
|
||||||
use tendermint::{
|
use tendermint::{
|
||||||
time::CanonicalInstant,
|
time::CanonicalInstant,
|
||||||
@@ -50,10 +52,7 @@ async fn invalid_valid_round() {
|
|||||||
async move {
|
async move {
|
||||||
let data = Data::Proposal(valid_round, TendermintBlock(vec![]));
|
let data = Data::Proposal(valid_round, TendermintBlock(vec![]));
|
||||||
let signed = signed_from_data::<N>(signer.clone().into(), signer_id, 0, 0, data).await;
|
let signed = signed_from_data::<N>(signer.clone().into(), signer_id, 0, 0, data).await;
|
||||||
(
|
(signed.clone(), TendermintTx::SlashEvidence(Evidence::InvalidValidRound(signed.encode())))
|
||||||
signed.clone(),
|
|
||||||
TendermintTx::SlashEvidence(Evidence::InvalidValidRound(borsh::to_vec(&signed).unwrap())),
|
|
||||||
)
|
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
@@ -71,8 +70,7 @@ async fn invalid_valid_round() {
|
|||||||
let mut random_sig = [0u8; 64];
|
let mut random_sig = [0u8; 64];
|
||||||
OsRng.fill_bytes(&mut random_sig);
|
OsRng.fill_bytes(&mut random_sig);
|
||||||
signed.sig = random_sig;
|
signed.sig = random_sig;
|
||||||
let tx =
|
let tx = TendermintTx::SlashEvidence(Evidence::InvalidValidRound(signed.encode()));
|
||||||
TendermintTx::SlashEvidence(Evidence::InvalidValidRound(borsh::to_vec(&signed).unwrap()));
|
|
||||||
|
|
||||||
// should fail
|
// should fail
|
||||||
assert!(verify_tendermint_tx::<N>(&tx, &validators, commit).is_err());
|
assert!(verify_tendermint_tx::<N>(&tx, &validators, commit).is_err());
|
||||||
@@ -92,10 +90,7 @@ async fn invalid_precommit_signature() {
|
|||||||
let signed =
|
let signed =
|
||||||
signed_from_data::<N>(signer.clone().into(), signer_id, 1, 0, Data::Precommit(precommit))
|
signed_from_data::<N>(signer.clone().into(), signer_id, 1, 0, Data::Precommit(precommit))
|
||||||
.await;
|
.await;
|
||||||
(
|
(signed.clone(), TendermintTx::SlashEvidence(Evidence::InvalidPrecommit(signed.encode())))
|
||||||
signed.clone(),
|
|
||||||
TendermintTx::SlashEvidence(Evidence::InvalidPrecommit(borsh::to_vec(&signed).unwrap())),
|
|
||||||
)
|
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
@@ -125,8 +120,7 @@ async fn invalid_precommit_signature() {
|
|||||||
let mut random_sig = [0u8; 64];
|
let mut random_sig = [0u8; 64];
|
||||||
OsRng.fill_bytes(&mut random_sig);
|
OsRng.fill_bytes(&mut random_sig);
|
||||||
signed.sig = random_sig;
|
signed.sig = random_sig;
|
||||||
let tx =
|
let tx = TendermintTx::SlashEvidence(Evidence::InvalidPrecommit(signed.encode()));
|
||||||
TendermintTx::SlashEvidence(Evidence::InvalidPrecommit(borsh::to_vec(&signed).unwrap()));
|
|
||||||
assert!(verify_tendermint_tx::<N>(&tx, &validators, commit).is_err());
|
assert!(verify_tendermint_tx::<N>(&tx, &validators, commit).is_err());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -144,32 +138,24 @@ async fn evidence_with_prevote() {
|
|||||||
// it should fail for all reasons.
|
// it should fail for all reasons.
|
||||||
let mut txs = vec![];
|
let mut txs = vec![];
|
||||||
txs.push(TendermintTx::SlashEvidence(Evidence::InvalidPrecommit(
|
txs.push(TendermintTx::SlashEvidence(Evidence::InvalidPrecommit(
|
||||||
borsh::to_vec(
|
signed_from_data::<N>(signer.clone().into(), signer_id, 0, 0, Data::Prevote(block_id))
|
||||||
&&signed_from_data::<N>(signer.clone().into(), signer_id, 0, 0, Data::Prevote(block_id))
|
.await
|
||||||
.await,
|
.encode(),
|
||||||
)
|
|
||||||
.unwrap(),
|
|
||||||
)));
|
)));
|
||||||
txs.push(TendermintTx::SlashEvidence(Evidence::InvalidValidRound(
|
txs.push(TendermintTx::SlashEvidence(Evidence::InvalidValidRound(
|
||||||
borsh::to_vec(
|
signed_from_data::<N>(signer.clone().into(), signer_id, 0, 0, Data::Prevote(block_id))
|
||||||
&signed_from_data::<N>(signer.clone().into(), signer_id, 0, 0, Data::Prevote(block_id))
|
.await
|
||||||
.await,
|
.encode(),
|
||||||
)
|
|
||||||
.unwrap(),
|
|
||||||
)));
|
)));
|
||||||
// Since these require a second message, provide this one again
|
// Since these require a second message, provide this one again
|
||||||
// ConflictingMessages can be fired for actually conflicting Prevotes however
|
// ConflictingMessages can be fired for actually conflicting Prevotes however
|
||||||
txs.push(TendermintTx::SlashEvidence(Evidence::ConflictingMessages(
|
txs.push(TendermintTx::SlashEvidence(Evidence::ConflictingMessages(
|
||||||
borsh::to_vec(
|
signed_from_data::<N>(signer.clone().into(), signer_id, 0, 0, Data::Prevote(block_id))
|
||||||
&signed_from_data::<N>(signer.clone().into(), signer_id, 0, 0, Data::Prevote(block_id))
|
.await
|
||||||
.await,
|
.encode(),
|
||||||
)
|
signed_from_data::<N>(signer.clone().into(), signer_id, 0, 0, Data::Prevote(block_id))
|
||||||
.unwrap(),
|
.await
|
||||||
borsh::to_vec(
|
.encode(),
|
||||||
&signed_from_data::<N>(signer.clone().into(), signer_id, 0, 0, Data::Prevote(block_id))
|
|
||||||
.await,
|
|
||||||
)
|
|
||||||
.unwrap(),
|
|
||||||
)));
|
)));
|
||||||
txs
|
txs
|
||||||
}
|
}
|
||||||
@@ -203,16 +189,16 @@ async fn conflicting_msgs_evidence_tx() {
|
|||||||
// non-conflicting data should fail
|
// non-conflicting data should fail
|
||||||
let signed_1 = signed_for_b_r(0, 0, Data::Proposal(None, TendermintBlock(vec![0x11]))).await;
|
let signed_1 = signed_for_b_r(0, 0, Data::Proposal(None, TendermintBlock(vec![0x11]))).await;
|
||||||
let tx = TendermintTx::SlashEvidence(Evidence::ConflictingMessages(
|
let tx = TendermintTx::SlashEvidence(Evidence::ConflictingMessages(
|
||||||
borsh::to_vec(&signed_1).unwrap(),
|
signed_1.encode(),
|
||||||
borsh::to_vec(&signed_1).unwrap(),
|
signed_1.encode(),
|
||||||
));
|
));
|
||||||
assert!(verify_tendermint_tx::<N>(&tx, &validators, commit).is_err());
|
assert!(verify_tendermint_tx::<N>(&tx, &validators, commit).is_err());
|
||||||
|
|
||||||
// conflicting data should pass
|
// conflicting data should pass
|
||||||
let signed_2 = signed_for_b_r(0, 0, Data::Proposal(None, TendermintBlock(vec![0x22]))).await;
|
let signed_2 = signed_for_b_r(0, 0, Data::Proposal(None, TendermintBlock(vec![0x22]))).await;
|
||||||
let tx = TendermintTx::SlashEvidence(Evidence::ConflictingMessages(
|
let tx = TendermintTx::SlashEvidence(Evidence::ConflictingMessages(
|
||||||
borsh::to_vec(&signed_1).unwrap(),
|
signed_1.encode(),
|
||||||
borsh::to_vec(&signed_2).unwrap(),
|
signed_2.encode(),
|
||||||
));
|
));
|
||||||
verify_tendermint_tx::<N>(&tx, &validators, commit).unwrap();
|
verify_tendermint_tx::<N>(&tx, &validators, commit).unwrap();
|
||||||
|
|
||||||
@@ -220,16 +206,16 @@ async fn conflicting_msgs_evidence_tx() {
|
|||||||
// (except for Precommit)
|
// (except for Precommit)
|
||||||
let signed_2 = signed_for_b_r(0, 1, Data::Proposal(None, TendermintBlock(vec![0x22]))).await;
|
let signed_2 = signed_for_b_r(0, 1, Data::Proposal(None, TendermintBlock(vec![0x22]))).await;
|
||||||
let tx = TendermintTx::SlashEvidence(Evidence::ConflictingMessages(
|
let tx = TendermintTx::SlashEvidence(Evidence::ConflictingMessages(
|
||||||
borsh::to_vec(&signed_1).unwrap(),
|
signed_1.encode(),
|
||||||
borsh::to_vec(&signed_2).unwrap(),
|
signed_2.encode(),
|
||||||
));
|
));
|
||||||
verify_tendermint_tx::<N>(&tx, &validators, commit).unwrap_err();
|
verify_tendermint_tx::<N>(&tx, &validators, commit).unwrap_err();
|
||||||
|
|
||||||
// Proposals for different block numbers should also fail as evidence
|
// Proposals for different block numbers should also fail as evidence
|
||||||
let signed_2 = signed_for_b_r(1, 0, Data::Proposal(None, TendermintBlock(vec![0x22]))).await;
|
let signed_2 = signed_for_b_r(1, 0, Data::Proposal(None, TendermintBlock(vec![0x22]))).await;
|
||||||
let tx = TendermintTx::SlashEvidence(Evidence::ConflictingMessages(
|
let tx = TendermintTx::SlashEvidence(Evidence::ConflictingMessages(
|
||||||
borsh::to_vec(&signed_1).unwrap(),
|
signed_1.encode(),
|
||||||
borsh::to_vec(&signed_2).unwrap(),
|
signed_2.encode(),
|
||||||
));
|
));
|
||||||
verify_tendermint_tx::<N>(&tx, &validators, commit).unwrap_err();
|
verify_tendermint_tx::<N>(&tx, &validators, commit).unwrap_err();
|
||||||
}
|
}
|
||||||
@@ -239,16 +225,16 @@ async fn conflicting_msgs_evidence_tx() {
|
|||||||
// non-conflicting data should fail
|
// non-conflicting data should fail
|
||||||
let signed_1 = signed_for_b_r(0, 0, Data::Prevote(Some([0x11; 32]))).await;
|
let signed_1 = signed_for_b_r(0, 0, Data::Prevote(Some([0x11; 32]))).await;
|
||||||
let tx = TendermintTx::SlashEvidence(Evidence::ConflictingMessages(
|
let tx = TendermintTx::SlashEvidence(Evidence::ConflictingMessages(
|
||||||
borsh::to_vec(&signed_1).unwrap(),
|
signed_1.encode(),
|
||||||
borsh::to_vec(&signed_1).unwrap(),
|
signed_1.encode(),
|
||||||
));
|
));
|
||||||
assert!(verify_tendermint_tx::<N>(&tx, &validators, commit).is_err());
|
assert!(verify_tendermint_tx::<N>(&tx, &validators, commit).is_err());
|
||||||
|
|
||||||
// conflicting data should pass
|
// conflicting data should pass
|
||||||
let signed_2 = signed_for_b_r(0, 0, Data::Prevote(Some([0x22; 32]))).await;
|
let signed_2 = signed_for_b_r(0, 0, Data::Prevote(Some([0x22; 32]))).await;
|
||||||
let tx = TendermintTx::SlashEvidence(Evidence::ConflictingMessages(
|
let tx = TendermintTx::SlashEvidence(Evidence::ConflictingMessages(
|
||||||
borsh::to_vec(&signed_1).unwrap(),
|
signed_1.encode(),
|
||||||
borsh::to_vec(&signed_2).unwrap(),
|
signed_2.encode(),
|
||||||
));
|
));
|
||||||
verify_tendermint_tx::<N>(&tx, &validators, commit).unwrap();
|
verify_tendermint_tx::<N>(&tx, &validators, commit).unwrap();
|
||||||
|
|
||||||
@@ -256,16 +242,16 @@ async fn conflicting_msgs_evidence_tx() {
|
|||||||
// (except for Precommit)
|
// (except for Precommit)
|
||||||
let signed_2 = signed_for_b_r(0, 1, Data::Prevote(Some([0x22; 32]))).await;
|
let signed_2 = signed_for_b_r(0, 1, Data::Prevote(Some([0x22; 32]))).await;
|
||||||
let tx = TendermintTx::SlashEvidence(Evidence::ConflictingMessages(
|
let tx = TendermintTx::SlashEvidence(Evidence::ConflictingMessages(
|
||||||
borsh::to_vec(&signed_1).unwrap(),
|
signed_1.encode(),
|
||||||
borsh::to_vec(&signed_2).unwrap(),
|
signed_2.encode(),
|
||||||
));
|
));
|
||||||
verify_tendermint_tx::<N>(&tx, &validators, commit).unwrap_err();
|
verify_tendermint_tx::<N>(&tx, &validators, commit).unwrap_err();
|
||||||
|
|
||||||
// Proposals for different block numbers should also fail as evidence
|
// Proposals for different block numbers should also fail as evidence
|
||||||
let signed_2 = signed_for_b_r(1, 0, Data::Prevote(Some([0x22; 32]))).await;
|
let signed_2 = signed_for_b_r(1, 0, Data::Prevote(Some([0x22; 32]))).await;
|
||||||
let tx = TendermintTx::SlashEvidence(Evidence::ConflictingMessages(
|
let tx = TendermintTx::SlashEvidence(Evidence::ConflictingMessages(
|
||||||
borsh::to_vec(&signed_1).unwrap(),
|
signed_1.encode(),
|
||||||
borsh::to_vec(&signed_2).unwrap(),
|
signed_2.encode(),
|
||||||
));
|
));
|
||||||
verify_tendermint_tx::<N>(&tx, &validators, commit).unwrap_err();
|
verify_tendermint_tx::<N>(&tx, &validators, commit).unwrap_err();
|
||||||
}
|
}
|
||||||
@@ -275,7 +261,7 @@ async fn conflicting_msgs_evidence_tx() {
|
|||||||
let signed_1 = signed_for_b_r(0, 0, Data::Proposal(None, TendermintBlock(vec![0x11]))).await;
|
let signed_1 = signed_for_b_r(0, 0, Data::Proposal(None, TendermintBlock(vec![0x11]))).await;
|
||||||
|
|
||||||
let signer_2 =
|
let signer_2 =
|
||||||
Signer::new(genesis, Zeroizing::new(<Ristretto as Ciphersuite>::F::random(&mut OsRng)));
|
Signer::new(genesis, Zeroizing::new(<Ristretto as WrappedGroup>::F::random(&mut OsRng)));
|
||||||
let signed_id_2 = signer_2.validator_id().await.unwrap();
|
let signed_id_2 = signer_2.validator_id().await.unwrap();
|
||||||
let signed_2 = signed_from_data::<N>(
|
let signed_2 = signed_from_data::<N>(
|
||||||
signer_2.into(),
|
signer_2.into(),
|
||||||
@@ -287,15 +273,14 @@ async fn conflicting_msgs_evidence_tx() {
|
|||||||
.await;
|
.await;
|
||||||
|
|
||||||
let tx = TendermintTx::SlashEvidence(Evidence::ConflictingMessages(
|
let tx = TendermintTx::SlashEvidence(Evidence::ConflictingMessages(
|
||||||
borsh::to_vec(&signed_1).unwrap(),
|
signed_1.encode(),
|
||||||
borsh::to_vec(&signed_2).unwrap(),
|
signed_2.encode(),
|
||||||
));
|
));
|
||||||
|
|
||||||
// update schema so that we don't fail due to invalid signature
|
// update schema so that we don't fail due to invalid signature
|
||||||
let signer_pub =
|
let signer_pub = <Ristretto as GroupIo>::read_G::<&[u8]>(&mut signer_id.as_slice()).unwrap();
|
||||||
<Ristretto as Ciphersuite>::read_G::<&[u8]>(&mut signer_id.as_slice()).unwrap();
|
|
||||||
let signer_pub_2 =
|
let signer_pub_2 =
|
||||||
<Ristretto as Ciphersuite>::read_G::<&[u8]>(&mut signed_id_2.as_slice()).unwrap();
|
<Ristretto as GroupIo>::read_G::<&[u8]>(&mut signed_id_2.as_slice()).unwrap();
|
||||||
let validators =
|
let validators =
|
||||||
Arc::new(Validators::new(genesis, vec![(signer_pub, 1), (signer_pub_2, 1)]).unwrap());
|
Arc::new(Validators::new(genesis, vec![(signer_pub, 1), (signer_pub_2, 1)]).unwrap());
|
||||||
|
|
||||||
@@ -307,8 +292,8 @@ async fn conflicting_msgs_evidence_tx() {
|
|||||||
let signed_1 = signed_for_b_r(0, 0, Data::Proposal(None, TendermintBlock(vec![]))).await;
|
let signed_1 = signed_for_b_r(0, 0, Data::Proposal(None, TendermintBlock(vec![]))).await;
|
||||||
let signed_2 = signed_for_b_r(0, 0, Data::Prevote(None)).await;
|
let signed_2 = signed_for_b_r(0, 0, Data::Prevote(None)).await;
|
||||||
let tx = TendermintTx::SlashEvidence(Evidence::ConflictingMessages(
|
let tx = TendermintTx::SlashEvidence(Evidence::ConflictingMessages(
|
||||||
borsh::to_vec(&signed_1).unwrap(),
|
signed_1.encode(),
|
||||||
borsh::to_vec(&signed_2).unwrap(),
|
signed_2.encode(),
|
||||||
));
|
));
|
||||||
assert!(verify_tendermint_tx::<N>(&tx, &validators, commit).is_err());
|
assert!(verify_tendermint_tx::<N>(&tx, &validators, commit).is_err());
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -8,7 +8,7 @@ use blake2::{Digest, Blake2b512};
|
|||||||
|
|
||||||
use ciphersuite::{
|
use ciphersuite::{
|
||||||
group::{Group, GroupEncoding},
|
group::{Group, GroupEncoding},
|
||||||
Ciphersuite,
|
*,
|
||||||
};
|
};
|
||||||
use dalek_ff_group::Ristretto;
|
use dalek_ff_group::Ristretto;
|
||||||
use schnorr::SchnorrSignature;
|
use schnorr::SchnorrSignature;
|
||||||
@@ -43,7 +43,7 @@ pub enum TransactionError {
|
|||||||
/// Data for a signed transaction.
|
/// Data for a signed transaction.
|
||||||
#[derive(Clone, PartialEq, Eq, Debug)]
|
#[derive(Clone, PartialEq, Eq, Debug)]
|
||||||
pub struct Signed {
|
pub struct Signed {
|
||||||
pub signer: <Ristretto as Ciphersuite>::G,
|
pub signer: <Ristretto as WrappedGroup>::G,
|
||||||
pub nonce: u32,
|
pub nonce: u32,
|
||||||
pub signature: SchnorrSignature<Ristretto>,
|
pub signature: SchnorrSignature<Ristretto>,
|
||||||
}
|
}
|
||||||
@@ -160,10 +160,10 @@ pub trait Transaction: 'static + Send + Sync + Clone + Eq + Debug + ReadWrite {
|
|||||||
/// Do not override this unless you know what you're doing.
|
/// Do not override this unless you know what you're doing.
|
||||||
///
|
///
|
||||||
/// Panics if called on non-signed transactions.
|
/// Panics if called on non-signed transactions.
|
||||||
fn sig_hash(&self, genesis: [u8; 32]) -> <Ristretto as Ciphersuite>::F {
|
fn sig_hash(&self, genesis: [u8; 32]) -> <Ristretto as WrappedGroup>::F {
|
||||||
match self.kind() {
|
match self.kind() {
|
||||||
TransactionKind::Signed(order, Signed { signature, .. }) => {
|
TransactionKind::Signed(order, Signed { signature, .. }) => {
|
||||||
<Ristretto as Ciphersuite>::F::from_bytes_mod_order_wide(
|
<Ristretto as WrappedGroup>::F::from_bytes_mod_order_wide(
|
||||||
&Blake2b512::digest(
|
&Blake2b512::digest(
|
||||||
[
|
[
|
||||||
b"Tributary Signed Transaction",
|
b"Tributary Signed Transaction",
|
||||||
@@ -182,8 +182,8 @@ pub trait Transaction: 'static + Send + Sync + Clone + Eq + Debug + ReadWrite {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub trait GAIN: FnMut(&<Ristretto as Ciphersuite>::G, &[u8]) -> Option<u32> {}
|
pub trait GAIN: FnMut(&<Ristretto as WrappedGroup>::G, &[u8]) -> Option<u32> {}
|
||||||
impl<F: FnMut(&<Ristretto as Ciphersuite>::G, &[u8]) -> Option<u32>> GAIN for F {}
|
impl<F: FnMut(&<Ristretto as WrappedGroup>::G, &[u8]) -> Option<u32>> GAIN for F {}
|
||||||
|
|
||||||
pub(crate) fn verify_transaction<F: GAIN, T: Transaction>(
|
pub(crate) fn verify_transaction<F: GAIN, T: Transaction>(
|
||||||
tx: &T,
|
tx: &T,
|
||||||
|
|||||||
@@ -21,7 +21,7 @@ thiserror = { version = "2", default-features = false, features = ["std"] }
|
|||||||
hex = { version = "0.4", default-features = false, features = ["std"] }
|
hex = { version = "0.4", default-features = false, features = ["std"] }
|
||||||
log = { version = "0.4", default-features = false, features = ["std"] }
|
log = { version = "0.4", default-features = false, features = ["std"] }
|
||||||
|
|
||||||
borsh = { version = "1", default-features = false, features = ["std", "derive", "de_strict_order"] }
|
parity-scale-codec = { version = "3", default-features = false, features = ["std", "derive"] }
|
||||||
|
|
||||||
futures-util = { version = "0.3", default-features = false, features = ["std", "async-await-macro", "sink", "channel"] }
|
futures-util = { version = "0.3", default-features = false, features = ["std", "async-await-macro", "sink", "channel"] }
|
||||||
futures-channel = { version = "0.3", default-features = false, features = ["std", "sink"] }
|
futures-channel = { version = "0.3", default-features = false, features = ["std", "sink"] }
|
||||||
|
|||||||
@@ -3,41 +3,33 @@ use std::{sync::Arc, collections::HashSet};
|
|||||||
|
|
||||||
use thiserror::Error;
|
use thiserror::Error;
|
||||||
|
|
||||||
use borsh::{BorshSerialize, BorshDeserialize};
|
use parity_scale_codec::{Encode, Decode};
|
||||||
|
|
||||||
use crate::{SignedMessageFor, SlashEvent, commit_msg};
|
use crate::{SignedMessageFor, SlashEvent, commit_msg};
|
||||||
|
|
||||||
/// An alias for a series of traits required for a type to be usable as a validator ID,
|
/// An alias for a series of traits required for a type to be usable as a validator ID,
|
||||||
/// automatically implemented for all types satisfying those traits.
|
/// automatically implemented for all types satisfying those traits.
|
||||||
pub trait ValidatorId:
|
pub trait ValidatorId:
|
||||||
Send + Sync + Clone + Copy + PartialEq + Eq + Hash + Debug + BorshSerialize + BorshDeserialize
|
Send + Sync + Clone + Copy + PartialEq + Eq + Hash + Debug + Encode + Decode
|
||||||
{
|
{
|
||||||
}
|
}
|
||||||
#[rustfmt::skip]
|
impl<V: Send + Sync + Clone + Copy + PartialEq + Eq + Hash + Debug + Encode + Decode> ValidatorId
|
||||||
impl<
|
for V
|
||||||
V: Send + Sync + Clone + Copy + PartialEq + Eq + Hash + Debug + BorshSerialize + BorshDeserialize,
|
|
||||||
> ValidatorId for V
|
|
||||||
{
|
{
|
||||||
}
|
}
|
||||||
|
|
||||||
/// An alias for a series of traits required for a type to be usable as a signature,
|
/// An alias for a series of traits required for a type to be usable as a signature,
|
||||||
/// automatically implemented for all types satisfying those traits.
|
/// automatically implemented for all types satisfying those traits.
|
||||||
pub trait Signature:
|
pub trait Signature: Send + Sync + Clone + PartialEq + Eq + Debug + Encode + Decode {}
|
||||||
Send + Sync + Clone + PartialEq + Eq + Debug + BorshSerialize + BorshDeserialize
|
impl<S: Send + Sync + Clone + PartialEq + Eq + Debug + Encode + Decode> Signature for S {}
|
||||||
{
|
|
||||||
}
|
|
||||||
impl<S: Send + Sync + Clone + PartialEq + Eq + Debug + BorshSerialize + BorshDeserialize> Signature
|
|
||||||
for S
|
|
||||||
{
|
|
||||||
}
|
|
||||||
|
|
||||||
// Type aliases which are distinct according to the type system
|
// Type aliases which are distinct according to the type system
|
||||||
|
|
||||||
/// A struct containing a Block Number, wrapped to have a distinct type.
|
/// A struct containing a Block Number, wrapped to have a distinct type.
|
||||||
#[derive(Clone, Copy, PartialEq, Eq, Hash, Debug, BorshSerialize, BorshDeserialize)]
|
#[derive(Clone, Copy, PartialEq, Eq, Hash, Debug, Encode, Decode)]
|
||||||
pub struct BlockNumber(pub u64);
|
pub struct BlockNumber(pub u64);
|
||||||
/// A struct containing a round number, wrapped to have a distinct type.
|
/// A struct containing a round number, wrapped to have a distinct type.
|
||||||
#[derive(Clone, Copy, PartialEq, Eq, Hash, Debug, BorshSerialize, BorshDeserialize)]
|
#[derive(Clone, Copy, PartialEq, Eq, Hash, Debug, Encode, Decode)]
|
||||||
pub struct RoundNumber(pub u32);
|
pub struct RoundNumber(pub u32);
|
||||||
|
|
||||||
/// A signer for a validator.
|
/// A signer for a validator.
|
||||||
@@ -135,7 +127,7 @@ impl<S: SignatureScheme> SignatureScheme for Arc<S> {
|
|||||||
/// A commit for a specific block.
|
/// A commit for a specific block.
|
||||||
///
|
///
|
||||||
/// The list of validators have weight exceeding the threshold for a valid commit.
|
/// The list of validators have weight exceeding the threshold for a valid commit.
|
||||||
#[derive(PartialEq, Debug, BorshSerialize, BorshDeserialize)]
|
#[derive(PartialEq, Debug, Encode, Decode)]
|
||||||
pub struct Commit<S: SignatureScheme> {
|
pub struct Commit<S: SignatureScheme> {
|
||||||
/// End time of the round which created this commit, used as the start time of the next block.
|
/// End time of the round which created this commit, used as the start time of the next block.
|
||||||
pub end_time: u64,
|
pub end_time: u64,
|
||||||
@@ -193,7 +185,7 @@ impl<W: Weights> Weights for Arc<W> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/// Simplified error enum representing a block's validity.
|
/// Simplified error enum representing a block's validity.
|
||||||
#[derive(Clone, Copy, PartialEq, Eq, Debug, Error, BorshSerialize, BorshDeserialize)]
|
#[derive(Clone, Copy, PartialEq, Eq, Debug, Error, Encode, Decode)]
|
||||||
pub enum BlockError {
|
pub enum BlockError {
|
||||||
/// Malformed block which is wholly invalid.
|
/// Malformed block which is wholly invalid.
|
||||||
#[error("invalid block")]
|
#[error("invalid block")]
|
||||||
@@ -205,20 +197,9 @@ pub enum BlockError {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/// Trait representing a Block.
|
/// Trait representing a Block.
|
||||||
pub trait Block:
|
pub trait Block: Send + Sync + Clone + PartialEq + Eq + Debug + Encode + Decode {
|
||||||
Send + Sync + Clone + PartialEq + Eq + Debug + BorshSerialize + BorshDeserialize
|
|
||||||
{
|
|
||||||
// Type used to identify blocks. Presumably a cryptographic hash of the block.
|
// Type used to identify blocks. Presumably a cryptographic hash of the block.
|
||||||
type Id: Send
|
type Id: Send + Sync + Copy + Clone + PartialEq + Eq + AsRef<[u8]> + Debug + Encode + Decode;
|
||||||
+ Sync
|
|
||||||
+ Copy
|
|
||||||
+ Clone
|
|
||||||
+ PartialEq
|
|
||||||
+ Eq
|
|
||||||
+ AsRef<[u8]>
|
|
||||||
+ Debug
|
|
||||||
+ BorshSerialize
|
|
||||||
+ BorshDeserialize;
|
|
||||||
|
|
||||||
/// Return the deterministic, unique ID for this block.
|
/// Return the deterministic, unique ID for this block.
|
||||||
fn id(&self) -> Self::Id;
|
fn id(&self) -> Self::Id;
|
||||||
|
|||||||
@@ -1,3 +1,5 @@
|
|||||||
|
#![expect(clippy::cast_possible_truncation)]
|
||||||
|
|
||||||
use core::fmt::Debug;
|
use core::fmt::Debug;
|
||||||
|
|
||||||
use std::{
|
use std::{
|
||||||
@@ -6,7 +8,7 @@ use std::{
|
|||||||
collections::{VecDeque, HashMap},
|
collections::{VecDeque, HashMap},
|
||||||
};
|
};
|
||||||
|
|
||||||
use borsh::{BorshSerialize, BorshDeserialize};
|
use parity_scale_codec::{Encode, Decode, IoReader};
|
||||||
|
|
||||||
use futures_channel::mpsc;
|
use futures_channel::mpsc;
|
||||||
use futures_util::{
|
use futures_util::{
|
||||||
@@ -41,14 +43,14 @@ pub fn commit_msg(end_time: u64, id: &[u8]) -> Vec<u8> {
|
|||||||
[&end_time.to_le_bytes(), id].concat()
|
[&end_time.to_le_bytes(), id].concat()
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, Copy, PartialEq, Eq, Hash, Debug, BorshSerialize, BorshDeserialize)]
|
#[derive(Clone, Copy, PartialEq, Eq, Hash, Debug, Encode, Decode)]
|
||||||
pub enum Step {
|
pub enum Step {
|
||||||
Propose,
|
Propose,
|
||||||
Prevote,
|
Prevote,
|
||||||
Precommit,
|
Precommit,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, Eq, Debug, BorshSerialize, BorshDeserialize)]
|
#[derive(Clone, Eq, Debug, Encode, Decode)]
|
||||||
pub enum Data<B: Block, S: Signature> {
|
pub enum Data<B: Block, S: Signature> {
|
||||||
Proposal(Option<RoundNumber>, B),
|
Proposal(Option<RoundNumber>, B),
|
||||||
Prevote(Option<B::Id>),
|
Prevote(Option<B::Id>),
|
||||||
@@ -90,7 +92,7 @@ impl<B: Block, S: Signature> Data<B, S> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, PartialEq, Eq, Debug, BorshSerialize, BorshDeserialize)]
|
#[derive(Clone, PartialEq, Eq, Debug, Encode, Decode)]
|
||||||
pub struct Message<V: ValidatorId, B: Block, S: Signature> {
|
pub struct Message<V: ValidatorId, B: Block, S: Signature> {
|
||||||
pub sender: V,
|
pub sender: V,
|
||||||
pub block: BlockNumber,
|
pub block: BlockNumber,
|
||||||
@@ -100,7 +102,7 @@ pub struct Message<V: ValidatorId, B: Block, S: Signature> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/// A signed Tendermint consensus message to be broadcast to the other validators.
|
/// A signed Tendermint consensus message to be broadcast to the other validators.
|
||||||
#[derive(Clone, PartialEq, Eq, Debug, BorshSerialize, BorshDeserialize)]
|
#[derive(Clone, PartialEq, Eq, Debug, Encode, Decode)]
|
||||||
pub struct SignedMessage<V: ValidatorId, B: Block, S: Signature> {
|
pub struct SignedMessage<V: ValidatorId, B: Block, S: Signature> {
|
||||||
pub msg: Message<V, B, S>,
|
pub msg: Message<V, B, S>,
|
||||||
pub sig: S,
|
pub sig: S,
|
||||||
@@ -117,18 +119,18 @@ impl<V: ValidatorId, B: Block, S: Signature> SignedMessage<V, B, S> {
|
|||||||
&self,
|
&self,
|
||||||
signer: &Scheme,
|
signer: &Scheme,
|
||||||
) -> bool {
|
) -> bool {
|
||||||
signer.verify(self.msg.sender, &borsh::to_vec(&self.msg).unwrap(), &self.sig)
|
signer.verify(self.msg.sender, &self.msg.encode(), &self.sig)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, Copy, PartialEq, Eq, Debug, BorshSerialize, BorshDeserialize)]
|
#[derive(Clone, Copy, PartialEq, Eq, Debug, Encode, Decode)]
|
||||||
pub enum SlashReason {
|
pub enum SlashReason {
|
||||||
FailToPropose,
|
FailToPropose,
|
||||||
InvalidBlock,
|
InvalidBlock,
|
||||||
InvalidProposer,
|
InvalidProposer,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, PartialEq, Eq, Debug, BorshSerialize, BorshDeserialize)]
|
#[derive(Clone, PartialEq, Eq, Debug, Encode, Decode)]
|
||||||
pub enum Evidence {
|
pub enum Evidence {
|
||||||
ConflictingMessages(Vec<u8>, Vec<u8>),
|
ConflictingMessages(Vec<u8>, Vec<u8>),
|
||||||
InvalidPrecommit(Vec<u8>),
|
InvalidPrecommit(Vec<u8>),
|
||||||
@@ -159,7 +161,7 @@ pub type SignedMessageFor<N> = SignedMessage<
|
|||||||
>;
|
>;
|
||||||
|
|
||||||
pub fn decode_signed_message<N: Network>(mut data: &[u8]) -> Option<SignedMessageFor<N>> {
|
pub fn decode_signed_message<N: Network>(mut data: &[u8]) -> Option<SignedMessageFor<N>> {
|
||||||
SignedMessageFor::<N>::deserialize_reader(&mut data).ok()
|
SignedMessageFor::<N>::decode(&mut data).ok()
|
||||||
}
|
}
|
||||||
|
|
||||||
fn decode_and_verify_signed_message<N: Network>(
|
fn decode_and_verify_signed_message<N: Network>(
|
||||||
@@ -339,7 +341,7 @@ impl<N: Network + 'static> TendermintMachine<N> {
|
|||||||
target: "tendermint",
|
target: "tendermint",
|
||||||
"proposer for block {}, round {round:?} was {} (me: {res})",
|
"proposer for block {}, round {round:?} was {} (me: {res})",
|
||||||
self.block.number.0,
|
self.block.number.0,
|
||||||
hex::encode(borsh::to_vec(&proposer).unwrap()),
|
hex::encode(proposer.encode()),
|
||||||
);
|
);
|
||||||
res
|
res
|
||||||
}
|
}
|
||||||
@@ -420,11 +422,7 @@ impl<N: Network + 'static> TendermintMachine<N> {
|
|||||||
// TODO: If the new slash event has evidence, emit to prevent a low-importance slash from
|
// TODO: If the new slash event has evidence, emit to prevent a low-importance slash from
|
||||||
// cancelling emission of high-importance slashes
|
// cancelling emission of high-importance slashes
|
||||||
if !self.block.slashes.contains(&validator) {
|
if !self.block.slashes.contains(&validator) {
|
||||||
log::info!(
|
log::info!(target: "tendermint", "Slashing validator {}", hex::encode(validator.encode()));
|
||||||
target: "tendermint",
|
|
||||||
"Slashing validator {}",
|
|
||||||
hex::encode(borsh::to_vec(&validator).unwrap()),
|
|
||||||
);
|
|
||||||
self.block.slashes.insert(validator);
|
self.block.slashes.insert(validator);
|
||||||
self.network.slash(validator, slash_event).await;
|
self.network.slash(validator, slash_event).await;
|
||||||
}
|
}
|
||||||
@@ -674,7 +672,7 @@ impl<N: Network + 'static> TendermintMachine<N> {
|
|||||||
self
|
self
|
||||||
.slash(
|
.slash(
|
||||||
msg.sender,
|
msg.sender,
|
||||||
SlashEvent::WithEvidence(Evidence::InvalidPrecommit(borsh::to_vec(&signed).unwrap())),
|
SlashEvent::WithEvidence(Evidence::InvalidPrecommit(signed.encode())),
|
||||||
)
|
)
|
||||||
.await;
|
.await;
|
||||||
Err(TendermintError::Malicious)?;
|
Err(TendermintError::Malicious)?;
|
||||||
@@ -745,10 +743,7 @@ impl<N: Network + 'static> TendermintMachine<N> {
|
|||||||
self.broadcast(Data::Prevote(None));
|
self.broadcast(Data::Prevote(None));
|
||||||
}
|
}
|
||||||
self
|
self
|
||||||
.slash(
|
.slash(msg.sender, SlashEvent::WithEvidence(Evidence::InvalidValidRound(msg.encode())))
|
||||||
msg.sender,
|
|
||||||
SlashEvent::WithEvidence(Evidence::InvalidValidRound(borsh::to_vec(&msg).unwrap())),
|
|
||||||
)
|
|
||||||
.await;
|
.await;
|
||||||
Err(TendermintError::Malicious)?;
|
Err(TendermintError::Malicious)?;
|
||||||
}
|
}
|
||||||
@@ -1039,7 +1034,7 @@ impl<N: Network + 'static> TendermintMachine<N> {
|
|||||||
|
|
||||||
while !messages.is_empty() {
|
while !messages.is_empty() {
|
||||||
self.network.broadcast(
|
self.network.broadcast(
|
||||||
SignedMessageFor::<N>::deserialize_reader(&mut messages)
|
SignedMessageFor::<N>::decode(&mut IoReader(&mut messages))
|
||||||
.expect("saved invalid message to DB")
|
.expect("saved invalid message to DB")
|
||||||
).await;
|
).await;
|
||||||
}
|
}
|
||||||
@@ -1064,7 +1059,7 @@ impl<N: Network + 'static> TendermintMachine<N> {
|
|||||||
} {
|
} {
|
||||||
if our_message {
|
if our_message {
|
||||||
assert!(sig.is_none());
|
assert!(sig.is_none());
|
||||||
sig = Some(self.signer.sign(&borsh::to_vec(&msg).unwrap()).await);
|
sig = Some(self.signer.sign(&msg.encode()).await);
|
||||||
}
|
}
|
||||||
let sig = sig.unwrap();
|
let sig = sig.unwrap();
|
||||||
|
|
||||||
@@ -1084,7 +1079,7 @@ impl<N: Network + 'static> TendermintMachine<N> {
|
|||||||
let message_tape_key = message_tape_key(self.genesis);
|
let message_tape_key = message_tape_key(self.genesis);
|
||||||
let mut txn = self.db.txn();
|
let mut txn = self.db.txn();
|
||||||
let mut message_tape = txn.get(&message_tape_key).unwrap_or(vec![]);
|
let mut message_tape = txn.get(&message_tape_key).unwrap_or(vec![]);
|
||||||
signed_msg.serialize(&mut message_tape).unwrap();
|
message_tape.extend(signed_msg.encode());
|
||||||
txn.put(&message_tape_key, message_tape);
|
txn.put(&message_tape_key, message_tape);
|
||||||
txn.commit();
|
txn.commit();
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,5 +1,7 @@
|
|||||||
use std::{sync::Arc, collections::HashMap};
|
use std::{sync::Arc, collections::HashMap};
|
||||||
|
|
||||||
|
use parity_scale_codec::Encode;
|
||||||
|
|
||||||
use crate::{ext::*, RoundNumber, Step, DataFor, SignedMessageFor, Evidence};
|
use crate::{ext::*, RoundNumber, Step, DataFor, SignedMessageFor, Evidence};
|
||||||
|
|
||||||
type RoundLog<N> = HashMap<<N as Network>::ValidatorId, HashMap<Step, SignedMessageFor<N>>>;
|
type RoundLog<N> = HashMap<<N as Network>::ValidatorId, HashMap<Step, SignedMessageFor<N>>>;
|
||||||
@@ -37,10 +39,7 @@ impl<N: Network> MessageLog<N> {
|
|||||||
target: "tendermint",
|
target: "tendermint",
|
||||||
"Validator sent multiple messages for the same block + round + step"
|
"Validator sent multiple messages for the same block + round + step"
|
||||||
);
|
);
|
||||||
Err(Evidence::ConflictingMessages(
|
Err(Evidence::ConflictingMessages(existing.encode(), signed.encode()))?;
|
||||||
borsh::to_vec(&existing).unwrap(),
|
|
||||||
borsh::to_vec(&signed).unwrap(),
|
|
||||||
))?;
|
|
||||||
}
|
}
|
||||||
return Ok(false);
|
return Ok(false);
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -4,7 +4,7 @@ use std::{
|
|||||||
time::{UNIX_EPOCH, SystemTime, Duration},
|
time::{UNIX_EPOCH, SystemTime, Duration},
|
||||||
};
|
};
|
||||||
|
|
||||||
use borsh::{BorshSerialize, BorshDeserialize};
|
use parity_scale_codec::{Encode, Decode};
|
||||||
|
|
||||||
use futures_util::sink::SinkExt;
|
use futures_util::sink::SinkExt;
|
||||||
use tokio::{sync::RwLock, time::sleep};
|
use tokio::{sync::RwLock, time::sleep};
|
||||||
@@ -89,7 +89,7 @@ impl Weights for TestWeights {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, PartialEq, Eq, Debug, BorshSerialize, BorshDeserialize)]
|
#[derive(Clone, PartialEq, Eq, Debug, Encode, Decode)]
|
||||||
struct TestBlock {
|
struct TestBlock {
|
||||||
id: TestBlockId,
|
id: TestBlockId,
|
||||||
valid: Result<(), BlockError>,
|
valid: Result<(), BlockError>,
|
||||||
|
|||||||
@@ -21,6 +21,7 @@ workspace = true
|
|||||||
zeroize = { version = "^1.5", default-features = false, features = ["std"] }
|
zeroize = { version = "^1.5", default-features = false, features = ["std"] }
|
||||||
rand_core = { version = "0.6", default-features = false, features = ["std"] }
|
rand_core = { version = "0.6", default-features = false, features = ["std"] }
|
||||||
|
|
||||||
|
scale = { package = "parity-scale-codec", version = "3", default-features = false, features = ["std", "derive"] }
|
||||||
borsh = { version = "1", default-features = false, features = ["std", "derive", "de_strict_order"] }
|
borsh = { version = "1", default-features = false, features = ["std", "derive", "de_strict_order"] }
|
||||||
|
|
||||||
blake2 = { version = "0.11.0-rc.0", default-features = false, features = ["alloc"] }
|
blake2 = { version = "0.11.0-rc.0", default-features = false, features = ["alloc"] }
|
||||||
@@ -29,7 +30,7 @@ dalek-ff-group = { path = "../../crypto/dalek-ff-group", default-features = fals
|
|||||||
dkg = { path = "../../crypto/dkg", default-features = false, features = ["std"] }
|
dkg = { path = "../../crypto/dkg", default-features = false, features = ["std"] }
|
||||||
schnorr = { package = "schnorr-signatures", path = "../../crypto/schnorr", default-features = false, features = ["std"] }
|
schnorr = { package = "schnorr-signatures", path = "../../crypto/schnorr", default-features = false, features = ["std"] }
|
||||||
|
|
||||||
serai-primitives = { path = "../../substrate/primitives", default-features = false, features = ["std"] }
|
serai-client = { path = "../../substrate/client", default-features = false, features = ["serai", "borsh"] }
|
||||||
|
|
||||||
serai-db = { path = "../../common/db" }
|
serai-db = { path = "../../common/db" }
|
||||||
serai-task = { path = "../../common/task", version = "0.1" }
|
serai-task = { path = "../../common/task", version = "0.1" }
|
||||||
|
|||||||
@@ -1,8 +1,11 @@
|
|||||||
|
#![expect(clippy::cast_possible_truncation)]
|
||||||
|
|
||||||
use std::collections::HashMap;
|
use std::collections::HashMap;
|
||||||
|
|
||||||
|
use scale::Encode;
|
||||||
use borsh::{BorshSerialize, BorshDeserialize};
|
use borsh::{BorshSerialize, BorshDeserialize};
|
||||||
|
|
||||||
use serai_primitives::{address::SeraiAddress, validator_sets::primitives::ExternalValidatorSet};
|
use serai_client::{primitives::SeraiAddress, validator_sets::primitives::ExternalValidatorSet};
|
||||||
|
|
||||||
use messages::sign::{VariantSignId, SignId};
|
use messages::sign::{VariantSignId, SignId};
|
||||||
|
|
||||||
@@ -13,7 +16,7 @@ use serai_cosign::CosignIntent;
|
|||||||
use crate::transaction::SigningProtocolRound;
|
use crate::transaction::SigningProtocolRound;
|
||||||
|
|
||||||
/// A topic within the database which the group participates in
|
/// A topic within the database which the group participates in
|
||||||
#[derive(Clone, Copy, PartialEq, Eq, Debug, BorshSerialize, BorshDeserialize)]
|
#[derive(Clone, Copy, PartialEq, Eq, Debug, Encode, BorshSerialize, BorshDeserialize)]
|
||||||
pub enum Topic {
|
pub enum Topic {
|
||||||
/// Vote to remove a participant
|
/// Vote to remove a participant
|
||||||
RemoveParticipant {
|
RemoveParticipant {
|
||||||
@@ -122,7 +125,7 @@ impl Topic {
|
|||||||
Topic::DkgConfirmation { attempt, round: _ } => Some({
|
Topic::DkgConfirmation { attempt, round: _ } => Some({
|
||||||
let id = {
|
let id = {
|
||||||
let mut id = [0; 32];
|
let mut id = [0; 32];
|
||||||
let encoded_set = borsh::to_vec(set).unwrap();
|
let encoded_set = set.encode();
|
||||||
id[.. encoded_set.len()].copy_from_slice(&encoded_set);
|
id[.. encoded_set.len()].copy_from_slice(&encoded_set);
|
||||||
VariantSignId::Batch(id)
|
VariantSignId::Batch(id)
|
||||||
};
|
};
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#![cfg_attr(docsrs, feature(doc_auto_cfg))]
|
#![cfg_attr(docsrs, feature(doc_cfg))]
|
||||||
#![doc = include_str!("../README.md")]
|
#![doc = include_str!("../README.md")]
|
||||||
#![deny(missing_docs)]
|
#![deny(missing_docs)]
|
||||||
|
|
||||||
@@ -8,9 +8,9 @@ use std::collections::HashMap;
|
|||||||
use ciphersuite::group::GroupEncoding;
|
use ciphersuite::group::GroupEncoding;
|
||||||
use dkg::Participant;
|
use dkg::Participant;
|
||||||
|
|
||||||
use serai_primitives::{
|
use serai_client::{
|
||||||
address::SeraiAddress,
|
primitives::SeraiAddress,
|
||||||
validator_sets::{ExternalValidatorSet, Slash},
|
validator_sets::primitives::{ExternalValidatorSet, Slash},
|
||||||
};
|
};
|
||||||
|
|
||||||
use serai_db::*;
|
use serai_db::*;
|
||||||
|
|||||||
@@ -6,15 +6,16 @@ use rand_core::{RngCore, CryptoRng};
|
|||||||
|
|
||||||
use blake2::{digest::typenum::U32, Digest, Blake2b};
|
use blake2::{digest::typenum::U32, Digest, Blake2b};
|
||||||
use ciphersuite::{
|
use ciphersuite::{
|
||||||
group::{ff::Field, Group, GroupEncoding},
|
group::{Group, GroupEncoding},
|
||||||
Ciphersuite,
|
*,
|
||||||
};
|
};
|
||||||
use dalek_ff_group::Ristretto;
|
use dalek_ff_group::Ristretto;
|
||||||
use schnorr::SchnorrSignature;
|
use schnorr::SchnorrSignature;
|
||||||
|
|
||||||
|
use scale::Encode;
|
||||||
use borsh::{BorshSerialize, BorshDeserialize};
|
use borsh::{BorshSerialize, BorshDeserialize};
|
||||||
|
|
||||||
use serai_primitives::{addess::SeraiAddress, validator_sets::MAX_KEY_SHARES_PER_SET};
|
use serai_client::{primitives::SeraiAddress, validator_sets::primitives::MAX_KEY_SHARES_PER_SET};
|
||||||
|
|
||||||
use messages::sign::VariantSignId;
|
use messages::sign::VariantSignId;
|
||||||
|
|
||||||
@@ -28,7 +29,7 @@ use tributary_sdk::{
|
|||||||
use crate::db::Topic;
|
use crate::db::Topic;
|
||||||
|
|
||||||
/// The round this data is for, within a signing protocol.
|
/// The round this data is for, within a signing protocol.
|
||||||
#[derive(Clone, Copy, PartialEq, Eq, Debug, BorshSerialize, BorshDeserialize)]
|
#[derive(Clone, Copy, PartialEq, Eq, Debug, Encode, BorshSerialize, BorshDeserialize)]
|
||||||
pub enum SigningProtocolRound {
|
pub enum SigningProtocolRound {
|
||||||
/// A preprocess.
|
/// A preprocess.
|
||||||
Preprocess,
|
Preprocess,
|
||||||
@@ -51,7 +52,7 @@ impl SigningProtocolRound {
|
|||||||
#[derive(Clone, Copy, PartialEq, Eq, Debug)]
|
#[derive(Clone, Copy, PartialEq, Eq, Debug)]
|
||||||
pub struct Signed {
|
pub struct Signed {
|
||||||
/// The signer.
|
/// The signer.
|
||||||
signer: <Ristretto as Ciphersuite>::G,
|
signer: <Ristretto as WrappedGroup>::G,
|
||||||
/// The signature.
|
/// The signature.
|
||||||
signature: SchnorrSignature<Ristretto>,
|
signature: SchnorrSignature<Ristretto>,
|
||||||
}
|
}
|
||||||
@@ -72,7 +73,7 @@ impl BorshDeserialize for Signed {
|
|||||||
|
|
||||||
impl Signed {
|
impl Signed {
|
||||||
/// Fetch the signer.
|
/// Fetch the signer.
|
||||||
pub(crate) fn signer(&self) -> <Ristretto as Ciphersuite>::G {
|
pub(crate) fn signer(&self) -> <Ristretto as WrappedGroup>::G {
|
||||||
self.signer
|
self.signer
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -85,10 +86,10 @@ impl Signed {
|
|||||||
impl Default for Signed {
|
impl Default for Signed {
|
||||||
fn default() -> Self {
|
fn default() -> Self {
|
||||||
Self {
|
Self {
|
||||||
signer: <Ristretto as Ciphersuite>::G::identity(),
|
signer: <Ristretto as WrappedGroup>::G::identity(),
|
||||||
signature: SchnorrSignature {
|
signature: SchnorrSignature {
|
||||||
R: <Ristretto as Ciphersuite>::G::identity(),
|
R: <Ristretto as WrappedGroup>::G::identity(),
|
||||||
s: <Ristretto as Ciphersuite>::F::ZERO,
|
s: <Ristretto as WrappedGroup>::F::ZERO,
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -241,20 +242,19 @@ impl TransactionTrait for Transaction {
|
|||||||
fn kind(&self) -> TransactionKind {
|
fn kind(&self) -> TransactionKind {
|
||||||
match self {
|
match self {
|
||||||
Transaction::RemoveParticipant { participant, signed } => TransactionKind::Signed(
|
Transaction::RemoveParticipant { participant, signed } => TransactionKind::Signed(
|
||||||
borsh::to_vec(&(b"RemoveParticipant".as_slice(), participant)).unwrap(),
|
(b"RemoveParticipant", participant).encode(),
|
||||||
signed.to_tributary_signed(0),
|
signed.to_tributary_signed(0),
|
||||||
),
|
),
|
||||||
|
|
||||||
Transaction::DkgParticipation { signed, .. } => TransactionKind::Signed(
|
Transaction::DkgParticipation { signed, .. } => {
|
||||||
borsh::to_vec(b"DkgParticipation".as_slice()).unwrap(),
|
TransactionKind::Signed(b"DkgParticipation".encode(), signed.to_tributary_signed(0))
|
||||||
signed.to_tributary_signed(0),
|
}
|
||||||
),
|
|
||||||
Transaction::DkgConfirmationPreprocess { attempt, signed, .. } => TransactionKind::Signed(
|
Transaction::DkgConfirmationPreprocess { attempt, signed, .. } => TransactionKind::Signed(
|
||||||
borsh::to_vec(b"DkgConfirmation".as_slice(), attempt).unwrap(),
|
(b"DkgConfirmation", attempt).encode(),
|
||||||
signed.to_tributary_signed(0),
|
signed.to_tributary_signed(0),
|
||||||
),
|
),
|
||||||
Transaction::DkgConfirmationShare { attempt, signed, .. } => TransactionKind::Signed(
|
Transaction::DkgConfirmationShare { attempt, signed, .. } => TransactionKind::Signed(
|
||||||
borsh::to_vec(b"DkgConfirmation".as_slice(), attempt).unwrap(),
|
(b"DkgConfirmation", attempt).encode(),
|
||||||
signed.to_tributary_signed(1),
|
signed.to_tributary_signed(1),
|
||||||
),
|
),
|
||||||
|
|
||||||
@@ -264,14 +264,13 @@ impl TransactionTrait for Transaction {
|
|||||||
Transaction::Batch { .. } => TransactionKind::Provided("Batch"),
|
Transaction::Batch { .. } => TransactionKind::Provided("Batch"),
|
||||||
|
|
||||||
Transaction::Sign { id, attempt, round, signed, .. } => TransactionKind::Signed(
|
Transaction::Sign { id, attempt, round, signed, .. } => TransactionKind::Signed(
|
||||||
borsh::to_vec(b"Sign".as_slice(), id, attempt).unwrap(),
|
(b"Sign", id, attempt).encode(),
|
||||||
signed.to_tributary_signed(round.nonce()),
|
signed.to_tributary_signed(round.nonce()),
|
||||||
),
|
),
|
||||||
|
|
||||||
Transaction::SlashReport { signed, .. } => TransactionKind::Signed(
|
Transaction::SlashReport { signed, .. } => {
|
||||||
borsh::to_vec(b"SlashReport".as_slice()).unwrap(),
|
TransactionKind::Signed(b"SlashReport".encode(), signed.to_tributary_signed(0))
|
||||||
signed.to_tributary_signed(0),
|
}
|
||||||
),
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -357,7 +356,7 @@ impl Transaction {
|
|||||||
&mut self,
|
&mut self,
|
||||||
rng: &mut R,
|
rng: &mut R,
|
||||||
genesis: [u8; 32],
|
genesis: [u8; 32],
|
||||||
key: &Zeroizing<<Ristretto as Ciphersuite>::F>,
|
key: &Zeroizing<<Ristretto as WrappedGroup>::F>,
|
||||||
) {
|
) {
|
||||||
fn signed(tx: &mut Transaction) -> &mut Signed {
|
fn signed(tx: &mut Transaction) -> &mut Signed {
|
||||||
#[allow(clippy::match_same_arms)] // This doesn't make semantic sense here
|
#[allow(clippy::match_same_arms)] // This doesn't make semantic sense here
|
||||||
@@ -381,13 +380,13 @@ impl Transaction {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Decide the nonce to sign with
|
// Decide the nonce to sign with
|
||||||
let sig_nonce = Zeroizing::new(<Ristretto as Ciphersuite>::F::random(rng));
|
let sig_nonce = Zeroizing::new(<Ristretto as WrappedGroup>::F::random(rng));
|
||||||
|
|
||||||
{
|
{
|
||||||
// Set the signer and the nonce
|
// Set the signer and the nonce
|
||||||
let signed = signed(self);
|
let signed = signed(self);
|
||||||
signed.signer = Ristretto::generator() * key.deref();
|
signed.signer = Ristretto::generator() * key.deref();
|
||||||
signed.signature.R = <Ristretto as Ciphersuite>::generator() * sig_nonce.deref();
|
signed.signature.R = <Ristretto as WrappedGroup>::generator() * sig_nonce.deref();
|
||||||
}
|
}
|
||||||
|
|
||||||
// Get the signature hash (which now includes `R || A` making it valid as the challenge)
|
// Get the signature hash (which now includes `R || A` making it valid as the challenge)
|
||||||
|
|||||||
@@ -17,15 +17,12 @@ rustdoc-args = ["--cfg", "docsrs"]
|
|||||||
workspace = true
|
workspace = true
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
std-shims = { path = "../../common/std-shims", version = "^0.1.1", default-features = false, optional = true }
|
std-shims = { path = "../../common/std-shims", version = "0.1.4", default-features = false }
|
||||||
|
|
||||||
rand_core = { version = "0.6", default-features = false }
|
|
||||||
|
|
||||||
zeroize = { version = "^1.5", default-features = false, features = ["derive"] }
|
zeroize = { version = "^1.5", default-features = false, features = ["derive"] }
|
||||||
subtle = { version = "^2.4", default-features = false }
|
subtle = { version = "^2.4", default-features = false }
|
||||||
|
|
||||||
digest = { version = "0.11.0-rc.0", default-features = false, features = ["block-api"] }
|
digest = { version = "0.11.0-rc.1", default-features = false }
|
||||||
transcript = { package = "flexible-transcript", path = "../transcript", version = "^0.3.2", default-features = false }
|
|
||||||
|
|
||||||
ff = { version = "0.13", default-features = false, features = ["bits"] }
|
ff = { version = "0.13", default-features = false, features = ["bits"] }
|
||||||
group = { version = "0.13", default-features = false }
|
group = { version = "0.13", default-features = false }
|
||||||
@@ -33,24 +30,18 @@ group = { version = "0.13", default-features = false }
|
|||||||
[dev-dependencies]
|
[dev-dependencies]
|
||||||
hex = { version = "0.4", default-features = false, features = ["std"] }
|
hex = { version = "0.4", default-features = false, features = ["std"] }
|
||||||
|
|
||||||
rand_core = { version = "0.6", default-features = false, features = ["std"] }
|
|
||||||
|
|
||||||
ff-group-tests = { version = "0.13", path = "../ff-group-tests" }
|
ff-group-tests = { version = "0.13", path = "../ff-group-tests" }
|
||||||
|
|
||||||
[features]
|
[features]
|
||||||
alloc = ["std-shims", "digest/alloc", "ff/alloc"]
|
alloc = ["zeroize/alloc", "digest/alloc", "ff/alloc"]
|
||||||
std = [
|
std = [
|
||||||
"alloc",
|
"alloc",
|
||||||
|
|
||||||
"std-shims/std",
|
"std-shims/std",
|
||||||
|
|
||||||
"rand_core/std",
|
|
||||||
|
|
||||||
"zeroize/std",
|
"zeroize/std",
|
||||||
"subtle/std",
|
"subtle/std",
|
||||||
|
|
||||||
"transcript/std",
|
|
||||||
|
|
||||||
"ff/std",
|
"ff/std",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
|||||||
@@ -21,7 +21,7 @@ rand_core = { version = "0.6", default-features = false }
|
|||||||
|
|
||||||
zeroize = { version = "^1.5", default-features = false, features = ["derive"] }
|
zeroize = { version = "^1.5", default-features = false, features = ["derive"] }
|
||||||
|
|
||||||
sha2 = { version = "0.11.0-rc.0", default-features = false }
|
sha2 = { version = "0.11.0-rc.2", default-features = false }
|
||||||
|
|
||||||
p256 = { version = "^0.13.1", default-features = false, features = ["arithmetic", "bits", "hash2curve"] }
|
p256 = { version = "^0.13.1", default-features = false, features = ["arithmetic", "bits", "hash2curve"] }
|
||||||
k256 = { version = "^0.13.1", default-features = false, features = ["arithmetic", "bits", "hash2curve"] }
|
k256 = { version = "^0.13.1", default-features = false, features = ["arithmetic", "bits", "hash2curve"] }
|
||||||
|
|||||||
@@ -1,11 +1,11 @@
|
|||||||
#![cfg_attr(docsrs, feature(doc_auto_cfg))]
|
#![cfg_attr(docsrs, feature(doc_cfg))]
|
||||||
#![cfg_attr(not(feature = "std"), no_std)]
|
#![cfg_attr(not(feature = "std"), no_std)]
|
||||||
|
|
||||||
use zeroize::Zeroize;
|
use zeroize::Zeroize;
|
||||||
|
|
||||||
use sha2::Sha512;
|
use sha2::Sha512;
|
||||||
|
|
||||||
use ciphersuite::Ciphersuite;
|
use ciphersuite::{WrappedGroup, Id, WithPreferredHash, GroupCanonicalEncoding};
|
||||||
|
|
||||||
pub use k256;
|
pub use k256;
|
||||||
pub use p256;
|
pub use p256;
|
||||||
@@ -18,17 +18,20 @@ macro_rules! kp_curve {
|
|||||||
$Ciphersuite: ident,
|
$Ciphersuite: ident,
|
||||||
$ID: literal
|
$ID: literal
|
||||||
) => {
|
) => {
|
||||||
impl Ciphersuite for $Ciphersuite {
|
impl WrappedGroup for $Ciphersuite {
|
||||||
type F = $lib::Scalar;
|
type F = $lib::Scalar;
|
||||||
type G = $lib::ProjectivePoint;
|
type G = $lib::ProjectivePoint;
|
||||||
type H = Sha512;
|
|
||||||
|
|
||||||
const ID: &'static [u8] = $ID;
|
|
||||||
|
|
||||||
fn generator() -> Self::G {
|
fn generator() -> Self::G {
|
||||||
$lib::ProjectivePoint::GENERATOR
|
$lib::ProjectivePoint::GENERATOR
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
impl Id for $Ciphersuite {
|
||||||
|
const ID: &'static [u8] = $ID;
|
||||||
|
}
|
||||||
|
impl WithPreferredHash for $Ciphersuite {
|
||||||
|
type H = Sha512;
|
||||||
|
}
|
||||||
|
impl GroupCanonicalEncoding for $Ciphersuite {}
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -1,30 +1,24 @@
|
|||||||
#![cfg_attr(docsrs, feature(doc_auto_cfg))]
|
#![cfg_attr(docsrs, feature(doc_cfg))]
|
||||||
#![doc = include_str!("lib.md")]
|
#![doc = include_str!("lib.md")]
|
||||||
#![cfg_attr(not(feature = "std"), no_std)]
|
#![cfg_attr(not(feature = "std"), no_std)]
|
||||||
|
|
||||||
use core::fmt::Debug;
|
use core::fmt::Debug;
|
||||||
#[cfg(feature = "alloc")]
|
|
||||||
#[allow(unused_imports)]
|
#[allow(unused_imports)]
|
||||||
use std_shims::prelude::*;
|
use std_shims::prelude::*;
|
||||||
#[cfg(feature = "alloc")]
|
|
||||||
use std_shims::io::{self, Read};
|
use std_shims::io::{self, Read};
|
||||||
|
|
||||||
use rand_core::{RngCore, CryptoRng};
|
use subtle::{CtOption, ConstantTimeEq, ConditionallySelectable};
|
||||||
|
|
||||||
use zeroize::Zeroize;
|
use zeroize::Zeroize;
|
||||||
use subtle::ConstantTimeEq;
|
|
||||||
|
|
||||||
pub use digest;
|
pub use digest;
|
||||||
use digest::{array::ArraySize, block_api::BlockSizeUser, OutputSizeUser, Digest, HashMarker};
|
use digest::{array::ArraySize, OutputSizeUser, Digest, HashMarker};
|
||||||
use transcript::SecureDigest;
|
|
||||||
|
|
||||||
pub use group;
|
pub use group;
|
||||||
use group::{
|
use group::{
|
||||||
ff::{Field, PrimeField, PrimeFieldBits},
|
ff::{PrimeField, PrimeFieldBits},
|
||||||
Group, GroupOps,
|
Group, GroupOps,
|
||||||
prime::PrimeGroup,
|
prime::PrimeGroup,
|
||||||
};
|
};
|
||||||
#[cfg(feature = "alloc")]
|
|
||||||
use group::GroupEncoding;
|
use group::GroupEncoding;
|
||||||
|
|
||||||
pub trait FromUniformBytes<T> {
|
pub trait FromUniformBytes<T> {
|
||||||
@@ -36,74 +30,115 @@ impl<const N: usize, F: group::ff::FromUniformBytes<N>> FromUniformBytes<[u8; N]
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Unified trait defining a ciphersuite around an elliptic curve.
|
/// A marker trait for fields which fleshes them out a bit more.
|
||||||
pub trait Ciphersuite:
|
pub trait F: PrimeField + PrimeFieldBits + Zeroize {}
|
||||||
|
impl<Fi: PrimeField + PrimeFieldBits + Zeroize> F for Fi {}
|
||||||
|
/// A marker trait for groups which fleshes them out a bit more.
|
||||||
|
pub trait G:
|
||||||
|
Group + GroupOps + GroupEncoding + PrimeGroup + ConstantTimeEq + ConditionallySelectable + Zeroize
|
||||||
|
{
|
||||||
|
}
|
||||||
|
impl<
|
||||||
|
Gr: Group
|
||||||
|
+ GroupOps
|
||||||
|
+ GroupEncoding
|
||||||
|
+ PrimeGroup
|
||||||
|
+ ConstantTimeEq
|
||||||
|
+ ConditionallySelectable
|
||||||
|
+ Zeroize,
|
||||||
|
> G for Gr
|
||||||
|
{
|
||||||
|
}
|
||||||
|
|
||||||
|
/// A `Group` type which has been wrapped into the current type.
|
||||||
|
///
|
||||||
|
/// This avoids having to re-implement all of the `Group` traits on the wrapper.
|
||||||
|
// TODO: Remove these bounds
|
||||||
|
pub trait WrappedGroup:
|
||||||
'static + Send + Sync + Clone + Copy + PartialEq + Eq + Debug + Zeroize
|
'static + Send + Sync + Clone + Copy + PartialEq + Eq + Debug + Zeroize
|
||||||
{
|
{
|
||||||
/// Scalar field element type.
|
/// Scalar field element type.
|
||||||
// This is available via G::Scalar yet `C::G::Scalar` is ambiguous, forcing horrific accesses
|
// This is available via `G::Scalar` yet `WG::G::Scalar` is ambiguous, forcing horrific accesses
|
||||||
type F: PrimeField
|
type F: F;
|
||||||
+ PrimeFieldBits
|
|
||||||
+ Zeroize
|
|
||||||
+ FromUniformBytes<<<Self::H as OutputSizeUser>::OutputSize as ArraySize>::ArrayType<u8>>;
|
|
||||||
/// Group element type.
|
/// Group element type.
|
||||||
type G: Group<Scalar = Self::F> + GroupOps + PrimeGroup + Zeroize + ConstantTimeEq;
|
type G: Group<Scalar = Self::F> + G;
|
||||||
/// Hash algorithm used with this curve.
|
|
||||||
// Requires BlockSizeUser so it can be used within Hkdf which requires that.
|
|
||||||
type H: Send + Clone + BlockSizeUser + Digest + HashMarker + SecureDigest;
|
|
||||||
|
|
||||||
/// ID for this curve.
|
|
||||||
const ID: &'static [u8];
|
|
||||||
|
|
||||||
/// Generator for the group.
|
/// Generator for the group.
|
||||||
// While group does provide this in its API, privacy coins may want to use a custom basepoint
|
|
||||||
fn generator() -> Self::G;
|
fn generator() -> Self::G;
|
||||||
|
}
|
||||||
|
impl<Gr: G<Scalar: F>> WrappedGroup for Gr {
|
||||||
|
type F = <Gr as Group>::Scalar;
|
||||||
|
type G = Gr;
|
||||||
|
fn generator() -> Self::G {
|
||||||
|
<Self::G as Group>::generator()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// An ID for an object.
|
||||||
|
pub trait Id {
|
||||||
|
// The ID.
|
||||||
|
const ID: &'static [u8];
|
||||||
|
}
|
||||||
|
|
||||||
|
/// A group with a preferred hash function.
|
||||||
|
pub trait WithPreferredHash:
|
||||||
|
WrappedGroup<
|
||||||
|
F: FromUniformBytes<<<Self::H as OutputSizeUser>::OutputSize as ArraySize>::ArrayType<u8>>,
|
||||||
|
>
|
||||||
|
{
|
||||||
|
type H: Send + Clone + Digest + HashMarker;
|
||||||
#[allow(non_snake_case)]
|
#[allow(non_snake_case)]
|
||||||
fn hash_to_F(data: &[u8]) -> Self::F {
|
fn hash_to_F(data: &[u8]) -> Self::F {
|
||||||
Self::F::from_uniform_bytes(&Self::H::digest(data).into())
|
Self::F::from_uniform_bytes(&Self::H::digest(data).into())
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/// Generate a random non-zero scalar.
|
/// A group which always encodes points canonically and supports decoding points while checking
|
||||||
#[allow(non_snake_case)]
|
/// they have a canonical encoding.
|
||||||
fn random_nonzero_F<R: RngCore + CryptoRng>(rng: &mut R) -> Self::F {
|
pub trait GroupCanonicalEncoding: WrappedGroup {
|
||||||
let mut res;
|
/// Decode a point from its canonical encoding.
|
||||||
while {
|
|
||||||
res = Self::F::random(&mut *rng);
|
|
||||||
res.ct_eq(&Self::F::ZERO).into()
|
|
||||||
} {}
|
|
||||||
res
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Read a canonical scalar from something implementing std::io::Read.
|
|
||||||
#[cfg(feature = "alloc")]
|
|
||||||
#[allow(non_snake_case)]
|
|
||||||
fn read_F<R: Read>(reader: &mut R) -> io::Result<Self::F> {
|
|
||||||
let mut encoding = <Self::F as PrimeField>::Repr::default();
|
|
||||||
reader.read_exact(encoding.as_mut())?;
|
|
||||||
|
|
||||||
// ff mandates this is canonical
|
|
||||||
let res = Option::<Self::F>::from(Self::F::from_repr(encoding))
|
|
||||||
.ok_or_else(|| io::Error::other("non-canonical scalar"));
|
|
||||||
encoding.as_mut().zeroize();
|
|
||||||
res
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Read a canonical point from something implementing std::io::Read.
|
|
||||||
///
|
///
|
||||||
/// The provided implementation is safe so long as `GroupEncoding::to_bytes` always returns a
|
/// Returns `None` if the point was invalid or not the encoding wasn't canonical.
|
||||||
/// canonical serialization.
|
///
|
||||||
#[cfg(feature = "alloc")]
|
/// If `<Self::G as GroupEncoding>::from_bytes` already only accepts canonical encodings, this
|
||||||
#[allow(non_snake_case)]
|
/// SHOULD be overriden with `<Self::G as GroupEncoding>::from_bytes(bytes)`.
|
||||||
fn read_G<R: Read>(reader: &mut R) -> io::Result<Self::G> {
|
fn from_canonical_bytes(bytes: &<Self::G as GroupEncoding>::Repr) -> CtOption<Self::G> {
|
||||||
let mut encoding = <Self::G as GroupEncoding>::Repr::default();
|
let res = Self::G::from_bytes(bytes).unwrap_or(Self::generator());
|
||||||
reader.read_exact(encoding.as_mut())?;
|
// Safe due to the bound points are always encoded canonically
|
||||||
|
let canonical = res.to_bytes().as_ref().ct_eq(bytes.as_ref());
|
||||||
let point = Option::<Self::G>::from(Self::G::from_bytes(&encoding))
|
CtOption::new(res, canonical)
|
||||||
.ok_or_else(|| io::Error::other("invalid point"))?;
|
|
||||||
if point.to_bytes().as_ref() != encoding.as_ref() {
|
|
||||||
Err(io::Error::other("non-canonical point"))?;
|
|
||||||
}
|
|
||||||
Ok(point)
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// `std::io` extensions for `GroupCanonicalEncoding.`
|
||||||
|
#[allow(non_snake_case)]
|
||||||
|
pub trait GroupIo: GroupCanonicalEncoding {
|
||||||
|
/// Read a canonical field element from something implementing `std::io::Read`.
|
||||||
|
fn read_F<R: Read>(reader: &mut R) -> io::Result<Self::F> {
|
||||||
|
let mut bytes = <Self::F as PrimeField>::Repr::default();
|
||||||
|
reader.read_exact(bytes.as_mut())?;
|
||||||
|
|
||||||
|
// `ff` mandates this is canonical
|
||||||
|
let res = Option::<Self::F>::from(Self::F::from_repr(bytes))
|
||||||
|
.ok_or_else(|| io::Error::other("non-canonical scalar"));
|
||||||
|
bytes.as_mut().zeroize();
|
||||||
|
|
||||||
|
res
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Read a canonical point from something implementing `std::io::Read`.
|
||||||
|
fn read_G<R: Read>(reader: &mut R) -> io::Result<Self::G> {
|
||||||
|
let mut bytes = <Self::G as GroupEncoding>::Repr::default();
|
||||||
|
reader.read_exact(bytes.as_mut())?;
|
||||||
|
|
||||||
|
let res = Option::<Self::G>::from(Self::from_canonical_bytes(&bytes))
|
||||||
|
.ok_or_else(|| io::Error::other("invalid point"))?;
|
||||||
|
bytes.as_mut().zeroize();
|
||||||
|
|
||||||
|
Ok(res)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
impl<Gr: GroupCanonicalEncoding> GroupIo for Gr {}
|
||||||
|
|
||||||
|
/// Unified trait defining a ciphersuite around an elliptic curve.
|
||||||
|
pub trait Ciphersuite: Id + WithPreferredHash + GroupCanonicalEncoding {}
|
||||||
|
impl<C: Id + WithPreferredHash + GroupCanonicalEncoding> Ciphersuite for C {}
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
[package]
|
[package]
|
||||||
name = "dalek-ff-group"
|
name = "dalek-ff-group"
|
||||||
version = "0.4.6"
|
version = "0.5.0"
|
||||||
description = "ff/group bindings around curve25519-dalek"
|
description = "ff/group bindings around curve25519-dalek"
|
||||||
license = "MIT"
|
license = "MIT"
|
||||||
repository = "https://github.com/serai-dex/serai/tree/develop/crypto/dalek-ff-group"
|
repository = "https://github.com/serai-dex/serai/tree/develop/crypto/dalek-ff-group"
|
||||||
@@ -22,15 +22,13 @@ subtle = { version = "^2.4", default-features = false }
|
|||||||
|
|
||||||
rand_core = { version = "0.6", default-features = false }
|
rand_core = { version = "0.6", default-features = false }
|
||||||
|
|
||||||
digest = { version = "0.10", default-features = false }
|
sha2 = { version = "0.11.0-rc.2", default-features = false, features = ["zeroize"] }
|
||||||
sha2 = { version = "0.11.0-rc.0", default-features = false }
|
blake2 = { version = "0.11.0-rc.2", default-features = false, features = ["zeroize"] }
|
||||||
|
|
||||||
prime-field = { path = "../prime-field", default-features = false }
|
prime-field = { path = "../prime-field", default-features = false }
|
||||||
ciphersuite = { version = "0.4.2", path = "../ciphersuite", default-features = false }
|
ciphersuite = { version = "0.4.2", path = "../ciphersuite", default-features = false }
|
||||||
|
|
||||||
crypto-bigint = { version = "0.5", default-features = false, features = ["zeroize"] }
|
curve25519-dalek = { version = ">= 4.0, < 4.2", default-features = false, features = ["zeroize", "digest", "group-bits", "precomputed-tables"] }
|
||||||
|
|
||||||
curve25519-dalek = { version = ">= 4.0, < 4.2", default-features = false, features = ["zeroize", "digest", "group", "precomputed-tables"] }
|
|
||||||
|
|
||||||
[dev-dependencies]
|
[dev-dependencies]
|
||||||
hex = "0.4"
|
hex = "0.4"
|
||||||
@@ -38,6 +36,6 @@ rand_core = { version = "0.6", default-features = false, features = ["std"] }
|
|||||||
ff-group-tests = { path = "../ff-group-tests" }
|
ff-group-tests = { path = "../ff-group-tests" }
|
||||||
|
|
||||||
[features]
|
[features]
|
||||||
alloc = ["zeroize/alloc", "digest/alloc", "prime-field/alloc", "ciphersuite/alloc", "curve25519-dalek/alloc"]
|
alloc = ["zeroize/alloc", "prime-field/alloc", "ciphersuite/alloc", "curve25519-dalek/alloc"]
|
||||||
std = ["alloc", "zeroize/std", "subtle/std", "rand_core/std", "digest/std", "prime-field/std", "ciphersuite/std"]
|
std = ["alloc", "zeroize/std", "subtle/std", "rand_core/std", "prime-field/std", "ciphersuite/std"]
|
||||||
default = ["std"]
|
default = ["std"]
|
||||||
|
|||||||
@@ -1,49 +1,48 @@
|
|||||||
use zeroize::Zeroize;
|
use zeroize::Zeroize;
|
||||||
|
|
||||||
use sha2::Sha512;
|
use sha2::Sha512;
|
||||||
|
use blake2::Blake2b512;
|
||||||
|
|
||||||
use ciphersuite::{group::Group, Ciphersuite};
|
use ::ciphersuite::{group::Group, *};
|
||||||
|
|
||||||
use crate::Scalar;
|
use crate::*;
|
||||||
|
|
||||||
macro_rules! dalek_curve {
|
|
||||||
(
|
|
||||||
$feature: literal,
|
|
||||||
|
|
||||||
$Ciphersuite: ident,
|
|
||||||
$Point: ident,
|
|
||||||
$ID: literal
|
|
||||||
) => {
|
|
||||||
use crate::$Point;
|
|
||||||
|
|
||||||
impl Ciphersuite for $Ciphersuite {
|
|
||||||
type F = Scalar;
|
|
||||||
type G = $Point;
|
|
||||||
type H = Sha512;
|
|
||||||
|
|
||||||
const ID: &'static [u8] = $ID;
|
|
||||||
|
|
||||||
fn generator() -> Self::G {
|
|
||||||
$Point::generator()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Ciphersuite for Ristretto.
|
/// Ciphersuite for Ristretto.
|
||||||
#[derive(Clone, Copy, PartialEq, Eq, Debug, Zeroize)]
|
#[derive(Clone, Copy, PartialEq, Eq, Debug, Zeroize)]
|
||||||
pub struct Ristretto;
|
pub struct Ristretto;
|
||||||
dalek_curve!("ristretto", Ristretto, RistrettoPoint, b"ristretto");
|
impl WrappedGroup for Ristretto {
|
||||||
#[test]
|
type F = Scalar;
|
||||||
fn test_ristretto() {
|
type G = RistrettoPoint;
|
||||||
ff_group_tests::group::test_prime_group_bits::<_, RistrettoPoint>(&mut rand_core::OsRng);
|
fn generator() -> Self::G {
|
||||||
|
<RistrettoPoint as Group>::generator()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
impl Id for Ristretto {
|
||||||
|
const ID: &[u8] = b"ristretto";
|
||||||
|
}
|
||||||
|
impl WithPreferredHash for Ristretto {
|
||||||
|
type H = Blake2b512;
|
||||||
|
}
|
||||||
|
impl GroupCanonicalEncoding for Ristretto {
|
||||||
|
fn from_canonical_bytes(bytes: &<Self::G as GroupEncoding>::Repr) -> CtOption<Self::G> {
|
||||||
|
Self::G::from_bytes(bytes)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Ciphersuite for Ed25519, inspired by RFC-8032.
|
/// Ciphersuite for Ed25519.
|
||||||
#[derive(Clone, Copy, PartialEq, Eq, Debug, Zeroize)]
|
#[derive(Clone, Copy, PartialEq, Eq, Debug, Zeroize)]
|
||||||
pub struct Ed25519;
|
pub struct Ed25519;
|
||||||
dalek_curve!("ed25519", Ed25519, EdwardsPoint, b"edwards25519");
|
impl WrappedGroup for Ed25519 {
|
||||||
#[test]
|
type F = Scalar;
|
||||||
fn test_ed25519() {
|
type G = EdwardsPoint;
|
||||||
ff_group_tests::group::test_prime_group_bits::<_, EdwardsPoint>(&mut rand_core::OsRng);
|
fn generator() -> Self::G {
|
||||||
|
<EdwardsPoint as Group>::generator()
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
impl Id for Ed25519 {
|
||||||
|
const ID: &[u8] = b"ed25519";
|
||||||
|
}
|
||||||
|
impl WithPreferredHash for Ed25519 {
|
||||||
|
type H = Sha512;
|
||||||
|
}
|
||||||
|
impl GroupCanonicalEncoding for Ed25519 {}
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
#![allow(deprecated)]
|
#![allow(deprecated)]
|
||||||
#![cfg_attr(docsrs, feature(doc_auto_cfg))]
|
#![cfg_attr(docsrs, feature(doc_cfg))]
|
||||||
#![no_std] // Prevents writing new code, in what should be a simple wrapper, which requires std
|
#![no_std] // Prevents writing new code, in what should be a simple wrapper, which requires std
|
||||||
#![doc = include_str!("../README.md")]
|
#![doc = include_str!("../README.md")]
|
||||||
#![allow(clippy::redundant_closure_call)]
|
#![allow(clippy::redundant_closure_call)]
|
||||||
@@ -7,33 +7,22 @@
|
|||||||
use core::{
|
use core::{
|
||||||
borrow::Borrow,
|
borrow::Borrow,
|
||||||
ops::{Deref, Add, AddAssign, Sub, SubAssign, Neg, Mul, MulAssign},
|
ops::{Deref, Add, AddAssign, Sub, SubAssign, Neg, Mul, MulAssign},
|
||||||
iter::{Iterator, Sum, Product},
|
iter::{Iterator, Sum},
|
||||||
hash::{Hash, Hasher},
|
|
||||||
};
|
};
|
||||||
|
|
||||||
use zeroize::Zeroize;
|
use zeroize::Zeroize;
|
||||||
use subtle::{ConstantTimeEq, ConditionallySelectable};
|
use subtle::{ConstantTimeEq, ConditionallySelectable};
|
||||||
|
|
||||||
use rand_core::RngCore;
|
use rand_core::RngCore;
|
||||||
use digest::{consts::U64, Digest, HashMarker};
|
|
||||||
|
|
||||||
use subtle::{Choice, CtOption};
|
use subtle::{Choice, CtOption};
|
||||||
|
|
||||||
pub use curve25519_dalek as dalek;
|
use curve25519_dalek::{
|
||||||
|
edwards::{EdwardsPoint as DEdwardsPoint, CompressedEdwardsY},
|
||||||
use dalek::{
|
|
||||||
constants::{self, BASEPOINT_ORDER},
|
|
||||||
scalar::Scalar as DScalar,
|
|
||||||
edwards::{EdwardsPoint as DEdwardsPoint, EdwardsBasepointTable, CompressedEdwardsY},
|
|
||||||
ristretto::{RistrettoPoint as DRistrettoPoint, RistrettoBasepointTable, CompressedRistretto},
|
|
||||||
};
|
};
|
||||||
pub use constants::{ED25519_BASEPOINT_TABLE, RISTRETTO_BASEPOINT_TABLE};
|
pub use curve25519_dalek::{Scalar, ristretto::RistrettoPoint};
|
||||||
|
|
||||||
use ::ciphersuite::group::{
|
use ::ciphersuite::group::{Group, GroupEncoding, prime::PrimeGroup};
|
||||||
ff::{Field, PrimeField, FieldBits, PrimeFieldBits, FromUniformBytes},
|
|
||||||
Group, GroupEncoding,
|
|
||||||
prime::PrimeGroup,
|
|
||||||
};
|
|
||||||
|
|
||||||
mod ciphersuite;
|
mod ciphersuite;
|
||||||
pub use crate::ciphersuite::{Ed25519, Ristretto};
|
pub use crate::ciphersuite::{Ed25519, Ristretto};
|
||||||
@@ -97,7 +86,41 @@ macro_rules! constant_time {
|
|||||||
}
|
}
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
pub(crate) use constant_time;
|
|
||||||
|
macro_rules! math_op_without_wrapping {
|
||||||
|
(
|
||||||
|
$Value: ident,
|
||||||
|
$Other: ident,
|
||||||
|
$Op: ident,
|
||||||
|
$op_fn: ident,
|
||||||
|
$Assign: ident,
|
||||||
|
$assign_fn: ident,
|
||||||
|
$function: expr
|
||||||
|
) => {
|
||||||
|
impl $Op<$Other> for $Value {
|
||||||
|
type Output = $Value;
|
||||||
|
fn $op_fn(self, other: $Other) -> Self::Output {
|
||||||
|
Self($function(self.0, other))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
impl $Assign<$Other> for $Value {
|
||||||
|
fn $assign_fn(&mut self, other: $Other) {
|
||||||
|
self.0 = $function(self.0, other);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
impl<'a> $Op<&'a $Other> for $Value {
|
||||||
|
type Output = $Value;
|
||||||
|
fn $op_fn(self, other: &'a $Other) -> Self::Output {
|
||||||
|
Self($function(self.0, other))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
impl<'a> $Assign<&'a $Other> for $Value {
|
||||||
|
fn $assign_fn(&mut self, other: &'a $Other) {
|
||||||
|
self.0 = $function(self.0, other);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
macro_rules! math_op {
|
macro_rules! math_op {
|
||||||
(
|
(
|
||||||
@@ -133,20 +156,12 @@ macro_rules! math_op {
|
|||||||
}
|
}
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
pub(crate) use math_op;
|
|
||||||
|
|
||||||
macro_rules! math {
|
|
||||||
($Value: ident, $Factor: ident, $add: expr, $sub: expr, $mul: expr) => {
|
|
||||||
math_op!($Value, $Value, Add, add, AddAssign, add_assign, $add);
|
|
||||||
math_op!($Value, $Value, Sub, sub, SubAssign, sub_assign, $sub);
|
|
||||||
math_op!($Value, $Factor, Mul, mul, MulAssign, mul_assign, $mul);
|
|
||||||
};
|
|
||||||
}
|
|
||||||
pub(crate) use math;
|
|
||||||
|
|
||||||
macro_rules! math_neg {
|
macro_rules! math_neg {
|
||||||
($Value: ident, $Factor: ident, $add: expr, $sub: expr, $mul: expr) => {
|
($Value: ident, $Factor: ident, $add: expr, $sub: expr, $mul: expr) => {
|
||||||
math!($Value, $Factor, $add, $sub, $mul);
|
math_op!($Value, $Value, Add, add, AddAssign, add_assign, $add);
|
||||||
|
math_op!($Value, $Value, Sub, sub, SubAssign, sub_assign, $sub);
|
||||||
|
math_op_without_wrapping!($Value, $Factor, Mul, mul, MulAssign, mul_assign, $mul);
|
||||||
|
|
||||||
impl Neg for $Value {
|
impl Neg for $Value {
|
||||||
type Output = Self;
|
type Output = Self;
|
||||||
@@ -157,187 +172,6 @@ macro_rules! math_neg {
|
|||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Wrapper around the dalek Scalar type.
|
|
||||||
#[derive(Clone, Copy, PartialEq, Eq, Default, Debug, Zeroize)]
|
|
||||||
pub struct Scalar(pub DScalar);
|
|
||||||
deref_borrow!(Scalar, DScalar);
|
|
||||||
constant_time!(Scalar, DScalar);
|
|
||||||
math_neg!(Scalar, Scalar, DScalar::add, DScalar::sub, DScalar::mul);
|
|
||||||
|
|
||||||
macro_rules! from_wrapper {
|
|
||||||
($uint: ident) => {
|
|
||||||
impl From<$uint> for Scalar {
|
|
||||||
fn from(a: $uint) -> Scalar {
|
|
||||||
Scalar(DScalar::from(a))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
from_wrapper!(u8);
|
|
||||||
from_wrapper!(u16);
|
|
||||||
from_wrapper!(u32);
|
|
||||||
from_wrapper!(u64);
|
|
||||||
from_wrapper!(u128);
|
|
||||||
|
|
||||||
impl Scalar {
|
|
||||||
pub fn pow(&self, other: Scalar) -> Scalar {
|
|
||||||
let mut table = [Scalar::ONE; 16];
|
|
||||||
table[1] = *self;
|
|
||||||
for i in 2 .. 16 {
|
|
||||||
table[i] = table[i - 1] * self;
|
|
||||||
}
|
|
||||||
|
|
||||||
let mut res = Scalar::ONE;
|
|
||||||
let mut bits = 0;
|
|
||||||
for (i, mut bit) in other.to_le_bits().iter_mut().rev().enumerate() {
|
|
||||||
bits <<= 1;
|
|
||||||
let mut bit = u8_from_bool(&mut bit);
|
|
||||||
bits |= bit;
|
|
||||||
bit.zeroize();
|
|
||||||
|
|
||||||
if ((i + 1) % 4) == 0 {
|
|
||||||
if i != 3 {
|
|
||||||
for _ in 0 .. 4 {
|
|
||||||
res *= res;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
let mut scale_by = Scalar::ONE;
|
|
||||||
#[allow(clippy::needless_range_loop)]
|
|
||||||
for i in 0 .. 16 {
|
|
||||||
#[allow(clippy::cast_possible_truncation)] // Safe since 0 .. 16
|
|
||||||
{
|
|
||||||
scale_by = <_>::conditional_select(&scale_by, &table[i], bits.ct_eq(&(i as u8)));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
res *= scale_by;
|
|
||||||
bits = 0;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
res
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Perform wide reduction on a 64-byte array to create a Scalar without bias.
|
|
||||||
pub fn from_bytes_mod_order_wide(bytes: &[u8; 64]) -> Scalar {
|
|
||||||
Self(DScalar::from_bytes_mod_order_wide(bytes))
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Derive a Scalar without bias from a digest via wide reduction.
|
|
||||||
pub fn from_hash<D: Digest<OutputSize = U64> + HashMarker>(hash: D) -> Scalar {
|
|
||||||
let mut output = [0u8; 64];
|
|
||||||
output.copy_from_slice(&hash.finalize());
|
|
||||||
let res = Scalar(DScalar::from_bytes_mod_order_wide(&output));
|
|
||||||
output.zeroize();
|
|
||||||
res
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl Field for Scalar {
|
|
||||||
const ZERO: Scalar = Scalar(DScalar::ZERO);
|
|
||||||
const ONE: Scalar = Scalar(DScalar::ONE);
|
|
||||||
|
|
||||||
fn random(rng: impl RngCore) -> Self {
|
|
||||||
Self(<DScalar as Field>::random(rng))
|
|
||||||
}
|
|
||||||
|
|
||||||
fn square(&self) -> Self {
|
|
||||||
Self(self.0.square())
|
|
||||||
}
|
|
||||||
fn double(&self) -> Self {
|
|
||||||
Self(self.0.double())
|
|
||||||
}
|
|
||||||
fn invert(&self) -> CtOption<Self> {
|
|
||||||
<DScalar as Field>::invert(&self.0).map(Self)
|
|
||||||
}
|
|
||||||
|
|
||||||
fn sqrt(&self) -> CtOption<Self> {
|
|
||||||
self.0.sqrt().map(Self)
|
|
||||||
}
|
|
||||||
|
|
||||||
fn sqrt_ratio(num: &Self, div: &Self) -> (Choice, Self) {
|
|
||||||
let (choice, res) = DScalar::sqrt_ratio(num, div);
|
|
||||||
(choice, Self(res))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl PrimeField for Scalar {
|
|
||||||
type Repr = [u8; 32];
|
|
||||||
|
|
||||||
const MODULUS: &'static str = <DScalar as PrimeField>::MODULUS;
|
|
||||||
|
|
||||||
const NUM_BITS: u32 = <DScalar as PrimeField>::NUM_BITS;
|
|
||||||
const CAPACITY: u32 = <DScalar as PrimeField>::CAPACITY;
|
|
||||||
|
|
||||||
const TWO_INV: Scalar = Scalar(<DScalar as PrimeField>::TWO_INV);
|
|
||||||
|
|
||||||
const MULTIPLICATIVE_GENERATOR: Scalar =
|
|
||||||
Scalar(<DScalar as PrimeField>::MULTIPLICATIVE_GENERATOR);
|
|
||||||
const S: u32 = <DScalar as PrimeField>::S;
|
|
||||||
|
|
||||||
const ROOT_OF_UNITY: Scalar = Scalar(<DScalar as PrimeField>::ROOT_OF_UNITY);
|
|
||||||
const ROOT_OF_UNITY_INV: Scalar = Scalar(<DScalar as PrimeField>::ROOT_OF_UNITY_INV);
|
|
||||||
|
|
||||||
const DELTA: Scalar = Scalar(<DScalar as PrimeField>::DELTA);
|
|
||||||
|
|
||||||
fn from_repr(bytes: [u8; 32]) -> CtOption<Self> {
|
|
||||||
<DScalar as PrimeField>::from_repr(bytes).map(Scalar)
|
|
||||||
}
|
|
||||||
fn to_repr(&self) -> [u8; 32] {
|
|
||||||
self.0.to_repr()
|
|
||||||
}
|
|
||||||
|
|
||||||
fn is_odd(&self) -> Choice {
|
|
||||||
self.0.is_odd()
|
|
||||||
}
|
|
||||||
|
|
||||||
fn from_u128(num: u128) -> Self {
|
|
||||||
Scalar(DScalar::from_u128(num))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl PrimeFieldBits for Scalar {
|
|
||||||
type ReprBits = [u8; 32];
|
|
||||||
|
|
||||||
fn to_le_bits(&self) -> FieldBits<Self::ReprBits> {
|
|
||||||
self.to_repr().into()
|
|
||||||
}
|
|
||||||
|
|
||||||
fn char_le_bits() -> FieldBits<Self::ReprBits> {
|
|
||||||
BASEPOINT_ORDER.to_bytes().into()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl FromUniformBytes<64> for Scalar {
|
|
||||||
fn from_uniform_bytes(bytes: &[u8; 64]) -> Self {
|
|
||||||
Self::from_bytes_mod_order_wide(bytes)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl Sum<Scalar> for Scalar {
|
|
||||||
fn sum<I: Iterator<Item = Scalar>>(iter: I) -> Scalar {
|
|
||||||
Self(DScalar::sum(iter))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl<'a> Sum<&'a Scalar> for Scalar {
|
|
||||||
fn sum<I: Iterator<Item = &'a Scalar>>(iter: I) -> Scalar {
|
|
||||||
Self(DScalar::sum(iter))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl Product<Scalar> for Scalar {
|
|
||||||
fn product<I: Iterator<Item = Scalar>>(iter: I) -> Scalar {
|
|
||||||
Self(DScalar::product(iter))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
impl<'a> Product<&'a Scalar> for Scalar {
|
|
||||||
fn product<I: Iterator<Item = &'a Scalar>>(iter: I) -> Scalar {
|
|
||||||
Self(DScalar::product(iter))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
macro_rules! dalek_group {
|
macro_rules! dalek_group {
|
||||||
(
|
(
|
||||||
$Point: ident,
|
$Point: ident,
|
||||||
@@ -347,9 +181,6 @@ macro_rules! dalek_group {
|
|||||||
$Table: ident,
|
$Table: ident,
|
||||||
|
|
||||||
$DCompressed: ident,
|
$DCompressed: ident,
|
||||||
|
|
||||||
$BASEPOINT_POINT: ident,
|
|
||||||
$BASEPOINT_TABLE: ident
|
|
||||||
) => {
|
) => {
|
||||||
/// Wrapper around the dalek Point type.
|
/// Wrapper around the dalek Point type.
|
||||||
///
|
///
|
||||||
@@ -363,9 +194,6 @@ macro_rules! dalek_group {
|
|||||||
constant_time!($Point, $DPoint);
|
constant_time!($Point, $DPoint);
|
||||||
math_neg!($Point, Scalar, $DPoint::add, $DPoint::sub, $DPoint::mul);
|
math_neg!($Point, Scalar, $DPoint::add, $DPoint::sub, $DPoint::mul);
|
||||||
|
|
||||||
/// The basepoint for this curve.
|
|
||||||
pub const $BASEPOINT_POINT: $Point = $Point(constants::$BASEPOINT_POINT);
|
|
||||||
|
|
||||||
impl Sum<$Point> for $Point {
|
impl Sum<$Point> for $Point {
|
||||||
fn sum<I: Iterator<Item = $Point>>(iter: I) -> $Point {
|
fn sum<I: Iterator<Item = $Point>>(iter: I) -> $Point {
|
||||||
Self($DPoint::sum(iter))
|
Self($DPoint::sum(iter))
|
||||||
@@ -396,7 +224,7 @@ macro_rules! dalek_group {
|
|||||||
Self($DPoint::identity())
|
Self($DPoint::identity())
|
||||||
}
|
}
|
||||||
fn generator() -> Self {
|
fn generator() -> Self {
|
||||||
$BASEPOINT_POINT
|
Self(<$DPoint as Group>::generator())
|
||||||
}
|
}
|
||||||
fn is_identity(&self) -> Choice {
|
fn is_identity(&self) -> Choice {
|
||||||
self.0.ct_eq(&$DPoint::identity())
|
self.0.ct_eq(&$DPoint::identity())
|
||||||
@@ -429,24 +257,6 @@ macro_rules! dalek_group {
|
|||||||
}
|
}
|
||||||
|
|
||||||
impl PrimeGroup for $Point {}
|
impl PrimeGroup for $Point {}
|
||||||
|
|
||||||
impl Mul<Scalar> for &$Table {
|
|
||||||
type Output = $Point;
|
|
||||||
fn mul(self, b: Scalar) -> $Point {
|
|
||||||
$Point(&b.0 * self)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Support being used as a key in a table
|
|
||||||
// While it is expensive as a key, due to the field operations required, there's frequently
|
|
||||||
// use cases for public key -> value lookups
|
|
||||||
#[allow(unknown_lints, renamed_and_removed_lints)]
|
|
||||||
#[allow(clippy::derived_hash_with_manual_eq, clippy::derive_hash_xor_eq)]
|
|
||||||
impl Hash for $Point {
|
|
||||||
fn hash<H: Hasher>(&self, state: &mut H) {
|
|
||||||
self.to_bytes().hash(state);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -456,24 +266,6 @@ dalek_group!(
|
|||||||
|point: DEdwardsPoint| point.is_torsion_free(),
|
|point: DEdwardsPoint| point.is_torsion_free(),
|
||||||
EdwardsBasepointTable,
|
EdwardsBasepointTable,
|
||||||
CompressedEdwardsY,
|
CompressedEdwardsY,
|
||||||
ED25519_BASEPOINT_POINT,
|
|
||||||
ED25519_BASEPOINT_TABLE
|
|
||||||
);
|
|
||||||
|
|
||||||
impl EdwardsPoint {
|
|
||||||
pub fn mul_by_cofactor(&self) -> EdwardsPoint {
|
|
||||||
EdwardsPoint(self.0.mul_by_cofactor())
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
dalek_group!(
|
|
||||||
RistrettoPoint,
|
|
||||||
DRistrettoPoint,
|
|
||||||
|_| true,
|
|
||||||
RistrettoBasepointTable,
|
|
||||||
CompressedRistretto,
|
|
||||||
RISTRETTO_BASEPOINT_POINT,
|
|
||||||
RISTRETTO_BASEPOINT_TABLE
|
|
||||||
);
|
);
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
@@ -494,21 +286,3 @@ prime_field::odd_prime_field_with_specific_repr!(
|
|||||||
false,
|
false,
|
||||||
crate::ThirtyTwoArray
|
crate::ThirtyTwoArray
|
||||||
);
|
);
|
||||||
|
|
||||||
impl FieldElement {
|
|
||||||
/// Create a FieldElement from a `crypto_bigint::U256`.
|
|
||||||
///
|
|
||||||
/// This will reduce the `U256` by the modulus, into a member of the field.
|
|
||||||
#[deprecated]
|
|
||||||
pub const fn from_u256(u256: &crypto_bigint::U256) -> Self {
|
|
||||||
FieldElement::from(&prime_field::crypto_bigint::U256::from_words(*u256.as_words()))
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Create a `FieldElement` from the reduction of a 512-bit number.
|
|
||||||
///
|
|
||||||
/// The bytes are interpreted in little-endian format.
|
|
||||||
#[deprecated]
|
|
||||||
pub fn wide_reduce(value: [u8; 64]) -> Self {
|
|
||||||
<FieldElement as ::ciphersuite::group::ff::FromUniformBytes<_>>::from_uniform_bytes(&value)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -21,21 +21,14 @@ zeroize = { version = "^1.5", default-features = false, features = ["zeroize_der
|
|||||||
|
|
||||||
thiserror = { version = "2", default-features = false }
|
thiserror = { version = "2", default-features = false }
|
||||||
|
|
||||||
std-shims = { version = "0.1", path = "../../common/std-shims", default-features = false }
|
std-shims = { version = "0.1", path = "../../common/std-shims", default-features = false, features = ["alloc"] }
|
||||||
|
|
||||||
borsh = { version = "1", default-features = false, features = ["derive", "de_strict_order"], optional = true }
|
|
||||||
|
|
||||||
ciphersuite = { path = "../ciphersuite", version = "^0.4.1", default-features = false, features = ["alloc"] }
|
ciphersuite = { path = "../ciphersuite", version = "^0.4.1", default-features = false, features = ["alloc"] }
|
||||||
|
|
||||||
[features]
|
[features]
|
||||||
std = [
|
std = [
|
||||||
"thiserror/std",
|
"thiserror/std",
|
||||||
|
|
||||||
"std-shims/std",
|
"std-shims/std",
|
||||||
|
|
||||||
"borsh?/std",
|
|
||||||
|
|
||||||
"ciphersuite/std",
|
"ciphersuite/std",
|
||||||
]
|
]
|
||||||
borsh = ["dep:borsh"]
|
|
||||||
default = ["std"]
|
default = ["std"]
|
||||||
|
|||||||
@@ -20,7 +20,7 @@ workspace = true
|
|||||||
zeroize = { version = "^1.5", default-features = false }
|
zeroize = { version = "^1.5", default-features = false }
|
||||||
rand_core = { version = "0.6", default-features = false }
|
rand_core = { version = "0.6", default-features = false }
|
||||||
|
|
||||||
std-shims = { version = "0.1", path = "../../../common/std-shims", default-features = false }
|
std-shims = { version = "0.1", path = "../../../common/std-shims", default-features = false, features = ["alloc"] }
|
||||||
|
|
||||||
ciphersuite = { path = "../../ciphersuite", version = "^0.4.1", default-features = false }
|
ciphersuite = { path = "../../ciphersuite", version = "^0.4.1", default-features = false }
|
||||||
dkg = { path = "../", version = "0.6", default-features = false }
|
dkg = { path = "../", version = "0.6", default-features = false }
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#![cfg_attr(docsrs, feature(doc_auto_cfg))]
|
#![cfg_attr(docsrs, feature(doc_cfg))]
|
||||||
#![doc = include_str!("../README.md")]
|
#![doc = include_str!("../README.md")]
|
||||||
#![no_std]
|
#![no_std]
|
||||||
|
|
||||||
@@ -10,12 +10,12 @@ use rand_core::{RngCore, CryptoRng};
|
|||||||
|
|
||||||
use ciphersuite::{
|
use ciphersuite::{
|
||||||
group::ff::{Field, PrimeField},
|
group::ff::{Field, PrimeField},
|
||||||
Ciphersuite,
|
GroupIo, Id,
|
||||||
};
|
};
|
||||||
pub use dkg::*;
|
pub use dkg::*;
|
||||||
|
|
||||||
/// Create a key via a dealer key generation protocol.
|
/// Create a key via a dealer key generation protocol.
|
||||||
pub fn key_gen<R: RngCore + CryptoRng, C: Ciphersuite>(
|
pub fn key_gen<R: RngCore + CryptoRng, C: GroupIo + Id>(
|
||||||
rng: &mut R,
|
rng: &mut R,
|
||||||
threshold: u16,
|
threshold: u16,
|
||||||
participants: u16,
|
participants: u16,
|
||||||
|
|||||||
@@ -23,7 +23,7 @@ rand_core = { version = "0.6", default-features = false, features = ["alloc"] }
|
|||||||
|
|
||||||
zeroize = { version = "^1.5", default-features = false, features = ["alloc", "zeroize_derive"] }
|
zeroize = { version = "^1.5", default-features = false, features = ["alloc", "zeroize_derive"] }
|
||||||
|
|
||||||
std-shims = { version = "0.1", path = "../../../common/std-shims", default-features = false }
|
std-shims = { version = "0.1", path = "../../../common/std-shims", default-features = false, features = ["alloc"] }
|
||||||
|
|
||||||
transcript = { package = "flexible-transcript", path = "../../transcript", version = "^0.3.2", default-features = false, features = ["recommended"] }
|
transcript = { package = "flexible-transcript", path = "../../transcript", version = "^0.3.2", default-features = false, features = ["recommended"] }
|
||||||
|
|
||||||
@@ -31,13 +31,13 @@ ciphersuite = { path = "../../ciphersuite", version = "^0.4.1", default-features
|
|||||||
multiexp = { path = "../../multiexp", version = "0.4", default-features = false }
|
multiexp = { path = "../../multiexp", version = "0.4", default-features = false }
|
||||||
|
|
||||||
generic-array = { version = "1", default-features = false, features = ["alloc"] }
|
generic-array = { version = "1", default-features = false, features = ["alloc"] }
|
||||||
blake2 = { version = "0.11.0-rc.0", default-features = false }
|
blake2 = { version = "0.11.0-rc.2", default-features = false }
|
||||||
rand_chacha = { version = "0.3", default-features = false }
|
rand_chacha = { version = "0.3", default-features = false }
|
||||||
|
|
||||||
generalized-bulletproofs = { git = "https://github.com/monero-oxide/monero-oxide", rev = "a6f8797007e768488568b821435cf5006517a962", default-features = false }
|
generalized-bulletproofs = { git = "https://github.com/monero-oxide/monero-oxide", rev = "dc1b3dbe436aae61ec363505052d4715d38ce1df", default-features = false }
|
||||||
ec-divisors = { git = "https://github.com/monero-oxide/monero-oxide", rev = "a6f8797007e768488568b821435cf5006517a962", default-features = false }
|
ec-divisors = { git = "https://github.com/monero-oxide/monero-oxide", rev = "dc1b3dbe436aae61ec363505052d4715d38ce1df", default-features = false }
|
||||||
generalized-bulletproofs-circuit-abstraction = { git = "https://github.com/monero-oxide/monero-oxide", rev = "a6f8797007e768488568b821435cf5006517a962", default-features = false }
|
generalized-bulletproofs-circuit-abstraction = { git = "https://github.com/monero-oxide/monero-oxide", rev = "dc1b3dbe436aae61ec363505052d4715d38ce1df", default-features = false }
|
||||||
generalized-bulletproofs-ec-gadgets = { git = "https://github.com/monero-oxide/monero-oxide", rev = "a6f8797007e768488568b821435cf5006517a962", default-features = false }
|
generalized-bulletproofs-ec-gadgets = { git = "https://github.com/monero-oxide/monero-oxide", rev = "dc1b3dbe436aae61ec363505052d4715d38ce1df", default-features = false }
|
||||||
|
|
||||||
dkg = { path = "..", default-features = false }
|
dkg = { path = "..", default-features = false }
|
||||||
|
|
||||||
@@ -52,7 +52,7 @@ rand = { version = "0.8", default-features = false, features = ["std"] }
|
|||||||
ciphersuite = { path = "../../ciphersuite", default-features = false, features = ["std"] }
|
ciphersuite = { path = "../../ciphersuite", default-features = false, features = ["std"] }
|
||||||
embedwards25519 = { path = "../../embedwards25519", default-features = false, features = ["std"] }
|
embedwards25519 = { path = "../../embedwards25519", default-features = false, features = ["std"] }
|
||||||
dalek-ff-group = { path = "../../dalek-ff-group", default-features = false, features = ["std"] }
|
dalek-ff-group = { path = "../../dalek-ff-group", default-features = false, features = ["std"] }
|
||||||
generalized-bulletproofs = { git = "https://github.com/monero-oxide/monero-oxide", rev = "a6f8797007e768488568b821435cf5006517a962", features = ["tests"] }
|
generalized-bulletproofs = { git = "https://github.com/monero-oxide/monero-oxide", rev = "dc1b3dbe436aae61ec363505052d4715d38ce1df", features = ["tests"] }
|
||||||
dkg-recovery = { path = "../recovery" }
|
dkg-recovery = { path = "../recovery" }
|
||||||
|
|
||||||
[features]
|
[features]
|
||||||
@@ -86,6 +86,5 @@ std = [
|
|||||||
]
|
]
|
||||||
secp256k1 = ["ciphersuite-kp256", "secq256k1"]
|
secp256k1 = ["ciphersuite-kp256", "secq256k1"]
|
||||||
ed25519 = ["dalek-ff-group", "embedwards25519"]
|
ed25519 = ["dalek-ff-group", "embedwards25519"]
|
||||||
ristretto = ["dalek-ff-group", "embedwards25519"]
|
|
||||||
tests = ["rand_core/getrandom"]
|
tests = ["rand_core/getrandom"]
|
||||||
default = ["std"]
|
default = ["std"]
|
||||||
|
|||||||
@@ -26,21 +26,9 @@ presented in section 4.2 is extended, with the following changes:
|
|||||||
just one round.
|
just one round.
|
||||||
|
|
||||||
For a gist of the verifiable encryption scheme, please see
|
For a gist of the verifiable encryption scheme, please see
|
||||||
https://gist.github.com/kayabaNerve/cfbde74b0660dfdf8dd55326d6ec33d7. Security
|
https://gist.github.com/kayabaNerve/cfbde74b0660dfdf8dd55326d6ec33d7. For
|
||||||
proofs are currently being worked on.
|
security proofs and audit information, please see
|
||||||
|
[here](../../../audits/crypto/dkg/evrf).
|
||||||
---
|
|
||||||
|
|
||||||
This library relies on an implementation of Bulletproofs and various
|
|
||||||
zero-knowledge gadgets. This library uses
|
|
||||||
[`generalized-bulletproofs`](https://docs.rs/generalized-bulletproofs),
|
|
||||||
[`generalized-bulletproofs-circuit-abstraction`](https://docs.rs/generalized-bulletproofs-circuit-abstraction),
|
|
||||||
and
|
|
||||||
[`generalized-bulletproofs-ec-gadgets`](https://docs.rs/generalized-bulletproofs-ec-gadgets)
|
|
||||||
from the Monero project's FCMP++ codebase. These libraries have received the
|
|
||||||
following audits in the past:
|
|
||||||
- https://github.com/kayabaNerve/monero-oxide/tree/fcmp++/audits/generalized-bulletproofs
|
|
||||||
- https://github.com/kayabaNerve/monero-oxide/tree/fcmp++/audits/fcmps
|
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
|
|||||||
@@ -17,7 +17,7 @@ type Blake2s256Keyed = Blake2sMac<U32>;
|
|||||||
|
|
||||||
use ciphersuite::{
|
use ciphersuite::{
|
||||||
group::{ff::FromUniformBytes, GroupEncoding},
|
group::{ff::FromUniformBytes, GroupEncoding},
|
||||||
Ciphersuite,
|
WrappedGroup, Id, GroupIo,
|
||||||
};
|
};
|
||||||
|
|
||||||
use ec_divisors::DivisorCurve;
|
use ec_divisors::DivisorCurve;
|
||||||
@@ -27,10 +27,10 @@ use generalized_bulletproofs_ec_gadgets::*;
|
|||||||
/// A pair of curves to perform the eVRF with.
|
/// A pair of curves to perform the eVRF with.
|
||||||
pub trait Curves {
|
pub trait Curves {
|
||||||
/// The towering curve, for which the resulting key is on.
|
/// The towering curve, for which the resulting key is on.
|
||||||
type ToweringCurve: Ciphersuite<F: FromUniformBytes<64>>;
|
type ToweringCurve: Id + GroupIo<F: FromUniformBytes<64>>;
|
||||||
/// The embedded curve which participants represent their public keys over.
|
/// The embedded curve which participants represent their public keys over.
|
||||||
type EmbeddedCurve: Ciphersuite<
|
type EmbeddedCurve: GroupIo<
|
||||||
G: DivisorCurve<FieldElement = <Self::ToweringCurve as Ciphersuite>::F>,
|
G: DivisorCurve<FieldElement = <Self::ToweringCurve as WrappedGroup>::F>,
|
||||||
>;
|
>;
|
||||||
/// The parameters to use the embedded curve with the discrete-log gadget.
|
/// The parameters to use the embedded curve with the discrete-log gadget.
|
||||||
type EmbeddedCurveParameters: DiscreteLogParameters;
|
type EmbeddedCurveParameters: DiscreteLogParameters;
|
||||||
@@ -49,14 +49,14 @@ impl<C: Curves> Generators<C> {
|
|||||||
pub fn new(max_threshold: u16, max_participants: u16) -> Generators<C> {
|
pub fn new(max_threshold: u16, max_participants: u16) -> Generators<C> {
|
||||||
let entropy = <Blake2s256Keyed as KeyInit>::new(&{
|
let entropy = <Blake2s256Keyed as KeyInit>::new(&{
|
||||||
let mut key = Array::<u8, <Blake2s256Keyed as KeySizeUser>::KeySize>::default();
|
let mut key = Array::<u8, <Blake2s256Keyed as KeySizeUser>::KeySize>::default();
|
||||||
let key_len = key.len().min(<C::ToweringCurve as Ciphersuite>::ID.len());
|
let key_len = key.len().min(<C::ToweringCurve as Id>::ID.len());
|
||||||
{
|
{
|
||||||
let key: &mut [u8] = key.as_mut();
|
let key: &mut [u8] = key.as_mut();
|
||||||
key[.. key_len].copy_from_slice(&<C::ToweringCurve as Ciphersuite>::ID[.. key_len])
|
key[.. key_len].copy_from_slice(&<C::ToweringCurve as Id>::ID[.. key_len])
|
||||||
}
|
}
|
||||||
key
|
key
|
||||||
})
|
})
|
||||||
.chain_update(<C::ToweringCurve as Ciphersuite>::generator().to_bytes())
|
.chain_update(<C::ToweringCurve as WrappedGroup>::generator().to_bytes())
|
||||||
.finalize()
|
.finalize()
|
||||||
.into_bytes();
|
.into_bytes();
|
||||||
let mut rng = ChaCha20Rng::from_seed(entropy.into());
|
let mut rng = ChaCha20Rng::from_seed(entropy.into());
|
||||||
@@ -71,7 +71,8 @@ impl<C: Curves> Generators<C> {
|
|||||||
h_bold.push(crate::sample_point::<C::ToweringCurve>(&mut rng));
|
h_bold.push(crate::sample_point::<C::ToweringCurve>(&mut rng));
|
||||||
}
|
}
|
||||||
Self(
|
Self(
|
||||||
BpGenerators::new(<C::ToweringCurve as Ciphersuite>::generator(), h, g_bold, h_bold).unwrap(),
|
BpGenerators::new(<C::ToweringCurve as WrappedGroup>::generator(), h, g_bold, h_bold)
|
||||||
|
.unwrap(),
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -95,13 +96,3 @@ impl Curves for Ed25519 {
|
|||||||
type EmbeddedCurve = embedwards25519::Embedwards25519;
|
type EmbeddedCurve = embedwards25519::Embedwards25519;
|
||||||
type EmbeddedCurveParameters = embedwards25519::Embedwards25519;
|
type EmbeddedCurveParameters = embedwards25519::Embedwards25519;
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Ristretto, and an elliptic curve defined over its scalar field (embedwards25519).
|
|
||||||
#[cfg(any(test, feature = "ristretto"))]
|
|
||||||
pub struct Ristretto;
|
|
||||||
#[cfg(any(test, feature = "ristretto"))]
|
|
||||||
impl Curves for Ristretto {
|
|
||||||
type ToweringCurve = dalek_ff_group::Ristretto;
|
|
||||||
type EmbeddedCurve = embedwards25519::Embedwards25519;
|
|
||||||
type EmbeddedCurveParameters = embedwards25519::Embedwards25519;
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#![cfg_attr(docsrs, feature(doc_auto_cfg))]
|
#![cfg_attr(docsrs, feature(doc_cfg))]
|
||||||
#![doc = include_str!("../README.md")]
|
#![doc = include_str!("../README.md")]
|
||||||
#![cfg_attr(not(feature = "std"), no_std)]
|
#![cfg_attr(not(feature = "std"), no_std)]
|
||||||
|
|
||||||
@@ -21,7 +21,7 @@ use ciphersuite::{
|
|||||||
ff::{Field, PrimeField},
|
ff::{Field, PrimeField},
|
||||||
Group, GroupEncoding,
|
Group, GroupEncoding,
|
||||||
},
|
},
|
||||||
Ciphersuite,
|
WrappedGroup, GroupIo,
|
||||||
};
|
};
|
||||||
use multiexp::multiexp_vartime;
|
use multiexp::multiexp_vartime;
|
||||||
|
|
||||||
@@ -49,7 +49,7 @@ mod tests;
|
|||||||
#[derive(Clone, PartialEq, Eq, Debug)]
|
#[derive(Clone, PartialEq, Eq, Debug)]
|
||||||
pub struct Participation<C: Curves> {
|
pub struct Participation<C: Curves> {
|
||||||
proof: Vec<u8>,
|
proof: Vec<u8>,
|
||||||
encrypted_secret_shares: HashMap<Participant, <C::ToweringCurve as Ciphersuite>::F>,
|
encrypted_secret_shares: HashMap<Participant, <C::ToweringCurve as WrappedGroup>::F>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<C: Curves> Participation<C> {
|
impl<C: Curves> Participation<C> {
|
||||||
@@ -79,7 +79,7 @@ impl<C: Curves> Participation<C> {
|
|||||||
|
|
||||||
let mut encrypted_secret_shares = HashMap::with_capacity(usize::from(n));
|
let mut encrypted_secret_shares = HashMap::with_capacity(usize::from(n));
|
||||||
for i in Participant::iter().take(usize::from(n)) {
|
for i in Participant::iter().take(usize::from(n)) {
|
||||||
encrypted_secret_shares.insert(i, <C::ToweringCurve as Ciphersuite>::read_F(reader)?);
|
encrypted_secret_shares.insert(i, <C::ToweringCurve as GroupIo>::read_F(reader)?);
|
||||||
}
|
}
|
||||||
|
|
||||||
Ok(Self { proof, encrypted_secret_shares })
|
Ok(Self { proof, encrypted_secret_shares })
|
||||||
@@ -151,14 +151,14 @@ pub enum VerifyResult<C: Curves> {
|
|||||||
pub struct Dkg<C: Curves> {
|
pub struct Dkg<C: Curves> {
|
||||||
t: u16,
|
t: u16,
|
||||||
n: u16,
|
n: u16,
|
||||||
evrf_public_keys: Vec<<C::EmbeddedCurve as Ciphersuite>::G>,
|
evrf_public_keys: Vec<<C::EmbeddedCurve as WrappedGroup>::G>,
|
||||||
verification_shares: HashMap<Participant, <C::ToweringCurve as Ciphersuite>::G>,
|
verification_shares: HashMap<Participant, <C::ToweringCurve as WrappedGroup>::G>,
|
||||||
#[allow(clippy::type_complexity)]
|
#[allow(clippy::type_complexity)]
|
||||||
encrypted_secret_shares: HashMap<
|
encrypted_secret_shares: HashMap<
|
||||||
Participant,
|
Participant,
|
||||||
HashMap<
|
HashMap<
|
||||||
Participant,
|
Participant,
|
||||||
([<C::EmbeddedCurve as Ciphersuite>::G; 2], <C::ToweringCurve as Ciphersuite>::F),
|
([<C::EmbeddedCurve as WrappedGroup>::G; 2], <C::ToweringCurve as WrappedGroup>::F),
|
||||||
>,
|
>,
|
||||||
>,
|
>,
|
||||||
}
|
}
|
||||||
@@ -167,7 +167,7 @@ impl<C: Curves> Dkg<C> {
|
|||||||
// Form the initial transcript for the proofs.
|
// Form the initial transcript for the proofs.
|
||||||
fn initial_transcript(
|
fn initial_transcript(
|
||||||
invocation: [u8; 32],
|
invocation: [u8; 32],
|
||||||
evrf_public_keys: &[<C::EmbeddedCurve as Ciphersuite>::G],
|
evrf_public_keys: &[<C::EmbeddedCurve as WrappedGroup>::G],
|
||||||
t: u16,
|
t: u16,
|
||||||
) -> [u8; 32] {
|
) -> [u8; 32] {
|
||||||
let mut transcript = Blake2s256::new();
|
let mut transcript = Blake2s256::new();
|
||||||
@@ -188,8 +188,8 @@ impl<C: Curves> Dkg<C> {
|
|||||||
generators: &Generators<C>,
|
generators: &Generators<C>,
|
||||||
context: [u8; 32],
|
context: [u8; 32],
|
||||||
t: u16,
|
t: u16,
|
||||||
evrf_public_keys: &[<C::EmbeddedCurve as Ciphersuite>::G],
|
evrf_public_keys: &[<C::EmbeddedCurve as WrappedGroup>::G],
|
||||||
evrf_private_key: &Zeroizing<<C::EmbeddedCurve as Ciphersuite>::F>,
|
evrf_private_key: &Zeroizing<<C::EmbeddedCurve as WrappedGroup>::F>,
|
||||||
) -> Result<Participation<C>, Error> {
|
) -> Result<Participation<C>, Error> {
|
||||||
let Ok(n) = u16::try_from(evrf_public_keys.len()) else {
|
let Ok(n) = u16::try_from(evrf_public_keys.len()) else {
|
||||||
Err(Error::TooManyParticipants { provided: evrf_public_keys.len() })?
|
Err(Error::TooManyParticipants { provided: evrf_public_keys.len() })?
|
||||||
@@ -202,7 +202,8 @@ impl<C: Curves> Dkg<C> {
|
|||||||
};
|
};
|
||||||
// This also ensures the private key is not 0, due to the prior check the identity point wasn't
|
// This also ensures the private key is not 0, due to the prior check the identity point wasn't
|
||||||
// present
|
// present
|
||||||
let evrf_public_key = <C::EmbeddedCurve as Ciphersuite>::generator() * evrf_private_key.deref();
|
let evrf_public_key =
|
||||||
|
<C::EmbeddedCurve as WrappedGroup>::generator() * evrf_private_key.deref();
|
||||||
if !evrf_public_keys.contains(&evrf_public_key) {
|
if !evrf_public_keys.contains(&evrf_public_key) {
|
||||||
Err(Error::NotAParticipant)?;
|
Err(Error::NotAParticipant)?;
|
||||||
};
|
};
|
||||||
@@ -231,7 +232,7 @@ impl<C: Curves> Dkg<C> {
|
|||||||
|
|
||||||
let mut encrypted_secret_shares = HashMap::with_capacity(usize::from(n));
|
let mut encrypted_secret_shares = HashMap::with_capacity(usize::from(n));
|
||||||
for (l, encryption_key) in Participant::iter().take(usize::from(n)).zip(encryption_keys) {
|
for (l, encryption_key) in Participant::iter().take(usize::from(n)).zip(encryption_keys) {
|
||||||
let share = polynomial::<<C::ToweringCurve as Ciphersuite>::F>(&coefficients, l);
|
let share = polynomial::<<C::ToweringCurve as WrappedGroup>::F>(&coefficients, l);
|
||||||
encrypted_secret_shares.insert(l, *share + *encryption_key);
|
encrypted_secret_shares.insert(l, *share + *encryption_key);
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -243,26 +244,26 @@ impl<C: Curves> Dkg<C> {
|
|||||||
#[allow(clippy::type_complexity)]
|
#[allow(clippy::type_complexity)]
|
||||||
fn verifiable_encryption_statements<C: Curves>(
|
fn verifiable_encryption_statements<C: Curves>(
|
||||||
rng: &mut (impl RngCore + CryptoRng),
|
rng: &mut (impl RngCore + CryptoRng),
|
||||||
coefficients: &[<C::ToweringCurve as Ciphersuite>::G],
|
coefficients: &[<C::ToweringCurve as WrappedGroup>::G],
|
||||||
encryption_key_commitments: &[<C::ToweringCurve as Ciphersuite>::G],
|
encryption_key_commitments: &[<C::ToweringCurve as WrappedGroup>::G],
|
||||||
encrypted_secret_shares: &HashMap<Participant, <C::ToweringCurve as Ciphersuite>::F>,
|
encrypted_secret_shares: &HashMap<Participant, <C::ToweringCurve as WrappedGroup>::F>,
|
||||||
) -> (
|
) -> (
|
||||||
<C::ToweringCurve as Ciphersuite>::F,
|
<C::ToweringCurve as WrappedGroup>::F,
|
||||||
Vec<(<C::ToweringCurve as Ciphersuite>::F, <C::ToweringCurve as Ciphersuite>::G)>,
|
Vec<(<C::ToweringCurve as WrappedGroup>::F, <C::ToweringCurve as WrappedGroup>::G)>,
|
||||||
) {
|
) {
|
||||||
let mut g_scalar = <C::ToweringCurve as Ciphersuite>::F::ZERO;
|
let mut g_scalar = <C::ToweringCurve as WrappedGroup>::F::ZERO;
|
||||||
let mut pairs = Vec::with_capacity(coefficients.len() + encryption_key_commitments.len());
|
let mut pairs = Vec::with_capacity(coefficients.len() + encryption_key_commitments.len());
|
||||||
|
|
||||||
// Push on the commitments to the polynomial being secret-shared
|
// Push on the commitments to the polynomial being secret-shared
|
||||||
for coefficient in coefficients {
|
for coefficient in coefficients {
|
||||||
// This uses `0` as we'll add to it later, given its fixed position
|
// This uses `0` as we'll add to it later, given its fixed position
|
||||||
pairs.push((<C::ToweringCurve as Ciphersuite>::F::ZERO, *coefficient));
|
pairs.push((<C::ToweringCurve as WrappedGroup>::F::ZERO, *coefficient));
|
||||||
}
|
}
|
||||||
|
|
||||||
for (i, encrypted_secret_share) in encrypted_secret_shares {
|
for (i, encrypted_secret_share) in encrypted_secret_shares {
|
||||||
let encryption_key_commitment = encryption_key_commitments[usize::from(u16::from(*i)) - 1];
|
let encryption_key_commitment = encryption_key_commitments[usize::from(u16::from(*i)) - 1];
|
||||||
|
|
||||||
let weight = <C::ToweringCurve as Ciphersuite>::F::random(&mut *rng);
|
let weight = <C::ToweringCurve as WrappedGroup>::F::random(&mut *rng);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
The encrypted secret share scaling `G`, minus the encryption key commitment, minus the
|
The encrypted secret share scaling `G`, minus the encryption key commitment, minus the
|
||||||
@@ -274,7 +275,7 @@ fn verifiable_encryption_statements<C: Curves>(
|
|||||||
pairs.push((weight, encryption_key_commitment));
|
pairs.push((weight, encryption_key_commitment));
|
||||||
// Calculate the commitment to the secret share via the commitments to the polynomial
|
// Calculate the commitment to the secret share via the commitments to the polynomial
|
||||||
{
|
{
|
||||||
let i = <C::ToweringCurve as Ciphersuite>::F::from(u64::from(u16::from(*i)));
|
let i = <C::ToweringCurve as WrappedGroup>::F::from(u64::from(u16::from(*i)));
|
||||||
(0 .. coefficients.len()).fold(weight, |exp, j| {
|
(0 .. coefficients.len()).fold(weight, |exp, j| {
|
||||||
pairs[j].0 += exp;
|
pairs[j].0 += exp;
|
||||||
exp * i
|
exp * i
|
||||||
@@ -300,7 +301,7 @@ impl<C: Curves> Dkg<C> {
|
|||||||
generators: &Generators<C>,
|
generators: &Generators<C>,
|
||||||
context: [u8; 32],
|
context: [u8; 32],
|
||||||
t: u16,
|
t: u16,
|
||||||
evrf_public_keys: &[<C::EmbeddedCurve as Ciphersuite>::G],
|
evrf_public_keys: &[<C::EmbeddedCurve as WrappedGroup>::G],
|
||||||
participations: &HashMap<Participant, Participation<C>>,
|
participations: &HashMap<Participant, Participation<C>>,
|
||||||
) -> Result<VerifyResult<C>, Error> {
|
) -> Result<VerifyResult<C>, Error> {
|
||||||
let Ok(n) = u16::try_from(evrf_public_keys.len()) else {
|
let Ok(n) = u16::try_from(evrf_public_keys.len()) else {
|
||||||
@@ -386,7 +387,7 @@ impl<C: Curves> Dkg<C> {
|
|||||||
{
|
{
|
||||||
let mut share_verification_statements_actual = HashMap::with_capacity(valid.len());
|
let mut share_verification_statements_actual = HashMap::with_capacity(valid.len());
|
||||||
if !{
|
if !{
|
||||||
let mut g_scalar = <C::ToweringCurve as Ciphersuite>::F::ZERO;
|
let mut g_scalar = <C::ToweringCurve as WrappedGroup>::F::ZERO;
|
||||||
let mut pairs = Vec::with_capacity(valid.len() * (usize::from(t) + evrf_public_keys.len()));
|
let mut pairs = Vec::with_capacity(valid.len() * (usize::from(t) + evrf_public_keys.len()));
|
||||||
for (i, (encrypted_secret_shares, data)) in &valid {
|
for (i, (encrypted_secret_shares, data)) in &valid {
|
||||||
let (this_g_scalar, mut these_pairs) = verifiable_encryption_statements::<C>(
|
let (this_g_scalar, mut these_pairs) = verifiable_encryption_statements::<C>(
|
||||||
@@ -417,9 +418,11 @@ impl<C: Curves> Dkg<C> {
|
|||||||
let sum_encrypted_secret_share = sum_encrypted_secret_shares
|
let sum_encrypted_secret_share = sum_encrypted_secret_shares
|
||||||
.get(j)
|
.get(j)
|
||||||
.copied()
|
.copied()
|
||||||
.unwrap_or(<C::ToweringCurve as Ciphersuite>::F::ZERO);
|
.unwrap_or(<C::ToweringCurve as WrappedGroup>::F::ZERO);
|
||||||
let sum_mask =
|
let sum_mask = sum_masks
|
||||||
sum_masks.get(j).copied().unwrap_or(<C::ToweringCurve as Ciphersuite>::G::identity());
|
.get(j)
|
||||||
|
.copied()
|
||||||
|
.unwrap_or(<C::ToweringCurve as WrappedGroup>::G::identity());
|
||||||
sum_encrypted_secret_shares.insert(*j, sum_encrypted_secret_share + enc_share);
|
sum_encrypted_secret_shares.insert(*j, sum_encrypted_secret_share + enc_share);
|
||||||
|
|
||||||
let j_index = usize::from(u16::from(*j)) - 1;
|
let j_index = usize::from(u16::from(*j)) - 1;
|
||||||
@@ -487,7 +490,7 @@ impl<C: Curves> Dkg<C> {
|
|||||||
for i in Participant::iter().take(usize::from(n)) {
|
for i in Participant::iter().take(usize::from(n)) {
|
||||||
verification_shares.insert(
|
verification_shares.insert(
|
||||||
i,
|
i,
|
||||||
(<C::ToweringCurve as Ciphersuite>::generator() * sum_encrypted_secret_shares[&i]) -
|
(<C::ToweringCurve as WrappedGroup>::generator() * sum_encrypted_secret_shares[&i]) -
|
||||||
sum_masks[&i],
|
sum_masks[&i],
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
@@ -506,9 +509,10 @@ impl<C: Curves> Dkg<C> {
|
|||||||
/// This will return _all_ keys belong to the participant.
|
/// This will return _all_ keys belong to the participant.
|
||||||
pub fn keys(
|
pub fn keys(
|
||||||
&self,
|
&self,
|
||||||
evrf_private_key: &Zeroizing<<C::EmbeddedCurve as Ciphersuite>::F>,
|
evrf_private_key: &Zeroizing<<C::EmbeddedCurve as WrappedGroup>::F>,
|
||||||
) -> Vec<ThresholdKeys<C::ToweringCurve>> {
|
) -> Vec<ThresholdKeys<C::ToweringCurve>> {
|
||||||
let evrf_public_key = <C::EmbeddedCurve as Ciphersuite>::generator() * evrf_private_key.deref();
|
let evrf_public_key =
|
||||||
|
<C::EmbeddedCurve as WrappedGroup>::generator() * evrf_private_key.deref();
|
||||||
let mut is = Vec::with_capacity(1);
|
let mut is = Vec::with_capacity(1);
|
||||||
for (i, evrf_key) in Participant::iter().zip(self.evrf_public_keys.iter()) {
|
for (i, evrf_key) in Participant::iter().zip(self.evrf_public_keys.iter()) {
|
||||||
if *evrf_key == evrf_public_key {
|
if *evrf_key == evrf_public_key {
|
||||||
@@ -518,14 +522,14 @@ impl<C: Curves> Dkg<C> {
|
|||||||
|
|
||||||
let mut res = Vec::with_capacity(is.len());
|
let mut res = Vec::with_capacity(is.len());
|
||||||
for i in is {
|
for i in is {
|
||||||
let mut secret_share = Zeroizing::new(<C::ToweringCurve as Ciphersuite>::F::ZERO);
|
let mut secret_share = Zeroizing::new(<C::ToweringCurve as WrappedGroup>::F::ZERO);
|
||||||
for shares in self.encrypted_secret_shares.values() {
|
for shares in self.encrypted_secret_shares.values() {
|
||||||
let (ecdh_commitments, encrypted_secret_share) = shares[&i];
|
let (ecdh_commitments, encrypted_secret_share) = shares[&i];
|
||||||
|
|
||||||
let mut ecdh = Zeroizing::new(<C::ToweringCurve as Ciphersuite>::F::ZERO);
|
let mut ecdh = Zeroizing::new(<C::ToweringCurve as WrappedGroup>::F::ZERO);
|
||||||
for point in ecdh_commitments {
|
for point in ecdh_commitments {
|
||||||
let (mut x, mut y) =
|
let (mut x, mut y) =
|
||||||
<C::EmbeddedCurve as Ciphersuite>::G::to_xy(point * evrf_private_key.deref()).unwrap();
|
<C::EmbeddedCurve as WrappedGroup>::G::to_xy(point * evrf_private_key.deref()).unwrap();
|
||||||
*ecdh += x;
|
*ecdh += x;
|
||||||
x.zeroize();
|
x.zeroize();
|
||||||
y.zeroize();
|
y.zeroize();
|
||||||
@@ -534,7 +538,7 @@ impl<C: Curves> Dkg<C> {
|
|||||||
}
|
}
|
||||||
debug_assert_eq!(
|
debug_assert_eq!(
|
||||||
self.verification_shares[&i],
|
self.verification_shares[&i],
|
||||||
<C::ToweringCurve as Ciphersuite>::G::generator() * secret_share.deref()
|
<C::ToweringCurve as WrappedGroup>::generator() * secret_share.deref()
|
||||||
);
|
);
|
||||||
|
|
||||||
res.push(
|
res.push(
|
||||||
|
|||||||
@@ -8,7 +8,7 @@ use zeroize::Zeroizing;
|
|||||||
use rand_core::{RngCore, CryptoRng, SeedableRng};
|
use rand_core::{RngCore, CryptoRng, SeedableRng};
|
||||||
use rand_chacha::ChaCha20Rng;
|
use rand_chacha::ChaCha20Rng;
|
||||||
|
|
||||||
use ciphersuite::{group::ff::Field, Ciphersuite};
|
use ciphersuite::{group::ff::Field, WrappedGroup};
|
||||||
|
|
||||||
use generalized_bulletproofs::{
|
use generalized_bulletproofs::{
|
||||||
Generators, BatchVerifier, PedersenCommitment, PedersenVectorCommitment,
|
Generators, BatchVerifier, PedersenCommitment, PedersenVectorCommitment,
|
||||||
@@ -28,8 +28,8 @@ mod tape;
|
|||||||
use tape::*;
|
use tape::*;
|
||||||
|
|
||||||
type EmbeddedPoint<C> = (
|
type EmbeddedPoint<C> = (
|
||||||
<<<C as Curves>::EmbeddedCurve as Ciphersuite>::G as DivisorCurve>::FieldElement,
|
<<<C as Curves>::EmbeddedCurve as WrappedGroup>::G as DivisorCurve>::FieldElement,
|
||||||
<<<C as Curves>::EmbeddedCurve as Ciphersuite>::G as DivisorCurve>::FieldElement,
|
<<<C as Curves>::EmbeddedCurve as WrappedGroup>::G as DivisorCurve>::FieldElement,
|
||||||
);
|
);
|
||||||
|
|
||||||
#[allow(non_snake_case)]
|
#[allow(non_snake_case)]
|
||||||
@@ -37,14 +37,15 @@ struct Circuit<
|
|||||||
'a,
|
'a,
|
||||||
C: Curves,
|
C: Curves,
|
||||||
CG: Iterator<
|
CG: Iterator<
|
||||||
Item = ChallengedGenerator<<C::ToweringCurve as Ciphersuite>::F, C::EmbeddedCurveParameters>,
|
Item = ChallengedGenerator<<C::ToweringCurve as WrappedGroup>::F, C::EmbeddedCurveParameters>,
|
||||||
>,
|
>,
|
||||||
> {
|
> {
|
||||||
curve_spec: &'a CurveSpec<<<C::EmbeddedCurve as Ciphersuite>::G as DivisorCurve>::FieldElement>,
|
curve_spec: &'a CurveSpec<<<C::EmbeddedCurve as WrappedGroup>::G as DivisorCurve>::FieldElement>,
|
||||||
circuit: &'a mut BpCircuit<C::ToweringCurve>,
|
circuit: &'a mut BpCircuit<C::ToweringCurve>,
|
||||||
challenge: DiscreteLogChallenge<<C::ToweringCurve as Ciphersuite>::F, C::EmbeddedCurveParameters>,
|
challenge:
|
||||||
|
DiscreteLogChallenge<<C::ToweringCurve as WrappedGroup>::F, C::EmbeddedCurveParameters>,
|
||||||
challenged_G:
|
challenged_G:
|
||||||
ChallengedGenerator<<C::ToweringCurve as Ciphersuite>::F, C::EmbeddedCurveParameters>,
|
ChallengedGenerator<<C::ToweringCurve as WrappedGroup>::F, C::EmbeddedCurveParameters>,
|
||||||
challenged_generators: &'a mut CG,
|
challenged_generators: &'a mut CG,
|
||||||
tape: Tape,
|
tape: Tape,
|
||||||
pedersen_commitment_tape: PedersenCommitmentTape,
|
pedersen_commitment_tape: PedersenCommitmentTape,
|
||||||
@@ -54,7 +55,7 @@ impl<
|
|||||||
'a,
|
'a,
|
||||||
C: Curves,
|
C: Curves,
|
||||||
CG: Iterator<
|
CG: Iterator<
|
||||||
Item = ChallengedGenerator<<C::ToweringCurve as Ciphersuite>::F, C::EmbeddedCurveParameters>,
|
Item = ChallengedGenerator<<C::ToweringCurve as WrappedGroup>::F, C::EmbeddedCurveParameters>,
|
||||||
>,
|
>,
|
||||||
> Circuit<'a, C, CG>
|
> Circuit<'a, C, CG>
|
||||||
{
|
{
|
||||||
@@ -92,7 +93,7 @@ impl<
|
|||||||
&self.challenge,
|
&self.challenge,
|
||||||
&challenged_generator,
|
&challenged_generator,
|
||||||
);
|
);
|
||||||
lincomb = lincomb.term(<C::ToweringCurve as Ciphersuite>::F::ONE, point.x());
|
lincomb = lincomb.term(<C::ToweringCurve as WrappedGroup>::F::ONE, point.x());
|
||||||
}
|
}
|
||||||
/*
|
/*
|
||||||
Constrain the sum of the two `x` coordinates to be equal to the value committed to in a
|
Constrain the sum of the two `x` coordinates to be equal to the value committed to in a
|
||||||
@@ -137,7 +138,7 @@ impl<
|
|||||||
&self.challenge,
|
&self.challenge,
|
||||||
&challenged_public_key,
|
&challenged_public_key,
|
||||||
);
|
);
|
||||||
lincomb = lincomb.term(<C::ToweringCurve as Ciphersuite>::F::ONE, point.x());
|
lincomb = lincomb.term(<C::ToweringCurve as WrappedGroup>::F::ONE, point.x());
|
||||||
debug_assert!(point_with_dlogs.next().is_none());
|
debug_assert!(point_with_dlogs.next().is_none());
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -152,20 +153,20 @@ impl<
|
|||||||
/// The result of proving.
|
/// The result of proving.
|
||||||
pub(super) struct ProveResult<C: Curves> {
|
pub(super) struct ProveResult<C: Curves> {
|
||||||
/// The coefficients for use in the DKG.
|
/// The coefficients for use in the DKG.
|
||||||
pub(super) coefficients: Vec<Zeroizing<<C::ToweringCurve as Ciphersuite>::F>>,
|
pub(super) coefficients: Vec<Zeroizing<<C::ToweringCurve as WrappedGroup>::F>>,
|
||||||
/// The masks to encrypt secret shares with.
|
/// The masks to encrypt secret shares with.
|
||||||
pub(super) encryption_keys: Vec<Zeroizing<<C::ToweringCurve as Ciphersuite>::F>>,
|
pub(super) encryption_keys: Vec<Zeroizing<<C::ToweringCurve as WrappedGroup>::F>>,
|
||||||
/// The proof itself.
|
/// The proof itself.
|
||||||
pub(super) proof: Vec<u8>,
|
pub(super) proof: Vec<u8>,
|
||||||
}
|
}
|
||||||
|
|
||||||
pub(super) struct Verified<C: Curves> {
|
pub(super) struct Verified<C: Curves> {
|
||||||
/// The commitments to the coefficients used within the DKG.
|
/// The commitments to the coefficients used within the DKG.
|
||||||
pub(super) coefficients: Vec<<C::ToweringCurve as Ciphersuite>::G>,
|
pub(super) coefficients: Vec<<C::ToweringCurve as WrappedGroup>::G>,
|
||||||
/// The ephemeral public keys to perform ECDHs with
|
/// The ephemeral public keys to perform ECDHs with
|
||||||
pub(super) ecdh_commitments: Vec<[<C::EmbeddedCurve as Ciphersuite>::G; 2]>,
|
pub(super) ecdh_commitments: Vec<[<C::EmbeddedCurve as WrappedGroup>::G; 2]>,
|
||||||
/// The commitments to the masks used to encrypt secret shares with.
|
/// The commitments to the masks used to encrypt secret shares with.
|
||||||
pub(super) encryption_key_commitments: Vec<<C::ToweringCurve as Ciphersuite>::G>,
|
pub(super) encryption_key_commitments: Vec<<C::ToweringCurve as WrappedGroup>::G>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<C: Curves> fmt::Debug for Verified<C> {
|
impl<C: Curves> fmt::Debug for Verified<C> {
|
||||||
@@ -175,7 +176,7 @@ impl<C: Curves> fmt::Debug for Verified<C> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
type GeneratorTable<C> = generalized_bulletproofs_ec_gadgets::GeneratorTable<
|
type GeneratorTable<C> = generalized_bulletproofs_ec_gadgets::GeneratorTable<
|
||||||
<<<C as Curves>::EmbeddedCurve as Ciphersuite>::G as DivisorCurve>::FieldElement,
|
<<<C as Curves>::EmbeddedCurve as WrappedGroup>::G as DivisorCurve>::FieldElement,
|
||||||
<C as Curves>::EmbeddedCurveParameters,
|
<C as Curves>::EmbeddedCurveParameters,
|
||||||
>;
|
>;
|
||||||
|
|
||||||
@@ -219,7 +220,7 @@ impl<C: Curves> Proof<C> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
fn circuit(
|
fn circuit(
|
||||||
curve_spec: &CurveSpec<<<C::EmbeddedCurve as Ciphersuite>::G as DivisorCurve>::FieldElement>,
|
curve_spec: &CurveSpec<<<C::EmbeddedCurve as WrappedGroup>::G as DivisorCurve>::FieldElement>,
|
||||||
evrf_public_key: EmbeddedPoint<C>,
|
evrf_public_key: EmbeddedPoint<C>,
|
||||||
coefficients: usize,
|
coefficients: usize,
|
||||||
ecdh_commitments: &[[EmbeddedPoint<C>; 2]],
|
ecdh_commitments: &[[EmbeddedPoint<C>; 2]],
|
||||||
@@ -281,7 +282,7 @@ impl<C: Curves> Proof<C> {
|
|||||||
fn sample_coefficients_evrf_points(
|
fn sample_coefficients_evrf_points(
|
||||||
seed: [u8; 32],
|
seed: [u8; 32],
|
||||||
coefficients: usize,
|
coefficients: usize,
|
||||||
) -> Vec<<C::EmbeddedCurve as Ciphersuite>::G> {
|
) -> Vec<<C::EmbeddedCurve as WrappedGroup>::G> {
|
||||||
let mut rng = ChaCha20Rng::from_seed(seed);
|
let mut rng = ChaCha20Rng::from_seed(seed);
|
||||||
let quantity = 2 * coefficients;
|
let quantity = 2 * coefficients;
|
||||||
let mut res = Vec::with_capacity(quantity);
|
let mut res = Vec::with_capacity(quantity);
|
||||||
@@ -293,28 +294,29 @@ impl<C: Curves> Proof<C> {
|
|||||||
|
|
||||||
/// Create the required tables for the generators.
|
/// Create the required tables for the generators.
|
||||||
fn generator_tables(
|
fn generator_tables(
|
||||||
coefficients_evrf_points: &[<C::EmbeddedCurve as Ciphersuite>::G],
|
coefficients_evrf_points: &[<C::EmbeddedCurve as WrappedGroup>::G],
|
||||||
participants: &[<<C as Curves>::EmbeddedCurve as Ciphersuite>::G],
|
participants: &[<<C as Curves>::EmbeddedCurve as WrappedGroup>::G],
|
||||||
) -> Vec<GeneratorTable<C>> {
|
) -> Vec<GeneratorTable<C>> {
|
||||||
let curve_spec = CurveSpec {
|
let curve_spec = CurveSpec {
|
||||||
a: <<C as Curves>::EmbeddedCurve as Ciphersuite>::G::a(),
|
a: <<C as Curves>::EmbeddedCurve as WrappedGroup>::G::a(),
|
||||||
b: <<C as Curves>::EmbeddedCurve as Ciphersuite>::G::b(),
|
b: <<C as Curves>::EmbeddedCurve as WrappedGroup>::G::b(),
|
||||||
};
|
};
|
||||||
|
|
||||||
let mut generator_tables =
|
let mut generator_tables =
|
||||||
Vec::with_capacity(1 + coefficients_evrf_points.len() + participants.len());
|
Vec::with_capacity(1 + coefficients_evrf_points.len() + participants.len());
|
||||||
{
|
{
|
||||||
let (x, y) =
|
let (x, y) = <C::EmbeddedCurve as WrappedGroup>::G::to_xy(
|
||||||
<C::EmbeddedCurve as Ciphersuite>::G::to_xy(<C::EmbeddedCurve as Ciphersuite>::generator())
|
<C::EmbeddedCurve as WrappedGroup>::generator(),
|
||||||
.unwrap();
|
)
|
||||||
|
.unwrap();
|
||||||
generator_tables.push(GeneratorTable::<C>::new(&curve_spec, x, y));
|
generator_tables.push(GeneratorTable::<C>::new(&curve_spec, x, y));
|
||||||
}
|
}
|
||||||
for generator in coefficients_evrf_points {
|
for generator in coefficients_evrf_points {
|
||||||
let (x, y) = <C::EmbeddedCurve as Ciphersuite>::G::to_xy(*generator).unwrap();
|
let (x, y) = <C::EmbeddedCurve as WrappedGroup>::G::to_xy(*generator).unwrap();
|
||||||
generator_tables.push(GeneratorTable::<C>::new(&curve_spec, x, y));
|
generator_tables.push(GeneratorTable::<C>::new(&curve_spec, x, y));
|
||||||
}
|
}
|
||||||
for generator in participants {
|
for generator in participants {
|
||||||
let (x, y) = <C::EmbeddedCurve as Ciphersuite>::G::to_xy(*generator).unwrap();
|
let (x, y) = <C::EmbeddedCurve as WrappedGroup>::G::to_xy(*generator).unwrap();
|
||||||
generator_tables.push(GeneratorTable::<C>::new(&curve_spec, x, y));
|
generator_tables.push(GeneratorTable::<C>::new(&curve_spec, x, y));
|
||||||
}
|
}
|
||||||
generator_tables
|
generator_tables
|
||||||
@@ -325,12 +327,12 @@ impl<C: Curves> Proof<C> {
|
|||||||
generators: &Generators<C::ToweringCurve>,
|
generators: &Generators<C::ToweringCurve>,
|
||||||
transcript: [u8; 32],
|
transcript: [u8; 32],
|
||||||
coefficients: usize,
|
coefficients: usize,
|
||||||
participant_public_keys: &[<<C as Curves>::EmbeddedCurve as Ciphersuite>::G],
|
participant_public_keys: &[<<C as Curves>::EmbeddedCurve as WrappedGroup>::G],
|
||||||
evrf_private_key: &Zeroizing<<<C as Curves>::EmbeddedCurve as Ciphersuite>::F>,
|
evrf_private_key: &Zeroizing<<<C as Curves>::EmbeddedCurve as WrappedGroup>::F>,
|
||||||
) -> Result<ProveResult<C>, AcProveError> {
|
) -> Result<ProveResult<C>, AcProveError> {
|
||||||
let curve_spec = CurveSpec {
|
let curve_spec = CurveSpec {
|
||||||
a: <<C as Curves>::EmbeddedCurve as Ciphersuite>::G::a(),
|
a: <<C as Curves>::EmbeddedCurve as WrappedGroup>::G::a(),
|
||||||
b: <<C as Curves>::EmbeddedCurve as Ciphersuite>::G::b(),
|
b: <<C as Curves>::EmbeddedCurve as WrappedGroup>::G::b(),
|
||||||
};
|
};
|
||||||
|
|
||||||
let coefficients_evrf_points = Self::sample_coefficients_evrf_points(transcript, coefficients);
|
let coefficients_evrf_points = Self::sample_coefficients_evrf_points(transcript, coefficients);
|
||||||
@@ -340,7 +342,7 @@ impl<C: Curves> Proof<C> {
|
|||||||
// Push a discrete logarithm onto the tape
|
// Push a discrete logarithm onto the tape
|
||||||
let discrete_log =
|
let discrete_log =
|
||||||
|vector_commitment_tape: &mut Vec<_>,
|
|vector_commitment_tape: &mut Vec<_>,
|
||||||
dlog: &ScalarDecomposition<<<C as Curves>::EmbeddedCurve as Ciphersuite>::F>| {
|
dlog: &ScalarDecomposition<<<C as Curves>::EmbeddedCurve as WrappedGroup>::F>| {
|
||||||
for coefficient in dlog.decomposition() {
|
for coefficient in dlog.decomposition() {
|
||||||
vector_commitment_tape.push(<_>::from(*coefficient));
|
vector_commitment_tape.push(<_>::from(*coefficient));
|
||||||
}
|
}
|
||||||
@@ -351,8 +353,8 @@ impl<C: Curves> Proof<C> {
|
|||||||
// Returns the point for which the claim was made.
|
// Returns the point for which the claim was made.
|
||||||
let discrete_log_claim =
|
let discrete_log_claim =
|
||||||
|vector_commitment_tape: &mut Vec<_>,
|
|vector_commitment_tape: &mut Vec<_>,
|
||||||
dlog: &ScalarDecomposition<<<C as Curves>::EmbeddedCurve as Ciphersuite>::F>,
|
dlog: &ScalarDecomposition<<<C as Curves>::EmbeddedCurve as WrappedGroup>::F>,
|
||||||
generator: <<C as Curves>::EmbeddedCurve as Ciphersuite>::G| {
|
generator: <<C as Curves>::EmbeddedCurve as WrappedGroup>::G| {
|
||||||
{
|
{
|
||||||
let divisor =
|
let divisor =
|
||||||
Zeroizing::new(dlog.scalar_mul_divisor(generator).normalize_x_coefficient());
|
Zeroizing::new(dlog.scalar_mul_divisor(generator).normalize_x_coefficient());
|
||||||
@@ -368,12 +370,12 @@ impl<C: Curves> Proof<C> {
|
|||||||
.y_coefficients
|
.y_coefficients
|
||||||
.first()
|
.first()
|
||||||
.copied()
|
.copied()
|
||||||
.unwrap_or(<C::ToweringCurve as Ciphersuite>::F::ZERO),
|
.unwrap_or(<C::ToweringCurve as WrappedGroup>::F::ZERO),
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
let dh = generator * dlog.scalar();
|
let dh = generator * dlog.scalar();
|
||||||
let (x, y) = <C::EmbeddedCurve as Ciphersuite>::G::to_xy(dh).unwrap();
|
let (x, y) = <C::EmbeddedCurve as WrappedGroup>::G::to_xy(dh).unwrap();
|
||||||
vector_commitment_tape.push(x);
|
vector_commitment_tape.push(x);
|
||||||
vector_commitment_tape.push(y);
|
vector_commitment_tape.push(y);
|
||||||
(dh, (x, y))
|
(dh, (x, y))
|
||||||
@@ -387,7 +389,7 @@ impl<C: Curves> Proof<C> {
|
|||||||
let mut coefficients = Vec::with_capacity(coefficients);
|
let mut coefficients = Vec::with_capacity(coefficients);
|
||||||
let evrf_public_key = {
|
let evrf_public_key = {
|
||||||
let evrf_private_key =
|
let evrf_private_key =
|
||||||
ScalarDecomposition::<<C::EmbeddedCurve as Ciphersuite>::F>::new(**evrf_private_key)
|
ScalarDecomposition::<<C::EmbeddedCurve as WrappedGroup>::F>::new(**evrf_private_key)
|
||||||
.expect("eVRF private key was zero");
|
.expect("eVRF private key was zero");
|
||||||
|
|
||||||
discrete_log(&mut vector_commitment_tape, &evrf_private_key);
|
discrete_log(&mut vector_commitment_tape, &evrf_private_key);
|
||||||
@@ -396,12 +398,12 @@ impl<C: Curves> Proof<C> {
|
|||||||
let (_, evrf_public_key) = discrete_log_claim(
|
let (_, evrf_public_key) = discrete_log_claim(
|
||||||
&mut vector_commitment_tape,
|
&mut vector_commitment_tape,
|
||||||
&evrf_private_key,
|
&evrf_private_key,
|
||||||
<<C as Curves>::EmbeddedCurve as Ciphersuite>::generator(),
|
<<C as Curves>::EmbeddedCurve as WrappedGroup>::generator(),
|
||||||
);
|
);
|
||||||
|
|
||||||
// Push the divisor for each point we use in the eVRF
|
// Push the divisor for each point we use in the eVRF
|
||||||
for pair in coefficients_evrf_points.chunks(2) {
|
for pair in coefficients_evrf_points.chunks(2) {
|
||||||
let mut coefficient = Zeroizing::new(<C::ToweringCurve as Ciphersuite>::F::ZERO);
|
let mut coefficient = Zeroizing::new(<C::ToweringCurve as WrappedGroup>::F::ZERO);
|
||||||
for point in pair {
|
for point in pair {
|
||||||
let (_, (dh_x, _)) =
|
let (_, (dh_x, _)) =
|
||||||
discrete_log_claim(&mut vector_commitment_tape, &evrf_private_key, *point);
|
discrete_log_claim(&mut vector_commitment_tape, &evrf_private_key, *point);
|
||||||
@@ -418,15 +420,16 @@ impl<C: Curves> Proof<C> {
|
|||||||
let mut ecdh_commitments = Vec::with_capacity(2 * participant_public_keys.len());
|
let mut ecdh_commitments = Vec::with_capacity(2 * participant_public_keys.len());
|
||||||
let mut ecdh_commitments_xy = Vec::with_capacity(participant_public_keys.len());
|
let mut ecdh_commitments_xy = Vec::with_capacity(participant_public_keys.len());
|
||||||
for participant_public_key in participant_public_keys {
|
for participant_public_key in participant_public_keys {
|
||||||
let mut ecdh_commitments_xy_i =
|
let mut ecdh_commitments_xy_i = [(
|
||||||
[(<C::ToweringCurve as Ciphersuite>::F::ZERO, <C::ToweringCurve as Ciphersuite>::F::ZERO);
|
<C::ToweringCurve as WrappedGroup>::F::ZERO,
|
||||||
2];
|
<C::ToweringCurve as WrappedGroup>::F::ZERO,
|
||||||
let mut encryption_key = Zeroizing::new(<C::ToweringCurve as Ciphersuite>::F::ZERO);
|
); 2];
|
||||||
|
let mut encryption_key = Zeroizing::new(<C::ToweringCurve as WrappedGroup>::F::ZERO);
|
||||||
for ecdh_commitments_xy_i_j_dest in &mut ecdh_commitments_xy_i {
|
for ecdh_commitments_xy_i_j_dest in &mut ecdh_commitments_xy_i {
|
||||||
let mut ecdh_ephemeral_secret;
|
let mut ecdh_ephemeral_secret;
|
||||||
loop {
|
loop {
|
||||||
ecdh_ephemeral_secret =
|
ecdh_ephemeral_secret =
|
||||||
Zeroizing::new(<C::EmbeddedCurve as Ciphersuite>::F::random(&mut *rng));
|
Zeroizing::new(<C::EmbeddedCurve as WrappedGroup>::F::random(&mut *rng));
|
||||||
// 0 would produce the identity, which isn't representable within the discrete-log proof.
|
// 0 would produce the identity, which isn't representable within the discrete-log proof.
|
||||||
if bool::from(!ecdh_ephemeral_secret.is_zero()) {
|
if bool::from(!ecdh_ephemeral_secret.is_zero()) {
|
||||||
break;
|
break;
|
||||||
@@ -434,7 +437,7 @@ impl<C: Curves> Proof<C> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
let ecdh_ephemeral_secret =
|
let ecdh_ephemeral_secret =
|
||||||
ScalarDecomposition::<<C::EmbeddedCurve as Ciphersuite>::F>::new(*ecdh_ephemeral_secret)
|
ScalarDecomposition::<<C::EmbeddedCurve as WrappedGroup>::F>::new(*ecdh_ephemeral_secret)
|
||||||
.expect("ECDH ephemeral secret zero");
|
.expect("ECDH ephemeral secret zero");
|
||||||
discrete_log(&mut vector_commitment_tape, &ecdh_ephemeral_secret);
|
discrete_log(&mut vector_commitment_tape, &ecdh_ephemeral_secret);
|
||||||
|
|
||||||
@@ -442,7 +445,7 @@ impl<C: Curves> Proof<C> {
|
|||||||
let (ecdh_commitment, ecdh_commitment_xy_i_j) = discrete_log_claim(
|
let (ecdh_commitment, ecdh_commitment_xy_i_j) = discrete_log_claim(
|
||||||
&mut vector_commitment_tape,
|
&mut vector_commitment_tape,
|
||||||
&ecdh_ephemeral_secret,
|
&ecdh_ephemeral_secret,
|
||||||
<<C as Curves>::EmbeddedCurve as Ciphersuite>::generator(),
|
<<C as Curves>::EmbeddedCurve as WrappedGroup>::generator(),
|
||||||
);
|
);
|
||||||
ecdh_commitments.push(ecdh_commitment);
|
ecdh_commitments.push(ecdh_commitment);
|
||||||
*ecdh_commitments_xy_i_j_dest = ecdh_commitment_xy_i_j;
|
*ecdh_commitments_xy_i_j_dest = ecdh_commitment_xy_i_j;
|
||||||
@@ -470,7 +473,7 @@ impl<C: Curves> Proof<C> {
|
|||||||
for chunk in vector_commitment_tape.chunks(generators_to_use) {
|
for chunk in vector_commitment_tape.chunks(generators_to_use) {
|
||||||
vector_commitments.push(PedersenVectorCommitment {
|
vector_commitments.push(PedersenVectorCommitment {
|
||||||
g_values: chunk.into(),
|
g_values: chunk.into(),
|
||||||
mask: <C::ToweringCurve as Ciphersuite>::F::random(&mut *rng),
|
mask: <C::ToweringCurve as WrappedGroup>::F::random(&mut *rng),
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -479,13 +482,13 @@ impl<C: Curves> Proof<C> {
|
|||||||
for coefficient in &coefficients {
|
for coefficient in &coefficients {
|
||||||
commitments.push(PedersenCommitment {
|
commitments.push(PedersenCommitment {
|
||||||
value: **coefficient,
|
value: **coefficient,
|
||||||
mask: <C::ToweringCurve as Ciphersuite>::F::random(&mut *rng),
|
mask: <C::ToweringCurve as WrappedGroup>::F::random(&mut *rng),
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
for enc_mask in &encryption_keys {
|
for enc_mask in &encryption_keys {
|
||||||
commitments.push(PedersenCommitment {
|
commitments.push(PedersenCommitment {
|
||||||
value: **enc_mask,
|
value: **enc_mask,
|
||||||
mask: <C::ToweringCurve as Ciphersuite>::F::random(&mut *rng),
|
mask: <C::ToweringCurve as WrappedGroup>::F::random(&mut *rng),
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -536,13 +539,13 @@ impl<C: Curves> Proof<C> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Prove the openings of the commitments were correct
|
// Prove the openings of the commitments were correct
|
||||||
let mut x = Zeroizing::new(<C::ToweringCurve as Ciphersuite>::F::ZERO);
|
let mut x = Zeroizing::new(<C::ToweringCurve as WrappedGroup>::F::ZERO);
|
||||||
for commitment in commitments {
|
for commitment in commitments {
|
||||||
*x += commitment.mask * transcript.challenge::<C::ToweringCurve>();
|
*x += commitment.mask * transcript.challenge::<C::ToweringCurve>();
|
||||||
}
|
}
|
||||||
|
|
||||||
// Produce a Schnorr PoK for the weighted-sum of the Pedersen commitments' blinding factors
|
// Produce a Schnorr PoK for the weighted-sum of the Pedersen commitments' blinding factors
|
||||||
let r = Zeroizing::new(<C::ToweringCurve as Ciphersuite>::F::random(&mut *rng));
|
let r = Zeroizing::new(<C::ToweringCurve as WrappedGroup>::F::random(&mut *rng));
|
||||||
transcript.push_point(&(generators.h() * r.deref()));
|
transcript.push_point(&(generators.h() * r.deref()));
|
||||||
let c = transcript.challenge::<C::ToweringCurve>();
|
let c = transcript.challenge::<C::ToweringCurve>();
|
||||||
transcript.push_scalar((c * x.deref()) + r.deref());
|
transcript.push_scalar((c * x.deref()) + r.deref());
|
||||||
@@ -557,14 +560,14 @@ impl<C: Curves> Proof<C> {
|
|||||||
verifier: &mut BatchVerifier<C::ToweringCurve>,
|
verifier: &mut BatchVerifier<C::ToweringCurve>,
|
||||||
transcript: [u8; 32],
|
transcript: [u8; 32],
|
||||||
coefficients: usize,
|
coefficients: usize,
|
||||||
participant_public_keys: &[<<C as Curves>::EmbeddedCurve as Ciphersuite>::G],
|
participant_public_keys: &[<<C as Curves>::EmbeddedCurve as WrappedGroup>::G],
|
||||||
evrf_public_key: <<C as Curves>::EmbeddedCurve as Ciphersuite>::G,
|
evrf_public_key: <<C as Curves>::EmbeddedCurve as WrappedGroup>::G,
|
||||||
proof: &[u8],
|
proof: &[u8],
|
||||||
) -> Result<Verified<C>, ()> {
|
) -> Result<Verified<C>, ()> {
|
||||||
let (mut transcript, ecdh_commitments, pedersen_commitments) = {
|
let (mut transcript, ecdh_commitments, pedersen_commitments) = {
|
||||||
let curve_spec = CurveSpec {
|
let curve_spec = CurveSpec {
|
||||||
a: <<C as Curves>::EmbeddedCurve as Ciphersuite>::G::a(),
|
a: <<C as Curves>::EmbeddedCurve as WrappedGroup>::G::a(),
|
||||||
b: <<C as Curves>::EmbeddedCurve as Ciphersuite>::G::b(),
|
b: <<C as Curves>::EmbeddedCurve as WrappedGroup>::G::b(),
|
||||||
};
|
};
|
||||||
|
|
||||||
let coefficients_evrf_points =
|
let coefficients_evrf_points =
|
||||||
@@ -600,9 +603,9 @@ impl<C: Curves> Proof<C> {
|
|||||||
ecdh_commitments.push(ecdh_commitments_i);
|
ecdh_commitments.push(ecdh_commitments_i);
|
||||||
// This inherently bans using the identity point, as it won't have an affine representation
|
// This inherently bans using the identity point, as it won't have an affine representation
|
||||||
ecdh_commitments_xy.push([
|
ecdh_commitments_xy.push([
|
||||||
<<C::EmbeddedCurve as Ciphersuite>::G as DivisorCurve>::to_xy(ecdh_commitments_i[0])
|
<<C::EmbeddedCurve as WrappedGroup>::G as DivisorCurve>::to_xy(ecdh_commitments_i[0])
|
||||||
.ok_or(())?,
|
.ok_or(())?,
|
||||||
<<C::EmbeddedCurve as Ciphersuite>::G as DivisorCurve>::to_xy(ecdh_commitments_i[1])
|
<<C::EmbeddedCurve as WrappedGroup>::G as DivisorCurve>::to_xy(ecdh_commitments_i[1])
|
||||||
.ok_or(())?,
|
.ok_or(())?,
|
||||||
]);
|
]);
|
||||||
}
|
}
|
||||||
@@ -610,7 +613,7 @@ impl<C: Curves> Proof<C> {
|
|||||||
let mut circuit = BpCircuit::verify();
|
let mut circuit = BpCircuit::verify();
|
||||||
Self::circuit(
|
Self::circuit(
|
||||||
&curve_spec,
|
&curve_spec,
|
||||||
<C::EmbeddedCurve as Ciphersuite>::G::to_xy(evrf_public_key).ok_or(())?,
|
<C::EmbeddedCurve as WrappedGroup>::G::to_xy(evrf_public_key).ok_or(())?,
|
||||||
coefficients,
|
coefficients,
|
||||||
&ecdh_commitments_xy,
|
&ecdh_commitments_xy,
|
||||||
&generator_tables.iter().collect::<Vec<_>>(),
|
&generator_tables.iter().collect::<Vec<_>>(),
|
||||||
|
|||||||
@@ -4,11 +4,11 @@ use zeroize::Zeroizing;
|
|||||||
use rand_core::OsRng;
|
use rand_core::OsRng;
|
||||||
use rand::seq::SliceRandom;
|
use rand::seq::SliceRandom;
|
||||||
|
|
||||||
use ciphersuite::{group::ff::Field, Ciphersuite};
|
use ciphersuite::{group::ff::Field, WrappedGroup};
|
||||||
use embedwards25519::Embedwards25519;
|
use embedwards25519::Embedwards25519;
|
||||||
|
|
||||||
use dkg_recovery::recover_key;
|
use dkg_recovery::recover_key;
|
||||||
use crate::{Participant, Curves, Generators, VerifyResult, Dkg, Ristretto};
|
use crate::{Participant, Curves, Generators, VerifyResult, Dkg, Ed25519};
|
||||||
|
|
||||||
mod proof;
|
mod proof;
|
||||||
|
|
||||||
@@ -17,14 +17,14 @@ const PARTICIPANTS: u16 = 5;
|
|||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn dkg() {
|
fn dkg() {
|
||||||
let generators = Generators::<Ristretto>::new(THRESHOLD, PARTICIPANTS);
|
let generators = Generators::<Ed25519>::new(THRESHOLD, PARTICIPANTS);
|
||||||
let context = [0; 32];
|
let context = [0; 32];
|
||||||
|
|
||||||
let mut priv_keys = vec![];
|
let mut priv_keys = vec![];
|
||||||
let mut pub_keys = vec![];
|
let mut pub_keys = vec![];
|
||||||
for i in 0 .. PARTICIPANTS {
|
for i in 0 .. PARTICIPANTS {
|
||||||
let priv_key = <Embedwards25519 as Ciphersuite>::F::random(&mut OsRng);
|
let priv_key = <Embedwards25519 as WrappedGroup>::F::random(&mut OsRng);
|
||||||
pub_keys.push(<Embedwards25519 as Ciphersuite>::generator() * priv_key);
|
pub_keys.push(<Embedwards25519 as WrappedGroup>::generator() * priv_key);
|
||||||
priv_keys.push((Participant::new(1 + i).unwrap(), Zeroizing::new(priv_key)));
|
priv_keys.push((Participant::new(1 + i).unwrap(), Zeroizing::new(priv_key)));
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -34,27 +34,15 @@ fn dkg() {
|
|||||||
for (i, priv_key) in priv_keys.iter().take(usize::from(THRESHOLD)) {
|
for (i, priv_key) in priv_keys.iter().take(usize::from(THRESHOLD)) {
|
||||||
participations.insert(
|
participations.insert(
|
||||||
*i,
|
*i,
|
||||||
Dkg::<Ristretto>::participate(
|
Dkg::<Ed25519>::participate(&mut OsRng, &generators, context, THRESHOLD, &pub_keys, priv_key)
|
||||||
&mut OsRng,
|
.unwrap(),
|
||||||
&generators,
|
|
||||||
context,
|
|
||||||
THRESHOLD,
|
|
||||||
&pub_keys,
|
|
||||||
priv_key,
|
|
||||||
)
|
|
||||||
.unwrap(),
|
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
let VerifyResult::Valid(dkg) = Dkg::<Ristretto>::verify(
|
let VerifyResult::Valid(dkg) =
|
||||||
&mut OsRng,
|
Dkg::<Ed25519>::verify(&mut OsRng, &generators, context, THRESHOLD, &pub_keys, &participations)
|
||||||
&generators,
|
.unwrap()
|
||||||
context,
|
else {
|
||||||
THRESHOLD,
|
|
||||||
&pub_keys,
|
|
||||||
&participations,
|
|
||||||
)
|
|
||||||
.unwrap() else {
|
|
||||||
panic!("verify didn't return VerifyResult::Valid")
|
panic!("verify didn't return VerifyResult::Valid")
|
||||||
};
|
};
|
||||||
|
|
||||||
@@ -80,7 +68,7 @@ fn dkg() {
|
|||||||
|
|
||||||
// TODO: Test for all possible combinations of keys
|
// TODO: Test for all possible combinations of keys
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
<<Ristretto as Curves>::ToweringCurve as Ciphersuite>::generator() *
|
<<Ed25519 as Curves>::ToweringCurve as WrappedGroup>::generator() *
|
||||||
*recover_key(&all_keys.values().cloned().collect::<Vec<_>>()).unwrap(),
|
*recover_key(&all_keys.values().cloned().collect::<Vec<_>>()).unwrap(),
|
||||||
group_key.unwrap()
|
group_key.unwrap()
|
||||||
);
|
);
|
||||||
|
|||||||
@@ -6,13 +6,13 @@ use zeroize::Zeroizing;
|
|||||||
|
|
||||||
use ciphersuite::{
|
use ciphersuite::{
|
||||||
group::{ff::Field, Group},
|
group::{ff::Field, Group},
|
||||||
Ciphersuite,
|
WrappedGroup,
|
||||||
};
|
};
|
||||||
|
|
||||||
use generalized_bulletproofs::{Generators, tests::insecure_test_generators};
|
use generalized_bulletproofs::{Generators, tests::insecure_test_generators};
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
Curves, Ristretto,
|
Curves, Ed25519,
|
||||||
proof::*,
|
proof::*,
|
||||||
tests::{THRESHOLD, PARTICIPANTS},
|
tests::{THRESHOLD, PARTICIPANTS},
|
||||||
};
|
};
|
||||||
@@ -20,9 +20,9 @@ use crate::{
|
|||||||
fn proof<C: Curves>() {
|
fn proof<C: Curves>() {
|
||||||
let generators = insecure_test_generators(&mut OsRng, 2048).unwrap();
|
let generators = insecure_test_generators(&mut OsRng, 2048).unwrap();
|
||||||
let embedded_private_key =
|
let embedded_private_key =
|
||||||
Zeroizing::new(<C::EmbeddedCurve as Ciphersuite>::F::random(&mut OsRng));
|
Zeroizing::new(<C::EmbeddedCurve as WrappedGroup>::F::random(&mut OsRng));
|
||||||
let ecdh_public_keys: [_; PARTICIPANTS as usize] =
|
let ecdh_public_keys: [_; PARTICIPANTS as usize] =
|
||||||
core::array::from_fn(|_| <C::EmbeddedCurve as Ciphersuite>::G::random(&mut OsRng));
|
core::array::from_fn(|_| <C::EmbeddedCurve as WrappedGroup>::G::random(&mut OsRng));
|
||||||
let time = Instant::now();
|
let time = Instant::now();
|
||||||
let res = Proof::<C>::prove(
|
let res = Proof::<C>::prove(
|
||||||
&mut OsRng,
|
&mut OsRng,
|
||||||
@@ -54,5 +54,5 @@ fn proof<C: Curves>() {
|
|||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn ristretto_proof() {
|
fn ristretto_proof() {
|
||||||
proof::<Ristretto>();
|
proof::<Ed25519>();
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -5,7 +5,7 @@ use rand_core::{RngCore, CryptoRng};
|
|||||||
|
|
||||||
use ciphersuite::{
|
use ciphersuite::{
|
||||||
group::{ff::PrimeField, Group, GroupEncoding},
|
group::{ff::PrimeField, Group, GroupEncoding},
|
||||||
Ciphersuite,
|
GroupIo,
|
||||||
};
|
};
|
||||||
|
|
||||||
use dkg::Participant;
|
use dkg::Participant;
|
||||||
@@ -13,7 +13,7 @@ use dkg::Participant;
|
|||||||
/// Sample a random, unbiased point on the elliptic curve with an unknown discrete logarithm.
|
/// Sample a random, unbiased point on the elliptic curve with an unknown discrete logarithm.
|
||||||
///
|
///
|
||||||
/// This keeps it simple by using rejection sampling.
|
/// This keeps it simple by using rejection sampling.
|
||||||
pub(crate) fn sample_point<C: Ciphersuite>(rng: &mut (impl RngCore + CryptoRng)) -> C::G {
|
pub(crate) fn sample_point<C: GroupIo>(rng: &mut (impl RngCore + CryptoRng)) -> C::G {
|
||||||
let mut repr = <C::G as GroupEncoding>::Repr::default();
|
let mut repr = <C::G as GroupEncoding>::Repr::default();
|
||||||
loop {
|
loop {
|
||||||
rng.fill_bytes(repr.as_mut());
|
rng.fill_bytes(repr.as_mut());
|
||||||
|
|||||||
@@ -23,7 +23,7 @@ rand_core = { version = "0.6", default-features = false }
|
|||||||
|
|
||||||
zeroize = { version = "^1.5", default-features = false, features = ["zeroize_derive"] }
|
zeroize = { version = "^1.5", default-features = false, features = ["zeroize_derive"] }
|
||||||
|
|
||||||
std-shims = { version = "0.1", path = "../../../common/std-shims", default-features = false }
|
std-shims = { version = "0.1", path = "../../../common/std-shims", default-features = false, features = ["alloc"] }
|
||||||
|
|
||||||
multiexp = { path = "../../multiexp", version = "0.4", default-features = false }
|
multiexp = { path = "../../multiexp", version = "0.4", default-features = false }
|
||||||
ciphersuite = { path = "../../ciphersuite", version = "^0.4.1", default-features = false }
|
ciphersuite = { path = "../../ciphersuite", version = "^0.4.1", default-features = false }
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#![cfg_attr(docsrs, feature(doc_auto_cfg))]
|
#![cfg_attr(docsrs, feature(doc_cfg))]
|
||||||
#![doc = include_str!("../README.md")]
|
#![doc = include_str!("../README.md")]
|
||||||
#![cfg_attr(not(feature = "std"), no_std)]
|
#![cfg_attr(not(feature = "std"), no_std)]
|
||||||
|
|
||||||
@@ -150,7 +150,7 @@ pub fn musig<C: Ciphersuite>(
|
|||||||
}
|
}
|
||||||
let group_key = multiexp::multiexp(&multiexp);
|
let group_key = multiexp::multiexp(&multiexp);
|
||||||
debug_assert_eq!(our_pub_key, verification_shares[¶ms.i()]);
|
debug_assert_eq!(our_pub_key, verification_shares[¶ms.i()]);
|
||||||
debug_assert_eq!(musig_key_vartime::<C>(context, keys).unwrap(), group_key);
|
debug_assert_eq!(musig_key_vartime::<C>(context, keys), Ok(group_key));
|
||||||
|
|
||||||
ThresholdKeys::new(
|
ThresholdKeys::new(
|
||||||
params,
|
params,
|
||||||
|
|||||||
@@ -4,7 +4,7 @@ use zeroize::Zeroizing;
|
|||||||
use rand_core::OsRng;
|
use rand_core::OsRng;
|
||||||
|
|
||||||
use dalek_ff_group::Ristretto;
|
use dalek_ff_group::Ristretto;
|
||||||
use ciphersuite::{group::ff::Field, Ciphersuite};
|
use ciphersuite::WrappedGroup;
|
||||||
|
|
||||||
use dkg_recovery::recover_key;
|
use dkg_recovery::recover_key;
|
||||||
use crate::*;
|
use crate::*;
|
||||||
@@ -17,21 +17,21 @@ pub fn test_musig() {
|
|||||||
let mut keys = vec![];
|
let mut keys = vec![];
|
||||||
let mut pub_keys = vec![];
|
let mut pub_keys = vec![];
|
||||||
for _ in 0 .. PARTICIPANTS {
|
for _ in 0 .. PARTICIPANTS {
|
||||||
let key = Zeroizing::new(<Ristretto as Ciphersuite>::F::random(&mut OsRng));
|
let key = Zeroizing::new(<Ristretto as WrappedGroup>::F::random(&mut OsRng));
|
||||||
pub_keys.push(<Ristretto as Ciphersuite>::generator() * *key);
|
pub_keys.push(<Ristretto as WrappedGroup>::generator() * *key);
|
||||||
keys.push(key);
|
keys.push(key);
|
||||||
}
|
}
|
||||||
|
|
||||||
const CONTEXT: [u8; 32] = *b"MuSig Test ";
|
const CONTEXT: [u8; 32] = *b"MuSig Test ";
|
||||||
|
|
||||||
// Empty signing set
|
// Empty signing set
|
||||||
musig::<Ristretto>(CONTEXT, Zeroizing::new(<Ristretto as Ciphersuite>::F::ZERO), &[])
|
musig::<Ristretto>(CONTEXT, Zeroizing::new(<Ristretto as WrappedGroup>::F::ZERO), &[])
|
||||||
.unwrap_err();
|
.unwrap_err();
|
||||||
// Signing set we're not part of
|
// Signing set we're not part of
|
||||||
musig::<Ristretto>(
|
musig::<Ristretto>(
|
||||||
CONTEXT,
|
CONTEXT,
|
||||||
Zeroizing::new(<Ristretto as Ciphersuite>::F::ZERO),
|
Zeroizing::new(<Ristretto as WrappedGroup>::F::ZERO),
|
||||||
&[<Ristretto as Ciphersuite>::generator()],
|
&[<Ristretto as WrappedGroup>::generator()],
|
||||||
)
|
)
|
||||||
.unwrap_err();
|
.unwrap_err();
|
||||||
|
|
||||||
@@ -48,7 +48,7 @@ pub fn test_musig() {
|
|||||||
|
|
||||||
verification_shares.insert(
|
verification_shares.insert(
|
||||||
these_keys.params().i(),
|
these_keys.params().i(),
|
||||||
<Ristretto as Ciphersuite>::generator() * **these_keys.original_secret_share(),
|
<Ristretto as WrappedGroup>::generator() * **these_keys.original_secret_share(),
|
||||||
);
|
);
|
||||||
|
|
||||||
assert_eq!(these_keys.group_key(), group_key);
|
assert_eq!(these_keys.group_key(), group_key);
|
||||||
@@ -63,7 +63,7 @@ pub fn test_musig() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
<Ristretto as Ciphersuite>::generator() *
|
<Ristretto as WrappedGroup>::generator() *
|
||||||
*recover_key(&created_keys.values().cloned().collect::<Vec<_>>()).unwrap(),
|
*recover_key(&created_keys.values().cloned().collect::<Vec<_>>()).unwrap(),
|
||||||
group_key
|
group_key
|
||||||
);
|
);
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#![cfg_attr(docsrs, feature(doc_auto_cfg))]
|
#![cfg_attr(docsrs, feature(doc_cfg))]
|
||||||
#![doc = include_str!("../README.md")]
|
#![doc = include_str!("../README.md")]
|
||||||
#![no_std]
|
#![no_std]
|
||||||
|
|
||||||
@@ -8,7 +8,7 @@ use alloc::vec::Vec;
|
|||||||
|
|
||||||
use zeroize::Zeroizing;
|
use zeroize::Zeroizing;
|
||||||
|
|
||||||
use ciphersuite::Ciphersuite;
|
use ciphersuite::{GroupIo, Id};
|
||||||
|
|
||||||
pub use dkg::*;
|
pub use dkg::*;
|
||||||
|
|
||||||
@@ -34,7 +34,7 @@ pub enum RecoveryError {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/// Recover a shared secret from a collection of `dkg::ThresholdKeys`.
|
/// Recover a shared secret from a collection of `dkg::ThresholdKeys`.
|
||||||
pub fn recover_key<C: Ciphersuite>(
|
pub fn recover_key<C: GroupIo + Id>(
|
||||||
keys: &[ThresholdKeys<C>],
|
keys: &[ThresholdKeys<C>],
|
||||||
) -> Result<Zeroizing<C::F>, RecoveryError> {
|
) -> Result<Zeroizing<C::F>, RecoveryError> {
|
||||||
let included = keys.iter().map(|keys| keys.params().i()).collect::<Vec<_>>();
|
let included = keys.iter().map(|keys| keys.params().i()).collect::<Vec<_>>();
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#![cfg_attr(docsrs, feature(doc_auto_cfg))]
|
#![cfg_attr(docsrs, feature(doc_cfg))]
|
||||||
#![doc = include_str!("../README.md")]
|
#![doc = include_str!("../README.md")]
|
||||||
#![cfg_attr(not(feature = "std"), no_std)]
|
#![cfg_attr(not(feature = "std"), no_std)]
|
||||||
|
|
||||||
@@ -17,12 +17,11 @@ use ciphersuite::{
|
|||||||
ff::{Field, PrimeField},
|
ff::{Field, PrimeField},
|
||||||
GroupEncoding,
|
GroupEncoding,
|
||||||
},
|
},
|
||||||
Ciphersuite,
|
GroupIo, Id,
|
||||||
};
|
};
|
||||||
|
|
||||||
/// The ID of a participant, defined as a non-zero u16.
|
/// The ID of a participant, defined as a non-zero u16.
|
||||||
#[derive(Clone, Copy, PartialEq, Eq, PartialOrd, Ord, Hash, Debug, Zeroize)]
|
#[derive(Clone, Copy, PartialEq, Eq, PartialOrd, Ord, Hash, Debug, Zeroize)]
|
||||||
#[cfg_attr(feature = "borsh", derive(borsh::BorshSerialize))]
|
|
||||||
pub struct Participant(u16);
|
pub struct Participant(u16);
|
||||||
impl Participant {
|
impl Participant {
|
||||||
/// Create a new Participant identifier from a u16.
|
/// Create a new Participant identifier from a u16.
|
||||||
@@ -129,18 +128,8 @@ pub enum DkgError {
|
|||||||
NotParticipating,
|
NotParticipating,
|
||||||
}
|
}
|
||||||
|
|
||||||
// Manually implements BorshDeserialize so we can enforce it's a valid index
|
|
||||||
#[cfg(feature = "borsh")]
|
|
||||||
impl borsh::BorshDeserialize for Participant {
|
|
||||||
fn deserialize_reader<R: io::Read>(reader: &mut R) -> io::Result<Self> {
|
|
||||||
Participant::new(u16::deserialize_reader(reader)?)
|
|
||||||
.ok_or_else(|| io::Error::other("invalid participant"))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Parameters for a multisig.
|
/// Parameters for a multisig.
|
||||||
#[derive(Clone, Copy, PartialEq, Eq, Debug, Zeroize)]
|
#[derive(Clone, Copy, PartialEq, Eq, Debug, Zeroize)]
|
||||||
#[cfg_attr(feature = "borsh", derive(borsh::BorshSerialize))]
|
|
||||||
pub struct ThresholdParams {
|
pub struct ThresholdParams {
|
||||||
/// Participants needed to sign on behalf of the group.
|
/// Participants needed to sign on behalf of the group.
|
||||||
t: u16,
|
t: u16,
|
||||||
@@ -210,16 +199,6 @@ impl ThresholdParams {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(feature = "borsh")]
|
|
||||||
impl borsh::BorshDeserialize for ThresholdParams {
|
|
||||||
fn deserialize_reader<R: io::Read>(reader: &mut R) -> io::Result<Self> {
|
|
||||||
let t = u16::deserialize_reader(reader)?;
|
|
||||||
let n = u16::deserialize_reader(reader)?;
|
|
||||||
let i = Participant::deserialize_reader(reader)?;
|
|
||||||
ThresholdParams::new(t, n, i).map_err(|e| io::Error::other(format!("{e:?}")))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// A method of interpolation.
|
/// A method of interpolation.
|
||||||
#[derive(Clone, PartialEq, Eq, Debug, Zeroize)]
|
#[derive(Clone, PartialEq, Eq, Debug, Zeroize)]
|
||||||
pub enum Interpolation<F: Zeroize + PrimeField> {
|
pub enum Interpolation<F: Zeroize + PrimeField> {
|
||||||
@@ -268,7 +247,7 @@ impl<F: Zeroize + PrimeField> Interpolation<F> {
|
|||||||
/// heap-allocated pointer to minimize copies on the stack (`ThresholdKeys`, the publicly exposed
|
/// heap-allocated pointer to minimize copies on the stack (`ThresholdKeys`, the publicly exposed
|
||||||
/// type).
|
/// type).
|
||||||
#[derive(Clone, PartialEq, Eq)]
|
#[derive(Clone, PartialEq, Eq)]
|
||||||
struct ThresholdCore<C: Ciphersuite> {
|
struct ThresholdCore<C: GroupIo + Id> {
|
||||||
params: ThresholdParams,
|
params: ThresholdParams,
|
||||||
group_key: C::G,
|
group_key: C::G,
|
||||||
verification_shares: HashMap<Participant, C::G>,
|
verification_shares: HashMap<Participant, C::G>,
|
||||||
@@ -276,7 +255,7 @@ struct ThresholdCore<C: Ciphersuite> {
|
|||||||
secret_share: Zeroizing<C::F>,
|
secret_share: Zeroizing<C::F>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<C: Ciphersuite> fmt::Debug for ThresholdCore<C> {
|
impl<C: GroupIo + Id> fmt::Debug for ThresholdCore<C> {
|
||||||
fn fmt(&self, fmt: &mut fmt::Formatter<'_>) -> fmt::Result {
|
fn fmt(&self, fmt: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||||
fmt
|
fmt
|
||||||
.debug_struct("ThresholdCore")
|
.debug_struct("ThresholdCore")
|
||||||
@@ -288,7 +267,7 @@ impl<C: Ciphersuite> fmt::Debug for ThresholdCore<C> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<C: Ciphersuite> Zeroize for ThresholdCore<C> {
|
impl<C: GroupIo + Id> Zeroize for ThresholdCore<C> {
|
||||||
fn zeroize(&mut self) {
|
fn zeroize(&mut self) {
|
||||||
self.params.zeroize();
|
self.params.zeroize();
|
||||||
self.group_key.zeroize();
|
self.group_key.zeroize();
|
||||||
@@ -302,7 +281,7 @@ impl<C: Ciphersuite> Zeroize for ThresholdCore<C> {
|
|||||||
|
|
||||||
/// Threshold keys usable for signing.
|
/// Threshold keys usable for signing.
|
||||||
#[derive(Clone, Debug, Zeroize)]
|
#[derive(Clone, Debug, Zeroize)]
|
||||||
pub struct ThresholdKeys<C: Ciphersuite> {
|
pub struct ThresholdKeys<C: GroupIo + Id> {
|
||||||
// Core keys.
|
// Core keys.
|
||||||
#[zeroize(skip)]
|
#[zeroize(skip)]
|
||||||
core: Arc<Zeroizing<ThresholdCore<C>>>,
|
core: Arc<Zeroizing<ThresholdCore<C>>>,
|
||||||
@@ -315,7 +294,7 @@ pub struct ThresholdKeys<C: Ciphersuite> {
|
|||||||
|
|
||||||
/// View of keys, interpolated and with the expected linear combination taken for usage.
|
/// View of keys, interpolated and with the expected linear combination taken for usage.
|
||||||
#[derive(Clone)]
|
#[derive(Clone)]
|
||||||
pub struct ThresholdView<C: Ciphersuite> {
|
pub struct ThresholdView<C: GroupIo + Id> {
|
||||||
interpolation: Interpolation<C::F>,
|
interpolation: Interpolation<C::F>,
|
||||||
scalar: C::F,
|
scalar: C::F,
|
||||||
offset: C::F,
|
offset: C::F,
|
||||||
@@ -326,7 +305,7 @@ pub struct ThresholdView<C: Ciphersuite> {
|
|||||||
verification_shares: HashMap<Participant, C::G>,
|
verification_shares: HashMap<Participant, C::G>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<C: Ciphersuite> fmt::Debug for ThresholdView<C> {
|
impl<C: GroupIo + Id> fmt::Debug for ThresholdView<C> {
|
||||||
fn fmt(&self, fmt: &mut fmt::Formatter<'_>) -> fmt::Result {
|
fn fmt(&self, fmt: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||||
fmt
|
fmt
|
||||||
.debug_struct("ThresholdView")
|
.debug_struct("ThresholdView")
|
||||||
@@ -341,7 +320,7 @@ impl<C: Ciphersuite> fmt::Debug for ThresholdView<C> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<C: Ciphersuite> Zeroize for ThresholdView<C> {
|
impl<C: GroupIo + Id> Zeroize for ThresholdView<C> {
|
||||||
fn zeroize(&mut self) {
|
fn zeroize(&mut self) {
|
||||||
self.scalar.zeroize();
|
self.scalar.zeroize();
|
||||||
self.offset.zeroize();
|
self.offset.zeroize();
|
||||||
@@ -357,7 +336,7 @@ impl<C: Ciphersuite> Zeroize for ThresholdView<C> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<C: Ciphersuite> ThresholdKeys<C> {
|
impl<C: GroupIo + Id> ThresholdKeys<C> {
|
||||||
/// Create a new set of ThresholdKeys.
|
/// Create a new set of ThresholdKeys.
|
||||||
pub fn new(
|
pub fn new(
|
||||||
params: ThresholdParams,
|
params: ThresholdParams,
|
||||||
@@ -632,7 +611,7 @@ impl<C: Ciphersuite> ThresholdKeys<C> {
|
|||||||
|
|
||||||
let mut verification_shares = HashMap::new();
|
let mut verification_shares = HashMap::new();
|
||||||
for l in (1 ..= n).map(Participant) {
|
for l in (1 ..= n).map(Participant) {
|
||||||
verification_shares.insert(l, <C as Ciphersuite>::read_G(reader)?);
|
verification_shares.insert(l, C::read_G(reader)?);
|
||||||
}
|
}
|
||||||
|
|
||||||
ThresholdKeys::new(
|
ThresholdKeys::new(
|
||||||
@@ -645,7 +624,7 @@ impl<C: Ciphersuite> ThresholdKeys<C> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<C: Ciphersuite> ThresholdView<C> {
|
impl<C: GroupIo + Id> ThresholdView<C> {
|
||||||
/// Return the scalar applied to this view.
|
/// Return the scalar applied to this view.
|
||||||
pub fn scalar(&self) -> C::F {
|
pub fn scalar(&self) -> C::F {
|
||||||
self.scalar
|
self.scalar
|
||||||
|
|||||||
@@ -19,8 +19,9 @@ workspace = true
|
|||||||
[dependencies]
|
[dependencies]
|
||||||
zeroize = { version = "1", default-features = false, features = ["zeroize_derive"] }
|
zeroize = { version = "1", default-features = false, features = ["zeroize_derive"] }
|
||||||
|
|
||||||
sha3 = { version = "0.11.0-rc.0", default-features = false }
|
sha3 = { version = "0.11.0-rc.2", default-features = false }
|
||||||
|
|
||||||
|
crypto-bigint = { version = "0.6", default-features = false, features = ["zeroize"] }
|
||||||
prime-field = { path = "../prime-field", default-features = false }
|
prime-field = { path = "../prime-field", default-features = false }
|
||||||
ciphersuite = { path = "../ciphersuite", default-features = false }
|
ciphersuite = { path = "../ciphersuite", default-features = false }
|
||||||
|
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
use zeroize::Zeroize;
|
use prime_field::subtle::CtOption;
|
||||||
|
|
||||||
use sha3::{
|
use sha3::{
|
||||||
digest::{
|
digest::{
|
||||||
@@ -8,9 +8,9 @@ use sha3::{
|
|||||||
Shake256,
|
Shake256,
|
||||||
};
|
};
|
||||||
|
|
||||||
use ciphersuite::{group::Group, Ciphersuite};
|
use ciphersuite::{group::GroupEncoding, Id, WithPreferredHash, GroupCanonicalEncoding};
|
||||||
|
|
||||||
use crate::{Scalar, Point};
|
use crate::Point;
|
||||||
|
|
||||||
/// Shake256, fixed to a 114-byte output, as used by Ed448.
|
/// Shake256, fixed to a 114-byte output, as used by Ed448.
|
||||||
#[derive(Clone, Default)]
|
#[derive(Clone, Default)]
|
||||||
@@ -49,21 +49,14 @@ impl FixedOutput for Shake256_114 {
|
|||||||
}
|
}
|
||||||
impl HashMarker for Shake256_114 {}
|
impl HashMarker for Shake256_114 {}
|
||||||
|
|
||||||
#[derive(Clone, Copy, PartialEq, Eq, Debug, Zeroize)]
|
impl Id for Point {
|
||||||
pub struct Ed448;
|
const ID: &[u8] = b"ed448";
|
||||||
impl Ciphersuite for Ed448 {
|
}
|
||||||
type F = Scalar;
|
impl WithPreferredHash for Point {
|
||||||
type G = Point;
|
|
||||||
type H = Shake256_114;
|
type H = Shake256_114;
|
||||||
|
}
|
||||||
const ID: &'static [u8] = b"ed448";
|
impl GroupCanonicalEncoding for Point {
|
||||||
|
fn from_canonical_bytes(bytes: &<Self::G as GroupEncoding>::Repr) -> CtOption<Self::G> {
|
||||||
fn generator() -> Self::G {
|
Self::G::from_bytes(bytes)
|
||||||
Point::generator()
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
|
||||||
fn test_ed448() {
|
|
||||||
ff_group_tests::group::test_prime_group_bits::<_, Point>(&mut rand_core::OsRng);
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#![cfg_attr(docsrs, feature(doc_auto_cfg))]
|
#![cfg_attr(docsrs, feature(doc_cfg))]
|
||||||
#![doc = include_str!("../README.md")]
|
#![doc = include_str!("../README.md")]
|
||||||
#![no_std]
|
#![no_std]
|
||||||
|
|
||||||
@@ -29,7 +29,6 @@ mod point;
|
|||||||
pub use point::Point;
|
pub use point::Point;
|
||||||
|
|
||||||
mod ciphersuite;
|
mod ciphersuite;
|
||||||
pub use crate::ciphersuite::Ed448;
|
|
||||||
|
|
||||||
pub(crate) fn u8_from_bool(bit_ref: &mut bool) -> u8 {
|
pub(crate) fn u8_from_bool(bit_ref: &mut bool) -> u8 {
|
||||||
use core::hint::black_box;
|
use core::hint::black_box;
|
||||||
|
|||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user