Project Overview
Aura is a fully peer-to-peer, private communication system that operates without dedicated servers. It uses a web-of-trust architecture to provide discovery, data availability, account recovery, and graceful async protocol evolution.
To accomplish this, Aura uses threshold cryptography so no single device holds complete keys. Network topology reflects social relationships, forming a web of trust that provides discovery, availability, and recovery. State converges through CRDT journals without central coordination. Session-typed choreographic protocols ensure safe multi-party execution.
How Aura Works
In Aura, all actors are authorities. An authority is an opaque cryptographic actor that may represent a person, a device group, or a shared context. External observers see only public keys and signed operations. This enables unlinkable participation across contexts.
State is append-only facts in journals. Each authority maintains its own journal. Shared contexts have journals written by multiple participants. Facts accumulate through CRDT merge and views are derived by reduction.
Side effects flow through explicit traits. Cryptography, storage, networking, and time are accessed only through effect handlers. This enables deterministic simulation and cross-platform portability.
Multi-party coordination uses session-typed choreographies. A global protocol specifies message flow. Each party's local behavior is projected from the global view.
Authorization passes through a layered guard chain. Before any message leaves, capabilities are verified, flow budgets are charged, and facts are committed atomically.
Aura separates key generation from agreement. Fast paths provide immediate usability while durable shared state is always consensus-finalized.
For the complete architecture, see System Architecture.
Documentation Index
The documents below cover theory, technical components, implementation guidance, and project organization.
1. Foundation
Theoretical Model establishes the formal calculus, algebraic types, and semilattice semantics underlying the system.
System Architecture describes the 8-layer architecture, effect patterns, and choreographic protocol structure.
Privacy and Information Flow Contract specifies consent-based privacy with trust boundaries, flow budgets, and leakage tracking.
Distributed Systems Contract defines safety and liveness guarantees, synchrony assumptions, and adversarial tolerance.
2. Core Systems
Cryptography documents primitives, key derivation, threshold signatures, and VSS schemes.
Identifiers and Boundaries defines the identifier types and their privacy-preserving properties.
Authority and Identity describes opaque authorities, commitment trees, and relational context structure.
Journal specifies fact-based journals, validation rules, and deterministic reduction flows.
Authorization covers capability semantics, Biscuit token integration, and guard chain authorization.
Effect System documents effect traits, handler design, and context propagation.
Runtime describes lifecycle management, guard chain execution, and service composition.
Consensus specifies single-shot agreement for non-monotone operations with witness attestation.
Operation Categories defines A/B/C operation tiers, K1/K2/K3 key generation, and agreement levels.
MPST and Choreography covers multi-party session types and choreographic protocol projection.
Transport and Information Flow specifies guard chain enforcement, secure channels, and flow receipts.
Aura Messaging Protocol (AMP) documents reliable async messaging with acknowledgment and ordering patterns.
Rendezvous Architecture covers context-scoped peer discovery and encrypted envelope exchange.
Relational Contexts specifies guardian bindings, recovery grants, and cross-authority journals.
Database Architecture defines the query layer using journals, Biscuit predicates, and CRDT views.
Social Architecture describes the three-tier model of messages, homes, and neighborhoods.
Distributed Maintenance Architecture covers snapshots, garbage collection, and system evolution.
CLI and Terminal User Interface specifies command-line and TUI interfaces for Aura operations.
Test Infrastructure Reference documents test fixtures, mock handlers, and scenario builders.
Simulation Infrastructure Reference covers deterministic simulation with virtual time and fault injection.
Formal Verification Reference describes Quint model checking and Lean theorem proving integration.
3. Developer Guides
Getting Started Guide provides a starting point for developers new to the codebase.
Effects and Handlers Guide covers the algebraic effect system, handler implementation, and platform support.
Choreography Development Guide explains choreographic protocol design, CRDTs, and distributed coordination.
Testing Guide covers test patterns, fixtures, conformance testing, and runtime harness.
Simulation Guide explains deterministic simulation for debugging and property verification.
Verification Guide documents Quint model checking and Lean proof workflows.
System Internals Guide covers guard chain internals, service patterns, and reactive scheduling.
Distributed Maintenance Guide covers operational concerns including snapshots and system upgrades.
4. Project Meta
UX Flow Coverage Report tracks harness and scenario coverage for user-visible flows.
Verification Coverage Report tracks formal verification status across Quint specs and Lean proofs.
Project Structure documents the 8-layer crate architecture and dependency relationships.
System Architecture
This document gives an intuitive overview of Aura's architecture. It covers the core abstractions, data flow patterns, and component interactions that define the system. Formal definitions live in Theoretical Model, an overview of the implementation can be found in Project Structure.
Overview
Aura is a peer-to-peer identity and communication system built on three pillars. Threshold cryptography distributes trust across multiple devices. Session-typed protocols ensure safe multi-party coordination. CRDT journals provide conflict-free replicated state.
The system operates without dedicated servers. Discovery, availability, and recovery are provided by the web of trust. Peers relay messages for each other based on social relationships. No single party can observe all communication or deny service.
Aura separates key generation from agreement. Fast paths provide immediate usability while durable shared state is always consensus-finalized.
flowchart TB
subgraph Authorities
A1[Authority A]
A2[Authority B]
end
subgraph State
J1[Journal A]
J2[Journal B]
JC[Context Journal]
end
subgraph Enforcement
GC[Guard Chain]
FB[Flow Budget]
end
subgraph Effects
EF[Effect System]
TR[Transport]
end
A1 --> J1
A2 --> J2
A1 & A2 --> JC
J1 & J2 & JC --> GC
GC --> FB
FB --> EF
EF --> TR
This diagram shows the primary data flow. Authorities own journals that store facts. The guard chain enforces authorization before any transport effect. The effect system provides the abstraction layer for all operations.
Every operation flows through the effect system. Every state change is replicated through journals. Every external action is authorized through guards. These three invariants define the architectural contract.
1. Dual Semilattice Model
Aura state consists of two complementary semilattices. Facts form a join-semilattice where information accumulates through the join operation. Capabilities form a meet-semilattice where authority restricts through the meet operation.
#![allow(unused)] fn main() { // Facts grow by join (⊔) struct Journal { facts: FactSet, // join-semilattice frontier: CapFrontier, // meet-semilattice } }
The Journal type keeps these dimensions separate. Facts can only grow. Capabilities can only shrink. This dual monotonicity provides convergence guarantees for replicated state.
Facts represent evidence that accumulates over time. Examples include signed operations, attestations, flow budget charges, and consensus commits. Once a fact is added, it cannot be removed. Garbage collection uses tombstones and reduction rather than deletion.
Capabilities represent authority that restricts over time. The system evaluates Biscuit tokens against policy to derive the current capability frontier. Delegation can only attenuate. No operation can widen capability scope. See Theoretical Model for formal definitions of these lattices.
2. Authority and Identity Architecture
An authority is an opaque cryptographic actor. External parties see only public keys and signed facts. Internal device structure is hidden. This abstraction provides unlinkability across contexts.
flowchart LR
subgraph External View
PK[Threshold Public Key]
end
subgraph Authority [Internal Structure]
direction TB
CT[Commitment Tree]
subgraph Devices
direction LR
D1[Device 1<br/>Share 1]
D2[Device 2<br/>Share 2]
D3[Device 3<br/>Share 3]
end
CT --> Devices
end
Devices -.->|2-of-3 signing| PK
style Authority fill:transparent,stroke:#888,stroke-dasharray: 5 5
This diagram shows authority structure. Externally, observers see only the threshold public key. Internally, the commitment tree tracks device membership. Each device holds a key share. Threshold signing combines shares without exposing internal structure.
2.1 Account authorities
Account authorities maintain device membership using commitment trees. The journal stores signed tree operations as facts. Reduction reconstructs the canonical tree state from accumulated facts.
#![allow(unused)] fn main() { pub struct AuthorityId(Uuid); }
AuthorityId is an opaque identifier for the authority namespace. It does not encode membership or reveal device count. Key derivation utilities provide context-scoped keys without exposing internal structure. See Identifiers and Boundaries for identifier semantics.
2.2 Relational contexts
Relational contexts are shared journals for cross-authority state. Each context has its own namespace and does not reveal participants to external observers.
#![allow(unused)] fn main() { pub struct ContextId(Uuid); }
ContextId identifies the shared namespace. Participation is expressed by writing relational facts. Profile data, nicknames, and relationship state live in context journals. See Authority and Identity for commitment tree details and Relational Contexts for context patterns.
2.3 Contextual identity
Identity is scoped to contexts. A device can participate in many contexts without linking them. Each context derives independent keys through deterministic key derivation. This prevents cross-context correlation by external observers.
3. Journal and State Reduction
The journal is the canonical state mechanism. All durable state is represented as facts in journals. Views are derived by reducing accumulated facts.
3.1 Journal namespaces
Journals are partitioned into namespaces. Authority namespaces store facts owned by a single authority. Context namespaces store facts shared across authorities participating in a relational context.
#![allow(unused)] fn main() { enum JournalNamespace { Authority(AuthorityId), Context(ContextId), } }
Namespace scoping provides isolation. Facts in one namespace cannot reference or affect facts in another namespace. Cross-namespace coordination requires explicit protocols.
3.2 Fact model
Facts are content-addressed immutable records. Each fact includes a type identifier, payload, attestation, and metadata. Facts are validated against type-specific rules before acceptance.
#![allow(unused)] fn main() { pub struct Fact { pub type_id: FactTypeId, pub payload: FactPayload, pub attestation: Attestation, pub metadata: FactMetadata, } }
Facts accumulate through CRDT merge. Duplicate facts are deduplicated by content hash. Conflicting facts are resolved by type-specific merge rules. The journal guarantees eventual consistency across replicas.
Attestations prove that an authority endorsed the fact. Threshold signatures require multiple devices to attest. Single-device signatures are used for local facts. The attestation type determines validation requirements.
3.3 State reduction
State reduction computes views from accumulated facts. Reducers are pure functions that transform fact sets into derived state. Reduction is deterministic and reproducible.
#![allow(unused)] fn main() { trait FactReducer { type State; fn reduce(facts: &FactSet) -> Self::State; } }
Reduction runs on demand or is cached for performance. Cached views are invalidated when new facts arrive. The reduction pipeline supports incremental updates for large fact sets. See Journal for the complete reduction architecture.
3.4 Flow budget facts
Flow budgets track message emission per context and peer. Only spent and epoch values are stored as facts. The limit is computed at runtime from capability evaluation.
#![allow(unused)] fn main() { pub struct FlowBudget { limit: u64, // derived from capabilities spent: u64, // stored as facts epoch: Epoch, // stored as facts } }
Budget charges are facts that increment spent. Epoch rotation resets counters through new epoch facts. This design keeps replicated state minimal while enabling runtime limit computation.
4. Effect System Architecture
Effect traits define async capabilities with explicit context. Handlers implement these traits for specific environments. The effect system provides the abstraction layer between application logic and runtime behavior.
flowchart TB
subgraph L3["Composite"]
direction LR
TRE[TreeEffects] ~~~ CHE[ChoreographyExt]
end
subgraph L2["Application"]
direction LR
JE[JournalEffects] ~~~ AE[AuthorizationEffects] ~~~ FE[FlowBudgetEffects] ~~~ LE[LeakageEffects]
end
subgraph L1["Infrastructure"]
direction LR
CE[CryptoEffects] ~~~ NE[NetworkEffects] ~~~ SE[StorageEffects] ~~~ TE[TimeEffects] ~~~ RE[RandomEffects]
end
L1 --> L2 --> L3
This diagram shows effect layering. Infrastructure effects wrap OS primitives. Application effects encode domain logic. Composite effects combine lower layers for convenience.
4.1 Effect trait classification
Infrastructure effects are implemented in aura-effects. These include CryptoEffects, NetworkEffects, StorageEffects, PhysicalTimeEffects, RandomEffects, and TraceEffects.
Application effects encode domain logic in domain crates. These include JournalEffects, AuthorizationEffects, FlowBudgetEffects, and LeakageEffects. Application effects compose infrastructure effects.
Composite effects are extension traits that combine multiple lower-level effects. These include TreeEffects for commitment tree operations and choreography extension traits for protocol execution.
4.2 Unified time system
The time system provides four domains for different use cases. PhysicalClock uses wall-clock time with optional uncertainty bounds. LogicalClock uses vector and Lamport clocks for causal ordering.
OrderClock uses opaque 32-byte tokens for deterministic ordering without temporal leakage. Range uses earliest and latest bounds for validity windows.
#![allow(unused)] fn main() { enum TimeStamp { PhysicalClock(PhysicalTime), LogicalClock(LogicalTime), OrderClock(OrderTime), Range(RangeTime), } }
Time access happens through PhysicalTimeEffects, LogicalClockEffects, and OrderClockEffects. Application code does not call system time directly. See Effect System for handler implementation patterns.
4.3 Context propagation
EffectContext is the operation scope that flows through async call chains. It carries authority id, context id, session id, execution mode, and metadata.
#![allow(unused)] fn main() { pub struct EffectContext { pub authority: AuthorityId, pub context: Option<ContextId>, pub session: SessionId, pub mode: ExecutionMode, pub metadata: ContextMetadata, } }
Context propagation ensures that all operations within a call chain share the same scope. Guards access context to make authorization decisions. Handlers access context to route operations to the correct namespace.
4.4 Impure function control
Application code must not call system time, randomness, or IO directly. These operations must flow through effect traits. This constraint enables deterministic testing and simulation.
#![allow(unused)] fn main() { async fn operation<E: PhysicalTimeEffects + RandomEffects>( effects: &E ) -> Result<Nonce> { let now = effects.current_time().await; let bytes = effects.random_bytes(32).await?; Ok(Nonce { now, bytes }) } }
The type signature makes dependencies explicit. Tests can inject mock handlers with controlled behavior. Simulations can replay exact sequences for debugging.
5. Guard Chain and Authorization
All transport sends pass through a guard chain before any network effect. The chain enforces authorization, budget accounting, journal coupling, and leakage tracking in a fixed sequence:
CapabilityGuard → FlowBudgetGuard → JournalCouplingGuard → LeakageTrackingGuard → TransportEffects
Each guard must succeed before the next executes. Failure at any guard blocks the send. This order enforces the charge-before-send invariant.
5.1 Guard responsibilities
CapabilityGuard evaluates Biscuit tokens against required capabilities. It verifies that the sender has authority to perform the requested operation. Biscuit caveats can restrict scope, time, or target.
FlowBudgetGuard charges flow budgets and emits receipts. It verifies that the sender has sufficient budget for the message cost. Budget charges are atomic with receipt generation.
JournalCouplingGuard commits facts alongside budget changes. It ensures that budget charges and other facts are durably recorded before the message leaves. This coupling provides atomicity.
LeakageTrackingGuard records privacy budget usage. It tracks information flow to different observer classes. Operations that exceed leakage budgets are blocked.
5.2 Receipts and accountability
Flow budget charges emit receipts. Receipts include context, sender, receiver, epoch, cost, and a hash chain link. Relays can verify that prior hops paid their budget cost.
#![allow(unused)] fn main() { pub struct Receipt { pub ctx: ContextId, pub src: AuthorityId, pub dst: AuthorityId, pub epoch: Epoch, pub cost: FlowCost, pub nonce: FlowNonce, pub prev: Hash32, pub sig: ReceiptSig, } }
The prev field links receipts in a per-hop chain. This chain provides accountability for multi-hop message forwarding. See Transport and Information Flow for receipt verification and Authorization for Biscuit integration.
6. Choreographic Protocols
Choreographies define global protocols using multi-party session types. The choreography! macro generates local session types and effect bridge helpers. Guard requirements are expressed through annotations.
6.1 Global protocol specification
A global type describes the entire protocol from a bird's-eye view. Each message specifies sender, receiver, payload type, and guard annotations.
#![allow(unused)] fn main() { choreography! { #[namespace = "key_rotation"] protocol KeyRotation { roles: Initiator, Witness1, Witness2; Initiator[guard_capability = "rotate", flow_cost = 10] -> Witness1, Witness2: Proposal(data: RotationData); Witness1[flow_cost = 5] -> Initiator: Vote(vote: bool); Witness2[flow_cost = 5] -> Initiator: Vote(vote: bool); } } }
The namespace attribute scopes the protocol. Annotations compile into guard requirements. The macro validates that the protocol is well-formed before generating code.
6.2 Projection and execution
Projection extracts each role's local view from the global type. The local view specifies what messages the role sends and receives. Execution interprets the local view against the effect system.
Production execution uses AuraChoreoEngine with the Telltale VM and the host bridge in Layer 6. Startup is manifest-driven and admitted by construction. Generated runner surfaces still exist for testing and migration support, but they are not the production execution model. See MPST and Choreography for projection rules and runtime details.
6.3 Annotation effects
Annotations drive guard chain behavior. guard_capability specifies required Biscuit capabilities. flow_cost specifies budget charges. journal_facts specifies facts to commit. leak specifies leakage budget allocation.
The guard chain interprets these annotations before each send. Annotation values are validated at compile time where possible. Runtime checks handle dynamic values.
7. Consensus and Agreement
Aura Consensus provides single-shot agreement for non-monotone operations. It produces CommitFact entries that are inserted into journals. The protocol is scoped to individual operations.
7.1 When consensus is needed
Monotone operations use CRDT merge without consensus. Facts accumulate through join. Capabilities restrict through meet. These operations converge without coordination.
Non-monotone operations require consensus. Examples include key rotation, membership changes, and authoritative state transitions. These operations cannot be safely executed with CRDT merge alone.
7.2 Operation categories
Operations are classified into categories A, B, and C. Category A uses CRDTs with immediate local effect. Category B shows pending state until agreement. Category C blocks until consensus completes.
#![allow(unused)] fn main() { enum OperationCategory { A, // CRDT, immediate B, // Pending until agreement C, // Blocking consensus } }
The category determines user experience and system behavior. See Operation Categories for classification rules and ceremony contracts.
7.3 Fast path and fallback
The fast path completes in one round trip when witnesses agree on prestate. Witnesses validate the operation, sign their shares, and return them to the initiator. The initiator aggregates shares into a threshold signature.
The fallback path activates when witnesses disagree or the initiator stalls. Bounded gossip propagates evidence until a quorum forms. Both paths yield the same CommitFact format.
7.4 CommitFact and journal integration
CommitFact represents a consensus decision. It includes prestate hash, operation hash, participant set, threshold signature, and timestamp.
#![allow(unused)] fn main() { pub struct CommitFact { pub consensus_id: ConsensusId, pub prestate: Hash32, pub operation: Hash32, pub participants: ParticipantSet, pub signature: ThresholdSignature, pub timestamp: ProvenancedTime, } }
CommitFacts are inserted into the relevant journal namespace. Reducers process CommitFacts to update derived state. The signature proves that a threshold of participants agreed.
The consensus_id binds the decision to a specific prestate and operation. This binding prevents reusing signatures across unrelated operations. Prestate binding ensures that consensus decisions apply to the expected state. See Consensus for protocol details.
8. Transport and Networking
Transport abstractions provide secure channels between authorities. The system does not assume persistent connections. Messages may be relayed through multiple hops.
8.1 SecureChannel abstraction
SecureChannel provides encrypted, authenticated communication between two authorities. Channels use context-scoped keys derived through DKD. Channel state is not stored in journals.
#![allow(unused)] fn main() { trait SecureChannel { async fn send(&self, msg: &[u8]) -> Result<()>; async fn recv(&self) -> Result<Vec<u8>>; fn peer(&self) -> AuthorityId; } }
Channels are established through rendezvous or direct connection. The abstraction hides transport details from application code. See Rendezvous Architecture for channel establishment.
8.2 Rendezvous and peer discovery
Rendezvous enables authorities to find each other without centralized directories. Rendezvous servers are untrusted relays that cannot read message content. Authorities publish encrypted envelopes that peers can retrieve.
The social topology provides routing hints. Home and neighborhood membership influences relay selection. Authorities prefer relays operated by trusted peers. Fallback uses public rendezvous servers when social relays are unavailable.
Envelope encryption ensures that rendezvous servers learn nothing about message content or recipients. The sender encrypts to the recipient's public key. The server sees only opaque blobs with timing metadata. See Social Architecture for topology details.
8.3 Asynchronous message patterns
AMP provides patterns for reliable asynchronous messaging. Messages may arrive out of order. Delivery may be delayed by offline peers. AMP handles acknowledgment, retry, and ordering.
Channels support both synchronous request-response and asynchronous fire-and-forget patterns. The pattern choice depends on operation requirements. See Asynchronous Message Patterns for implementation details.
9. Crate Architecture
Aura uses eight layers with strict dependency ordering. Dependencies flow downward. No crate imports from a higher layer.
flowchart TB
L1[Layer 1: Foundation<br/>aura-core]
L2[Layer 2: Specification<br/>aura-journal, aura-authorization, aura-mpst]
L3[Layer 3: Implementation<br/>aura-effects, aura-composition]
L4[Layer 4: Orchestration<br/>aura-protocol, aura-guards, aura-consensus]
L5[Layer 5: Feature<br/>aura-chat, aura-recovery, aura-social]
L6[Layer 6: Runtime<br/>aura-agent, aura-app, aura-simulator]
L7[Layer 7: Interface<br/>aura-terminal]
L8[Layer 8: Testing<br/>aura-testkit, aura-harness]
L1 --> L2 --> L3 --> L4 --> L5 --> L6 --> L7
L8 -.-> L1 & L2 & L3 & L4 & L5 & L6
This diagram shows dependency flow. Testing crates can depend on any layer for test support.
9.1 Layer descriptions
Layer 1, Foundation — aura-core with effect traits, identifiers, and cryptographic utilities.
Layer 2, Specification — domain crates defining semantics without runtime. No OS access, no Tokio. Facts use DAG-CBOR.
Layer 3, Implementation — aura-effects for production handlers, aura-composition for handler assembly.
Layer 4, Orchestration — multi-party coordination via aura-protocol, aura-guards, aura-consensus, aura-amp, aura-anti-entropy.
Layer 5, Feature — end-to-end protocols. Each crate declares OPERATION_CATEGORIES. Runtime caches live in Layer 6.
Layer 6, Runtime — aura-agent for assembly, aura-app for portable logic, aura-simulator for deterministic simulation.
Layer 7, Interface — aura-terminal for CLI and TUI entry points.
Layer 8, Testing — aura-testkit, aura-quint, and aura-harness for test infrastructure.
9.2 Code location guidance
The layer determines where code belongs. Stateless single-party operations go in Layer 3. Multi-party coordination goes in Layer 4. Complete protocols go in Layer 5. Runtime assembly goes in Layer 6.
Effect traits are defined only in aura-core. Infrastructure handlers live in aura-effects. Mock handlers live in aura-testkit. This separation keeps the dependency graph clean.
For practical guidance on effects and handlers, see Effects Guide. For choreography development, see Choreography Guide. For complete crate breakdown and dependency graph, see Project Structure.
10. Security Model
Aura's security model eliminates single points of trust. No central server holds keys or can read messages. Trust is distributed across devices and social relationships.
10.1 Threshold cryptography
Account authorities use threshold signatures. Key operations require a threshold of devices to participate. Compromising fewer than the threshold reveals nothing about the key.
FROST provides the threshold signature scheme. DKG distributes key shares without a trusted dealer. Key rotation and resharing maintain security as devices join or leave. See Cryptographic Architecture for implementation details.
The threshold is configurable per authority. A 2-of-3 threshold balances security and availability for typical users. Higher thresholds provide stronger security at the cost of requiring more devices to be online.
10.2 Capability-based authorization
Authorization uses Biscuit tokens with cryptographic attenuation. Capabilities can only be restricted, never expanded. Delegation chains are verifiable without contacting the issuer.
The guard chain enforces capabilities at runtime. Biscuit Datalog queries check predicates against facts. See Authorization for token structure and evaluation.
10.3 Context isolation
Contexts provide information flow boundaries. Keys are derived per-context. Facts are scoped to namespaces. Cross-context flow requires explicit bridge protocols.
Leakage tracking monitors information flow to observer classes. Operations that exceed privacy budgets are blocked. See Privacy and Information Flow for the leakage model.
11. Reactive State Management
The system uses reactive signals for state propagation. Journal changes trigger signal updates. UI components subscribe to signals. This pattern decouples state producers from consumers.
11.1 Signal architecture
ReactiveEffects defines the signal interface. ReactiveHandler implements batched processing with configurable windows. Signals carry typed state that observers can query.
#![allow(unused)] fn main() { trait ReactiveEffects { async fn emit<T: Signal>(&self, signal: T); async fn subscribe<T: Signal>(&self) -> Receiver<T>; } }
Signal emission is non-blocking. Handlers batch rapid updates to reduce overhead. Subscribers receive the latest state when they poll.
11.2 Journal to UI flow
Journal fact changes flow through reducers to signals. Reducers compute derived state from facts. Signals expose that state to UI observers. The flow is unidirectional and predictable.
The aura-app crate defines application signals. The aura-terminal crate consumes these signals for rendering. This separation keeps UI concerns out of core logic.
12. Error Handling
Errors are unified through the AuraError type in aura-core. Domain crates define specific error variants. Effects propagate errors through Result types.
11.1 Error classification
Errors are classified by recoverability. Transient errors may succeed on retry. Permanent errors indicate invalid operations. System errors indicate infrastructure failures.
#![allow(unused)] fn main() { pub enum ErrorKind { Transient, // Retry may succeed Permanent, // Invalid operation System, // Infrastructure failure } }
Error classification guides retry behavior and user feedback. Transient errors trigger automatic retry with backoff. Permanent errors are reported to the user. System errors may require recovery procedures.
11.2 Error propagation
Effects propagate errors through Result types. Handlers convert low-level errors to domain errors. The error chain preserves context for debugging. Logging captures error details without exposing sensitive data.
#![allow(unused)] fn main() { async fn operation<E: JournalEffects>(effects: &E) -> Result<(), AuraError> { effects.merge_facts(facts) .await .map_err(|e| AuraError::journal(e, "merge failed"))?; Ok(()) } }
Error context includes the operation name and relevant identifiers. Stack traces are available in debug builds. Production builds log structured error data for monitoring.
11.3 Consensus and recovery
Consensus failures have specific handling. Fast path failures fall back to gossip. Network partitions delay but do not corrupt state. Recovery procedures restore operation after failures.
The journal provides durability. Uncommitted facts are replayed after restart. Committed facts are immutable. This design simplifies recovery logic.
Device recovery uses guardian protocols. Guardians hold encrypted recovery shares. A threshold of guardians can restore account access. See Relational Contexts for recovery patterns.
Theoretical Model
This document establishes the complete mathematical foundation for Aura's distributed system architecture. It presents the formal calculus, algebraic/session types, and semilattice semantics that underlie all system components.
Overview
Aura's theoretical foundation rests on four mathematical pillars:
- Aura Calculus () provides the core computational model for communication, state, and trust.
- Algebraic Types structure state as semilattices with monotonic properties.
- Multi-Party Session Types specify choreographic protocols with safety guarantees.
- CRDT Semantics enable conflict-free replication with convergence proofs.
The combination forms a privacy-preserving, spam-resistant, capability-checked distributed λ-calculus that enforces information flow budget across all operations.
Shared Terms and Notation
This section defines shared terminology and notation for core contracts. Use this section when writing Theoretical Model, Privacy and Information Flow Contract, and Distributed Systems Contract.
Core Entities
| Symbol / Term | Type | Description |
|---|---|---|
| , | AuthorityId | Authority identifiers |
ContextId | Context identifier | |
Epoch | Epoch number | |
| authority | — | An account authority |
| context | — | A relational or authority namespace keyed by ContextId |
| peer authority | — | A remote authority in a context |
| member | — | An authority in a home's threshold authority set |
| participant | — | An authority granted home access but not in the threshold set |
| moderator | — | A member with moderation designation for a home |
Access and Topology
| Term | Description |
|---|---|
Full | Same-home access (0 hops) |
Partial | 1-hop neighborhood access |
Limited | 2-hop-or-greater and disconnected access |
| 1-hop link | Direct home-to-home neighborhood edge |
| n-hop | Multi-edge path (2-hop, 3-hop, etc.) |
| Shared Storage | Community storage pool |
| pinned | Fact attribute for retained content |
Flow Budget and Receipts
Use for a flow budget for context and peer authority .
| Field | Type / Storage | Description |
|---|---|---|
| FlowBudget | ||
spent | Replicated fact | Monotone counter of consumed budget |
epoch | Replicated fact | Current epoch identifier |
limit | Derived at runtime | From capability evaluation and policy |
| Receipt | ||
ctx | ContextId | Context scope |
src | AuthorityId | Sending authority |
dst | AuthorityId | Receiving authority |
epoch | Epoch | Validity window |
cost | u32 | Budget charge |
nonce | u64 | Replay prevention |
prev_hash | Hash32 | Links receipts in per-hop chain |
sig | Signature | Cryptographic proof |
Observers, Time, and Delay
| Symbol | Category | Description |
|---|---|---|
| Observer | Relationship (direct relationship observer) | |
| Observer | Group (group member observer) | |
| Observer | Neighbor (neighborhood observer) | |
| Observer | External (external/network observer) | |
| CRDT/state | Local state deltas (NOT network delay) | |
| Network | Network delay bounds under partial synchrony | |
GST | Timing | Global Stabilization Time |
Leakage tuple order:
Invariant Naming
Use InvariantXxx names for implementation and proof references.
If a prose alias exists, include it once, then reference the invariant name.
Guard Chain Order
| Order | Component | Role |
|---|---|---|
| 1 | CapGuard | Capability verification |
| 2 | FlowGuard | Budget enforcement |
| 3 | JournalCoupler | Fact commitment |
| 4 | TransportEffects | Message transmission |
This order defines ChargeBeforeSend.
1. Aura Calculus ()
1.1 Syntax
We define programs as effectful, session-typed processes operating over semilattice-structured state.
Terms:
Facts (Join-Semilattice):
Capabilities (Meet-Semilattice):
Contexts:
Contexts are opaque UUIDs representing authority journals or relational contexts. Keys for transport sessions and DKD outputs are scoped to these identifiers. The identifier itself never leaks participants. See Identifiers and Boundaries for canonical definitions.
Messages:
Message extraction functions (used by operational rules):
A process configuration:
This represents a running session with fact-state , capability frontier derived from verified Biscuit tokens and local policy, and privacy context .
1.2 Judgments
Under typing context , expression has type and may perform effects .
Effect set:
1.3 Operational Semantics
State evolution:
Capability-guarded actions:
Each side effect or message action carries a required capability predicate .
The function attn applies the Biscuit token's caveats to the local frontier. Biscuit attenuation never widens authority. The operation remains meet-monotone even though the token data lives outside the journal.
Context isolation:
No reduction may combine messages of distinct contexts:
Here is shorthand for the budget predicate derived from journal facts and Biscuit-imposed limits for context :
Implementations realize this by merging a FlowBudget charge fact before send (see §2.3 and §5.3) while evaluating Biscuit caveats inside the guard chain. The side condition is enforced by the same monotone laws as other effects even though capability data itself is not stored in the CRDT.
1.4 Algebraic Laws (Invariants)
- Monotonic Growth:
- Monotonic Restriction:
- Safety: Every side effect requires .
- Context Separation: For any two contexts , no observable trace relates their internal state unless a bridge protocol is typed for .
- Compositional Confluence: and
2. Core Algebraic Types
2.1 Foundation Objects
#![allow(unused)] fn main() { // Capabilities describe Biscuit caveats. They form a meet-semilattice but are evaluated outside the CRDT. type Cap // partially ordered set (≤), with meet ⊓ and top ⊤ type Policy // same carrier as Cap, representing sovereign policy // Facts are join-semilattice elements (accumulation only grows them). type Fact // partially ordered set (≤), with join ⊔ and bottom ⊥ // Journal state is only a Cv/Δ/CmRDT over facts. struct Journal { facts: Fact, // Cv/Δ/CmRDT carrier with ⊔ } // Context identifiers are opaque (authority namespace or relational context). struct ContextId(Uuid); struct Epoch(u64); // monotone, context-scoped struct FlowBudget { limit: u64, spent: u64, epoch: Epoch }; struct Receipt { ctx: ContextId, src: AuthorityId, dst: AuthorityId, epoch: Epoch, cost: u32, nonce: u64, prev_hash: Hash32, sig: Signature }; // Typed messages carry effects and proofs under a context. struct Msg<Ctx, Payload, Version> { ctx: Ctx, // ContextId chosen by relationship payload: Payload, // typed by protocol role/state ver: Version, // semantic version nego auth: AuthTag, // signatures/MACs/AEAD tags biscuit: Option<Biscuit>, // optional attenuated capability } }
These type definitions establish the foundation for Aura's formal model. The Cap and Policy types form meet-semilattices for capability evaluation. The Fact type forms a join-semilattice for accumulating evidence. The Journal contains only facts and inherits join-semilattice properties. Messages are typed and scoped to contexts to ensure isolation.
The type Cap represents the evaluation lattice used by Biscuit. Refinement operations (caveats, delegation) can only reduce authority through the meet operation.
The type Fact represents facts as join-semilattice elements. Accumulation operations can only add information through the join operation.
Journals replicate only facts. Capability evaluations run locally by interpreting Biscuit tokens plus policy. This keeps authorization independent of the replicated CRDT while preserving the same meet monotonicity at runtime.
Contexts (ContextId) define privacy partitions. Messages never cross partition boundaries without explicit protocol support. See Identifiers and Boundaries for precise identifier semantics and Relational Contexts for implementation patterns.
2.2 Content Addressing Contract
All Aura artifacts are identified by the hash of their canonical encoding. This includes facts, snapshot blobs, cache metadata, and upgrade manifests.
Structures are serialized using canonical CBOR with sorted maps and deterministic integer width. The helper function hash_canonical(bytes) computes digests when needed.
Once a digest is published, the bytes for that artifact cannot change. New content requires a new digest and a new fact in the journal.
Snapshots and upgrade bundles stored outside the journal are referenced solely by their digest. Downloaders verify the digest before accepting the payload. Journal merges compare digests and reject mismatches before updating state. See Distributed Maintenance Architecture for the complete fact-to-state pipeline.
2.3 Effect Signatures
Core effect families provide the runtime contract:
-- Read/append mergeable state
class JournalEffects (m : Type → Type) where
read_facts : m Fact
merge_facts : Fact → m Unit
-- Biscuit verification + guard evaluation
class AuthorizationEffects (m : Type → Type) where
evaluate_guard : Biscuit → CapabilityPredicate → m Cap
derive_cap : ContextId → m Cap -- cached policy frontier
-- Cryptography and key mgmt (abstracted to swap FROST, AEAD, DR, etc.)
class CryptoEffects (m : Type → Type) where
sign_threshold : Bytes → m SigWitness
aead_seal : K_box → Plain → m Cipher
aead_open : K_box → Cipher → m (Option Plain)
commitment_step : ContextId → m ContextId
-- Transport (unified)
class TransportEffects (m : Type → Type) where
send : PeerId → Msg Ctx P V → m Unit
recv : m (Msg Ctx Any V)
connect : PeerId → m Channel
These effect signatures define the interface between protocols and the runtime. The JournalEffects family handles state operations. The AuthorizationEffects family verifies Biscuit tokens and fuses them with local policy. The CryptoEffects family handles cryptographic operations. The TransportEffects family handles network communication.
2.4 Guards and Observability Invariants
Every observable side effect is mediated by a guard chain fully described in Authorization:
- CapGuard:
- FlowGuard:
headroom(ctx, cost)wherecharge(ctx, peer, cost, epoch)succeeds and yields aReceipt - JournalCoupler: commit of attested facts is atomic with the send
Named invariants used across documents:
- Charge-Before-Send: FlowGuard must succeed before any transport send.
- No-Observable-Without-Charge: there is no event without a preceding successful .
- Deterministic-Replenishment:
limit(ctx)is computed deterministically from capability evaluation. The valuespent(stored as journal facts) is join-monotone. Epochs gate resets.
-- Time & randomness for simulation/proofs
class TimeEffects (m : Type → Type) where
now : m Instant
sleep : Duration → m Unit
class RandEffects (m : Type → Type) where
sample : Dist → m Val
-- Privacy budgets (relationship, group, neighbor, external observers)
class LeakageEffects (m : Type → Type) where
record_leakage : ObserverClass → Number → m Unit
remaining_budget : ObserverClass → m Number
The TimeEffects and RandEffects families support simulation and testing. The LeakageEffects family enforces privacy budget constraints.
The LeakageEffects implementation is the runtime hook that enforces the annotations introduced in the session grammar. The system wires it through the effect system so choreographies cannot exceed configured budgets.
Information Flow Budgets (Spam + Privacy)
Each context and peer authority pair carries a flow budget to couple spam resistance with privacy guarantees. This pair is written as . Keys and receipts are authority-scoped. Devices are internal to an authority and never appear in flow-budget state.
#![allow(unused)] fn main() { struct FlowBudget { spent: u64, // monotone counter (join = max) limit: u64, // capability-style guard (meet = min) } // Logical key: (ContextId, AuthorityId) -> FlowBudget // Receipts: (ctx, src_authority, dst_authority, epoch, cost, nonce) }
The FlowBudget struct tracks message emission through two components:
- The
spentfield is stored in the journal as facts and increases through join operations - The
limitfield is derived from capability evaluation (Biscuit tokens + local policy) and decreases through meet operations
Only spent counters live in the journal as facts, inheriting join-semilattice properties. The limit is computed at runtime by evaluating Biscuit tokens and local policy through the capability meet-semilattice, consistent with the principle that capabilities are evaluated outside the CRDT.
Sending a message deducts a fixed flow_cost from the local budget before the effect executes. If , the effect runtime blocks the send.
Budget charge facts (incrementing spent) are emitted to the journal during send operations. The limit value is deterministically computed from the current capability frontier, ensuring all replicas with the same tokens and policy derive the same limit. Receipts bind to AuthorityId for both src/dst and use nonce chains per (ctx, src, epoch) to maintain charge-before-send proofs without leaking device structure.
Multi-hop forwarding charges budgets hop-by-hop. Relays attach a signed Receipt that proves the previous hop still had headroom. Receipts are scoped to the same context so they never leak to unrelated observers.
2.5 Semantic Laws
Join laws apply to facts. These operations are associative, commutative, and idempotent. If and after we have , then with respect to the facts partial order.
Meet laws apply to capabilities. These operations are associative, commutative, and idempotent. Applying Biscuit caveats corresponds to multiplying by under and never increases authority.
Cap-guarded effects enforce non-interference. For any effect guarded by capability predicate , executing from is only permitted if .
Context isolation prevents cross-context flow. If two contexts are not explicitly bridged by a typed protocol, no flows into .
3. Multi-Party Session Type Algebra
3.1 Global Type Grammar (G)
The global choreography type describes the entire protocol from a bird's-eye view. Aura extends vanilla MPST with capability guards, journal coupling, and leakage budgets:
Conventions:
- means "role checks , applies , records leakage , sends to , then continues with ."
- performs the same sequence for broadcasts.
- means "execute and concurrently."
- means "role decides which branch to take, affecting all participants."
- binds recursion variable in .
Note on : the journal delta may include budget-charge updates (incrementing spent for the active epoch) and receipt acknowledgments. Projection ensures these updates occur before any transport effect so "no observable without charge" holds operationally.
Note on : check_caps and refine_caps are implemented via AuthorizationEffects. Sends verify Biscuit chains (with optional cached evaluations) before touching transport. Receives cache new tokens by refining the local capability frontier. Neither operation mutates the journal. All capability semantics stay outside the CRDT.
3.2 Local Type Grammar (L)
After projection, each role executes a local session type (binary protocol) augmented with effect sequencing:
3.3 Projection Function ()
The projection function extracts role 's local view from global choreography :
By convention, an annotation at a global step induces per-side deltas and . Unless otherwise specified by a protocol, we take (symmetric journal updates applied at both endpoints).
Point-to-point projection:
- If (sender):
- If (receiver):
- Otherwise:
Broadcast projection:
- If (sender):
- Otherwise (receiver):
Parallel composition:
where is the merge operator (sequential interleaving if no conflicts)
Choice projection:
- If (decider):
- If (observer):
Recursion projection:
- If :
- If :
Base cases:
3.4 Duality and Safety
For binary session types, duality ensures complementary behavior:
Property: If Alice's local type is , then Bob's local type is for their communication to be type-safe.
3.5 Session Type Safety Guarantees
The projection process ensures:
- Deadlock Freedom: No circular dependencies in communication
- Type Safety: Messages have correct types at send/receive
- Communication Safety: Every send matches a receive
- Progress: Protocols always advance (no livelocks)
- Agreement: All participants agree on the chosen branch and protocol state (modulo permitted interleavings of independent actions)
3.6 Turing Completeness vs Safety Restrictions
The MPST algebra is Turing complete when recursion () is unrestricted. However, well-typed programs intentionally restrict expressivity to ensure critical safety properties:
- Termination: Protocols that always complete (no infinite loops)
- Deadlock Freedom: No circular waiting on communication
- Progress: Protocols always advance to next state
Telltale-backed session runtimes balance expressivity and safety through guarded recursion constructs.
3.7 Free Algebra View (Choreography as Initial Object)
You can think of the choreography language as a small set of protocol-building moves:
Generators:
- and
Taken together, these moves form a free algebra: the language carries just enough structure to compose protocols, but no extra operational behavior. The effect runtime is the target algebra that gives these moves concrete meaning.
Projection (from a global protocol to each role) followed by interpretation (running it against the effect runtime) yields one canonical way to execute any choreography.
The "free" (initial) property is what keeps this modular. Because the choreographic layer only expresses structure, any effect runtime that respects those composition laws admits exactly one interpretation of a given protocol. This allows swapping or layering handlers without changing choreographies.
The system treats computation and communication symmetrically. A step is the same transform whether it happens locally or across the network. If the sender and receiver are the same role, the projection collapses the step into a local effect call. If they differ, it becomes a message exchange with the same surrounding journal/guard/leak actions. Protocol authors write global transforms, the interpreter decides local versus remote at time of projection.
3.8 Algebraic Effects and the Interpreter
Aura treats protocol execution as interpretation over an algebraic effect interface. After projecting a global choreography to each role, a polymorphic interpreter walks the role's AST and dispatches each operation to AuraEffectSystem via explicit effect handlers. The core actions are exactly the ones defined by the calculus and effect signatures in this document: merge (facts grow by ), refine (caps shrink by ), send/recv (context-scoped communication), and leakage/budget metering. The interpreter enforces the lattice laws and guard predicates while executing these actions in the order dictated by the session type.
Because the interface is algebraic, there is a single semantics regardless of execution strategy. This enables two interchangeable modes:
Static compilation: choreographies lower to direct effect calls with zero runtime overhead.
Dynamic interpretation: choreographies execute through the runtime interpreter for flexibility and tooling.
Both preserve the same program structure and checks. The choice becomes an implementation detail. This also captures the computation/communication symmetry. A choreographic step describes a typed transform. If the sender and receiver are the same role, projection collapses the step to a local effect invocation. If they differ, the interpreter performs a network send/receive with the same surrounding merge/check_caps/refine/record_leak sequence. Protocol authors reason about transforms. The interpreter decides locality at projection time.
4. CRDT Semantic Foundations
4.1 CRDT Type System
Aura implements four CRDT variants to handle different consistency requirements.
#![allow(unused)] fn main() { // State-based (CvRDT) - Full state synchronization pub trait JoinSemilattice: Clone { fn join(&self, other: &Self) -> Self; } pub trait Bottom { fn bottom() -> Self; } pub trait CvState: JoinSemilattice + Bottom {} // Delta CRDTs - Incremental state synchronization pub trait Delta: Clone { fn join_delta(&self, other: &Self) -> Self; } pub trait DeltaProduce<S> { fn delta_from(old: &S, new: &S) -> Self; } // Operation-based (CmRDT) - Causal operation broadcast pub trait CausalOp { type Id: Clone; type Ctx: Clone; fn id(&self) -> Self::Id; fn ctx(&self) -> &Self::Ctx; } pub trait CmApply<Op> { fn apply(&mut self, op: Op); } pub trait Dedup<I> { fn seen(&self, id: &I) -> bool; fn mark_seen(&mut self, id: I); } // Meet-based CRDTs - Constraint propagation pub trait MeetSemilattice: Clone { fn meet(&self, other: &Self) -> Self; } pub trait Top { fn top() -> Self; } pub trait MvState: MeetSemilattice + Top {} }
The type system enforces mathematical properties. CvState types must satisfy associativity, commutativity, and idempotency for join operations. MvState types satisfy the same laws for meet operations.
4.2 Convergence Proofs
Each CRDT type provides specific convergence guarantees:
CvRDT Convergence: Under eventual message delivery, all replicas converge to the least upper bound of their updates.
Proof sketch: Let states be replica states after all updates. The final state is . By associativity and commutativity of join, this value is unique regardless of message ordering.
Delta-CRDT Convergence: Equivalent to CvRDT but with bandwidth optimization through incremental updates.
CmRDT Convergence: Under causal message delivery and deduplication, all replicas apply the same set of operations.
Proof sketch: Causal delivery ensures operations are applied in a consistent order respecting dependencies. Deduplication prevents double-application. Commutativity ensures that concurrent operations can be applied in any order with the same result.
Meet-CRDT Convergence: All replicas converge to the greatest lower bound of their constraints.
4.3 Authority-Specific Applications
Authority journals use join-semilattice facts for evidence accumulation. Facts include AttestedOp records proving commitment tree operations occurred. Multiple attestations of the same operation join to the same fact.
Relational context journals use both join operations for shared facts and meet operations for consensus constraints. The prestate model ensures all authorities agree on initial conditions before applying operations.
Biscuit capability evaluation uses meet operations outside the CRDT. Token caveats restrict authority through intersection without affecting replicated state.
4.4 Message Schemas for Authority Synchronization
#![allow(unused)] fn main() { // Tagged message types for different synchronization patterns #[derive(Clone)] pub enum SyncMsgKind { FullState, Delta, Op, Constraint } pub type AuthorityStateMsg = (AuthorityFacts, SyncMsgKind); pub type ContextStateMsg = (ContextFacts, SyncMsgKind); pub type FactDelta = (Vec<Fact>, SyncMsgKind); #[derive(Clone)] pub struct AuthenticatedOp { pub op: TreeOp, pub attestation: ThresholdSig, pub ctx: ContextId } // Anti-entropy protocol support pub type FactDigest = Vec<Hash32>; pub type MissingFacts = Vec<Fact>; }
These message types support different synchronization strategies. Authority namespaces primarily use delta synchronization for efficiency. Relational context namespaces use operation-based synchronization to preserve consensus ordering.
4.5 Authority Synchronization Protocols
Authority Fact Synchronization: Authorities exchange facts using delta-based gossip for efficiency.
Relational context consensus: Contexts use operation-based synchronization to preserve consensus ordering.
Anti-Entropy Recovery: Peers detect missing facts through digest comparison and request specific items.
These protocols ensure eventual consistency across authority boundaries while maintaining context isolation. Facts are scoped to their originating authority or relational context.
4.6 Implementation Verification
Aura provides property verification for CRDT implementations through several mechanisms.
Property-Based Testing: Implementations include QuickCheck properties verifying semilattice laws. Tests generate random sequences of operations and verify associativity, commutativity, and idempotency.
Convergence Testing: Integration tests simulate network partitions and verify that replicas converge after partition healing. Tests measure convergence time and verify final state consistency.
Formal Verification: Critical CRDT implementations include formal proofs using the Quint specification language. Proofs verify that implementation behavior matches mathematical specifications.
Safety Guarantees: Session type projection ensures communication safety and deadlock freedom. Semilattice laws guarantee convergence under eventual message delivery. Meet operations ensure that capability restrictions are monotonic and cannot be bypassed.
5. Information Flow Contract (Privacy + Spam)
5.1 Privacy Layers
For any trace of observable messages:
- Unlinkability:
- Non-amplification: Information visible to observer class is monotone in authorized capabilities:
- Leakage Bound: For each observer , .
- Flow Budget Soundness (Named):
- Charge-Before-Send
- No-Observable-Without-Charge
- Deterministic-Replenishment
- Convergence: Within a fixed epoch and after convergence, across replicas .
5.2 Web-of-Trust Model
Let where vertices are accounts. Edges carry relationship contexts and delegation fragments.
- Each edge defines a pairwise context with derived keys
- Delegations are meet-closed elements , scoped to contexts
- The effective capability at is:
WoT invariants:
- Compositionality: Combining multiple delegations uses (never widens)
- Local sovereignty: is always in the meet. can only reduce authority further
- Projection: For any protocol projection to , guard checks refer to
5.3 Flow Budget Contract
The unified information-flow budget regulates emission rate/volume and observable leakage. The budget system combines join-semilattice facts (for spent counters) with meet-semilattice capability evaluation (for limit computation). For any context and peer :
- Charge-Before-Send: A send or forward is permitted only if a budget charge succeeds first. If charging fails, the step blocks locally and emits no network observable.
- No-Observable-Without-Charge: For any trace , there is no event labeled without a preceding successful charge for in the same epoch.
- Receipt soundness: A relay accepts a packet only with a valid per-hop
Receipt(context-scoped, epoch-bound, signed) and sufficient local headroom. Otherwise it drops locally. - Deterministic limit computation: is computed deterministically from Biscuit tokens and local policy via meet operations. is stored as journal facts and is join-monotone. Upon epoch rotation, resets through new epoch facts.
- Context scope: Budget facts and receipts are scoped to . They neither leak nor apply across distinct contexts (non-interference).
- Composition with caps: A transport effect requires both and (see §1.3). Either guard failing blocks the effect.
- Convergence bound: Within a fixed epoch and after convergence, across replicas , where each replica's limit is computed from its local capability evaluation.
6. Application Model
Every distributed protocol is defined as a multi-party session type with role projections:
When executed, each role instantiates a handler:
Handlers compose algebraically over by distributing operations over semilattice state transitions. This yields an effect runtime capable of:
- key-ceremony coordination (threshold signatures)
- gossip and rendezvous (context-isolated send/recv)
- distributed indexing (merge facts, meet constraints)
- garbage collection (join-preserving retractions)
7. Interpretation
Under this calculus, we can make the following interpretation:
The Semilattice Layer
The join-semilattice (Facts) captures evidence and observations (trust and information flow). Examples: delegations/attestations, quorum proofs, ceremony transcripts, flow receipts, and monotone spent counters.
The meet-semilattice (Capabilities) captures enforcement limits and constraints (trust and information flow). Examples: the sovereign policy lattice, Biscuit token caveats, leak bounds, and consent gates. See Authorization for implementation details. Flow budget limits are derived from capability evaluation, not stored as facts. This lattice is evaluated locally rather than stored in the journal, but it obeys the same algebra.
Effective authority and headroom are computed from both lattices:
The Session-Typed Process Layer
This layer guarantees communication safety and progress. It projects global types with annotations into local programs, ensuring deadlock freedom, communication safety, branch agreement, and aligning capability checks, journal updates, and leakage accounting with each send/recv.
The Effect Handler Layer
The Effect Handler system provides operational semantics and composability. It realizes merge/refine/send/recv as algebraic effects, enforces lattice monotonicity ( facts, caps), guard predicates, and budget/leakage metering, and composes via explicit dependency injection across crypto, storage, and transport layers.
The Privacy Contract
The privacy contract defines which transitions are observationally equivalent. Under context isolation and budgeted leakage, traces that differ only by in-context reorderings or by merges/refinements preserving observer-class budgets and effective capabilities are indistinguishable. No cross-context flow occurs without a typed bridge.
Together, these form a privacy-preserving, capability-checked distributed λ-calculus.
System Contracts
- Privacy and Information Flow Contract - Unified information-flow budgets, privacy layers, leakage tracking, and adversary analysis
- Distributed Systems Contract - Safety, liveness, consistency guarantees, synchrony assumptions, and failure handling
See Also
- Aura System Architecture - Implementation patterns and runtime layering
- Journal - Fact storage and CRDT semantics
- Authorization - Biscuit evaluation and guard chaining
- Identifiers and Boundaries - Canonical identifier definitions
- Distributed Maintenance Architecture - Reducer pipelines for authorities and contexts
Privacy and Information Flow Contract
This contract specifies Aura's privacy and information-flow model. It defines privacy boundaries, leakage budgets, and enforcement mechanisms. Privacy boundaries align with social relationships rather than technical perimeters. Violations occur when information crosses trust boundaries without consent. Acceptable flows consume explicitly budgeted headroom.
This document complements Distributed Systems Contract, which covers safety, liveness, and consistency. Together these contracts define the full set of invariants protocol authors must respect.
Formal verification of these properties uses Quint model checking (verification/quint/) and Lean 4 theorem proofs (verification/lean/). See Verification Coverage Report for current status.
1. Scope
The contract applies to information flows across privacy boundaries:
- Flow budgets: Per-context per-peer spending limits enforced by FlowGuard
- Leakage tracking: Metadata exposure accounting by observer class
- Context isolation: Separation of identities and journals across contexts
- Receipt chains: Multi-hop forwarding accountability
- Epoch boundaries: Temporal isolation of budget and receipt state
Related specifications: Authorization, Transport and Information Flow, and Theoretical Model. Shared notation appears in Theoretical Model.
1.1 Terminology Alignment
This contract uses shared terminology from Theoretical Model.
- Home role terms:
Member,Participant,Moderator(only members can be moderators) - Access-level terms:
Full,Partial,Limited - Storage terms:
Shared Storageandallocation - Pinning term:
pinnedas a fact attribute
1.2 Assumptions
- Cryptographic primitives are secure at configured key sizes.
- Local runtimes enforce guard-chain ordering before transport sends.
- Epoch updates and budget facts eventually propagate through anti-entropy.
- Optional privacy enhancements such as Tor and cover traffic are correctly configured when enabled.
1.3 Non-goals
- This contract does not guarantee traffic-analysis resistance against global passive adversaries without optional privacy overlays.
- This contract does not define social policy decisions such as who should trust whom.
2. Privacy Philosophy
Traditional privacy systems force users to choose between complete isolation and complete exposure. Aura recognizes that privacy is relational. Sharing information with trusted parties is not a privacy violation, it's the foundation of meaningful collaboration.
2.1 Core Principles
- Consensual disclosure: Joining a group or establishing a relationship implies consent to share coordination information
- Contextual identity: Deterministic Key Derivation presents different identities in different contexts, and only relationship parties can link them
- Neighborhood visibility: Gossip neighbors observe encrypted envelope metadata, bounded by flow budgets and context isolation
2.2 Privacy Layers
| Layer | Protection | Mechanism |
|---|---|---|
| Identity | Context-specific keys | DKD: derive(root, app_id, context_label) |
| Relationship | Graph opacity | No central directory, out-of-band establishment |
| Group | Membership hiding | Threshold operations, group-scoped identity |
| Content | End-to-end encryption | AES-256-GCM, HPKE, per-message keys |
| Metadata | Rate/volume bounds | Flow budgets, fixed-size envelopes, batching |
Verified by: Aura.Proofs.KeyDerivation, authorization.qnt
3. Flow Budget System
3.1 Budget Structure
For each context and peer pair, the journal records charge facts that contribute to a flow budget:
FlowBudget {
limit: u64, // derived at runtime from Biscuit + policy
spent: u64, // replicated fact (merge = max)
epoch: Epoch, // replicated fact
}
Only spent and epoch appear as journal facts. The limit is computed at runtime by intersecting Biscuit-derived capabilities with sovereign policy.
3.2 Limit Computation
The limit for a context and peer is computed as:
limit(ctx, peer) = base(ctx) ⊓ policy(ctx) ⊓ role(ctx, peer)
⊓ relay_factor(ctx) ⊓ peer_health(peer)
Each term is a lattice element. Merges occur via meet (⊓), ensuring convergence and preventing widening.
| Term | Source | Purpose |
|---|---|---|
base(ctx) | Context class | Default headroom |
policy(ctx) | Sovereign settings | Account-level limits |
role(ctx, peer) | Biscuit token | Per-peer role attenuation |
relay_factor(ctx) | Network topology | Hub amplification mitigation |
peer_health(peer) | Liveness monitoring | Overload protection |
3.3 Charge-Before-Send
Every transport observable is preceded by guard evaluation:
- CapGuard: Verify Biscuit authorization
- FlowGuard: Charge
costto(context, peer)budget - JournalCoupler: Atomically commit charge fact and protocol deltas
- Transport: Emit packet only after successful charge
If spent + cost > limit, the send is blocked locally with no observable behavior.
Invariants:
spent ≤ limitat all times (InvariantFlowBudgetNonNegative)- Charging never increases available budget (
monotonic_decrease) - Guard chain order is fixed (
guardChainOrder) - Attenuation only narrows, never widens (
attenuationOnlyNarrows)
Verified by: Aura.Proofs.FlowBudget, authorization.qnt, transport.qnt
3.4 Multi-Hop Enforcement
For forwarding, each hop independently executes the guard chain:
- Relay validates upstream receipt before forwarding
- Relay charges its own budget before emitting
- Receipt facts are scoped to
(context, epoch)with chained hashes - Downstream peers can audit budget usage via receipt chain
Because spent is monotone (merge = max), convergence holds even if later hops fail.
Verified by: transport.qnt (InvariantSentMessagesHaveFacts)
3.5 Receipts and Epochs
Per-hop receipts are required for forwarding and bound to the epoch:
Receipt { ctx, src, dst, epoch, cost, nonce, prev_hash, sig }
This schema is the canonical receipt shape used across core contracts.
- Acceptance window: Current epoch only
- Rotation trigger: Journal fact
Epoch(ctx)increments - On rotation:
spent(ctx, *)resets, and old receipts become invalid
Invariants:
- Receipts only valid within their epoch (
InvariantReceiptValidityWindow) - Old epoch receipts cannot be replayed (
InvariantCrossEpochReplayPrevention)
Verified by: epochs.qnt
4. Leakage Tracking
4.1 Observer Classes
Information leakage is tracked per observer class:
| Class | Visibility | Budget Scope |
|---|---|---|
Relationship | Full context content | Consensual, no budget |
Group | Group-scoped content | Group dimension |
Neighbor | Encrypted envelope metadata | Per-hop budget |
External | Network-level patterns | Tor + cover traffic |
4.2 Leakage Budget
Each observer class has a leakage budget separate from flow budgets:
LeakageBudget {
observer: DeviceId,
leakage_type: LeakageType, // Metadata, Timing, Participation
limit: u64,
spent: u64,
refresh_interval: Duration,
}
Leakage is charged before any operation that exposes information to the observer class.
4.3 Policy Modes
| Policy | Behavior | Use Case |
|---|---|---|
Deny | Reject if no explicit budget | Production (secure default) |
DefaultBudget(n) | Fall back to n units | Transition period |
LegacyPermissive | Allow unlimited | Migration only |
Verified by: Aura.Proofs.ContextIsolation
5. Privacy Boundaries
5.1 Relationship Boundary
Within a direct relationship, both parties have consented to share coordination information:
- Visible: Context-specific identity, online status, message content
- Hidden: Activity in other contexts, identity linkage across contexts
- Enforcement: DKD ensures unique identity per context
5.2 Neighborhood Boundary
Gossip neighbors forward encrypted traffic:
- Visible: Envelope size (fixed), rotating rtags, timing patterns
- Hidden: Content, ultimate sender/receiver, rtag-to-identity mapping
- Enforcement: Flow budgets, onion routing, cover traffic
5.3 Group Boundary
Group participants share group-scoped information:
- Visible: Member identities (within group), group content, threshold operations
- Hidden: Member identities outside group, other group memberships
- Enforcement: Group-specific DKD identity, k-anonymity for sensitive operations
5.4 External Boundary
External observers have no relationship with you:
- With Tor: Only encrypted Tor traffic visible
- Without Tor: ISP sees connections to Aura nodes only
- Enforcement: Fixed-size envelopes, no central directory, flow budgets
6. Time Domain Semantics
Time handling affects privacy through leakage:
| Variant | Purpose | Privacy Impact |
|---|---|---|
PhysicalClock | Guard charging, receipts, cooldowns | Leaks wall-clock via receipts |
LogicalClock | CRDT causality, journal ordering | No wall-clock leakage |
OrderClock | Privacy-preserving total order | Opaque tokens (no temporal meaning) |
Range | Validity windows, disputes | Bounded uncertainty from physical |
- Cross-domain comparisons require explicit
TimeStamp::compare(policy) - Physical time obtained only through
PhysicalTimeEffects(never directSystemTime::now()) - Privacy mode (
ignorePhysical = true) hides physical timestamps
Verified by: Aura.Proofs.TimeSystem
7. Adversarial Model
7.1 Direct Relationship
A party in a direct relationship sees everything within that context by consent.
- Cannot: Link identity across contexts, access undisclosed contexts
- Attack vector: Social engineering, context correlation
- Mitigation: UI clearly indicates active context
7.2 Group Insider
A group member sees all group activity by consent.
- Cannot: Determine member identities outside group, access other groups
- Attack vector: Threshold signing timing correlation
- Mitigation: k-anonymity, random delays in signing rounds
7.3 Gossip Neighbor
Devices forwarding your traffic observe encrypted metadata.
- Cannot: Decrypt content, identify ultimate sender/receiver, link rtags to identities
- Attack vector: Traffic correlation through sustained observation
- Mitigation: Onion routing, cover traffic, batching, rtag rotation
7.4 Network Observer
An ISP-level adversary sees IP connections and packet timing.
- With Tor: Only Tor usage visible
- Without Tor: Connections to known Aura nodes visible
- Attack vector: Confirmation attacks, traffic correlation
- Mitigation: Tor integration, fixed-size envelopes, cover traffic
7.5 Compromised Device
A single compromised device reveals its key share and synced journal state.
- Cannot: Perform threshold operations alone, derive account root key
- Attack vector: Compromise M-of-N devices for full control
- Mitigation: Threshold cryptography, device revocation via resharing, epoch invalidation
8. Privacy Metrics
| Metric | Target | Measurement |
|---|---|---|
| Identity linkability | < 5% confidence | identity_linkability_score(ctx_a, ctx_b) |
| Relationship inference (neighbor) | < 10% confidence | relationship_inference_confidence |
| Relationship inference (external) | < 1% confidence | relationship_inference_confidence |
| Group membership inference | ≤ 1/k (k-anonymity) | group_membership_inference |
| Timing entropy | > 4 bits | H(actual_send_time | observed_traffic) |
| Activity detection | ± 10% of base rate | P(user_active | traffic) |
Tests instantiate adversary observers and measure inference confidence against these bounds.
9. Cover Traffic
Cover traffic is an optional enhancement layered on mandatory flow-budget enforcement:
- Adaptive: Matches real usage patterns (e.g., 20 messages/hour during work hours)
- Group-leveraged: Groups naturally provide steady traffic rates
- Scheduled slots: Real messages inserted into fixed intervals
- Indistinguishable: Only recipient can distinguish real from cover by decryption attempt
Target: P(real | observed) ≈ 0.5
10. Hub Node Mitigation
Hub nodes with high connectivity observe metadata for many relationships:
| Mitigation | Mechanism |
|---|---|
| Route selection | Minimize fraction observed by any single node |
| Hub tracking | System identifies high-degree nodes |
| Privacy routing | Users can avoid hubs at cost of longer routes |
| Per-hop budgets | Bound forwarding rate per context |
| Decoy envelopes | Optional dummy traffic |
11. Implementation Requirements
11.1 Key Derivation
- Use HKDF with domain separation
- Path:
(account_root, app_id, context_label, "aura.key.derive.v1") - Never reuse keys across contexts
Verified by: Aura.Proofs.KeyDerivation
11.2 Envelope Format
- Fixed-size with random padding
- Encrypted and authenticated
- Onion-routed through multiple hops
- Rtags rotate on negotiated schedule
11.3 Guard Chain
- All transport calls pass through FlowGuard
- Charge failure branches locally with no packet emitted
- Multi-hop forwarding attaches and validates per-hop receipts
Verified by: authorization.qnt (chargeBeforeSend, spentWithinLimit)
11.4 Secure Storage
- Use platform secure storage (Keychain, Secret Service, Keystore)
- Never store keys in plaintext files
- Audit logs for security-critical operations in journal
11.5 Error-Channel Privacy Requirements
- Guard failures must return bounded, typed errors only.
- Error payloads must not include raw context payload, peer identity material, or decrypted content.
- Remote peers must not infer whether failure came from capability checks, budget checks, or local storage faults beyond allowed protocol-level status codes.
12. Verification Coverage
This contract's guarantees are formally verified: Canonical invariant names are indexed in Project Structure. Canonical proof and coverage status are indexed in Verification Coverage Report.
| Property | Tool | Location |
|---|---|---|
| Flow budget monotonicity | Lean | Aura.Proofs.FlowBudget |
| Key derivation isolation | Lean | Aura.Proofs.KeyDerivation |
| Context isolation | Lean | Aura.Proofs.ContextIsolation |
| Guard chain ordering | Quint | authorization.qnt |
| Budget invariants | Quint | authorization.qnt, transport.qnt |
| Epoch validity | Quint | epochs.qnt |
See Verification Coverage Report for metrics and Aura Formal Verification for the Quint-Lean correspondence map.
13. References
Distributed Systems Contract covers safety, liveness, and consistency.
Theoretical Model covers the formal calculus and semilattice laws.
Aura System Architecture describes runtime layering and the guard chain.
Authorization covers CapGuard, FlowGuard, and Biscuit integration.
Transport and Information Flow documents transport semantics and receipts.
Relational Contexts documents cross-authority state and context isolation.
Verification Coverage Report tracks formal verification status.
14. Implementation References
| Component | Location |
|---|---|
| Guard chain | crates/aura-guards/src/guards/ |
| Flow budget | crates/aura-protocol/src/flow_budget/ |
| Context isolation | crates/aura-relational/src/privacy/ |
| Privacy testing | crates/aura-testkit/src/privacy/ |
| Transport patterns | crates/aura-transport/ |
Distributed Systems Contract
This contract specifies Aura's distributed systems model. It defines the safety, liveness, and consistency guarantees provided by the architecture. It also documents the synchrony assumptions, latency expectations, and adversarial capabilities the system tolerates. This contract complements Privacy and Information Flow Contract, which focuses on metadata and privacy budgets. Together these contracts define the full set of invariants protocol authors must respect.
Formal verification of these properties uses Quint model checking (verification/quint/) and Lean 4 theorem proofs (verification/lean/). See Verification Coverage Report for current status.
1. Scope
The contract applies to the following aspects of the system.
Effect handlers and protocols operate within the 8-layer architecture described in Aura System Architecture. Journals and reducers are covered by this contract. The journal specification appears in Authority and Identity and Journal. Aura Consensus is documented in Consensus.
Relational contexts and rendezvous flows fall under this contract. Relational contexts are specified in Relational Contexts. Transport semantics appear in Transport and Information Flow. Rendezvous flows are detailed in Rendezvous Architecture. Shared notation appears in Theoretical Model.
1.1 Terminology Alignment
This contract uses shared terminology from Theoretical Model.
- Consensus role terms:
witnessfor consensus attestation,signerfor FROST share holders - Social-role terms:
Member,Participant,Moderator - Access terms:
AccessLevel(Full,Partial,Limited) - Topology terms:
1-hopandn-hoppaths
1.2 Assumptions
- Cryptographic primitives remain secure at configured parameters.
- Partial synchrony eventually holds after GST, with bounded
Δ_net. - Honest participants execute the guard chain in the required order.
- Anti-entropy exchange eventually delivers missing facts to connected peers.
1.3 Non-goals
- This contract does not provide global linearizability across all operations.
- This contract does not guarantee progress during permanent partitions.
- This contract does not guarantee metadata secrecy without privacy controls defined in Privacy and Information Flow Contract.
2. Safety Guarantees
2.1 Journal CRDT Properties
Facts merge via set union and never retract. Journals satisfy the semilattice laws:
- Commutativity:
merge(j1, j2) ≡ merge(j2, j1) - Associativity:
merge(merge(j1, j2), j3) ≡ merge(j1, merge(j2, j3)) - Idempotence:
merge(j, j) ≡ j
Reduction is deterministic. Identical fact sets produce identical states. No two facts share the same nonce within a namespace (InvariantNonceUnique).
Verified by: Aura.Proofs.Journal, journal/core.qnt
2.2 Charge-Before-Send
Every transport observable is preceded by CapGuard, FlowGuard, and JournalCoupler. See Runtime and Authorization. No packet is emitted without a successful charge. Guard evaluation is pure over a prepared snapshot and yields commands that an interpreter executes, so the chain never blocks on embedded I/O.
Flow budgets satisfy monotonicity: charging never increases available budget (monotonic_decrease). Charging the exact remaining amount results in zero budget (exact_charge).
Verified by: Aura.Proofs.FlowBudget, authorization.qnt, transport.qnt
2.3 Consensus Agreement
For any pair (cid, prestate_hash) there is at most one commit fact (InvariantUniqueCommitPerInstance). Fallback gossip plus FROST signatures prevent divergent commits. Byzantine witnesses cannot force multiple commits for the same instance.
Commits require threshold participation (InvariantCommitRequiresThreshold). Equivocating witnesses are excluded from threshold calculations (InvariantEquivocatorsExcluded).
Verified by: Aura.Proofs.Consensus.Agreement, consensus/core.qnt
2.3.1 Runtime Byzantine Admission
Before consensus-backed ceremonies execute, Aura validates runtime theorem-pack capability profiles and rejects missing requirements at admission time. This prevents silent downgrade of BFT assumptions.
Capability keys are mapped to threat-model assumptions as follows:
byzantine_envelope: Byzantine threshold/fault model assumptions (f < n/3), authenticated participants, evidence validity, conflict exclusion.termination_bounded: bounded progress/termination assumptions for long-running protocol lanes.mixed_determinism: declared native/wasm determinism envelope and conformance constraints.reconfiguration: coherent delegation/link assumptions for live topology changes.
Successful admissions are recorded as ByzantineSafetyAttestation payloads attached to consensus outputs (CommitFact, DKG transcript commits). Admission mismatches fail closed and surface redacted capability references for debugging.
2.3.2 Quantitative Termination Bounds
Category C choreography executions are bounded by deterministic step budgets derived from Telltale weighted measures:
W = 2 * sum(depth(local_type)) + sum(buffer_sizes)max_steps = ceil(k_sigma * W * budget_multiplier)
Aura enforces this budget per execution and fails with BoundExceeded when the bound is crossed. This removes wall-clock coupling from safety/liveness enforcement and keeps bound checks replay-deterministic across native and wasm conformance lanes.
2.3.3 Threshold Parameterization
Consensus parameterization uses n participants, threshold t, and Byzantine budget f.
Threshold signatures require t distinct valid shares.
Safety requires f < t.
Byzantine quorum-style assumptions additionally require f < n/3.
Profiles must declare which bound is active for a given ceremony.
2.3.4 Runtime Parity Guarantees and Limits
Aura uses telltale parity lanes to compare runtime artifacts under declared conformance envelopes. These checks provide implementation conformance signals. They do not replace domain theorem claims from Quint and Lean.
Guarantees:
- parity lanes use canonical surfaces (
observable,scheduler_step,effect) - mismatch reports include first mismatch location for deterministic triage
- strict and envelope-bounded classifications are explicit in report artifacts
Limits:
- parity checks validate observed runtime behavior, not full state-space reachability
- parity checks depend on scenario and seed coverage
- parity success does not imply new consensus or CRDT theorem coverage
2.4 Evidence CRDT
The evidence system tracks votes and equivocations as a grow-only CRDT:
- Monotonicity: Votes and equivocator sets only grow under merge
- Commit preservation:
mergepreserves existing commit facts - Semilattice laws: Evidence merge is commutative, associative, and idempotent
Verified by: Aura.Proofs.Consensus.Evidence, consensus/core.qnt
2.5 Equivocation Detection
The system detects witnesses who vote for conflicting results:
- Soundness: Detection only reports actual equivocation (no false positives)
- Completeness: All equivocations are detectable given sufficient evidence
- Honest safety: Honest witnesses are never falsely accused
Types like HasEquivocated and HasEquivocatedInSet exclude conflicting shares from consensus. See Consensus.
Verified by: Aura.Proofs.Consensus.Equivocation, consensus/adversary.qnt
2.6 FROST Threshold Signatures
Threshold signatures satisfy binding and consistency properties:
- Share binding: Shares are cryptographically bound to
(consensus_id, result_id, prestate_hash) - Threshold requirement: Aggregation requires at least k shares from distinct signers
- Session consistency: All shares in an aggregation have the same session
- Determinism: Same shares always produce the same signature
Verified by: Aura.Proofs.Consensus.Frost, consensus/frost.qnt
2.7 Context Isolation
Messages scoped to ContextId never leak into other contexts. Contexts may be explicitly bridged through typed protocols only. See Theoretical Model. Each authority maintains separate journals per context to enforce this isolation.
Verified by: transport.qnt (InvariantContextIsolation)
2.8 Transport Layer
Beyond context isolation, transport satisfies:
- Flow budget non-negativity: Spent never exceeds limit (
InvariantFlowBudgetNonNegative) - Sequence monotonicity: Message sequence numbers strictly increase (
InvariantSequenceMonotonic) - Fact backing: Every sent message has a corresponding journal fact (
InvariantSentMessagesHaveFacts)
Verified by: transport.qnt
2.9 Deterministic Reduction Order
Commitment tree operations resolve conflicts using the stable ordering described in Authority and Identity. This ordering is derived from the cryptographic identifiers and facts stored in the journal. Conflicts are always resolved in the same way across all replicas.
2.10 Receipt Chain
Multi-hop forwarding requires signed receipts. Downstream peers reject messages lacking a chain rooted in their relational context. See Transport and Information Flow. This prevents unauthorized message propagation.
3. Protocol-Specific Guarantees
3.1 DKG and Resharing
Distributed key generation and resharing satisfy:
- Threshold bounds:
1 ≤ t ≤ nwhere t is threshold and n is participant count - Phase consistency: Commitment counts match protocol phase
- Share timing: Shares distributed only after commitment verification
Verified by: keys/dkg.qnt, keys/resharing.qnt
3.2 Invitation Flows
Invitation lifecycle satisfies authorization invariants:
- Sender authority: Only sender can cancel an invitation
- Receiver authority: Only receiver can accept or decline
- Single resolution: No invitation resolved twice
- Terminal immutability: Terminal status (accepted/declined/cancelled/expired) is permanent
- Fact backing: Accepted invitations have corresponding journal facts
- Ceremony gating: Ceremonies only initiated for accepted invitations
Verified by: invitation.qnt
3.3 Epoch Validity
Epochs enforce temporal boundaries:
- Receipt validity window: Receipts only valid within their epoch
- Replay prevention: Old epoch receipts cannot be replayed in new epochs
Verified by: epochs.qnt
3.4 Cross-Protocol Safety
Concurrent protocol execution (e.g., Recovery∥Consensus) satisfies:
- No deadlock: Interleaved execution always makes progress
- Revocation enforcement: Revoked devices excluded from all protocols
Verified by: interaction.qnt
4. Liveness Guarantees
4.1 Fast-Path Consensus
Fast-path consensus completes in 2×Δ_net (two message delays) when all witnesses are online. Responses are gathered synchronously before committing.
4.2 Fallback Consensus
Fallback consensus eventually completes under partial synchrony with bounded message delays. The fallback timeout is T_fallback = 2×Δ_net in formal verification (implementations may use up to 3×Δ_net for margin). Gossip ensures progress if a majority of witnesses re-transmit proposals. See Consensus for timeout configuration.
Verified by: Aura.Proofs.Consensus.Liveness, consensus/liveness.qnt
4.3 Anti-Entropy
Journals converge under eventual delivery. Periodic syncs or reorder-resistant CRDT merges reconcile fact sets even after partitions. Vector clocks accurately track causal dependencies (InvariantVectorClockConsistent). Authorities exchange their complete fact journals with neighbors to ensure diverged state is healed.
Verified by: journal/anti_entropy.qnt
4.4 Rendezvous
Offer and answer envelopes flood gossip neighborhoods. Secure channels can be established as long as at least one bidirectional path remains between parties. See Rendezvous Architecture.
4.5 Flow Budgets
Flow-budget progress is conditional.
If derived limit(ctx, peer) > 0 in a future epoch and epoch updates converge, FlowGuard eventually grants headroom.
Budget exhaustion remains temporary only under these assumptions.
Liveness requires that each authority eventually receives messages from its immediate neighbors. This is the eventual delivery assumption. Liveness also requires that clocks do not drift unboundedly. This is necessary for epoch rotation and receipt expiry.
5. Time System
Aura uses a unified TimeStamp with domain-specific comparison:
- Reflexivity:
compare(policy, t, t) = eq - Transitivity:
compare(policy, a, b) = lt ∧ compare(policy, b, c) = lt → compare(policy, a, c) = lt - Privacy: Physical time hidden when
ignorePhysical = true
Time variants include PhysicalClock (wall time), LogicalClock (vector/Lamport), OrderClock (opaque ordering tokens), and Range (validity windows).
Verified by: Aura.Proofs.TimeSystem
6. Synchrony and Timing Model
Aura assumes partial synchrony. There exists a bound Δ_net on message delay and processing time once the network stabilizes (Global Stabilization Time, GST). This bound is possibly unknown before stabilization occurs.
| Parameter | Value | Notes |
|---|---|---|
T_fallback | 2×Δ_net (verification) / 2-3×Δ_net (production) | Fallback consensus timeout |
| Gossip interval | 250-500ms | Periodic message exchange |
| GST model | Unknown until stabilization | Partial synchrony assumption |
Before stabilization, timeouts may be conservative. Handlers must tolerate jitter but assume eventual delivery of periodic messages.
Epoch rotation relies on loosely synchronized clocks. The journal serves as the source of truth. Authorities treat an epoch change as effective once they observe it in the journal. This design avoids hard synchronization requirements.
7. Latency Expectations
| Operation | Typical Bound (Δ_net assumptions) |
|---|---|
| Threshold tree update (fast path) | ≤ 2×Δ_net (two message delays) |
| Fallback consensus | ≤ T_fallback after GST |
| Rendezvous offer propagation | O(log N) gossip hops |
| FlowGuard charge | Single local transaction (<10 ms) |
| Anti-entropy reconciliation | k × gossip period (k depends on fanout) |
8. Adversarial Model
8.1 Network Adversary
A network adversary controls a subset of transport links. It can delay or drop packets but cannot break cryptography. It cannot forge receipts without FlowGuard grants. Receipts are protected by signatures and epoch binding.
The network adversary may compromise up to f authorities per consensus instance within the active threshold profile.
Safety requires the bounds declared in §2.3.3 (f < t, and f < n/3 when that profile is selected).
8.2 Byzantine Witness
A Byzantine witness may equivocate during consensus. The system detects equivocation via the evidence CRDT (see §2.5). Equivocators are excluded from threshold calculations.
A Byzantine witness cannot cause multiple commits. Threshold signature verification rejects tampered results. The t of t-of-n threshold signatures prevents this attack. Honest majority can always commit (InvariantHonestMajorityCanCommit).
Verified by: Aura.Proofs.Consensus.Adversary, consensus/adversary.qnt
8.3 Malicious Relay
A malicious relay may drop or delay envelopes. It cannot forge flow receipts because receipts require cryptographic signatures. It cannot read payloads because of context-specific encryption.
Downstream peers detect misbehavior via missing receipts or inconsistent budget charges. The transport layer detects relay failures automatically.
8.4 Device Compromise
A compromised device reveals its share and journal copy. It cannot reconstitute the account without meeting the branch policy. Recovery relies on relational contexts as described in Relational Contexts.
Device compromise is recoverable because the threshold prevents a single device from acting unilaterally. Guardians can revoke the compromised device and issue a new one. Compromised nonces are excluded from future consensus (InvariantCompromisedNoncesExcluded).
9. Consistency Model
Journals eventually converge after replicas exchange all facts. This is eventual consistency. Authorities that have seen the same fact set arrive at identical states.
Operations guarded by Aura Consensus achieve operation-scoped agreement. Once a commit fact is accepted, all honest replicas agree on that operation result. See Consensus. This does not define a single global linearizable log for all operations.
Causal delivery is not enforced at the transport layer. Choreographies enforce ordering via session types instead. See Multi-party Session Types and Choreography.
Each authority's view of its own journal is monotone. Once it observes a fact locally, it will never un-see it. This is monotonic read-after-write consistency.
10. Failure Handling
Timeouts trigger fallback consensus. See Consensus for T_fallback guidelines. Fallback consensus allows the system to make progress during temporary network instability.
Partition recovery relies on anti-entropy. Authorities merge fact sets when connectivity returns. The journal is the single source of truth for state.
Budget exhaustion causes local blocking. Protocols must implement backoff or wait for epoch rotation. Budgets are described in Privacy and Information Flow Contract.
Guard-chain failures return local errors. These errors include AuthorizationDenied, InsufficientBudget, and JournalCommitFailed. Protocol authors must handle these errors without leaking information. Proper error handling is critical for security.
10.1 Error-Channel Privacy Requirements
- Runtime errors must use bounded enums and redacted payloads.
- Error paths must not include plaintext message content, raw capability tokens, or cross-context identifiers.
- Remote peers may observe protocol-level status outcomes only, not internal guard-stage diagnostics.
11. Deployment Guidance
Configure witness sets using the parameter bounds declared in §2.3.3 for the active ceremony profile.
Tune gossip fanout and timeout parameters based on observed round-trip times and network topology. Conservative parameters ensure liveness under poor conditions.
Monitor receipt acceptance rates, consensus backlog, and budget utilization. See Distributed Maintenance Architecture for monitoring guidance. Early detection of synchrony violations prevents cascading failures.
12. Verification Coverage
This contract's guarantees are formally verified: Canonical invariant names are indexed in Project Structure. Canonical proof and coverage status are indexed in Verification Coverage Report.
| Layer | Tool | Location |
|---|---|---|
| Model checking | Quint + Apalache | verification/quint/ |
| Theorem proofs | Lean 4 | verification/lean/ |
| Differential testing | Rust + Lean oracle | crates/aura-testkit/ |
| ITF conformance | Quint traces | verification/quint/traces/ |
See Verification Coverage Report for metrics and Aura Formal Verification for the Quint-Lean correspondence map.
13. References
Aura System Architecture describes runtime layering.
Authorization describes guard chain ordering.
Theoretical Model covers the formal calculus and semilattice laws.
Authority and Identity documents reduction ordering.
Journal and Distributed Maintenance Architecture cover fact storage and convergence.
Relational Contexts documents cross-authority state.
Consensus describes fast path and fallback consensus.
Transport and Information Flow documents transport semantics.
Authorization covers CapGuard and FlowGuard sequencing.
Verification Coverage Report tracks formal verification status.
Cryptography
This document describes the cryptographic architecture in Aura. It defines layer responsibilities, code organization patterns, security invariants, and compliance requirements for cryptographic operations.
1. Overview
Aura's cryptographic architecture follows the 8-layer system design with strict separation of concerns.
- Layer 1 (
aura-core): Type wrappers, trait definitions, pure functions - Layer 3 (
aura-effects): Production implementations with real crypto libraries - Layer 8 (
aura-testkit): Mock implementations for deterministic testing
This separation ensures that cryptographic operations are auditable, testable, and maintainable. Security review focuses on a small number of files rather than scattered usage throughout the codebase.
2. Layer Responsibilities
2.1 Layer 1: aura-core
The aura-core crate provides cryptographic foundations without direct side effects.
Type wrappers live in crates/aura-core/src/crypto/ed25519.rs.
#![allow(unused)] fn main() { pub struct Ed25519SigningKey(pub [u8; 32]); pub struct Ed25519VerifyingKey(pub [u8; 32]); pub struct Ed25519Signature(pub [u8; 64]); }
These wrappers use fixed-size arrays for type safety and delegate to ed25519_dalek internally. They expose a stable API independent of the underlying library. They enable future algorithm migration without changing application code. They provide type safety across crate boundaries.
Effect trait definitions live in crates/aura-core/src/effects/. The CryptoCoreEffects trait inherits from RandomCoreEffects and provides core cryptographic operations.
#![allow(unused)] fn main() { #[async_trait] pub trait CryptoCoreEffects: RandomCoreEffects + Send + Sync { // Key derivation async fn hkdf_derive(&self, ikm: &[u8], salt: &[u8], info: &[u8], output_len: u32) -> Result<Vec<u8>, CryptoError>; async fn derive_key(&self, master_key: &[u8], context: &KeyDerivationContext) -> Result<Vec<u8>, CryptoError>; // Ed25519 signatures async fn ed25519_generate_keypair(&self) -> Result<(Vec<u8>, Vec<u8>), CryptoError>; async fn ed25519_sign(&self, message: &[u8], private_key: &[u8]) -> Result<Vec<u8>, CryptoError>; async fn ed25519_verify(&self, message: &[u8], signature: &[u8], public_key: &[u8]) -> Result<bool, CryptoError>; // Utility methods fn is_simulated(&self) -> bool; fn crypto_capabilities(&self) -> Vec<String>; fn constant_time_eq(&self, a: &[u8], b: &[u8]) -> bool; fn secure_zero(&self, data: &mut [u8]); } }
The CryptoExtendedEffects trait provides additional operations with default implementations that return errors:
#![allow(unused)] fn main() { #[async_trait] pub trait CryptoExtendedEffects: CryptoCoreEffects + Send + Sync { // Unified signing API async fn generate_signing_keys(&self, threshold: u16, max_signers: u16) -> Result<SigningKeyGenResult, CryptoError>; async fn generate_signing_keys_with(&self, method: KeyGenerationMethod, threshold: u16, max_signers: u16) -> Result<SigningKeyGenResult, CryptoError>; async fn sign_with_key(&self, message: &[u8], key_package: &[u8], mode: SigningMode) -> Result<Vec<u8>, CryptoError>; async fn verify_signature(&self, message: &[u8], signature: &[u8], public_key_package: &[u8], mode: SigningMode) -> Result<bool, CryptoError>; // FROST threshold signatures async fn frost_generate_keys(&self, threshold: u16, max_signers: u16) -> Result<FrostKeyGenResult, CryptoError>; async fn frost_generate_nonces(&self, key_package: &[u8]) -> Result<Vec<u8>, CryptoError>; async fn frost_create_signing_package(&self, message: &[u8], nonces: &[Vec<u8>], participants: &[u16], public_key_package: &[u8]) -> Result<FrostSigningPackage, CryptoError>; async fn frost_sign_share(&self, signing_package: &FrostSigningPackage, key_share: &[u8], nonces: &[u8]) -> Result<Vec<u8>, CryptoError>; async fn frost_aggregate_signatures(&self, signing_package: &FrostSigningPackage, signature_shares: &[Vec<u8>]) -> Result<Vec<u8>, CryptoError>; async fn frost_verify(&self, message: &[u8], signature: &[u8], group_public_key: &[u8]) -> Result<bool, CryptoError>; async fn ed25519_public_key(&self, private_key: &[u8]) -> Result<Vec<u8>, CryptoError>; // Symmetric encryption async fn chacha20_encrypt(&self, plaintext: &[u8], key: &[u8; 32], nonce: &[u8; 12]) -> Result<Vec<u8>, CryptoError>; async fn chacha20_decrypt(&self, ciphertext: &[u8], key: &[u8; 32], nonce: &[u8; 12]) -> Result<Vec<u8>, CryptoError>; async fn aes_gcm_encrypt(&self, plaintext: &[u8], key: &[u8; 32], nonce: &[u8; 12]) -> Result<Vec<u8>, CryptoError>; async fn aes_gcm_decrypt(&self, ciphertext: &[u8], key: &[u8; 32], nonce: &[u8; 12]) -> Result<Vec<u8>, CryptoError>; // Key rotation and conversion async fn frost_rotate_keys(&self, old_shares: &[Vec<u8>], old_threshold: u16, new_threshold: u16, new_max_signers: u16) -> Result<FrostKeyGenResult, CryptoError>; async fn convert_ed25519_to_x25519_public(&self, ed25519_public_key: &[u8]) -> Result<[u8; 32], CryptoError>; async fn convert_ed25519_to_x25519_private(&self, ed25519_private_key: &[u8]) -> Result<[u8; 32], CryptoError>; } pub trait CryptoEffects: CryptoCoreEffects + CryptoExtendedEffects {} }
The core trait provides key derivation and Ed25519 signatures. The extended trait provides unified signing that routes between single-signer and threshold modes, FROST threshold operations, symmetric encryption, and key conversion. Hashing is not included because it is a pure operation. Use aura_core::hash::hash() for synchronous hashing instead.
The RandomCoreEffects trait provides cryptographically secure random number generation.
#![allow(unused)] fn main() { #[async_trait] pub trait RandomCoreEffects: Send + Sync { async fn random_bytes(&self, len: usize) -> Vec<u8>; async fn random_bytes_32(&self) -> [u8; 32]; async fn random_u64(&self) -> u64; } #[async_trait] pub trait RandomExtendedEffects: RandomCoreEffects + Send + Sync { async fn random_range(&self, min: u64, max: u64) -> u64; async fn random_uuid(&self) -> Uuid; } }
The core trait provides basic random generation. The extended trait adds range and UUID generation with default implementations. All randomness flows through these traits for testability and simulation.
Pure functions in crates/aura-core/src/crypto/ implement hash functions, signature verification, and other deterministic operations. These require no side effects and can be called directly.
2.2 Layer 3: aura-effects
The aura-effects crate contains the only production implementations that directly use cryptographic libraries.
The production handler lives in crates/aura-effects/src/crypto.rs. RealCryptoHandler can operate with OS entropy (production) or with a seed (deterministic testing). It implements all methods from CryptoCoreEffects and RandomCoreEffects.
The following direct imports are allowed in Layer 3:
ed25519_dalekfrost_ed25519chacha20poly1305aes_gcmgetrandomrand_core::OsRngrand_chachahkdf
2.3 Threshold Lifecycle (K1/K2/K3) and Transcript Binding
Aura separates key generation from agreement/finality:
- K1: Single-signer (Ed25519). No DKG required.
- K2: Dealer-based DKG. A trusted coordinator produces dealer packages.
- K3: Consensus-finalized DKG. The BFT-DKG transcript is finalized by consensus.
Transcript hashing uses the following rules:
- All DKG transcripts are hashed using canonical DAG‑CBOR encoding.
DkgTranscriptCommitbindstranscript_hash,prestate_hash, andoperation_hash.
Dealer packages (K2) follow these rules:
- Deterministic dealer packages are acceptable in trusted settings.
- Dealer packages must include encrypted shares for every participant.
BFT‑DKG (K3) follows these rules:
- A transcript is only usable once consensus finalizes the commit fact.
- All K3 ceremonies must reference the finalized transcript (hash or blob ref).
2.4 Layer 8: aura-testkit
The aura-testkit crate provides mock implementations for deterministic testing.
The mock handler lives in crates/aura-testkit/src/stateful_effects/crypto.rs. MockCryptoHandler uses a seed and counter for deterministic behavior, enabling reproducible test results, simulation of edge cases, and faster test execution.
3. Code Organization Patterns
3.1 Correct Usage
Application code should use effect traits.
#![allow(unused)] fn main() { async fn authenticate<E: CryptoCoreEffects>(effects: &E, private_key: &[u8], data: &[u8]) -> Result<Vec<u8>, CryptoError> { effects.ed25519_sign(data, private_key).await } }
This pattern ensures all cryptographic operations flow through the effect system. The generic constraint allows both production and mock implementations.
Application code should use aura-core wrappers for type safety.
#![allow(unused)] fn main() { use aura_core::crypto::ed25519::{Ed25519SigningKey, Ed25519VerifyingKey, Ed25519Signature}; fn verify_authority(key: &Ed25519VerifyingKey, data: &[u8], sig: &Ed25519Signature) -> Result<(), AuraError> { key.verify(data, sig) } }
The wrapper types provide a stable API and enable algorithm migration without changing application code.
3.2 Incorrect Usage
Do not import crypto libraries directly in application code.
#![allow(unused)] fn main() { // INCORRECT: Direct crypto library import use ed25519_dalek::{SigningKey, VerifyingKey}; // BAD // INCORRECT: Direct randomness use rand_core::OsRng; // BAD (outside Layer 3) let mut rng = OsRng; }
Direct imports bypass the effect system and break testability. They also scatter cryptographic usage throughout the codebase.
3.3 Randomness Patterns
All randomness should flow through RandomCoreEffects.
#![allow(unused)] fn main() { async fn generate_nonce<E: RandomCoreEffects>(effects: &E) -> [u8; 12] { let bytes = effects.random_bytes(12).await; bytes.try_into().expect("12 bytes") } }
For encryption in feature crates, use parameter injection.
#![allow(unused)] fn main() { pub struct EncryptionRandomness { nonce: [u8; 12], padding: Vec<u8>, } pub fn encrypt_with_randomness(data: &[u8], key: &[u8], randomness: EncryptionRandomness) -> Vec<u8> { // Deterministic given the randomness parameter } }
This pattern enables deterministic testing by externalizing randomness.
4. Allowed Locations
The following locations may directly use cryptographic libraries.
| Location | Allowed Libraries | Purpose |
|---|---|---|
aura-core/src/crypto/* | ed25519_dalek, frost_ed25519 | Type wrappers |
aura-core/src/types/authority.rs | ed25519_dalek | Authority trait types |
aura-effects/src/* | All crypto libs | Production handlers |
aura-effects/src/noise.rs | snow | Noise Protocol implementation |
aura-testkit/* | All crypto libs | Test infrastructure |
**/tests/*, *_test.rs | OsRng | Test-only randomness |
#[cfg(test)] modules | OsRng | Test-only randomness |
5. Security Invariants
The cryptographic architecture maintains these invariants.
- All production crypto operations flow through
RealCryptoHandler - Security review focuses on Layer 3 handlers, not scattered usage
- All crypto is controllable via mock handlers for testing
- Private keys remain in wrapper types, not exposed as raw bytes
- Production randomness comes from OS entropy via
OsRng
6. Signing Modes
Aura supports two signing modes to handle different account configurations.
6.1 SigningMode Enum
#![allow(unused)] fn main() { pub enum SigningMode { SingleSigner, // Standard Ed25519 for 1-of-1 Threshold, // FROST for m-of-n where m >= 2 } }
The SingleSigner mode is used for new user onboarding with single device accounts. It is also used for bootstrap scenarios before multi-device setup and for simple personal accounts that do not need threshold security.
The Threshold mode is used for multi-device accounts such as 2-of-3 or 3-of-5 configurations. It is also used for guardian-protected accounts and group decisions requiring multiple approvals.
6.2 Why Two Modes?
FROST requires at least 2 signers. For 1-of-1 configurations, we use standard Ed25519.
Ed25519 uses the same curve as FROST and produces compatible signatures for verification, however single signature does not require nonce coordination or aggregation.
6.3 API Usage
The unified API handles mode selection automatically.
#![allow(unused)] fn main() { // Generate keys - mode is determined by threshold let keys = crypto.generate_signing_keys(threshold, max_signers).await?; // keys.mode == SingleSigner if (1, 1), Threshold otherwise // Sign with the key package (single-signer only) let signature = crypto.sign_with_key(message, &key_package, keys.mode).await?; // Verify the signature let valid = crypto.verify_signature(message, &signature, &keys.public_key_package, keys.mode).await?; }
For threshold mode where m >= 2, the sign_with_key() method returns an error. Threshold signing requires the full FROST protocol flow with nonce coordination across signers.
6.4 Storage Separation
Single-signer and threshold keys use different storage paths.
signing_keys/<authority>/<epoch>/1 # SingleSignerKeyPackage (Ed25519)
frost_keys/<authority>/<epoch>/<index> # FROST KeyPackage
The storage location is managed by SecureStorageEffects. The authority is the AuthorityId in display format. The epoch is the current key epoch. The index is the signer index for FROST keys.
7. FROST and Threshold Signatures
Aura provides a unified threshold signing architecture for all scenarios requiring m-of-n signatures where m >= 2.
7.1 Architecture Layers
The trait definition lives in aura-core/src/effects/threshold.rs.
#![allow(unused)] fn main() { #[async_trait] pub trait ThresholdSigningEffects: Send + Sync { async fn bootstrap_authority(&self, authority: &AuthorityId) -> Result<PublicKeyPackage, ThresholdSigningError>; async fn sign(&self, context: SigningContext) -> Result<ThresholdSignature, ThresholdSigningError>; async fn threshold_config(&self, authority: &AuthorityId) -> Option<ThresholdConfig>; async fn threshold_state(&self, authority: &AuthorityId) -> Option<ThresholdState>; async fn has_signing_capability(&self, authority: &AuthorityId) -> bool; async fn public_key_package(&self, authority: &AuthorityId) -> Option<PublicKeyPackage>; async fn rotate_keys(&self, authority: &AuthorityId, new_threshold: u16, new_total_participants: u16, participants: &[ParticipantIdentity]) -> Result<(u64, Vec<Vec<u8>>, PublicKeyPackage), ThresholdSigningError>; async fn commit_key_rotation(&self, authority: &AuthorityId, new_epoch: u64) -> Result<(), ThresholdSigningError>; async fn rollback_key_rotation(&self, authority: &AuthorityId, failed_epoch: u64) -> Result<(), ThresholdSigningError>; } }
The trait provides methods for bootstrapping authorities, signing operations, querying configurations and state, checking capabilities, and key rotation lifecycle management.
Context types live in aura-core/src/threshold/context.rs.
#![allow(unused)] fn main() { pub struct SigningContext { pub authority: AuthorityId, pub operation: SignableOperation, pub approval_context: ApprovalContext, } pub enum SignableOperation { TreeOp(TreeOp), RecoveryApproval { target: AuthorityId, new_root: TreeCommitment }, GroupProposal { group: AuthorityId, action: GroupAction }, Message { domain: String, payload: Vec<u8> }, OTAActivation { ceremony_id: [u8; 32], upgrade_hash: [u8; 32], prestate_hash: [u8; 32], activation_epoch: Epoch, ready: bool }, } pub enum ApprovalContext { SelfOperation, RecoveryAssistance { recovering: AuthorityId, session_id: String }, GroupDecision { group: AuthorityId, proposal_id: String }, ElevatedOperation { operation_type: String, value_context: Option<String> }, } }
The SignableOperation enum defines what is being signed. Its OTA activation variant should be interpreted as scoped activation approval evidence. activation_epoch is meaningful only when the chosen scope actually owns an epoch fence. The ApprovalContext enum provides context for audit and display purposes.
The service implementation lives in aura-agent/src/runtime/services/threshold_signing.rs. ThresholdSigningService manages per-authority signing state and key storage using SecureStorageEffects for key material persistence.
Low-level primitives live in aura-core/src/crypto/tree_signing.rs. This module defines FROST types and pure coordination logic. It re-exports frost_ed25519 types for type safety.
The handler in aura-effects/src/crypto.rs implements FROST key generation and signing. This is the only location with direct frost_ed25519 library calls.
7.2 Serialized Size Invariants (FROST)
Aura treats the postcard serialization of FROST round-one data as canonical and fixed-size. This prevents malleability and makes invalid encodings unrepresentable at the boundary.
SigningNonces(secret) must serialize to exactly 138 bytesSigningCommitments(public) must serialize to exactly 69 bytes
These sizes are enforced in aura-core/src/crypto/tree_signing.rs and mirrored in aura-core/src/constants.rs. If the upstream FROST or postcard encoding changes, update the constants and add/adjust tests to lock in the new canonical sizes.
7.3 Lifecycle Taxonomy (Key Generation vs Agreement)
Aura separates key generation from agreement/finality:
- K1: Local/Single-Signer (no DKG)
- K2: Dealer-Based DKG (trusted coordinator)
- K3: Quorum/BFT-DKG (consensus-finalized transcript)
Agreement modes are orthogonal:
- A1: Provisional (usable immediately, not final)
- A2: Coordinator Soft-Safe (bounded divergence + convergence cert)
- A3: Consensus-Finalized (unique, durable, non-forkable)
Leader selection (lottery/round seed/fixed coordinator) and pipelining are orthogonal optimizations, not agreement modes.
7.4 Usage Pattern
The recommended pattern uses AppCore for high-level operations.
#![allow(unused)] fn main() { // Sign a tree operation let attested_op = app_core.sign_tree_op(&tree_op).await?; // Bootstrap 1-of-1 keys for single-device accounts (uses Ed25519) let public_key = app_core.bootstrap_signing_keys().await?; }
For direct trait usage, import and call the service.
#![allow(unused)] fn main() { use aura_core::effects::ThresholdSigningEffects; let context = SigningContext::self_tree_op(authority, tree_op); let signature = signing_service.sign(context).await?; }
7.5 Design Rationale
The unified trait abstraction enables a consistent interface across multi-device, guardian, and group scenarios. It provides proper audit context via ApprovalContext for UX and logging. It enables testability via mock implementations in aura-testkit. It provides key isolation with secure storage integration.
7.6 FROST Minimum Threshold
FROST requires threshold >= 2. Calling frost_generate_keys(1, 1) returns an error. For single-signer scenarios, use generate_signing_keys(1, 1) which routes to Ed25519 automatically.
8. Future Considerations
8.1 Algorithm Migration
The wrapper pattern enables algorithm migration.
- Update wrappers in
aura-core/src/crypto/ - Update handler in
aura-effects/src/crypto.rs - Application code remains unchanged
8.2 HSM Integration
Hardware Security Module support would require a new HsmCryptoHandler implementing CryptoCoreEffects. Runtime selection between RealCryptoHandler and HsmCryptoHandler would be needed. Application code would require no changes.
See Also
- Effect System for effect trait patterns
- Project Structure for 8-layer architecture
- Effects and Handlers Guide for handler implementation guidance
Identifiers and Boundaries
This reference defines the identifiers that appear in Aura documents. Every other document should reuse these definitions instead of restating partial variants. Each identifier preserves structural privacy by design.
1. Authority Identifiers
| Identifier | Type | Purpose |
|---|---|---|
AuthorityId | Uuid | Journal namespace for an authority. Does not leak operator or membership metadata. All public keys, commitment trees, and attested operations reduce under this namespace. |
DeviceId | Uuid | Device within a threshold account. Each device holds shares of the root key. Visible only inside the authority namespace. |
LocalDeviceId | u32 | Compact internal device identifier for efficiency. Never appears in cross-authority communication. |
GuardianId | Uuid | Social recovery guardian. Does not reveal the guardian's own authority structure. |
AccountId | Uuid | Legacy identifier being replaced by AuthorityId. Exists for backward compatibility. |
2. Context Identifiers
| Identifier | Type | Purpose |
|---|---|---|
ContextId | Uuid | Relational context or derived subcontext. Opaque on the wire, appears only inside encrypted envelopes and receipts. Never encodes participant lists or roles. Flow budgets and leakage metrics scope to (ContextId, peer) pairs. |
SessionId | Uuid | Choreographic protocol execution instance. Pairs a ContextId with a nonce. Not long-lived; expires when protocol completes or times out. |
DkdContextId | { app_label: String, fingerprint: [u8; 32] } | Deterministic Key Derivation context. Combines application label with fingerprint to scope derived keys across application boundaries. |
3. Communication Identifiers
| Identifier | Type | Purpose |
|---|---|---|
ChannelId | Hash32 | AMP messaging substream scoped under a relational context. Opaque; does not reveal membership or topology. |
RelayId | [u8; 32] | Pairwise communication context derived from X25519 keys. Foundation for RID message contexts. |
GroupId | [u8; 32] | Threshold group communication context derived from group membership. Foundation for GID message contexts. |
MessageContext | enum { Relay, Group, DkdContext } | Unifies the three privacy context types. Enforces mutual exclusivity; cross-partition routing requires explicit bridge operations. |
ConnectionId | Uuid | Network connection identifier with privacy-preserving properties. Does not encode endpoint information. |
4. Content Identifiers
| Identifier | Type | Purpose |
|---|---|---|
ContentId | { hash: Hash32, size: Option<u64> } | Hash of canonical content bytes (files, documents, encrypted payloads, CRDT state). Any party can verify integrity by hashing and comparing. |
ChunkId | { hash: Hash32, sequence: Option<u32> } | Storage-layer chunk identifier. Multiple chunks may comprise a single ContentId. Enables content-addressable storage with deduplication. |
Hash32 | [u8; 32] | Raw 32-byte Blake3 cryptographic hash. Foundation for all content addressing. Provides collision and preimage resistance. |
DataId | String | Stored data chunk identifier with type prefixes (data:uuid, encrypted:uuid). Enables heterogeneous storage addressing. |
5. Journal Identifiers
| Identifier | Type | Purpose |
|---|---|---|
FactId | u64 | Lightweight reference to journal facts. Enables efficient queries without cloning fact content. Internal to journal layer. |
EventId | Uuid | Event identifier within the effect API system. Used in audit logs and debugging. |
OperationId | Uuid | Operation tracking identifier. |
6. Consensus Identifiers
| Identifier | Type | Purpose |
|---|---|---|
ConsensusId | Hash32 | Consensus instance identifier derived from prestate hash, operation hash, and nonce. Binds operations to prestates through hash commitment. See Consensus. |
FrostParticipantId | NonZeroU16 | Threshold signing participant. Must be non-zero for FROST protocol compatibility. Scoped to signing sessions. |
7. Social Topology Identifiers
| Identifier | Type | Purpose |
|---|---|---|
HomeId | [u8; 32] | Home in the urban social topology. Each user resides in exactly one home. See Social Architecture. |
NeighborhoodId | [u8; 32] | Neighborhood (collection of homes with 1-hop link relationships). Enables governance and traversal policies. |
8. Tree Identifiers
| Identifier | Type | Purpose |
|---|---|---|
LeafId | u32 | Leaf node in the commitment tree. Stable across tree modifications and epoch rotations. See Authority and Identity. |
ProposalId | Hash32 | Snapshot proposal identifier. Enables proposal deduplication and verification during tree operations. |
9. Accountability Structures
Receipt
Receipt is the accountability record emitted by FlowGuard. It contains context, source authority, destination authority, epoch, cost, nonce, chained hash, and signature. Receipts prove that upstream participants charged their budget before forwarding. No receipt includes device identifiers or user handles.
Fields: ctx: ContextId, src: AuthorityId, dst: AuthorityId, epoch: Epoch, cost: FlowCost, nonce: FlowNonce, prev: Hash32, sig: ReceiptSig.
See Transport and Information Flow for receipt propagation.
10. Derived Keys
Aura derives per-context cryptographic keys from reduced account state and ContextId. Derived keys never surface on the wire. They only exist inside effect handlers to encrypt payloads, generate commitment tree secrets, or run DKD.
The derivation inputs never include device identifiers. Derived keys inherit the privacy guarantees of AuthorityId and ContextId. The derivation function uses derive(account_root, app_id, context_label) and is deterministic but irreversible.
See Also
Authority and Identity describes the authority model in detail. Relational Contexts covers cross-authority relationships. Transport and Information Flow documents receipt chains and flow budgets. Social Architecture defines homes and neighborhoods.
Authority and Identity
This document describes the architecture of authorities and identity in Aura. It defines the authority model, account state machine, commitment tree structure, and relational identity model. Identity emerges through shared contexts rather than as a global property of keys.
1. Authority Model
An authority is a cryptographic actor represented by a public key. An authority hides its internal structure. An authority may contain one or more devices. An authority is the smallest unit that can sign facts or capabilities.
An authority has an internal journal namespace. The journal namespace stores facts relevant to that authority. The authority derives its state from deterministic reduction of that fact set.
Devices are not exclusive to a single authority. A single device may hold threshold shares for multiple authorities at the same time. Joining a new authority adds signing capability for that authority without removing any existing authority memberships.
1.1 Authority Types and Membership Terms
Aura uses three authority types:
User: individual authorityHome: group authority that acceptsUserauthorities as membersNeighborhood: group authority that acceptsHomeauthorities as members
Home membership terminology is:
Member: authority in the home's threshold authority setParticipant: authority granted access to the home without threshold membershipModerator: optional designation granted to a member for moderation operations
AuthorityId (see Identifiers and Boundaries) selects the journal namespace associated with the authority. The identifier does not encode structure or membership. The authority publishes its current public key and root commitment inside its own journal.
Authorities can interact with other authorities through Relational Contexts. These interactions do not change the authority's internal structure. The authority remains isolated except where relational state is explicitly shared.
2. Account Authorities
An account authority is an authority with long term state. An account maintains device membership through its commitment tree. An account contains its own journal namespace. An account evolves through attested operations stored as facts.
The commitment tree defines device membership and threshold policies. The journal stores facts that represent signed tree operations. The reduction function reconstructs the canonical tree state from the accumulated fact set.
An account authority exposes a single public key derived from the commitment tree root. The authority never exposes device structure. The account state changes only when an attested operation appears in the journal.
Aura supports multiple key generation methods for account authorities. K1 uses single-signer local key generation. K2 uses dealer-based DKG with a trusted coordinator. K3 uses quorum/BFT DKG with consensus-finalized transcripts. These are orthogonal to agreement modes (A1 provisional, A2 coordinator soft-safe, A3 consensus-finalized). Durable shared authority state must be finalized via A3.
#![allow(unused)] fn main() { pub trait Authority: Send + Sync { fn authority_id(&self) -> AuthorityId; fn public_key(&self) -> Ed25519VerifyingKey; fn root_commitment(&self) -> Hash32; async fn sign_operation(&self, operation: &[u8]) -> Result<Signature>; fn get_threshold(&self) -> u16; fn active_device_count(&self) -> usize; } }
The Authority trait provides the external interface for authority operations. The trait exposes only the public key, root commitment, and signing capabilities. The internal device structure remains hidden from external consumers.
An account authority derives context specific keys using deterministic key derivation. These derived authorities represent application scoped identities. See Effects and Handlers Guide for implementation examples.
3. Account State Machine
The TreeState structure represents the materialized state of an account at a specific epoch.
#![allow(unused)] fn main() { pub struct TreeState { pub epoch: Epoch, pub root_commitment: TreeHash32, pub branches: BTreeMap<NodeIndex, BranchNode>, pub leaves: BTreeMap<LeafId, LeafNode>, leaf_commitments: BTreeMap<LeafId, TreeHash32>, tree_topology: TreeTopology, branch_signing_keys: BTreeMap<NodeIndex, BranchSigningKey>, } }
The epoch field is monotonically increasing. The root_commitment is the hash of the entire tree structure. The branches map stores branch nodes by index. The leaves map stores leaf nodes by ID.
The leaf_commitments map caches leaf commitment hashes. The tree_topology tracks parent-child relationships. The branch_signing_keys map stores FROST group public keys for verification.
For the authority-internal reducer implementation (AuthorityTreeState), topology is explicitly materialized as ordered binary edges with parent pointers. Active leaves are canonically sorted by LeafId, paired in stable order, and assigned deterministic branch indices (NodeIndex(0) root, then breadth-first). This allows path-local commitment recomputation for non-structural updates while preserving deterministic replay semantics.
TreeState is derived state and is never stored directly in the journal. It is computed on-demand from the OpLog via the reduction function.
#![allow(unused)] fn main() { pub struct TreeStateSummary { epoch: Epoch, commitment: Hash32, threshold: u16, device_count: u32, } }
The TreeStateSummary provides a public view with only epoch, commitment, threshold, and device count. This summary hides internal device structure for external consumers while the full TreeState is used for internal operations.
4. Commitment Tree Structure
A commitment tree contains branch nodes and leaf nodes. A leaf node represents a device or guardian inside the account. A branch node represents a subpolicy with threshold requirements. The root node defines the account-level threshold policy.
#![allow(unused)] fn main() { pub enum NodeKind { Leaf(LeafNode), Branch, } }
This type defines leaf and branch variants. The Leaf variant contains a LeafNode with device information. The Branch variant is a marker indicating an internal node.
#![allow(unused)] fn main() { pub struct LeafNode { pub leaf_id: LeafId, pub device_id: DeviceId, pub role: LeafRole, pub public_key: LeafPublicKey, pub meta: LeafMetadata, } pub enum LeafRole { Device, Guardian, } }
The LeafNode structure stores device information required for threshold signing. The leaf_id is a stable identifier across tree modifications. The device_id identifies the device within the authority. The role distinguishes devices from guardians.
#![allow(unused)] fn main() { pub struct BranchNode { pub node: NodeIndex, pub policy: Policy, pub commitment: TreeHash32, } }
The BranchNode structure stores policy data for internal nodes. The node field is the branch index. The policy defines the threshold requirement. The commitment is the cryptographic hash of the branch structure.
#![allow(unused)] fn main() { pub struct BranchSigningKey { pub group_public_key: [u8; 32], pub key_epoch: Epoch, } }
A BranchSigningKey stores the FROST group public key for threshold signing at a branch node. The key_epoch tracks when the key was established via DKG. Signing keys are updated when membership changes under the branch or when policy changes affect the signing group.
5. Tree Topology
The TreeTopology structure tracks parent-child relationships for efficient navigation.
#![allow(unused)] fn main() { pub struct TreeTopology { parent_pointers: BTreeMap<NodeIndex, NodeIndex>, children_pointers: BTreeMap<NodeIndex, BTreeSet<NodeIndex>>, leaf_parents: BTreeMap<LeafId, NodeIndex>, root_node: Option<NodeIndex>, } }
The parent_pointers map links nodes to their parents. The children_pointers map links parents to their children. The leaf_parents map links leaves to their parent branches. The root_node tracks the root of the tree.
This structure enables efficient path-to-root traversal and affected node computation during commitment updates.
graph TD
R["Root Commitment<br/>(policy)"]
L1["Branch<br/>(threshold)"]
L2["Branch<br/>(threshold)"]
A["Leaf A<br/>(device)"]
B["Leaf B<br/>(device)"]
C["Leaf C<br/>(guardian)"]
D["Leaf D<br/>(device)"]
R --> L1
R --> L2
L1 --> A
L1 --> B
L2 --> C
L2 --> D
Each node commitment is computed over its ordered children plus its policy metadata. The root commitment is used in key derivation and verification. Leaves represent device shares. Branches represent threshold policies.
6. Policies
A branch node contains a threshold policy. A policy describes the number of required signatures for authorization.
#![allow(unused)] fn main() { pub enum Policy { Any, Threshold { m: u16, n: u16 }, All, } }
The Any policy accepts one signature from any device under that branch. The Threshold policy requires m signatures out of n devices. The All policy requires all devices under the branch.
Policies form a meet semilattice where the meet operation selects the stricter of two policies.
#![allow(unused)] fn main() { impl Policy { pub fn required_signers(&self, child_count: usize) -> Result<u16, PolicyError> { match self { Policy::Any => Ok(1), Policy::All => Ok(child_count as u16), Policy::Threshold { m, n } => { if child_count as u16 != *n { return Err(PolicyError::ChildCountMismatch { expected: *n, actual: child_count as u16 }); } Ok(*m) } } } } }
The required_signers method derives the concrete threshold from the policy given the current child count. It returns an error if the child count does not match the policy's expected total. This is used during signature verification to determine how many signers must have participated.
7. Tree Operations
Tree operations modify the commitment tree. Each operation references a parent epoch and parent commitment. Each operation is signed through threshold signing.
#![allow(unused)] fn main() { pub enum TreeOpKind { AddLeaf { leaf: LeafNode, under: NodeIndex }, RemoveLeaf { leaf: LeafId, reason: u8 }, ChangePolicy { node: NodeIndex, new_policy: Policy }, RotateEpoch { affected: Vec<NodeIndex> }, } }
The AddLeaf operation inserts a new leaf under a branch. The RemoveLeaf operation removes an existing leaf with a reason code. The ChangePolicy operation updates the policy of a branch. The RotateEpoch operation increments the epoch for affected nodes and invalidates derived context keys.
#![allow(unused)] fn main() { pub struct TreeOp { pub parent_epoch: Epoch, pub parent_commitment: TreeHash32, pub op: TreeOpKind, pub version: u16, } }
The TreeOp structure binds an operation to its parent state. The parent_epoch and parent_commitment prevent replay attacks. The version field enables protocol upgrades.
#![allow(unused)] fn main() { pub struct AttestedOp { pub op: TreeOp, pub agg_sig: Vec<u8>, pub signer_count: u16, } }
The agg_sig field stores the FROST aggregate signature. The signer_count records how many devices contributed. The signature validates under the parent root commitment. Devices refuse to sign if the local tree state does not match.
8. Tree Operation Verification
Tree operations use a two-phase verification model that separates cryptographic verification from state consistency checking.
8.1 Cryptographic Verification
The verify_attested_op function performs cryptographic signature checking only.
#![allow(unused)] fn main() { pub fn verify_attested_op( attested: &AttestedOp, signing_key: &BranchSigningKey, threshold: u16, current_epoch: Epoch, ) -> Result<(), VerificationError>; }
Verification checks that the signer count meets the threshold requirement. It computes the binding message including the group public key. It verifies the FROST aggregate signature. Verification is self-contained and can be performed offline.
8.2 State Consistency Check
The check_attested_op function performs full verification plus TreeState consistency.
#![allow(unused)] fn main() { pub fn check_attested_op<S: TreeStateView>( state: &S, attested: &AttestedOp, target_node: NodeIndex, ) -> Result<(), CheckError>; }
Check verifies the operation cryptographically. It ensures the signing key exists for the target node. It validates that the operation epoch and parent commitment match state.
8.3 TreeStateView Trait
#![allow(unused)] fn main() { pub trait TreeStateView { fn get_signing_key(&self, node: NodeIndex) -> Option<&BranchSigningKey>; fn get_policy(&self, node: NodeIndex) -> Option<&Policy>; fn child_count(&self, node: NodeIndex) -> usize; fn current_epoch(&self) -> Epoch; fn current_commitment(&self) -> TreeHash32; } }
This trait abstracts over TreeState for verification. It enables verification without a direct dependency on the journal crate.
8.4 Binding Message Security
#![allow(unused)] fn main() { pub fn compute_binding_message( attested: &AttestedOp, current_epoch: Epoch, group_public_key: &[u8; 32], ) -> Vec<u8>; }
The binding message contains a domain separator, parent epoch and commitment for replay prevention, protocol version, current epoch, group public key, and serialized operation content. Including the group public key ensures signatures are bound to a specific signing group.
8.5 Error Types
Verification produces two error categories. VerificationError covers cryptographic issues: missing signing keys, insufficient signers, signature failures, epoch mismatches, and parent commitment mismatches. CheckError covers state consistency issues: verification failures, key epoch mismatches, and missing nodes or policies.
9. Reduction and Conflict Resolution
The account journal is a join semilattice. It stores AttestedOp facts. All replicas merge fact sets using set union. The commitment tree state is recovered using deterministic reduction.
Reduction applies the following rules. Group operations by parent state using ParentKey (epoch + commitment tuple). Select a single winner using a deterministic ordering based on operation hash. Discard superseded operations. Apply winners in parent epoch order.
Conflicts arise when multiple operations reference the same parent epoch and commitment. The reduction algorithm resolves conflicts using a total order on operations. The winning operation applies. Losing operations are ignored.
The reduction algorithm builds a DAG from operations, performs topological sort with tie-breaking, and applies operations sequentially to build the final TreeState.
Conflict resolution uses operation hash as the tie-breaker. When multiple operations share the same parent, they are sorted by hash and the maximum hash wins. This ensures deterministic winner selection across all replicas.
graph TD
OpLog["OpLog<br/>(CRDT OR-set of AttestedOp)"]
Verify["verify_attested_op()"]
Check["check_attested_op()"]
Reduce["reduce()"]
TreeState["TreeState<br/>(derived on-demand)"]
OpLog -->|cryptographic| Verify
Verify -->|state consistency| Check
Check -->|valid ops| Reduce
Reduce -->|derives| TreeState
This diagram shows the data flow from OpLog to TreeState. Operations are verified cryptographically first. Then they are checked for state consistency. Valid operations are reduced to produce the materialized TreeState.
10. Epochs and Derived Keys
The epoch is an integer stored in the tree state that scopes deterministic key derivation. Derived keys depend on the current epoch. Rotation invalidates previous derived keys. The RotateEpoch operation updates the epoch for selected subtrees.
Epochs also scope flow budgets and context presence tickets. All context identities must refresh when the epoch changes.
Derived context keys bind relationship data to the account state. The deterministic key derivation function uses the commitment tree root commitment and epoch. This ensures that all devices compute the same context keys. Derived keys do not modify the tree state.
11. Operators and Devices
An operator controls an authority by operating its devices. An operator is not represented in the protocol. Devices are internal to the authority and hold share material required for signing.
Devices produce partial signatures during threshold signing. The operator coordinates these partial signatures to produce the final signature.
The commitment tree manages device membership. The AddLeaf and RemoveLeaf operations modify device presence in the authority. Device identifiers do not appear outside the authority. No external party can link devices to authorities.
When a device is enrolled or replaced, Aura performs session delegation alongside tree updates. Runtime delegation transfers active protocol session ownership to the receiving device authority and records a SessionDelegation fact for auditability. The commitment tree update (AddLeaf/RemoveLeaf) remains the source of truth for membership, while delegation preserves in-flight protocol continuity during migration.
12. Relational Identity Model
Aura defines identity as contextual and relational. Identity exists only inside a specific relationship and does not exist globally. Authorities represent cryptographic actors rather than people. Identity emerges when two authorities form a shared context.
A shared context exists inside a relational context. A relational context stores relational facts that define how two authorities relate. Profile data may appear in a relational context if both authorities choose to share it. This profile data is scoped to that context.
ContextId (see Identifiers and Boundaries) identifies a relational context. It does not encode membership. It does not reveal which authorities participate. The context stores only the relational facts required by the participants.
Identity inside a context may include nickname suggestions or other profile attributes. These values are private to that context. See Identifiers and Boundaries for context isolation mechanisms. Nicknames (local mappings) allow a device to associate multiple authorities with a single local contact.
13. Authority Relationships
Authorities interact through relational contexts to create shared state. Relational contexts do not modify authority structure. Each relational context has its own journal. Facts in the relational context reference commitments of participating authorities.
Authorities may form long lived or ephemeral relationships. These relationships do not affect global identity. The authority model ensures that each relationship remains isolated.
graph TD
A[Authority A] --> C(Context)
B[Authority B] --> C(Context)
This diagram shows two authorities interacting through a relational context. The context holds the relational facts that define the relationship. Neither authority exposes its internal structure to the other.
14. Privacy and Isolation
Authorities reveal no internal structure and contexts do not reveal participants. Identity exists only where authorities choose to share information. Nicknames remain local to devices. There is no global identifier for people or devices.
Every relationship is private to its participants. Each relationship forms its own identity layer. Authorities can operate in many contexts without linking those contexts together.
15. Interaction with Consensus
Consensus is used when a tree operation must have strong agreement across a committee. Consensus produces a commit fact containing a threshold signature. This fact becomes an attested operation in the journal.
Consensus is used when multiple devices must agree on the same prestate. Simple device-initiated changes may use local threshold signing. The account journal treats both cases identically.
Consensus references the root commitment and epoch of the account. This binds the commit fact to the current state.
16. Security Properties
The commitment tree provides fork resistance. Devices refuse to sign under mismatched parent commitments. The reduction function ensures that all replicas converge. Structural opacity hides device membership from external parties.
The threshold signature scheme prevents unauthorized updates. All operations must be signed by the required number of devices. An attacker cannot forge signatures or bypass policies.
The tree design ensures that no external party can identify device structure. The only visible values are the epoch and the root commitment.
17. Implementation Architecture
17.1 Critical Invariants
The implementation enforces these rules.
- TreeState is never stored in the journal. It is always derived on-demand via reduction.
- OpLog is the only persisted tree data. All tree state can be recovered from the operation log.
- Reduction is deterministic across all replicas. The same OpLog always produces the same TreeState.
- DeviceId is authority-internal only. It is never exposed in public APIs.
17.2 Data Flow
graph TD
A["Tree Operation Initiated"]
B["TreeOperationProcessor validates"]
C["Convert to AttestedOp fact"]
D["Journal stores fact"]
E["verify_attested_op (crypto)"]
F["check_attested_op (state)"]
G["reduce processes valid facts"]
H["apply_operation builds TreeState"]
I["TreeState materialized"]
A --> B
B --> C
C --> D
D --> E
E --> F
F --> G
G --> H
H --> I
This diagram shows the complete lifecycle of a tree operation from initiation to materialization.
17.3 Key Module Locations
Implementation files are in aura-journal/src/commitment_tree/.
state.rs: TreeState structure and TreeStateView implementationreduction.rs: Deterministic reduction algorithmoperations.rs: TreeOperationProcessor and TreeStateQueryapplication.rs: Operation application to TreeStatecompaction.rs: Garbage collection of superseded operationsattested_ops.rs: AttestedOp fact handling
Verification code is in aura-core/src/tree/verification.rs. Type definitions are in aura-core/src/tree/types.rs and aura-core/src/tree/policy.rs.
See Also
- Journal System for fact semantics and reduction flows
- Relational Contexts for cross-authority relationship management
- Consensus for threshold signing and agreement
- Identifiers and Boundaries for context isolation mechanisms
- Cryptographic Architecture for FROST and signing modes
- Effects and Handlers Guide for practical implementation patterns
- Privacy and Information Flow for privacy guarantees
Journal
This document describes the journal architecture and state reduction system in Aura. It explains how journals implement CRDT semantics, how facts are structured, and how reduction produces deterministic state for account authorities and relational contexts. It describes the reduction pipeline, flow budget semantics, and integration with the effect system. It defines the invariants that ensure correctness. See Maintenance for the end-to-end snapshot and garbage collection pipeline.
Hybrid Journal Model (Facts + Capabilities)
Aura’s journal state is a composite of:
- Fact Journal: the canonical, namespaced CRDT of immutable facts.
- Capability Frontier: the current capability lattice for the namespace.
- Composite journal view: the runtime carries facts and capabilities together when evaluating effects.
The fact journal is stored and merged as a semilattice. Capabilities are refined via meet. The runtime always treats these as orthogonal dimensions of state.
1. Journal Namespaces
Aura maintains a separate journal namespace for each authority and each relational context. A journal namespace stores all facts relevant to the entity it represents. A namespace is identified by an AuthorityId (see Authority and Identity) or a ContextId and no namespace shares state with another. Identifier definitions appear in Identifiers and Boundaries.
A journal namespace evolves through fact insertion. Facts accumulate monotonically. No fact is removed except through garbage collection rules that preserve logical meaning.
#![allow(unused)] fn main() { pub struct Journal { pub namespace: JournalNamespace, pub facts: BTreeSet<Fact>, } pub enum JournalNamespace { Authority(AuthorityId), Context(ContextId), } }
This type defines a journal as a namespaced set of facts. The namespace identifies whether this journal tracks an authority's commitment tree or a relational context. The journal is a join semilattice under set union where merging two journals produces a new journal containing all facts from both inputs. Journals with different namespaces cannot be merged.
2. Fact Model
Facts represent immutable events or operations that contribute to the state of a namespace. Facts have ordering tokens, timestamps, and content. Facts do not contain device identifiers used for correctness.
#![allow(unused)] fn main() { pub struct Fact { pub order: OrderTime, pub timestamp: TimeStamp, pub content: FactContent, } pub enum FactContent { AttestedOp(AttestedOp), Relational(RelationalFact), Snapshot(SnapshotFact), RendezvousReceipt { envelope_id: [u8; 32], authority_id: AuthorityId, timestamp: TimeStamp, signature: Vec<u8>, }, } }
The order field provides an opaque, privacy-preserving total order for deterministic fact ordering in the BTreeSet. The timestamp field provides semantic time information for application logic. Facts implement Ord via the OrderTime field. Do not use TimeStamp for cross-domain indexing or total ordering; use OrderTime or consensus/session sequencing.
This model supports account operations, relational context operations, snapshots, and rendezvous receipts. Each fact is self contained. Facts are validated before insertion into a namespace.
2.2 Protocol-Level vs Domain-Level Relational Facts
RelationalFact has only two variants:
Protocol(ProtocolRelationalFact): Protocol-level facts that must live inaura-journalbecause reduction semantics depend on them.Generic { .. }: Extensibility hook for domain facts (DomainFact+FactReducer).
Criteria for ProtocolRelationalFact (all must hold):
- Reduction-coupled: the fact directly affects core reduction invariants in
reduce_context()(not just a view). - Cross-domain: the fact’s semantics are shared across multiple protocols or layers.
- Non-derivable: the state cannot be reconstructed purely via
FactReducer+RelationalFact::Generic.
If a fact does not meet all three criteria, it must be implemented as a domain fact and stored via RelationalFact::Generic.
Enforcement:
- All protocol facts are defined in
crates/aura-journal/src/protocol_facts.rs. - Any new protocol fact requires a doc update in this section and a matching reduction rule.
2.1 Domain Fact Contract (Checklist + Lint)
Domain facts are the extensibility mechanism for Layer 2 crates. Every domain fact must follow this contract to ensure cross-replica determinism and schema stability:
- Type ID: define a
*_FACT_TYPE_IDconstant (unique, registered incrates/aura-agent/src/fact_types.rs). - Schema version: specify a schema version (via
#[domain_fact(schema_version = N)]or*_FACT_SCHEMA_VERSION). - Canonical encoding: use
#[derive(DomainFact)]or explicitencode_domain_fact/VersionedMessagehelpers. - Context derivation: declare
context/context_fnforDomainFactor implement a stablecontext_id()derivation. - Reducer registration: provide a
FactReducerand register it in the central registry (crates/aura-agent/src/fact_registry.rs).
3. Semilattice Structure
Journals use a join semilattice. The semilattice uses set union as the join operator with partial order defined by subset inclusion. The journal never removes facts during merge. Every merge operation increases or preserves the fact set.
The join semilattice ensures convergence across replicas. Any two replicas that exchange facts eventually converge to identical fact sets. All replicas reduce the same fact set to the same state.
The merge operation asserts namespace equality, unions the fact sets, and returns the combined journal. The result is monotonic and convergent.
4. Reduction Pipeline
Aura maintains two replicated state machines. Account journals describe commitment trees for authorities. Relational context journals describe cross-authority coordination. Both use the same fact-only semilattice and deterministic reducers.
flowchart TD
A[Ledger append] --> B[Journal merge];
B --> C[Group by parent];
C --> D[Resolve conflicts via max hash];
D --> E[Apply operations in topological order];
E --> F[Recompute commitments bottom-up];
Ledger append writes facts durably. Journal merge unions the fact set. Reducers group operations by parent commitment, resolve conflicts deterministically using max hash tie-breaking, and then apply winners in topological order. The final step recomputes commitments bottom-up which downstream components treat as the canonical state.
4.1 Fact Production
Account operations originate from local threshold signing or Aura Consensus. Relational context operations always run through Aura Consensus because multiple authorities must agree on the prestate. Each successful operation produces an AttestedOp fact. Receipts that must be retained for accountability are stored as RendezvousReceipt facts scoped to the context that emitted them. Implementations must only emit facts after verifying signatures and parent commitments.
4.2 Determinism Invariants
The reduction pipeline maintains strict determinism:
- No HashMap iteration: All maps use BTreeMap for consistent ordering
- No system time: OrderTime tokens provide opaque ordering
- No floating point: All arithmetic uses exact integer/fixed-point
- Pure functions only: Reducers have no side effects
These properties are verified by test_reduction_determinism() which confirms all fact permutations produce identical state.
5. Account Journal Reduction
Account journals store attested operations for commitment tree updates. Reduction computes a TreeStateSummary (epoch, commitment, threshold, device count) from the fact set. See Authority and Identity for structure details. The summary is a lightweight public view that hides internal device structure. For the full internal representation with branches, leaves, and topology, see TreeState in aura-journal::commitment_tree.
Internally, AuthorityTreeState materializes explicit branch topology to support incremental commitment updates. It keeps ordered branch children, branch/leaf parent pointers, and branch depth metadata. The topology is deterministic: active leaves are sorted by LeafId, paired in stable order, and materialized with NodeIndex(0) as root followed by breadth-first branch assignment. This guarantees identical branch structure for identical active leaf sets across replicas.
Commitment recomputation for non-structural mutations is path-local. Aura marks affected branches dirty, collects their paths to root, then recomputes only those branches bottom-up. Structural mutations (AddLeaf, RemoveLeaf) currently use deterministic rebuild of topology and branch commitments (correctness-first baseline). Merkle proof paths are derived from the same materialized topology so proof generation and commitment caches share one source of truth.
Maintenance note: do not change topology construction or branch indexing rules without updating replay fixtures and root commitment fixtures. Determinism depends on stable leaf sort order, stable pair composition, and stable branch index assignment.
Reduction follows these steps:
- Identify all
AttestedOpfacts. - Group operations by their referenced parent state (epoch + commitment).
- For concurrent operations (same parent), select winner using max hash tie-breaking:
H(op)comparison. - Apply winners in topological order respecting parent dependencies.
- Recompute commitments bottom-up after each operation.
The max hash conflict resolution (max_by_key(hash_op)) ensures determinism by selecting a single winner from concurrent operations. The result is a single TreeStateSummary for the account derived by extracting AttestedOp facts and applying them in deterministic order.
6. RelationalContext Journal Reduction
Relational contexts store relational facts. These facts reference authority commitments. Reduction produces a RelationalState that captures the current relationship between authorities.
#![allow(unused)] fn main() { pub struct RelationalState { pub bindings: Vec<RelationalBinding>, pub flow_budgets: BTreeMap<(AuthorityId, AuthorityId, u64), u64>, pub leakage_budget: LeakageBudget, pub channel_epochs: BTreeMap<ChannelId, ChannelEpochState>, } pub struct RelationalBinding { pub binding_type: RelationalBindingType, pub context_id: ContextId, pub data: Vec<u8>, } }
This structure represents the reduced relational state. It contains relational bindings, flow budget tracking between authorities, leakage budget totals for privacy accounting, and AMP channel epoch state for message ratcheting.
Reduction processes the following protocol fact types wrapped in Protocol(...):
GuardianBindingmaps toRelationalBindingfor guardian relationshipsRecoveryGrantcreates recovery permission bindings between authoritiesConsensusstores generic bindings with consensus metadataAmpChannelCheckpointanchors channel state snapshots for AMP messagingAmpProposedChannelEpochBumpandAmpCommittedChannelEpochBumptrack channel epoch transitionsAmpChannelPolicydefines channel-specific policy overrides for skip windowsDkgTranscriptCommitstores consensus-finalized DKG transcriptsConvergenceCertrecords soft-safe convergence certificatesReversionFacttracks explicit reversion eventsRotateFactmarks lifecycle rotation or upgrade events
Domain-specific facts use Generic { context_id, binding_type, binding_data } and are reduced by registered FactReducer implementations.
Reduction verifies that relational facts reference valid authority commitments and applies them in dependency order.
7. Flow Budgets
Flow budgets track message sending allowances between authorities. The FlowBudget structure uses semilattice semantics for distributed convergence:
#![allow(unused)] fn main() { pub struct FlowBudget { pub limit: u64, pub spent: u64, pub epoch: Epoch, } impl FlowBudget { pub fn merge(&self, other: &Self) -> Self { Self { limit: self.limit.min(other.limit), spent: self.spent.max(other.spent), epoch: if self.epoch.value() >= other.epoch.value() { self.epoch } else { other.epoch }, } } } }
The spent field uses join-semilattice (max) because charges only increase. The limit field uses meet-semilattice (min) because the most restrictive limit wins. The epoch field advances monotonically. Spent resets on epoch rotation.
Flow budget tracking operates at the runtime layer via FlowBudgetManager. The RelationalState includes a flow_budgets map for CRDT-based replication of budget state across replicas.
8. Receipts and Accountability
Receipts reference the current epoch commitment so reducers can reject stale receipts automatically. The RendezvousReceipt variant in FactContent stores accountability proofs (envelope ID, authority, timestamp, signature). Receipts are stored as relational facts scoped to the emitting context. This coupling ensures that receipt validity follows commitment tree epochs.
9. Snapshots and Garbage Collection
Snapshots summarize all prior facts. A snapshot fact contains a state hash, the list of superseded facts, and a sequence number. A snapshot establishes a high water mark. Facts older than the snapshot can be pruned.
#![allow(unused)] fn main() { pub struct SnapshotFact { pub state_hash: Hash32, pub superseded_facts: Vec<OrderTime>, pub sequence: u64, } }
Garbage collection removes pruned facts while preserving logical meaning. Pruning does not change the result of reduction. The GC algorithm uses safety margins to prevent premature pruning:
- Default skip window: 1024 generations
- Safety margin:
skip_window / 2 - Pruning boundary:
max_generation - (2 * skip_window) - safety_margin
Helper functions (compute_checkpoint_pruning_boundary, can_prune_checkpoint, can_prune_proposed_bump) determine what can be safely pruned based on generation boundaries.
10. Journal Effects Integration
The effect system provides journal operations through JournalEffects. This trait handles persistence, merging, and flow budget tracking:
#![allow(unused)] fn main() { #[async_trait] pub trait JournalEffects: Send + Sync { async fn merge_facts(&self, target: Journal, delta: Journal) -> Result<Journal, AuraError>; async fn refine_caps(&self, target: Journal, refinement: Journal) -> Result<Journal, AuraError>; async fn get_journal(&self) -> Result<Journal, AuraError>; async fn persist_journal(&self, journal: &Journal) -> Result<(), AuraError>; async fn get_flow_budget(&self, context: &ContextId, peer: &AuthorityId) -> Result<FlowBudget, AuraError>; async fn update_flow_budget(&self, context: &ContextId, peer: &AuthorityId, budget: &FlowBudget) -> Result<FlowBudget, AuraError>; async fn charge_flow_budget(&self, context: &ContextId, peer: &AuthorityId, cost: FlowCost) -> Result<FlowBudget, AuraError>; } }
The effect layer writes facts to persistent storage. Replica synchronization loads facts through effect handlers into journal memory. The effect layer guarantees durability but does not affect CRDT merge semantics.
At the choreography runtime boundary, journal coupling is always driven by guard/effect commands. Generated choreography annotations and runtime guard checks emit EffectCommand values, and JournalCoupler ensures journal commits occur before transport observables. In VM execution, AuraVmEffectHandler emits envelopes, but journal writes still flow through the same EffectInterpreter + JournalEffects path.
11. AttestedOp Structure
AttestedOp exists in two layers with different levels of detail:
Layer 1 (aura-core) - Full operation metadata:
#![allow(unused)] fn main() { pub struct AttestedOp { pub op: TreeOp, pub agg_sig: Vec<u8>, pub signer_count: u16, } pub struct TreeOp { pub parent_epoch: Epoch, pub parent_commitment: TreeHash32, pub op: TreeOpKind, pub version: u16, } }
Layer 2 (aura-journal) - Flattened for journal storage:
#![allow(unused)] fn main() { pub struct AttestedOp { pub tree_op: TreeOpKind, pub parent_commitment: Hash32, pub new_commitment: Hash32, pub witness_threshold: u16, pub signature: Vec<u8>, } }
The aura-core version includes epoch and version for full verification. The aura-journal version includes computed commitments for efficient reduction.
12. Invariants
The journal and reduction architecture satisfy several invariants:
- Convergence: All replicas reach the same state when they have the same facts
- Idempotence: Repeated merges or reductions do not change state
- Determinism: Reduction produces identical output for identical input across all replicas
- No HashMap iteration: Uses BTreeMap for deterministic ordering
- No system time: Uses OrderTime tokens for ordering
- No floating point: All arithmetic is exact
These invariants guarantee correct distributed behavior. They also support offline operation with eventual consistency. They form the foundation for Aura's account and relational context state machines.
13. Fact Validation Pipeline
Every fact inserted into a journal must be validated before merge. The following steps outline the required checks and the effect traits responsible for each fact type:
13.1 AttestedOp Facts
Checks
- Verify the threshold signature (
agg_sig) using the two-phase verification model fromaura-core::tree::verification:verify_attested_op(): Cryptographic signature check againstBranchSigningKeystored in TreeStatecheck_attested_op(): Full verification plus state consistency (epoch, parent commitment)
- Ensure the referenced parent state exists locally; otherwise request missing facts.
- Confirm the operation is well-formed (e.g.,
AddLeafindexes a valid parent node).
See Tree Operation Verification for details on the verify/check model and binding message security.
Responsible Effects
CryptoEffectsfor FROST signature verification viaverify_attested_op().JournalEffectsfor parent lookup, state consistency viacheck_attested_op(), and conflict detection.StorageEffectsto persist the fact once validated.
13.2 Relational Facts
Checks
- Validate that each authority commitment referenced in the fact matches the current reduced state (
AuthorityState::root_commitment). - Verify Aura Consensus proofs if present (guardian bindings, recovery grants).
- Enforce application-specific invariants (e.g., no duplicate guardian bindings).
Responsible Effects
AuthorizationEffects/RelationalEffectsfor context membership checks.CryptoEffectsfor consensus proof verification.JournalEffectsfor context-specific merge.
13.3 FlowBudget Facts
Checks
- Ensure
spentdeltas are non-negative and reference the active epoch for the(ContextId, peer)pair. - Reject facts that would decrease the recorded
spent(monotone requirement). - Validate receipt signatures associated with the charge (see
109_transport_and_information_flow.md).
Responsible Effects
FlowBudgetEffects(or FlowGuard) produce the fact and enforce monotonicity before inserting.JournalEffectsgate insertion to prevent stale epochs from updating headroom.
13.4 Snapshot Facts
Checks
- Confirm the snapshot
state_hashmatches the hash of all facts below the snapshot. - Ensure no newer snapshot already exists for the namespace (check
sequencenumber). - Verify that pruning according to the snapshot does not remove facts still referenced by receipts or pending consensus operations.
Responsible Effects
JournalEffectscompute and validate snapshot digests.StorageEffectspersist the snapshot atomically with pruning metadata.
By clearly separating validation responsibilities, runtime authors know which effect handlers must participate before a fact mutation is committed. This structure keeps fact semantics consistent across authorities and contexts.
14. Consistency Metadata Schema
Facts carry consistency metadata for tracking agreement level, propagation status, and acknowledgments. See Operation Categories for the full consistency metadata type definitions.
14.1 Fact Schema Fields
The base Fact structure (section 2) is extended with consistency metadata fields (added via serde defaults for backwards compatibility):
| Field | Type | Purpose |
|---|---|---|
agreement | Agreement | Finalization level (A1 Provisional, A2 SoftSafe, A3 Finalized) |
propagation | Propagation | Anti-entropy sync status (Local, Syncing, Complete, Failed) |
ack_tracked | bool | Opt-in flag for per-peer acknowledgment tracking |
14.2 Ack Storage Table
For facts with ack_tracked = true, acknowledgments are stored in a separate table:
| fact_id | peer_id | acked_at |
|---|---|---|
| msg-001 | alice_authority_id | 2024-01-15T10:30:00Z |
| msg-001 | bob_authority_id | 2024-01-15T10:30:05Z |
14.3 Journal API for Consistency
The Journal API provides methods for acknowledgment tracking: record_ack records an acknowledgment from a peer, get_acks retrieves all acknowledgments for a fact, and gc_ack_tracking garbage collects acknowledgment data based on policy.
See Also
- Database Architecture for query system and Datalog integration
- Operation Categories for consistency metadata types
- Maintenance for snapshot and garbage collection pipeline
- Relational Contexts for context journal structure
- Authority and Identity for commitment tree operations
Authorization
Overview
Aura authorizes every observable action through Biscuit capability evaluation combined with sovereign policy and flow budgets. The authorization pipeline spans AuthorizationEffects, the guard chain, and receipt accounting. This document describes the data flow and integration points.
Biscuit Capability Model
Biscuit tokens encode attenuation chains. Each attenuation step applies additional caveats that shrink authority through meet composition. Aura stores Biscuit material outside the replicated CRDT. Local runtimes evaluate tokens at send time and cache the resulting lattice element for the active ContextId.
Cached entries expire on epoch change or when policy revokes a capability. Policy data always participates in the meet. A token can only reduce authority relative to local policy.
flowchart LR
A[frontier, token] -->|verify signature| B[parsed token]
B -->|apply caveats| C[frontier ∩ caveats]
C -->|apply policy| D[result ∩ policy]
D -->|return| E[Cap element]
This algorithm produces a meet-monotone capability frontier. Step 1 ensures provenance. Steps 2 and 3 ensure evaluation never widens authority. Step 4 feeds the guard chain with a cached outcome.
Guard Chain
Authorization evaluation feeds the transport guard chain. All documents reference this section to avoid divergence.
flowchart LR
A[Send request] --> B[CapGuard]
B --> C[FlowGuard]
C --> D[JournalCoupler]
D --> E[Transport send]
This diagram shows the guard chain sequence. CapGuard performs Biscuit evaluation. FlowGuard charges the budget. JournalCoupler commits facts before transport.
Guard evaluation is pure and synchronous over a prepared GuardSnapshot. CapGuard reads the cached frontier and any inline Biscuit token already present in the snapshot. FlowGuard and JournalCoupler emit EffectCommand items rather than executing I/O directly. An async interpreter executes those commands in production or simulation.
Only after all guards pass does transport emit a packet. Any failure returns locally and leaves no observable side effect. DKG payloads require proportional budget charges before any transport send.
Telltale Integration
Aura uses Telltale runtime admission and VM guard checkpoints. Runtime admission gates whether a runtime profile may execute. VM acquire and release guards gate per-session resource leases inside VM execution. The Aura guard chain remains the authoritative policy and accounting path for application sends.
Failure handling is layered. Admission failure rejects engine startup. VM acquire deny blocks the guarded VM action. Aura guard-chain failure denies transport and returns deterministic effect errors.
Runtime Capability Admission
Aura uses a dedicated admission surface for theorem-pack and runtime capability checks before choreography execution. RuntimeCapabilityEffects in aura-core defines capability inventory queries and admission checks. RuntimeCapabilityHandler in aura-effects stores a boot-time immutable capability snapshot. The aura-protocol::admission module declares protocol requirements and maps them to capability keys.
Current protocol capability keys include byzantine_envelope for consensus ceremony admission, termination_bounded for sync epoch-rotation admission, reconfiguration for dynamic topology transfer paths, and mixed_determinism for cross-target mixed lanes.
Execution order is runtime capability admission first, then VM profile gates, then the Aura guard chain. Admission diagnostics must respect Aura privacy constraints. Production runtime paths must not emit plaintext capability inventory events. Admission failures use redacted capability references.
Failure Handling and Caching
Runtimes cache evaluated capability frontiers per context and predicate with an epoch tag. Cache entries invalidate when journal policy facts change or when the epoch rotates.
CapGuard failures return AuthorizationError::Denied without charging flow or touching the journal. FlowGuard failures return FlowError::InsufficientBudget without emitting transport traffic. JournalCoupler failures surface as JournalError::CommitAborted and instruct the protocol to retry after reconciling journal state.
This isolation keeps the guard chain deterministic and side-channel free.
Biscuit Token Workflow
Biscuit tokens provide cryptographically verifiable, attenuated delegation chains. The typical workflow creates root authority via TokenAuthority::new(authority_id). Issue tokens via authority.create_token(recipient_authority_id). Attenuate for delegation via BiscuitTokenManager::attenuate_read(). Authorize via BiscuitAuthorizationBridge::authorize(&token, operation, &resource_scope).
flowchart TB
A[Root Authority Token] -->|append caveats| B[Device Token<br/>read + write]
B -->|check operation = read| C[Read-Only Token]
C -->|check resource starts_with public/| D[Public Read-Only Token]
This diagram shows token attenuation. Each block appends restrictions to the chain. Attenuation preserves the cryptographic signature chain while reducing authority.
Biscuit tokens are secure through cryptographic signature chains that prevent forgery. They support offline verification without contacting the issuer. Epoch rotation provides revocation by invalidating old tokens.
Guard Chain Integration
Biscuit authorization integrates with the guard chain in three phases. Cryptographic verification calls bridge.authorize(token, operation, resource_scope) to perform Datalog evaluation. Guard evaluation prepares a GuardSnapshot asynchronously and then calls guards.evaluate(&snapshot, &request) synchronously. Effect execution interprets the EffectCommand items from the guard outcome.
If any phase fails, the operation returns an error without observable side effects.
Authorization Scenarios
Biscuit tokens handle all authorization scenarios through cryptographic verification. Local device operations use device tokens with full capabilities. Cross-authority delegation uses attenuated tokens with resource restrictions. Policy enforcement integrates sovereign policy into Datalog evaluation.
API access control uses scoped tokens with operation restrictions. Guardian recovery uses guardian tokens with recovery capabilities. Storage operations use storage-scoped tokens with path restrictions. Relaying and forwarding use context tokens with relay permissions.
Performance and Caching
Biscuit token authorization has predictable performance characteristics. Signature verification requires O(chain length) cryptographic operations. Authorization evaluation takes O(facts × rules) time for Datalog evaluation. Attenuation costs O(1) to append blocks.
Token results are cacheable with epoch-based invalidation. Cache authorization results per authority, token hash, and resource scope. Invalidate cache on epoch rotation or policy update.
Security Model
Cryptographic signature verification prevents token forgery. Epoch scoping limits token lifetime and replay attacks. Attenuation preserves security while growing verification cost proportional to chain length. Root key compromise invalidates all derived tokens.
Authority-based ResourceScope prevents cross-authority access. Local sovereign policy integration provides an additional security layer. Guard chain isolation ensures authorization failures leak no sensitive information.
Implementation References
The Cap type in aura-core/src/domain/journal.rs wraps serialized Biscuit tokens with optional root key storage. The Cap::meet() implementation computes capability intersection. Tokens from the same issuer return the more attenuated token. Tokens from different issuers return bottom.
BiscuitAuthorizationBridge in aura-guards/src/authorization.rs handles guard chain integration. TokenAuthority and BiscuitTokenManager in aura-authorization/src/biscuit_token.rs handle token creation and attenuation. ResourceScope in aura-core/src/types/scope.rs defines authority-centric resource patterns.
See Transport and Information Flow for flow budget details. See Journal for fact commit semantics.
Effect System
Overview
Aura uses algebraic effects to abstract system capabilities. Effect traits define abstract interfaces for cryptography, storage, networking, time, and randomness. Handlers implement these traits with concrete behavior. Context propagation ensures consistent execution across async boundaries.
This document covers effect trait design, handler patterns, and the context model. See Runtime for lifecycle management, service composition, and guard chain execution.
Effect Traits
Aura defines effect traits as abstract interfaces for system capabilities. Core traits expose essential functionality. Extended traits expose optional operations and coordinated behaviors. Each trait is independent and does not assume global state.
Core traits include CryptoCoreEffects, NetworkCoreEffects, StorageCoreEffects, time domain traits, RandomCoreEffects, and JournalEffects. Extended traits include CryptoExtendedEffects, NetworkExtendedEffects, StorageExtendedEffects, RandomExtendedEffects, and system-level traits such as SystemEffects and ChoreographicEffects.
#![allow(unused)] fn main() { #[async_trait] pub trait CryptoCoreEffects: RandomCoreEffects + Send + Sync { async fn ed25519_sign( &self, message: &[u8], private_key: &[u8], ) -> Result<Vec<u8>, CryptoError>; async fn ed25519_verify( &self, message: &[u8], signature: &[u8], public_key: &[u8], ) -> Result<bool, CryptoError>; async fn hkdf_derive( &self, ikm: &[u8], salt: &[u8], info: &[u8], output_len: u32, ) -> Result<Vec<u8>, CryptoError>; } }
This example shows a core effect trait for cryptographic operations. Traits contain async methods for compatibility with async runtimes. Extension traits add optional capabilities without forcing all handlers to implement them. The hash() function is intentionally pure in aura-core::hash rather than an effect because it is deterministic and side-effect-free.
Time Traits
The legacy monolithic TimeEffects trait is replaced by domain-specific traits. PhysicalTimeEffects returns wall-clock time with uncertainty and provides sleep operations. LogicalClockEffects advances and reads causal vector clocks and Lamport scalars. OrderClockEffects produces opaque total order tokens without temporal meaning.
Callers select the domain appropriate to their semantics. Guards and transport use physical time. CRDT operations use logical clocks. Privacy-preserving ordering uses order tokens.
Cross-domain comparisons are explicit via TimeStamp::compare(policy). Total ordering across domains must use OrderTime or consensus sequencing. Direct SystemTime::now() or chrono usage is forbidden outside effect implementations.
Threshold Signing
Aura provides a unified ThresholdSigningEffects trait in aura-core/src/effects/threshold.rs for all threshold signing scenarios. The trait supports multi-device personal signing, guardian recovery approvals, and group operation approvals.
The trait uses a unified SigningContext that pairs a SignableOperation with an ApprovalContext. This design allows the same FROST signing machinery to handle all scenarios with proper audit context. The ThresholdSigningService in aura-agent provides the production implementation.
Key components include ThresholdSigningEffects for async signing operations, lifecycle traits for provisional and consensus modes, and AppCore.sign_tree_op() for high-level signing. See Cryptography for detailed threshold signature architecture.
When to Create Effect Traits
Create new effect traits when abstracting OS or external system integration. Use them when defining domain-specific operations that multiple implementations might provide. They isolate side effects for testing and simulation. They enable deterministic simulation of complex behaviors.
Follow YAGNI principles. Defer abstraction when only one implementation exists. Avoid abstractions that add complexity without clear benefit. Do not abstract without concrete need.
Application-specific effect traits should remain in their application layer. Do not move CliEffects or ConfigEffects from aura-terminal to aura-core when only one implementation exists. The aura-core crate provides infrastructure effects. Application layers compose these into domain-specific abstractions.
Database Effects
Database operations use existing effect traits rather than a dedicated DatabaseEffects layer. JournalEffects in aura-core provides fact insertion for monotone operations. Non-monotone operations use aura-consensus protocols driven by session types and the guard chain.
Reactive queries are handled via QueryEffects and ReactiveEffects. The coordination pattern follows two orthogonal dimensions described in Database Architecture. Authority scope determines single versus cross-authority operations. Agreement level determines monotone versus consensus operations.
Handler Design
Effect handlers implement effect traits. Stateless handlers execute operations without internal state. Stateful handlers coordinate multiple effects or maintain internal caches. Typed handlers implement concrete effect traits. Type-erased handlers allow dynamic dispatch through the effect executor.
Handlers do not store global state. All required inputs flow through method parameters. This avoids hidden dependencies and enables deterministic testing.
Unified Encrypted Storage
Aura uses StorageEffects as the single persistence interface in application code. The production runtime wires StorageEffects through a unified encryption-at-rest wrapper. FilesystemStorageHandler provides raw bytes persistence. RealSecureStorageHandler uses Keychain or TPM for master-key persistence.
EncryptedStorage implements StorageEffects by encrypting and decrypting transparently. It generates or loads the master key on first use. Runtime assembly remains synchronous.
#![allow(unused)] fn main() { use aura_effects::{ EncryptedStorage, EncryptedStorageConfig, FilesystemStorageHandler, RealCryptoHandler, RealSecureStorageHandler, }; use std::sync::Arc; let secure = Arc::new(RealSecureStorageHandler::with_base_path(base_path.clone())); let storage = EncryptedStorage::new( FilesystemStorageHandler::from_path(base_path.clone()), Arc::new(RealCryptoHandler::new()), secure, EncryptedStorageConfig::default(), ); }
This example shows the encryption wrapper assembly. RealCryptoHandler lives in aura-effects and implements CryptoCoreEffects. Storage configuration controls encryption enablement and opaque naming. Application code uses StorageEffects without knowledge of encryption details.
Context Model
The effect system propagates an EffectContext through async tasks. The context carries authority identity, context scope, session identification, execution mode, and metadata. No ambient state exists.
#![allow(unused)] fn main() { pub struct EffectContext { authority_id: AuthorityId, context_id: ContextId, session_id: SessionId, execution_mode: ExecutionMode, metadata: HashMap<String, String>, } }
This structure defines the operation-scoped effect context. The context flows through all effect calls and identifies which authority, context, and session the operation belongs to. The execution_mode controls handler selection for production versus test environments.
Context propagation uses scoped execution. A task local stores the current context. Nested tasks inherit the context. This ensures consistent behavior across async boundaries.
ReactiveEffects Trait
The ReactiveEffects trait provides type-safe signal-based state management. Signals are phantom-typed identifiers that reference reactive state. The phantom type ensures compile-time type safety.
#![allow(unused)] fn main() { pub struct Signal<T> { id: SignalId, _phantom: PhantomData<T>, } #[async_trait] pub trait ReactiveEffects: Send + Sync { async fn read<T>(&self, signal: &Signal<T>) -> Result<T, ReactiveError> where T: Clone + Send + Sync + 'static; async fn emit<T>(&self, signal: &Signal<T>, value: T) -> Result<(), ReactiveError> where T: Clone + Send + Sync + 'static; fn subscribe<T>(&self, signal: &Signal<T>) -> SignalStream<T> where T: Clone + Send + Sync + 'static; async fn register<T>(&self, signal: &Signal<T>, initial: T) -> Result<(), ReactiveError> where T: Clone + Send + Sync + 'static; } }
The trait defines four core operations for reactive state. The read method returns the current value. The emit method updates the value. The subscribe method returns a stream of changes. The register method initializes a signal with a default value. See Runtime for reactive scheduling implementation.
QueryEffects Trait
The QueryEffects trait provides typed Datalog queries with capability-based authorization. Queries implement the Query trait which defines typed access to journal facts.
#![allow(unused)] fn main() { pub trait Query: Send + Sync + Clone + 'static { type Result: Clone + Send + Sync + Default + 'static; fn to_datalog(&self) -> DatalogProgram; fn required_capabilities(&self) -> Vec<QueryCapability>; fn dependencies(&self) -> Vec<FactPredicate>; fn parse(bindings: DatalogBindings) -> Result<Self::Result, QueryParseError>; fn query_id(&self) -> String; } #[async_trait] pub trait QueryEffects: Send + Sync { async fn query<Q: Query>(&self, query: &Q) -> Result<Q::Result, QueryError>; async fn query_raw(&self, program: &DatalogProgram) -> Result<DatalogBindings, QueryError>; fn subscribe<Q: Query>(&self, query: &Q) -> QuerySubscription<Q::Result>; async fn check_capabilities(&self, caps: &[QueryCapability]) -> Result<(), QueryError>; async fn invalidate(&self, predicate: &FactPredicate); } }
The Query trait converts queries to Datalog programs and defines capability requirements. The QueryEffects trait executes queries and manages subscriptions. Query isolation levels control consistency requirements. See Database Architecture for complete query system documentation.
Determinism Rules
Effect boundaries determine native and WASM conformance parity. Protocol code must follow these rules to ensure deterministic execution.
The pure transition core requires identical outputs given the same input stream. No hidden state may affect observable behavior. All state must flow through explicit effect calls. Non-determinism is permitted only through explicit algebraic effects. Time comes from time traits. Randomness comes from RandomEffects. Storage comes from StorageEffects.
Conformance lanes compare logical steps rather than wall-clock timing. Tests must not depend on execution speed. Time-dependent behavior uses simulated time through effect handlers. Conformance artifacts use canonical encoding with deterministic field ordering.
Session-Local VM Bridge Effects
Production choreography execution uses a narrow synchronous bridge trait at the Aura and Telltale boundary. VmBridgeEffects in aura-core exposes only immediate session-local queue and snapshot operations. It does not expose async transport, storage, or journal methods.
This split exists because Telltale host callbacks are synchronous. The callback path may enqueue outbound payloads, record blocked receive edges, consume branch choices, and snapshot scheduler signals. It must not perform network I/O or journal work directly.
Async host work resumes outside the VM step boundary in Layer 6 runtime services. vm_host_bridge observes VmBridgeEffects state, performs transport and guard-chain work, and injects completed results back into the VM. This preserves deterministic VM progression while keeping Aura's runtime async.
Layer Placement
The effect system spans several crates with strict dependency boundaries. aura-core defines effect traits, identifiers, and core data structures. It contains no implementations.
aura-effects contains stateless and single-party effect handlers. It provides default implementations for cryptography, storage, networking, and randomness. aura-protocol contains orchestrated and multi-party behavior. It bridges session types to effect calls.
aura-agent assembles handlers into runnable systems. It configures effect pipelines for production environments. aura-simulator provides deterministic execution with simulated time, networking, and controlled failure injection.
Performance
Aura includes several performance optimizations. Parallel initialization reduces startup time. Caching handlers reduce repeated computation. Buffer pools reduce memory allocation. The effect system avoids OS threads for WASM compatibility.
#![allow(unused)] fn main() { let builder = EffectSystemBuilder::new() .with_handler(Arc::new(RealCryptoHandler)) .with_parallel_init(); }
This snippet shows parallel initialization of handlers. The builder pattern allows flexible handler composition. Lazy initialization creates handlers on first use. Async tasks and cooperative scheduling provide efficient execution.
Testing Support
The effect system supports deterministic testing through mock handlers. A simulated runtime provides control over time and network behavior. The simulator exposes primitives to inject delays or failures.
#![allow(unused)] fn main() { let system = TestRuntime::new() .with_mock_crypto() .with_deterministic_time() .build(); }
This snippet creates a test runtime with mock handlers for all effects. It provides deterministic time and network control. Tests use in-memory storage and mock networking to execute protocols without side effects. See Test Infrastructure Reference for test patterns.
Consensus
This document describes the architecture of Aura Consensus. It defines the problem model, protocol phases, data structures, and integration with journals. It explains how consensus provides single-shot agreement for non-monotone operations such as account updates or relational context operations.
1. Problem Model
Aura uses consensus only for operations that cannot be expressed as monotone growth. Consensus produces a commit fact. The commit fact is inserted into one or more journals and drives deterministic reduction. Aura does not maintain a global log. Consensus operates in the scope of an authority or a relational context.
Consensus is single-shot. It agrees on a single operation and a single prestate. Commit facts are immutable and merge by join in journal namespaces.
A consensus instance uses a context-scoped committee. The committee contains witnesses selected by the authority or relational context. Committee members may be offline. The protocol completes even under partitions.
Consensus finalization is the single source of durable shared state. Fast-path coordination (provisional or soft-safe) may run in parallel for liveness, but its outputs must be superseded by commit facts. For BFT-DKG, consensus finalizes a transcript and emits DkgTranscriptCommit facts.
1.2 Witness Terminology
Aura uses witness (not validator) for consensus attestation participants.
witness: attests that an operation is valid for a specific prestate and contributes consensus evidence.signer: contributes threshold signature shares in FROST.
Keeping witness and signer distinct avoids conflating consensus attestation responsibilities with cryptographic share-generation roles.
1.3 Consensus is NOT Linearizable by Default
Aura Consensus is single-shot agreement, not log-based linearization.
Each consensus instance independently agrees on:
- A single operation
- A single prestate
- Produces a single commit fact
Consensus does NOT provide:
- Global operation ordering
- Sequential linearization across instances
- Automatic operation dependencies
To sequence operations, use session types (docs/108_mpst_and_choreography.md) executed through Aura’s Telltale-backed choreography runtime (execute_as runners or VM backend):
#![allow(unused)] fn main() { use aura_mpst::{choreography, Role}; #[choreography] async fn sequential_device_updates<C: EffectContext>( ctx: &C, account: Role<Account>, witnesses: Vec<Role<Witness>>, ) -> Result<(), AuraError> { // Session type enforces ordering: // 1. Update policy (must complete first) let policy_commit = consensus_single_shot( ctx, account.clone(), witnesses.clone(), TreeOp::UpdatePolicy { new_policy }, prestate1.hash(), ).await?; // 2. Remove device (uses policy_commit as prestate) // Session type prevents op2 from starting until op1 completes let prestate2 = account.read_tree_state(ctx).await?; assert_eq!(prestate2.hash(), policy_commit.result_id.prestate_hash); let remove_commit = consensus_single_shot( ctx, account, witnesses, TreeOp::RemoveLeaf { target: device_id }, prestate2.hash(), ).await?; Ok(()) } }
Cross-reference: See docs/113_database.md §8 for database transaction integration.
2. Core Protocol
Aura Consensus has two paths. The fast path completes in one round trip. The fallback path uses epidemic gossip and a threshold race. Both paths produce the same commit fact once enough matching witness shares exist.
The fast path uses direct communication. The initiator broadcasts an execute message. Witnesses run the operation against the prestate. Witnesses return FROST shares. The initiator aggregates shares and produces a threshold signature. FROST primitives are in aura-core::crypto::tree_signing.
The fallback path triggers when witnesses disagree or when the initiator stalls. Witnesses exchange share proposals using bounded fanout gossip. Any witness that assembles a valid threshold signature broadcasts a complete commit fact.
3. Common Structures and Notation
Consensus uses the following core concepts and notation throughout the protocol.
Core Variables
cid : ConsensusId- consensus instance identifierOp- operation being agreed on (application-defined)prestate- local pre-state for this instance (e.g., journal snapshot)prestate_hash = H(prestate)- hash of prestaterid = H(Op, prestate)- result identifiert- threshold; adversary controls< tkey sharesW- finite set of witnesses for this consensus instance
Per-Instance Tracking
Each consensus participant maintains per-cid state. The decided flags prevent double-voting. Proposals are keyed by (rid, prestate_hash) to group matching shares.
Equivocation Detection
Equivocation occurs when the same witness signs two different rid values under the same (cid, prestate_hash). This violates safety. The protocol detects and excludes equivocating shares.
4. Data Structures
Consensus instances use identifiers for operations and results.
#![allow(unused)] fn main() { /// Consensus instance identifier (derived from prestate and operation) pub struct ConsensusId(pub Hash32); impl ConsensusId { pub fn new(prestate_hash: Hash32, operation_hash: Hash32, nonce: u64) -> Self { // Hash of domain separator, prestate, operation, and nonce } } }
This structure identifies consensus instances. The result is identified by H(Op, prestate) computed inline.
A commit fact contains full consensus evidence including the operation, threshold signature, and participant list. Participants are recorded as AuthorityId values to preserve device privacy.
#![allow(unused)] fn main() { /// From crates/aura-consensus/src/types.rs pub struct CommitFact { pub consensus_id: ConsensusId, pub prestate_hash: Hash32, pub operation_hash: Hash32, pub operation_bytes: Vec<u8>, pub threshold_signature: ThresholdSignature, pub group_public_key: Option<PublicKeyPackage>, pub participants: Vec<AuthorityId>, pub threshold: u16, pub timestamp: ProvenancedTime, pub fast_path: bool, pub byzantine_attestation: Option<ByzantineSafetyAttestation>, } }
The commit fact is the output of consensus. Every peer merges CommitFact into its journal CRDT. A peer finalizes when it accepts a valid threshold signature and inserts the corresponding CommitFact.
The commit fact is inserted into the appropriate journal namespace. This includes account journals for account updates and relational context journals for cross-authority operations.
Ordering of committed facts relies on OrderTime (or session/consensus sequencing), not on timestamps. ProvenancedTime is semantic metadata and must not be used for cross-domain total ordering or indexing.
4.1 Byzantine Admission and Attestation
Consensus-backed ceremonies (including BFT-DKG finalization) are admitted only after runtime capability checks for the consensus profile. Aura currently enforces theorem-pack/runtime capability requirements such as byzantine_envelope before executing DKG and threshold-signing paths.
At admission, Aura captures a CapabilitySnapshot and records a ByzantineSafetyAttestation with:
- protocol id (
aura.consensus) - required capability keys for the profile
- snapshot of admitted/not-admitted runtime capabilities
- optional runtime evidence references
The attestation is attached to CommitFact and DkgTranscriptCommit. Capability mismatches fail before ceremony execution and emit redacted capability references for operational diagnostics.
5. Prestate Model
Consensus binds operations to explicit prestates. A prestate hash commits to the current reduced state of all participants.
#![allow(unused)] fn main() { let prestate_hash = H(C_auth1, C_auth2, C_context); }
This value includes root commitments of participating authorities. It may also include the current relational context commitment. Witnesses verify that their local reduced state matches the prestate hash. This prevents forks.
The result identifier binds the operation to the prestate.
#![allow(unused)] fn main() { let rid = H(Op, prestate); }
Witnesses treat matching (rid, prestate_hash) pairs as belonging to the same bucket. The protocol groups shares by these pairs to detect agreement.
6. Fast Path Protocol
The fast path optimistically assumes agreement. It completes in one round trip when all participants share the same prestate.
Messages
Execute(cid, Op, prestate_hash, evidΔ)- initiator requests operation executionWitnessShare(cid, rid, share, prestate_hash, evidΔ)- witness returns threshold shareCommit(cid, rid, sig, attesters, evidΔ)- initiator broadcasts commit factStateMismatch(cid, expected_pre_hash, actual_pre_hash, evidΔ)- optional debugging signal
Each message carries evidence delta evidΔ for the consensus instance. Evidence propagates along with protocol messages.
Initiator Protocol
The initiator i coordinates the fast path.
State at initiator i:
cid : ConsensusId // fresh per instance
Op : Operation
shares : Map[WitnessId -> (rid, share, prestate_hash)]
decided : Map[ConsensusId -> Bool] // initially decided[cid] = false
W : Set[WitnessId]
t : Nat
The initiator maintains per-witness shares and tracks decision status.
Start Operation
Start(cid, Op):
prestate := ReadState()
prestate_hash := H(prestate)
decided[cid] := false
shares := {}
For all w in W:
Send Execute(cid, Op, prestate_hash, EvidenceDelta(cid)) to w
The initiator reads local state and broadcasts the operation to all witnesses.
Process Witness Shares
On WitnessShare(cid, rid, share, prestate_hash, evidΔ) from w:
MergeEvidence(cid, evidΔ)
If decided[cid] = false and w not in shares:
shares[w] := (rid, share, prestate_hash)
// collect all shares for this specific (rid, prestate_hash)
Hset := { (w', s') in shares | s'.rid = rid
∧ s'.prestate_hash = prestate_hash }
If |Hset| ≥ t:
sig := CombineShares( { s'.share | (_, s') in Hset } )
attesters := { w' | (w', _) in Hset }
CommitFact(cid, rid, sig, attesters)
For all v in W:
Send Commit(cid, rid, sig, attesters, EvidenceDelta(cid)) to v
decided[cid] := true
The initiator collects shares. When t shares agree on (rid, prestate_hash), it combines them into a threshold signature.
Witness Protocol
Each witness w responds to execute requests.
State at witness w:
proposals : Map[(rid, prestate_hash) -> Set[(WitnessId, Share)]]
decided : Map[ConsensusId -> Bool]
timers : Map[ConsensusId -> TimerHandle]
W : Set[WitnessId]
t : Nat
Witnesses maintain proposals and fallback timers.
Process Execute Request
On Execute(cid, Op, prestate_hash, evidΔ) from i:
MergeEvidence(cid, evidΔ)
if decided.get(cid, false) = true:
return
prestate := ReadState()
if H(prestate) != prestate_hash:
Send StateMismatch(cid, prestate_hash, H(prestate), EvidenceDelta(cid)) to i
StartTimer(cid, T_fallback)
return
rid := H(Op, prestate)
share := ProduceShare(cid, rid)
proposals[(rid, prestate_hash)] := { (w, share) }
Send WitnessShare(cid, rid, share, prestate_hash, EvidenceDelta(cid)) to i
StartTimer(cid, T_fallback)
Witnesses verify prestate agreement before computing shares. Mismatches trigger fallback timers.
Process Commit
On Commit(cid, rid, sig, attesters, evidΔ) from any v:
if decided.get(cid, false) = false
and VerifyThresholdSig(rid, sig, attesters):
MergeEvidence(cid, evidΔ)
CommitFact(cid, rid, sig, attesters)
decided[cid] := true
StopTimer(cid)
Witnesses accept valid commit facts from any source.
7. Evidence Propagation
Evidence tracks equivocation and accountability information for consensus instances. The system uses CRDT-based incremental propagation. Evidence deltas attach to all protocol messages. Witnesses merge incoming evidence automatically.
Equivocation occurs when a witness signs conflicting result IDs for the same prestate. The protocol detects this pattern and generates cryptographic proofs. Each proof records both conflicting signatures with witness identity and timestamp.
#![allow(unused)] fn main() { pub struct EquivocationProof { pub witness: AuthorityId, pub consensus_id: ConsensusId, pub prestate_hash: Hash32, pub first_result_id: Hash32, pub second_result_id: Hash32, pub timestamp_ms: u64, } }
The proof structure preserves both conflicting votes. This enables accountability and slashing logic in higher layers.
Evidence deltas propagate via incremental synchronization. Each delta contains only new proofs since the last exchange. Timestamps provide watermark-based deduplication. The delta structure is lightweight and merges idempotently.
#![allow(unused)] fn main() { pub struct EvidenceDelta { pub consensus_id: ConsensusId, pub equivocation_proofs: Vec<EquivocationProof>, pub timestamp_ms: u64, } }
Every consensus message includes an evidence delta field. Coordinators attach deltas when broadcasting execute and commit messages. Witnesses attach deltas when sending shares. Receivers merge incoming deltas into local evidence trackers. This piggybacking ensures evidence propagates without extra round trips.
8. Fallback Protocol
Fallback activates when the fast path stalls. Triggers include witness disagreement or initiator failure. Fallback uses leaderless gossip to complete consensus.
Messages
Conflict(cid, conflicts, evidΔ)- initiator seeds fallback with known conflictsAggregateShare(cid, proposals, evidΔ)- witnesses exchange proposal setsThresholdComplete(cid, rid, sig, attesters, evidΔ)- any witness broadcasts completion
Equivocation Detection
Equivocation means signing two different rid values under the same (cid, prestate_hash).
HasEquivocated(proposals, witness, cid, pre_hash, new_rid):
For each ((rid, ph), S) in proposals:
if ph = pre_hash
and rid ≠ new_rid
and witness ∈ { w' | (w', _) in S }:
return true
return false
The protocol excludes equivocating shares from threshold computation.
Fallback Gossip
Witnesses periodically exchange proposals with random peers.
OnPeriodic(cid) for fallback gossip when decided[cid] = false:
peers := SampleRandomSubset(W \ {w}, k)
For each p in peers:
Send AggregateShare(cid, proposals, EvidenceDelta(cid)) to p
Gossip fanout k controls redundancy. Typical values are 3-5.
Threshold Checking
Witnesses check for threshold completion after each update.
CheckThreshold(cid):
if decided.get(cid, false) = true:
return
For each ((rid, pre_hash), S) in proposals:
if |S| ≥ t:
sig := CombineShares({ sh | (_, sh) in S })
if VerifyThresholdSig(rid, sig):
attesters := { w' | (w', _) in S }
CommitFact(cid, rid, sig, attesters)
For all v in W:
Send ThresholdComplete(cid, rid, sig, attesters, EvidenceDelta(cid)) to v
decided[cid] := true
return
Any witness reaching threshold broadcasts the commit fact. The first valid threshold signature wins.
9. Integration with Journals
Consensus emits commit facts. Journals merge commit facts using set union. Reduction interprets commit facts as confirmed non-monotone events.
Account journals integrate commit facts that represent tree operations. Relational context journals integrate commit facts that represent guardian bindings or recovery grants.
Reduction remains deterministic. Commit facts simply appear as additional facts in the semilattice.
A commit fact is monotone even when the event it represents is non-monotone. This supports convergence.
9.1 Context Isolation and Guard Integration
Every consensus instance binds to a context identifier. The context provides isolation for capability checking and flow budget enforcement. Authority-scoped ceremonies derive context from authority identity. Relational ceremonies use explicit context identifiers from the relational binding.
The protocol integrates with the guard chain at message send boundaries. Guards evaluate before constructing messages. The guard chain sequences capability checks, flow budget charges, leakage tracking, and journal coupling. Journal facts commit at the runtime bridge layer after guard evaluation completes.
#![allow(unused)] fn main() { // Protocol evaluates guards before sending let guard = SignShareGuard::new(context_id, coordinator); let guard_result = guard.evaluate(effects).await?; if !guard_result.authorized { return Err(AuraError::permission_denied("Guard denied SignShare")); } // Construct and send message after guard approval let message = ConsensusMessage::SignShare { ... }; }
The dependency injection pattern passes effect systems as parameters. Protocol methods accept generic effect trait bounds. This enables testing with mock effects and production use with real implementations. The pattern avoids storing effects in protocol state.
10. Integration Points
Consensus is used for account tree operations that require strong agreement. Examples include membership changes and policy changes when local signing is insufficient.
Consensus is also used for relational context operations. Guardian bindings use consensus to bind authority state. Recovery grants use consensus to approve account modifications. Application specific relational contexts may also use consensus.
11. FROST Threshold Signatures
Consensus uses FROST to produce threshold signatures. Each witness holds a secret share. Witnesses compute partial signatures. The initiator or fallback proposer aggregates the shares. The final signature verifies under the group public key stored in the commitment tree. See Authority and Identity for details on the TreeState structure.
#![allow(unused)] fn main() { /// From crates/aura-consensus/src/messages.rs pub enum ConsensusMessage { SignShare { consensus_id: ConsensusId, /// The result_id (hash of execution result) this witness computed result_id: Hash32, share: PartialSignature, /// Optional commitment for the next consensus round (pipelining optimization) next_commitment: Option<NonceCommitment>, /// Epoch for commitment validation epoch: Epoch, }, // ... other message variants } }
Witness shares validate only for the current consensus instance. They cannot be replayed. Witnesses generate only one share per (consensus_id, prestate_hash).
The attester set in the commit fact contains only devices that contributed signing shares. This provides cryptographic proof of participation.
11.1 Integration with Tree Operation Verification
When consensus produces a commit fact for a tree operation, the resulting AttestedOp is verified using the two-phase model from aura-core::tree::verification:
- Verification (
verify_attested_op): Cryptographic check against theBranchSigningKeystored in TreeState - Check (
check_attested_op): Full verification plus state consistency validation
The binding message includes the group public key to prevent signature reuse attacks. This ensures an attacker cannot substitute a different signing key and replay a captured signature. See Tree Operation Verification for details.
#![allow(unused)] fn main() { // Threshold derived from policy at target node let threshold = state.get_policy(target_node)?.required_signers(child_count); // Signing key from TreeState let signing_key = state.get_signing_key(target_node)?; // Verify against stored key and policy-derived threshold verify_attested_op(&attested_op, signing_key, threshold, current_epoch)?; }
The verification step ensures signature authenticity. The check step validates state consistency. Both steps must pass before applying tree operations.
11.2 Type-Safe Share Collection
The implementation provides type-level guarantees for threshold signature aggregation. Share collection uses sealed and unsealed types to prevent combining signatures before reaching threshold. The type system enforces this invariant at compile time.
#![allow(unused)] fn main() { pub struct LinearShareSet { shares: BTreeMap<AuthorityId, PartialSignature>, sealed: bool, } pub struct ThresholdShareSet { shares: BTreeMap<AuthorityId, PartialSignature>, } }
The unsealed type accepts new shares via insertion. When threshold is reached the set seals into the threshold type. Only the threshold type provides the combine method. This prevents calling aggregation before sufficient shares exist.
The current protocol uses hash map based tracking for signature collection. The type-safe approach exists but awaits integration. Future work will replace runtime threshold checks with compile-time proofs. The sealed type pattern eliminates an entire class of bugs.
12. FROST Commitment Pipeline Optimization
The pipelined commitment optimization reduces steady-state consensus from 2 RTT (round-trip times) to 1 RTT by bundling next-round nonce commitments with current-round signature shares.
12.1 Overview
The FROST pipelining optimization improves consensus performance by bundling next-round nonce commitments with current-round signature shares. This allows the coordinator to start the next consensus round immediately without waiting for a separate nonce commitment phase.
Standard FROST Consensus (2 RTT):
- Execute → NonceCommit (1 RTT): Coordinator sends execute request, witnesses respond with nonce commitments
- SignRequest → SignShare (1 RTT): Coordinator sends aggregated nonces, witnesses respond with signature shares
Pipelined FROST Consensus (1 RTT) (after warm-up):
- Execute+SignRequest → SignShare+NextCommitment (1 RTT):
- Coordinator sends execute request with cached commitments from previous round
- Witnesses respond with signature share AND next-round nonce commitment
12.2 Core Components
WitnessState (aura-consensus/src/witness.rs) manages persistent nonce state for each witness:
#![allow(unused)] fn main() { pub struct WitnessState { /// Witness identifier witness_id: AuthorityId, /// Current epoch to detect when cached commitments become stale epoch: Epoch, /// Precomputed nonce for the next consensus round next_nonce: Option<(NonceCommitment, NonceToken)>, /// Active consensus instances this witness is participating in active_instances: HashMap<ConsensusId, WitnessInstance>, } }
Key methods:
get_next_commitment(): Returns cached commitment if valid for current epochtake_nonce(): Consumes cached nonce for use in current roundset_next_nonce(): Stores new nonce for future useinvalidate(): Clears cached state on epoch change
Message Schema Updates: The SignShare message now includes optional next-round commitment:
#![allow(unused)] fn main() { SignShare { consensus_id: ConsensusId, share: PartialSignature, /// Optional commitment for the next consensus round (pipelining optimization) next_commitment: Option<NonceCommitment>, /// Epoch for commitment validation epoch: Epoch, } }
ConsensusProtocol (aura-consensus/src/protocol.rs) integrates the pipelining optimization via FrostConsensusOrchestrator and WitnessTracker. The pipelining logic is distributed across these components rather than isolated in a separate orchestrator.
Key methods:
run_consensus(): Determines fast path vs slow path based on cached commitmentscan_use_fast_path(): Checks if sufficient cached commitments availablehandle_epoch_change(): Invalidates all cached state on epoch rotation
12.3 Epoch Safety
All cached commitments are bound to epochs to prevent replay attacks:
- Epoch Binding: Each commitment is tied to a specific epoch
- Automatic Invalidation: Epoch changes invalidate all cached commitments
- Validation: Witnesses reject commitments from wrong epochs
#![allow(unused)] fn main() { // Epoch change invalidates all cached nonces if self.epoch != current_epoch { self.next_nonce = None; self.epoch = current_epoch; return None; } }
12.4 Fallback Handling
The system gracefully falls back to 2 RTT when:
- Insufficient Cached Commitments: Less than threshold witnesses have cached nonces
- Epoch Change: All cached commitments become invalid
- Witness Failures: Missing or invalid next_commitment in responses
- Initial Bootstrap: First round after startup (no cached state)
#![allow(unused)] fn main() { if has_quorum { // Fast path: 1 RTT using cached commitments self.run_fast_path(...) } else { // Slow path: 2 RTT standard consensus self.run_slow_path(...) } }
12.5 Performance Impact
Latency Reduction:
- Before: 2 RTT per consensus
- After: 1 RTT per consensus (steady state)
- Improvement: 50% latency reduction
Message Count:
- Before: 4 messages per witness (Execute, NonceCommit, SignRequest, SignShare)
- After: 2 messages per witness (Execute+SignRequest, SignShare+NextCommitment)
- Improvement: 50% message reduction
Trade-offs:
- Memory: Small overhead for caching one nonce per witness
- Complexity: Additional state management and epoch tracking
- Bootstrap: First round still requires 2 RTT
12.6 Implementation Guidelines
Adding Pipelining to New Consensus Operations:
- Update message schema: Add
next_commitmentandepochfields to response messages - Generate next nonce: During signature generation, also generate next-round nonce
- Cache management: Store next nonce in
WitnessStatefor future use - Epoch handling: Always validate epoch before using cached commitments
Example Witness Implementation:
#![allow(unused)] fn main() { pub async fn handle_sign_request<R: RandomEffects + ?Sized>( &mut self, consensus_id: ConsensusId, aggregated_nonces: Vec<NonceCommitment>, current_epoch: Epoch, random: &R, ) -> Result<ConsensusMessage> { // Generate signature share let share = self.create_signature_share(consensus_id, aggregated_nonces)?; // Generate or retrieve next-round commitment let next_commitment = if let Some((commitment, _)) = self.witness_state.take_nonce(current_epoch) { // Use cached nonce Some(commitment) } else { // Generate fresh nonce let (nonces, commitment) = self.generate_nonce(random).await?; let token = NonceToken::from(nonces); // Cache for future self.witness_state.set_next_nonce(commitment.clone(), token, current_epoch); Some(commitment) }; Ok(ConsensusMessage::SignShare { consensus_id, share, next_commitment, epoch: current_epoch, }) } }
12.7 Security Considerations
- Nonce Reuse Prevention: Each nonce is used exactly once and tied to specific epoch
- Epoch Isolation: Nonces from different epochs cannot be mixed
- Forward Security: Epoch rotation provides natural forward security boundary
- Availability: Fallback ensures consensus continues even without optimization
12.8 Testing Strategy
Unit Tests:
- Epoch invalidation logic
- Nonce caching and retrieval
- Message serialization with new fields
Integration Tests:
- Fast path vs slow path selection
- Epoch transition handling
- Performance measurement
Simulation Tests:
- Network delay impact on 1 RTT vs 2 RTT
- Behavior under partial failures
- Convergence properties
12.9 Future Enhancements
- Adaptive Thresholds: Dynamically adjust quorum requirements based on cached state
- Predictive Caching: Pre-generate multiple rounds of nonces during idle time
- Compression: Batch multiple commitments in single message
- Cross-Context Optimization: Share cached state across related consensus contexts
13. Fallback Protocol Details
Fallback Trigger
The initiator can proactively trigger fallback when it detects conflicts.
Fallback_Trigger_Initiator(cid):
// Extract conflicts from shares (same prestate_hash, different rid)
conflicts := Map[(rid, prestate_hash) -> Set[(w, share)]]
For each (w, (rid, share, prestate_hash)) in shares:
conflicts[(rid, prestate_hash)] :=
conflicts.get((rid, prestate_hash), ∅) ∪ { (w, share) }
For all w in W:
Send Conflict(cid, conflicts, EvidenceDelta(cid)) to w
This optimization seeds fallback with known conflicts. Witnesses also start fallback independently on timeout.
Fallback Message Handlers
Witnesses process fallback messages to accumulate shares.
On Conflict(cid, conflicts, evidΔ) from any peer:
MergeEvidence(cid, evidΔ)
For each ((rid, pre_hash), S) in conflicts:
if not HasEquivocatedInSet(proposals, S, pre_hash):
proposals[(rid, pre_hash)] :=
proposals.get((rid, pre_hash), ∅) ∪ S
CheckThreshold(cid)
Fallback_Start(cid)
Conflict messages bootstrap the fallback phase. The witness merges non-equivocating shares.
On AggregateShare(cid, proposals', evidΔ) from any peer:
MergeEvidence(cid, evidΔ)
For each ((rid, pre_hash), S') in proposals':
For each (w', sh') in S':
if not HasEquivocated(proposals, w', cid, pre_hash, rid):
proposals[(rid, pre_hash)] :=
proposals.get((rid, pre_hash), ∅) ∪ { (w', sh') }
CheckThreshold(cid)
Aggregate shares spread proposal sets through gossip. Each witness checks for threshold after updates. Threshold-complete messages carry the final aggregated signature; once validated, the witness commits and stops its timers.
14. Safety Guarantees
Consensus satisfies agreement. At most one commit fact forms for a given (cid, prestate_hash). The threshold signature ensures authenticity. No attacker can forge a threshold signature without the required number of shares.
Consensus satisfies validity. A commit fact references a result computed from the agreed prestate. All honest witnesses compute identical results. Malformed shares are rejected.
Consensus satisfies deterministic convergence. Evidence merges through CRDT join. All nodes accept the same commit fact.
Formal Verification Status
These properties are formally verified through complementary approaches:
-
Lean 4 Proofs (
verification/lean/Aura/Consensus/):Agreement.agreement: Unique commit per consensus instanceValidity.validity: Committed values bound to prestatesValidity.prestate_binding_unique: Hash collision resistance for prestate bindingEquivocation.detection_soundness: Conflicting signatures are detectable
-
Quint Model Checking (
verification/quint/protocol_consensus*.qnt):AllInvariants: Combined safety properties pass 1000-sample model checkingInvariantByzantineThreshold: Byzantine witnesses bounded below thresholdInvariantEquivocationDetected: Equivocation detection correctnessInvariantProgressUnderSynchrony: Liveness under partial synchrony
See verification/quint/ and verification/lean/ for verification artifacts and run notes.
Binding Message Security
The binding message for tree operations includes the group public key. This prevents key substitution attacks where an attacker captures a valid signature and replays it with a different key they control. The full binding includes:
- Domain separator for domain isolation
- Parent epoch and commitment for replay prevention
- Protocol version for upgrade safety
- Current epoch for temporal binding
- Group public key for signing group binding
- Serialized operation content
This security property is enforced by compute_binding_message() in the verification module.
15. Liveness Guarantees
The fast path completes in one round trip when the initiator and witnesses are online. The fallback path provides eventual completion under asynchrony. Gossip ensures that share proposals propagate across partitions.
The protocol does not rely on stable leaders. No global epoch boundaries exist. Witnesses can rejoin by merging the journal fact set.
Fallback State Diagram
stateDiagram-v2
[*] --> FastPath
FastPath --> FallbackPending: timeout or conflict
FallbackPending --> FallbackGossip: send Conflict/AggregateShare
FallbackGossip --> Completed: threshold signature assembled
FastPath --> Completed: Commit broadcast
Completed --> [*]
- FastPath: initiator collecting shares.
- FallbackPending: timer expired or conflicting
(rid, prestate_hash)observed. - FallbackGossip: witnesses exchange proposal sets until
CheckThresholdsucceeds. - Completed: commit fact accepted and timers stopped.
Design Trade-offs: Latency vs Availability
The two protocol paths optimize for different objectives:
| Property | Fast Path | Slow Path |
|---|---|---|
| Primary Goal | Latency (speed) | Availability (robustness) |
| Completion Time | 2δ (network speed) | Eventual (no tight bound) |
| Synchrony Assumption | Yes (GST reached) | No (works during partitions) |
| Coordination | Leader-driven (initiator) | Leaderless gossip |
| Failure Tolerance | Requires all/most online | Tolerates n-k failures |
Why leaderless gossip for fallback?
We intentionally sacrifice tight timing bounds (the theoretical 2Δ from optimal protocols) for:
-
Partition tolerance: Any connected component with k honest witnesses can complete independently. No leader election needed across partitions.
-
No single point of failure: If the initiator fails, any witness can drive completion. View-based protocols require leader handoff.
-
Simpler protocol: No view numbers, no leader election, no synchronization barriers. Witnesses simply gossip until threshold is reached.
-
Natural CRDT integration: Evidence propagates as a CRDT set. Gossip naturally merges evidence without coordination.
What we can prove:
- Fast path: Commits within 2δ when all witnesses online (verified:
TemporalFastPathBound) - Slow path: Commits when any connected component has k honest witnesses (not time-bounded)
- Safety: Always holds regardless of network conditions
What we cannot prove (by design):
- Slow path completion in fixed time (would require synchrony assumption)
- Termination under permanent partition (FLP impossibility)
Parameter Guidance
T_fallback: set to roughly 2-3 times the median witness RTT. Too aggressive causes unnecessary fallback. Too lax delays recovery when an initiator stalls.- Gossip fanout
f: see the fanout adequacy analysis below for principled selection. - Gossip interval: 250-500 ms is a good default. Ensure at least a few gossip rounds occur before timers retrigger. This gives the network time to converge.
These parameters should be tuned per deployment. The ranges above keep fallback responsive without overwhelming the network.
Fanout Adequacy for Slow Path
The slow path requires the gossip network to be connected for evidence propagation. The key question is how many gossip peers (fanout f) are needed.
Random Graph Connectivity Theory
For a random graph G(n, p) to be connected with high probability (w.h.p.), the classical Erdős-Rényi result states:
p ≥ (1 + ε) × ln(n) / n for any ε > 0
Translating to gossip: each node selects f random peers, giving edge probability p ≈ f/n. For connectivity:
f ≥ c × ln(n) where c ≈ 1.1-1.2 for practical certainty
Recommended Fanout by Witness Count
| Witnesses (n) | Minimum Fanout (f) | Recommended | Notes |
|---|---|---|---|
| 3 | 2 | 2 | All must connect (fully connected) |
| 5 | 2 | 3 | ln(5) ≈ 1.6, add margin |
| 7 | 3 | 3-4 | ln(7) ≈ 1.9 |
| 10 | 3 | 4 | ln(10) ≈ 2.3 |
| 15 | 4 | 4-5 | ln(15) ≈ 2.7 |
| 21 | 4 | 5 | ln(21) ≈ 3.0 |
| 50 | 5 | 6 | ln(50) ≈ 3.9 |
Failure Tolerance
With fanout f and n witnesses, the graph remains connected under random node failures as long as:
- At least
f + 1nodes remain (sufficient edges for spanning tree) - Failed nodes are randomly distributed (not targeted attacks)
For k-of-n threshold where k ≤ n - f:
- The
khonest witnesses form a connected subgraph w.h.p. - Each honest witness can reach at least one other honest witness via gossip
- Evidence propagates to all honest witnesses in O(log(n)) gossip rounds
Adversarial Considerations
Random graph analysis assumes non-adversarial peer selection. Against Byzantine adversaries:
- Eclipse attacks: If adversary controls gossip peer selection, connectivity guarantees fail
- Mitigation: Use deterministic peer selection based on witness IDs (e.g., consistent hashing)
- Stronger bound: For Byzantine tolerance, use
f ≥ 2t + 1where t is Byzantine threshold
Verified Properties (Quint)
The following properties are model-checked in protocol_consensus_liveness.qnt:
PropertyQuorumSufficient: k honest witnesses in same connected component can progressPropertyPartitionTolerance: largest connected component with k honest succeedsAvailabilityGuarantee: combined availability invariant
With adequate fanout, these properties ensure slow path completion when quorum exists.
16. Relation to FROST
Consensus uses FROST as the final stage of agreement. FROST ensures that only one threshold signature exists per result. FROST signatures provide strong cryptographic proof.
Consensus and FROST remain separate. Consensus orchestrates communication. FROST proves final agreement. Combining the two yields a single commit fact.
17. Operation Categories
Not all operations require consensus. Aura classifies operations into three categories based on their security requirements and execution timing.
17.1 Category A: Optimistic Operations
Operations that can proceed immediately without consensus. These use CRDT facts with eventual consistency.
Characteristics:
- Immediate local effect
- Background sync via anti-entropy
- Failure shows indicator, doesn't block functionality
- Partial success is acceptable
Examples:
- Send message (within established context)
- Create channel (within established relational context)
- Update channel topic
- Block/unblock contact
- Pin message
These operations work because the cryptographic context already exists. Keys derive deterministically from shared journal state. No new agreement is needed.
17.2 Category B: Deferred Operations
Operations that apply locally but require agreement for finalization. Effect is pending until confirmed.
Characteristics:
- Immediate local effect shown as "pending"
- Background ceremony for agreement
- Failure triggers rollback with user notification
- Multi-moderator operations use this pattern
Examples:
- Change channel permissions (requires moderator consensus)
- Remove channel member (may be contested)
- Transfer channel ownership
- Rename channel
17.3 Category C: Consensus-Gated Operations
Operations where partial state is dangerous. These block until consensus completes.
Characteristics:
- Operation does NOT proceed until consensus achieved
- Partial state would be dangerous or irrecoverable
- User must wait for confirmation
Examples:
- Guardian rotation (key shares distributed atomically)
- Recovery execution (account state replacement)
- OTA hard fork activation (breaking protocol change)
- Device revocation (security-critical removal)
- Add contact / Create group (establishes cryptographic context)
- Add member to group (changes group encryption keys)
These use Aura Consensus as described in this document.
17.4 Decision Tree
Does this operation establish or modify cryptographic relationships?
│
├─ YES: Does the user need to wait for completion?
│ │
│ ├─ YES (new context, key changes) → Category C (Consensus)
│ │ Examples: add contact, create group, guardian rotation
│ │
│ └─ NO (removal from existing context) → Category B (Deferred)
│ Examples: remove from group, revoke device
│
└─ NO: Does this affect other users' access or policies?
│
├─ YES: Is this high-security or irreversible?
│ │
│ ├─ YES → Category B (Deferred)
│ │ Examples: transfer ownership, delete channel
│ │
│ └─ NO → Category A (Optimistic)
│ Examples: pin message, update topic
│
└─ NO → Category A (Optimistic)
Examples: send message, create channel, block contact
17.5 Key Insight
Ceremonies establish shared cryptographic context. Operations within that context are cheap.
Once a relational context exists (established via Category C invitation ceremony), channels and messages within that context are Category A. The expensive part is establishing WHO is in the relationship. Once established, operations WITHIN the relationship derive keys deterministically from shared state.
See work/optimistic.md for detailed design and effect policy integration.
18. BFT‑DKG Transcript Finalization (K3)
Consensus is also used to finalize BFT-DKG transcripts. The output of the DKG is a DkgTranscriptCommit fact that is consensus-finalized and then merged into the relevant journal.
Inputs
DkgConfig:epoch,threshold,max_signers,participants,membership_hash.DealerPackage[]: one package per dealer, containing encrypted shares for all participants and a deterministic commitment.prestate_hashandoperation_hash: bind the transcript to the authority/context state and the intended ceremony operation.
Transcript formation
- Validate
DkgConfigand dealer packages (unique dealers, complete share sets). - Assemble the transcript deterministically.
- Hash the transcript using canonical DAG-CBOR encoding.
Finalize with consensus
- Build
DkgTranscriptCommitwith:transcript_hashblob_ref(optional, if stored out‑of‑line)prestate_hashoperation_hashconfigandparticipants
- Run consensus over the commit fact itself.
- Insert both
CommitFactevidence and theDkgTranscriptCommitfact into the authority or context journal.
The transcript commit is the single durable artifact used to bootstrap all subsequent threshold operations. Any K3 ceremony that depends on keys must bind to this commit by reference (direct hash or blob ref).
19. Decentralized Coordinator Selection (Lottery)
Coordinator-based fast paths (A2) require a deterministic, decentralized selection mechanism so every participant can independently derive the same leader without extra coordination.
Round seed
round_seedis a 32‑byte value shared by all participants for the round.- Sources:
- VRF output (preferred when available).
- Trusted oracle or beacon.
- Initiator‑provided seed (acceptable in trusted settings).
Selection rule
score_i = H("AURA_COORD_LOTTERY" || round_seed || authority_id_i)
winner = argmin_i score_i
Fencing + safety
- The coordinator must hold a monotonic fencing token (
coord_epoch). - Proposals are rejected if
coord_epochdoes not advance or ifprestate_hashmismatches local state.
Convergence
- Coordinators emit a
ConvergenceCertonce a quorum acks the proposal. - Fast-path results remain soft-safe until a consensus
CommitFactis merged.
20. Summary
Aura Consensus produces monotone commit facts that represent non-monotone operations. It integrates with journals through set union. It uses FROST threshold signatures and CRDT evidence structures. It provides agreement, validity, and liveness. It supports authority updates and relational operations. It requires no global log and no central coordinator.
The helper HasEquivocatedInSet excludes conflict batches that contain conflicting signatures from the same witness. Fallback_Start transitions the local state machine into fallback mode and arms gossip timers. Implementations must provide these utilities alongside the timers described earlier.
See Also
- Operation Categories - When consensus is required and ceremony lifecycle
- Relational Contexts - Consensus integration for relational operations
Operation Categories
This document defines the three-tier classification system for distributed operations in Aura. It specifies the ceremony contract for Category C operations, the consistency metadata types for each category, and the decision framework for categorizing new operations. The core insight is that not all operations require consensus. Many can proceed optimistically with background reconciliation.
1. Overview
Operations in Aura fall into three categories based on their effect timing and security requirements.
| Category | Name | Effect Timing | When Used |
|---|---|---|---|
| A | Optimistic | Immediate local effect | Low-risk operations within established contexts |
| B | Deferred | Pending until confirmed | Medium-risk policy/membership changes |
| C | Consensus-Gated | Blocked until ceremony completes | Cryptographic context establishment |
Agreement modes are orthogonal to categories. Operations can use provisional or soft-safe fast paths, but any durable shared state must be consensus-finalized (A3). See Consensus for the fast-path and finalization taxonomy.
1.1 Key Generation Methods
Aura separates key generation from agreement:
| Code | Method | Description |
|---|---|---|
| K1 | Single-signer | No DKG required. Local key generation. |
| K2 | Dealer-based DKG | Trusted coordinator distributes shares. |
| K3 | Consensus-finalized DKG | BFT-DKG with transcript commit. |
| DKD | Distributed key derivation | Multi-party derivation without DKG. |
1.2 Agreement Levels
| Code | Level | Description |
|---|---|---|
| A1 | Provisional | Usable immediately but not final. |
| A2 | Coordinator Soft-Safe | Bounded divergence with convergence certificate. |
| A3 | Consensus-Finalized | Unique, durable, non-forkable. |
Fast paths (A1/A2) are provisional. Durable shared state must be finalized by A3.
1.3 The Key Architectural Insight
Ceremonies establish shared cryptographic context. Operations within that context are cheap.
Ceremony (Category C) Optimistic Operations (Category A)
───────────────────── ─────────────────────────────────
• Runs once per relationship • Within established context
• Establishes ContextId + shared roots • Derive keys from context
• Creates relational context journal • Just emit CRDT facts
• All future encryption derives here • No new agreement needed
2. Category A: Optimistic Operations
Category A operations have immediate local effect via CRDT fact emission. Background sync via anti-entropy propagates facts to peers. Failure shows a status indicator but does not block functionality. Partial success is acceptable.
2.1 Examples
| Operation | Immediate Action | Background Sync | On Failure |
|---|---|---|---|
| Create channel | Show channel, enable messaging | Fact syncs to members | Show "unsynced" badge |
| Send message | Display in chat immediately | Delivery receipts | Show "undelivered" indicator |
| Add contact (within context) | Show in list | Mutual acknowledgment | Show "pending" status |
| Block contact | Hide from view immediately | Propagate to context | Already effective locally |
| Update profile | Show changes immediately | Propagate to contacts | Show sync indicator |
| React to message | Show reaction | Fact syncs | Show "pending" |
2.2 Implementation Pattern
#![allow(unused)] fn main() { async fn create_channel_optimistic(&mut self, config: ChannelConfig) -> ChannelId { let channel_id = ChannelId::derive(&config); self.emit_fact(ChatFact::ChannelCheckpoint { channel_id, epoch: 0, base_gen: 0, window: 1024, }).await; channel_id } }
This pattern emits a fact into the existing relational context journal. The channel is immediately usable. Key derivation uses KDF(ContextRoot, ChannelId, epoch).
2.3 Why This Works
Category A operations work because encryption keys already exist (derived from established context), facts are CRDTs (eventual consistency is sufficient), no coordination is needed (shared state already agreed upon), and the worst case is delay rather than a security issue.
3. Category B: Deferred Operations
Category B operations have local effect pending until agreement is reached. The UI shows intent immediately with a "pending" indicator. Operations may require approval from capability holders. Automatic rollback occurs on rejection.
3.1 Examples
| Operation | Immediate Action | Agreement Required | On Rejection |
|---|---|---|---|
| Change channel permissions | Show "pending" | Moderator approval | Revert, notify |
| Remove channel member | Show "pending removal" | Moderator consensus | Keep member |
| Transfer ownership | Show "pending transfer" | Recipient acceptance | Cancel transfer |
| Rename channel | Show "pending rename" | Member acknowledgment | Keep old name |
| Archive channel | Show "pending archive" | Moderator approval | Stay active |
3.2 Implementation Pattern
#![allow(unused)] fn main() { async fn change_permissions_deferred( &mut self, channel_id: ChannelId, changes: PermissionChanges, ) -> ProposalId { let proposal = Proposal { operation: Operation::ChangePermissions { channel_id, changes }, requires_approval_from: vec![CapabilityRequirement::Role("moderator")], threshold: ApprovalThreshold::Any, timeout_ms: 24 * 60 * 60 * 1000, }; let proposal_id = self.emit_proposal(proposal).await; proposal_id } }
This pattern creates a proposal that does not apply the effect yet. The UI shows "pending" state. The effect applies when threshold approvals are received. Auto-revert occurs on timeout or rejection.
3.3 Approval Thresholds
#![allow(unused)] fn main() { pub enum ApprovalThreshold { Any, Unanimous, Threshold { required: u32 }, Percentage { percent: u8 }, } }
Any requires any single holder of the required capability. Unanimous requires all holders to approve. Threshold requires k-of-n approval. Percentage requires a percentage of holders.
4. Category C: Consensus-Gated Operations
Category C operations do NOT proceed until a ceremony completes. Partial state would be dangerous or irrecoverable. The user must wait for confirmation. These operations use choreographic protocols with session types, executed through Aura’s Telltale runtime.
4.1 Examples
| Operation | Why Blocking Required | Risk if Optimistic |
|---|---|---|
| Add contact (new relationship) | Creates cryptographic context | No shared keys possible |
| Create group | Multi-party key agreement | Inconsistent member views |
| Add member to group | Changes group keys | Forward secrecy violation |
| Device enrollment | Key shares distributed atomically | Partial enrollment unusable |
| Guardian rotation | Key shares distributed atomically | Partial rotation unusable |
| Recovery execution | Account state replacement | Partial recovery corruption |
| OTA hard fork | Scope-bound breaking protocol change | Explicit partition or rejected incompatible sessions outside the cutover scope |
| Device revocation | Security-critical removal | Attacker acts first |
4.2 Implementation Pattern
#![allow(unused)] fn main() { async fn add_contact(&mut self, invitation: Invitation) -> Result<ContactId> { let ceremony_id = self.ceremony_executor .initiate_invitation_ceremony(invitation) .await?; loop { match self.ceremony_executor.get_status(&ceremony_id)? { CeremonyStatus::Committed => { return Ok(ContactId::from_ceremony(&ceremony_id)); } CeremonyStatus::Aborted { reason } => { return Err(AuraError::ceremony_failed(reason)); } _ => { tokio::time::sleep(POLL_INTERVAL).await; } } } } }
This pattern blocks until the ceremony completes. The user sees progress UI during execution. Context is established only on successful commit.
5. Ceremony Contract
All Category C ceremonies follow a shared contract that ensures atomic commit/abort semantics.
5.1 Ceremony Phases
-
Compute prestate: Derive a stable prestate hash from the authority/context state being modified. Include the current epoch and effective participant set.
-
Propose operation: Define the operation being performed. Compute an operation hash bound to the proposal parameters.
-
Enter pending epoch: Generate new key material at a pending epoch without invalidating the old epoch. Store metadata for commit or rollback.
-
Collect responses: Send invitations/requests to participants. Participants respond using their full runtimes. Responses must be authenticated and recorded as facts.
-
Commit or abort: If acceptance/threshold conditions are met, commit the pending epoch and emit resulting facts. Otherwise abort, emit an abort fact with a reason, and leave the prior epoch active.
5.2 Ceremony Properties
All Category C ceremonies implement:
-
Prestate Binding:
CeremonyId = H(prestate_hash, operation_hash, nonce)prevents concurrent ceremonies on same state and ensures exactly-once semantics. -
Atomic Commit/Abort: Either fully committed or no effect. No partial state possible.
-
Epoch Isolation: Uncommitted key packages are inert. No explicit rollback needed on abort.
-
Session Types: Protocol compliance enforced at compile time via choreographic projection and enforced at runtime via Telltale adapter/VM execution.
5.3 Per-Ceremony Policy Matrix
Authority and Device Ceremonies
| Ceremony | Key Gen | Agreement | Fallback | Notes |
|---|---|---|---|---|
| Authority bootstrap | K1 | A3 | None | Local, immediate |
| Device enrollment | K2 | A1→A2→A3 | A1/A2 | Provisional → soft-safe → finalize |
| Device MFA rotation | K3 | A2→A3 | A2 | Consensus-finalized keys |
| Device removal | K3 | A2→A3 | A2 | Remove via rotation |
Guardian Ceremonies
| Ceremony | Key Gen | Agreement | Fallback | Notes |
|---|---|---|---|---|
| Guardian setup/rotation | K3 | A2→A3 | A2 | Consensus-finalized for durability |
| Recovery approval | — | A2→A3 | A2 | Soft-safe approvals → consensus |
| Recovery execution | — | A2→A3 | A2 | Consensus-finalized commit |
Channel and Group Ceremonies
| Ceremony | Key Gen | Agreement | Fallback | Notes |
|---|---|---|---|---|
| AMP channel epoch bump | — | A1→A2→A3 | A1/A2 | Proposed → cert → commit |
| AMP channel bootstrap | — | A1→A2→A3 | A1/A2 | Provisional → group key rotation |
| Group/Block creation | K3 | A1→A2→A3 | A1/A2 | Provisional bootstrap → consensus |
| Rendezvous secure-channel | — | A1→A2→A3 | A1/A2 | Provisional → consensus |
Other Ceremonies
| Ceremony | Key Gen | Agreement | Fallback | Notes |
|---|---|---|---|---|
| Invitation (contact/channel/guardian) | — | A3 | None | Consensus-finalized only |
| OTA activation | — | A2→A3 | A2 | Threshold-signed → consensus |
| DKD ceremony | DKD | A2→A3 | A2 | Multi-party derivation → commit |
5.4 Bootstrap Exception
When creating a new group/channel before the group key ceremony completes, Aura allows a bootstrap epoch using a trusted-dealer key (K2/A1). The dealer distributes a bootstrap key with the channel invite, enabling immediate encrypted messaging. This is explicitly provisional and superseded by the consensus-finalized group key (K3/A3) once the ceremony completes.
6. Consistency Metadata
Each operation category has a purpose-built status type for tracking consistency.
6.1 Core Types
#![allow(unused)] fn main() { pub enum Agreement { Provisional, SoftSafe { cert: Option<ConvergenceCert> }, Finalized { consensus_id: ConsensusId }, } pub enum Propagation { Local, Syncing { peers_reached: u16, peers_known: u16 }, Complete, Failed { retry_at: PhysicalTime, retry_count: u32, error: String }, } pub struct Acknowledgment { pub acked_by: Vec<AckRecord>, } }
Agreement indicates the finalization level (A1/A2/A3). Propagation tracks anti-entropy sync status. Acknowledgment tracks explicit per-peer delivery confirmation.
6.2 Category A: OptimisticStatus
#![allow(unused)] fn main() { pub struct OptimisticStatus { pub agreement: Agreement, pub propagation: Propagation, pub acknowledgment: Option<Acknowledgment>, } }
Use cases include send message, create channel, update profile, and react to message.
UI patterns:
◐Sending: propagation == Local✓Sent: propagation == Complete✓✓Delivered: acknowledgment.count() >= expected.len()◆Finalized: agreement == Finalized
6.3 Category B: DeferredStatus
#![allow(unused)] fn main() { pub struct DeferredStatus { pub proposal_id: ProposalId, pub state: ProposalState, pub approvals: ApprovalProgress, pub applied_agreement: Option<Agreement>, pub expires_at: PhysicalTime, } pub enum ProposalState { Pending, Approved, Rejected { reason: String, by: AuthorityId }, Expired, Superseded { by: ProposalId }, } }
Use cases include change permissions, remove member, transfer ownership, and archive channel.
6.4 Category C: CeremonyStatus
#![allow(unused)] fn main() { pub struct CeremonyStatus { pub ceremony_id: CeremonyId, pub state: CeremonyState, pub responses: Vec<ParticipantResponse>, pub prestate_hash: Hash32, pub committed_agreement: Option<Agreement>, } pub enum CeremonyState { Preparing, PendingEpoch { pending_epoch: Epoch, required_responses: u16, received_responses: u16 }, Committing, Committed { consensus_id: ConsensusId, committed_at: PhysicalTime }, Aborted { reason: String, aborted_at: PhysicalTime }, Superseded { by: CeremonyId, reason: SupersessionReason }, } }
Use cases include add contact, create group, guardian rotation, device enrollment, and recovery.
When a ceremony commits successfully, committed_agreement is set to Agreement::Finalized with the consensus ID, indicating A3 durability.
6.5 Unified Consistency Type
For cross-category queries and generic handling:
#![allow(unused)] fn main() { pub struct Consistency { pub category: OperationCategory, pub agreement: Agreement, pub propagation: Propagation, pub acknowledgment: Option<Acknowledgment>, } pub enum OperationCategory { Optimistic, Deferred { proposal_id: ProposalId }, Ceremony { ceremony_id: CeremonyId }, } }
7. Ceremony Supersession
When a new ceremony replaces an old one, Aura emits explicit supersession facts that propagate via anti-entropy.
7.1 Supersession Reasons
#![allow(unused)] fn main() { pub enum SupersessionReason { PrestateStale, NewerRequest, ExplicitCancel, Timeout, Precedence, } }
PrestateStale indicates the prestate changed while the ceremony was pending. NewerRequest indicates an explicit newer request from the same initiator. ExplicitCancel indicates manual cancellation by an authorized participant. Timeout indicates the ceremony exceeded its validity window. Precedence indicates a concurrent ceremony won via conflict resolution.
7.2 Supersession Facts
Each ceremony fact enum includes a CeremonySuperseded variant:
#![allow(unused)] fn main() { CeremonySuperseded { superseded_ceremony_id: String, superseding_ceremony_id: String, reason: String, trace_id: Option<String>, timestamp_ms: u64, } }
7.3 CeremonyTracker API
The CeremonyTracker in aura-agent maintains supersession records for auditability:
| Method | Purpose |
|---|---|
supersede(old_id, new_id, reason) | Record a supersession event |
check_supersession_candidates(prestate_hash, op_type) | Find stale ceremonies |
get_supersession_chain(ceremony_id) | Get full supersession history |
is_superseded(ceremony_id) | Check if ceremony was replaced |
Supersession facts propagate via the existing anti-entropy mechanism. Peers receiving a CeremonySuperseded fact update their local ceremony state accordingly.
8. Decision Tree
Use this tree to categorize new operations:
Does this operation establish or modify cryptographic relationships?
│
├─ YES: Does the user need to wait for completion?
│ │
│ ├─ YES (new context, key changes) → Category C (Blocking Ceremony)
│ │ Examples: add contact, create group, guardian rotation
│ │
│ └─ NO (removal from existing context) → Category B (Deferred)
│ Examples: remove from group (epoch rotation in background)
│
└─ NO: Does this affect other users' access or policies?
│
├─ YES: Is this high-security or irreversible?
│ │
│ ├─ YES → Category B (Deferred)
│ │ Examples: transfer ownership, delete channel, kick member
│ │
│ └─ NO → Category A (Optimistic)
│ Examples: pin message, update topic
│
└─ NO → Category A (Optimistic)
Examples: send message, create channel, block contact
9. UI Feedback Patterns
9.1 Category A: Instant Result with Sync Indicator
┌───────────────────────────────────┐
│ You: Hello everyone! ◆ ✓✓ │ ← Finalized + Delivered
│ You: Check this out ✓✓ │ ← Delivered (not yet finalized)
│ You: Another thought ✓ │ ← Sent
│ You: New idea ◐ │ ← Sending
└───────────────────────────────────┘
Effect already applied. Indicators show delivery status (◐ → ✓ → ✓✓ → ✓✓ blue) and finalization (◆ appears when A3 consensus achieved).
9.2 Category B: Pending Indicator
┌─────────────────────────────────────────────────────────────────────┐
│ Channel: #project │
├─────────────────────────────────────────────────────────────────────┤
│ Pending: Remove Carol (waiting for Bob to confirm) │
├─────────────────────────────────────────────────────────────────────┤
│ Members: │
│ Alice (moderator) ✓ │
│ Bob (moderator) ✓ │
│ Carol ✓ ← Still has access until confirmed │
└─────────────────────────────────────────────────────────────────────┘
Proposal shown. Effect NOT applied yet.
9.3 Category C: Blocking Wait
┌─────────────────────────────────────────────────────────────────────┐
│ Adding Bob to group... │
│ │
│ ✓ Invitation sent │
│ ✓ Bob accepted │
│ ◐ Deriving group keys... │
│ ○ Ready │
│ │
│ [Cancel] │
└─────────────────────────────────────────────────────────────────────┘
User waits. Cannot proceed until ceremony completes.
10. Effect Policy Configuration
Operations use configurable policies that reference the capability system:
#![allow(unused)] fn main() { pub struct EffectPolicy { pub operation: OperationType, pub timing: EffectTiming, pub security_level: SecurityLevel, } pub enum EffectTiming { Immediate, Deferred { requires_approval_from: Vec<CapabilityRequirement>, timeout_ms: u64, threshold: ApprovalThreshold, }, Blocking { ceremony: CeremonyType, }, } }
10.1 Context-Specific Overrides
Contexts can override default policies:
#![allow(unused)] fn main() { // Strict security channel: unanimous moderator approval for kicks channel.set_effect_policy(RemoveFromChannel, EffectTiming::Deferred { requires_approval_from: vec![CapabilityRequirement::Role("moderator")], timeout_ms: 48 * 60 * 60 * 1000, threshold: ApprovalThreshold::Unanimous, }); // Casual channel: any moderator can kick immediately channel.set_effect_policy(RemoveFromChannel, EffectTiming::Immediate); }
11. Full Operation Matrix
| Operation | Category | Effect Timing | Security | Notes |
|---|---|---|---|---|
| Within Established Context | ||||
| Send message | A | Immediate | Low | Keys already derived |
| Create channel | A | Immediate | Low | Just facts into context |
| Update topic | A | Immediate | Low | CRDT, last-write-wins |
| React to message | A | Immediate | Low | Local expression |
| Local Authority | ||||
| Block contact | A | Immediate | Low | Your decision |
| Mute channel | A | Immediate | Low | Local preference |
| Policy Changes | ||||
| Change permissions | B | Deferred | Medium | Others affected |
| Kick from channel | B | Deferred | Medium | Affects access |
| Archive channel | B | Deferred | Low-Med | Reversible |
| High Risk | ||||
| Transfer ownership | B | Deferred | High | Irreversible |
| Delete channel | B | Deferred | High | Data loss |
| Remove from context | B | Deferred | High | Affects encryption |
| Cryptographic | ||||
| Add contact | C | Blocking | Critical | Creates context |
| Create group | C | Blocking | Critical | Multi-party keys |
| Add group member | C | Blocking | Critical | Changes group keys |
| Device enrollment | C | Blocking | Critical | DeviceEnrollment choreography |
| Guardian rotation | C | Blocking | Critical | Key shares |
| Recovery execution | C | Blocking | Critical | Account state |
| Device revocation | C | Blocking | Critical | Security response |
12. Common Mistakes to Avoid
Mistake 1: Making Everything Category C
Wrong: "Adding a channel member requires ceremony"
Right: If the member is already in the relational context, it is Category A. Just emit a fact. Only if they need to join the context first is it Category C.
Mistake 2: Forgetting Context Existence
Wrong: Trying to create a channel before establishing relationship
Right: Contact invitation (Category C) must complete before any channel operations (Category A) are possible.
Mistake 3: Optimistic Key Operations
Wrong: "User can start using new guardians while ceremony runs"
Right: Guardian changes affect key shares. Partial state means unusable keys. Must be Category C.
Mistake 4: Blocking on Low-Risk Operations
Wrong: "Wait for all members to confirm before showing channel"
Right: Channel creation is optimistic. Show immediately, sync status later.
See Also
- Consensus for fast path and fallback consensus
- Journal for fact semantics and reduction flows
- AMP Protocol for channel encryption and key derivation
- Relational Contexts for context vs channel distinction
- Choreography Guide for session types in Category C
- Transport for sync status tracking
- Effect System for effect policies
MPST and Choreography
This document describes the architecture of choreographic protocols in Aura. It explains how global protocols are defined, projected, and executed. It defines the structure of local session types, the integration with the Effect System, and the use of guard chains and journal coupling.
1. DSL and Projection
Aura defines global protocols using the choreography! macro. The macro parses a global specification into an abstract syntax tree. The macro produces code that represents the protocol as a choreographic structure. The source of truth for protocols is a .choreo file stored next to the Rust module that loads it.
Projection converts the global protocol into per-role local session types. Each local session type defines the exact sequence of sends and receives for a single role. Projection eliminates deadlocks and ensures that communication structure is correct.
#![allow(unused)] fn main() { choreography!(include_str!("example.choreo")); }
Example file: example.choreo
module example exposing (Example)
protocol Example =
roles A, B
A -> B : Msg(data: Vec<u8>)
B -> A : Ack(code: u32)
This snippet defines a global protocol with two roles. Projection produces a local type for A and a local type for B. Each local type enforces the required ordering at compile time.
2. Local Session Types
Local session types describe the allowed actions for a role. Each send and receive is represented as a typed operation. Local types prevent protocol misuse by ensuring that nodes follow the projected sequence.
Local session types embed type-level guarantees. These guarantees prevent message ordering errors. They prevent unmatched sends or receives. Each protocol execution must satisfy the session type.
#![allow(unused)] fn main() { type A_Local = Send<B, Msg, Receive<B, Ack, End>>; }
This example shows the projected type for role A. The type describes that A must send Msg to B and then receive Ack.
3. Runtime Integration
Aura executes production choreographies through the Telltale VM. The choreography! macro emits the global type, projected local types, role metadata, and composition metadata that the runtime uses to build VM code images. AuraChoreoEngine in crates/aura-agent/src/runtime/choreo_engine.rs is the production runtime surface.
Generated runners still expose role-specific execution helpers. Aura keeps those helpers for tests, focused migration utilities, and narrow tooling paths. They are not the production execution boundary.
Generated runtime artifacts also carry the data that production startup needs:
provide_messagefor outbound payloadsselect_branchfor choice decisions- protocol id and determinism policy reference
- required capability keys
- link and delegation constraints
- operational-envelope selection inputs
These values are sourced from runtime state such as params, journal facts, UI inputs, and manifest-driven admission state.
Aura has one production choreography backend:
- VM backend (
AuraChoreoEngine) for admitted Telltale VM execution, replay, and parity checks.
Direct generated-runner execution is test and migration support only.
Production runtime ownership is fragment-scoped. The admitted unit is one VM fragment derived from the generated CompositionManifest. A manifest without link metadata yields one protocol fragment. A manifest with link metadata yields one fragment per linked bundle.
delegate and link define how ownership moves. Local runtime services claim fragment ownership through AuraEffectSystem. Runtime transfer goes through ReconfigurationManager. The runtime rejects ambiguous local ownership before a transfer reaches the VM.
VmBridgeEffects is the synchronous host boundary for one fragment. VM callbacks use it for session-local payload queues, blocked receive snapshots, and scheduler signals. Async transport, journal, and storage work stay outside the callback path in the host bridge loop.
4. Choreography Annotations and Effect Commands
Choreographies support annotations that modify runtime behavior. The choreography! macro extracts these annotations and generates EffectCommand sequences. This follows the choreography-first architecture where choreographic annotations are the canonical source of truth for guard requirements.
Supported Annotations
| Annotation | Description | Generated Effect |
|---|---|---|
guard_capability = "cap" | Capability requirement | StoreMetadata (audit trail) |
flow_cost = N | Flow budget charge | ChargeBudget |
journal_facts = "fact" | Journal fact recording | StoreMetadata (fact key) |
journal_merge = true | Request journal merge | StoreMetadata (merge flag) |
audit_log = "event" | Audit trail entry | StoreMetadata (audit key) |
leak = "External" | Leakage tracking | RecordLeakage |
Annotation Syntax
#![allow(unused)] fn main() { // Single annotation A[guard_capability = "sync"] -> B: SyncMsg; // Multiple annotations A[guard_capability = "sync", flow_cost = 10, journal_facts = "sync_started"] -> B: SyncMsg; // Leakage annotation (multiple syntaxes supported) A[leak = "External,Neighbor"] -> B: PublicMsg; A[leak: External] -> B: PublicMsg; }
Protocol Artifact Requirements
Aura binds choreography bundles to runtime capability requirements through generated CompositionManifest metadata and aura-protocol::admission. Production startup uses open_manifest_vm_session_admitted(...) so protocol id, capability requirements, determinism policy reference, and link constraints come from one canonical manifest.
Admission also resolves the runtime execution envelope. Cooperative protocols stay on the canonical VM path. Replay-deterministic and envelope-bounded protocols select the threaded runtime path only when the required runtime capabilities and envelope artifacts are present.
Current registry mappings:
aura.consensus->byzantine_envelopeaura.sync.epoch_rotation->termination_boundedaura.dkg.ceremony->byzantine_envelope+termination_boundedaura.recovery.grant->termination_bounded
For lower-level VM tests, AuraChoreoEngine::open_session_admitted(...) remains available. Production services should use open_manifest_vm_session_admitted(...) instead.
Dynamic Reconfiguration (@link + delegation)
Aura treats protocol reconfiguration as a first-class choreography concern. Static composition uses @link metadata in choreographies so compatible bundles can be merged at compile time. Runtime transfer uses delegation receipts and session-footprint coherence checks.
@linkannotations are parsed byaura-mpstand validated byaura-macrosfor import/export compatibility.ReconfigurationManagerinaura-agentexecutes link/delegate operations and verifies coherence after each delegation.- Delegation writes a
SessionDelegationFactaudit record to the relational journal.
This model is used by device migration and guardian handoff flows: session ownership moves without restarting the full choreography, while invariants remain checkable from persisted facts.
The runtime now treats linked composition metadata as an ownership boundary as well as a composition boundary. Fragment keys derive from protocol id or link bundle id. This keeps runtime ownership aligned with Telltale's composition model.
Protocol Evolution Compatibility Policy
Aura classifies choreography evolution by comparing baseline and candidate
projections with async_subtype(new_local, old_local) for every shared role.
Compatibility rule:
- Compatible if
async_subtypesucceeds for every shared role. - Breaking if
async_subtypefails for any shared role.
Version bump rules:
- Patch: no projection-shape change (comments, docs, metadata-only updates).
- Minor: projection changes are present, but compatibility check passes for all shared roles.
- Major: any compatibility failure, role-set change, or protocol namespace replacement.
Reviewer decision table:
| Change type | Default classification | Required action |
|---|---|---|
| Additive branch (existing roles unchanged) | Minor if async-subtype passes | Run compatibility checker; bump minor on pass, major on fail |
| Payload narrowing/widening | Minor or Major based on async-subtype result | Treat checker as source of truth; bump major on any role failure |
| Role added/removed/renamed | Major | Bump major and treat as new protocol contract |
| Namespace change with same semantics | Major (new contract surface) | Register as new protocol namespace and keep old baseline until migration completes |
Operational notes:
- Compatibility checks are tooling/CI gates, never runtime hot-path logic.
- New protocol files without a baseline are not auto-classified; reviewers must explicitly approve initial versioning.
- Baseline/current checks are implemented by
scripts/check/protocol-compat.shandjust ci-protocol-compat.
Quantitative Termination Budgets
Aura derives deterministic step budgets from Telltale's weighted measure:
W = 2 * sum(depth(local_type)) + sum(buffer_sizes)
Runtime execution applies:
max_steps = ceil(k_sigma(protocol) * W * budget_multiplier)
Implementation points:
aura-mpst::terminationcomputes local-type depth, buffer weight, and weighted measure.aura-protocol::terminationdefines protocol classes (consensus,sync,dkg,recovery), calibratedk_sigmafactors, and expected budget ranges.AuraChoreoEngineconstructs aTerminationBudgetat run start, checks budget consumption after every VM scheduler step, emits deterministicBoundExceededfailures, warns at >80% utilization, and logs multiplier divergence from computed bounds.
Effect Command Generation
The macro generates an effect_bridge module containing:
#![allow(unused)] fn main() { pub mod effect_bridge { use aura_core::effects::guard::{EffectCommand, EffectInterpreter}; /// Convert annotations to effect commands pub fn annotation_to_commands(ctx: &EffectContext, annotation: ...) -> Vec<EffectCommand>; /// Execute commands through interpreter pub async fn execute_commands<I: EffectInterpreter>( interpreter: &I, ctx: &EffectContext, annotations: Vec<...>, ) -> Result<Vec<EffectResult>, String>; } }
Macro Output Contracts & Stability
The choreography! macro emits a stable, minimal surface intended for runtime integration.
Consumers should rely only on the contracts below. All other generated items are internal and
may change without notice.
Stable contracts:
effect_bridge::EffectBridgeContext(context, authority, peer, timestamp)effect_bridge::annotation_to_commands(...)effect_bridge::annotations_to_commands(...)effect_bridge::create_context(...)effect_bridge::execute_commands(...)
Stability rules:
- The above names and signatures are stable within a minor release series.
- Additional helper functions may be added, but existing signatures will not change without a schema/version bump in the macro output.
- Generated role enums and protocol-specific modules are not part of the stable API surface.
Integration with Effect Interpreters
Generated EffectCommand sequences execute through:
- Production:
ProductionEffectInterpreter(aura-effects) - Simulation:
SimulationEffectInterpreter(aura-simulator) - Testing:
BorrowedEffectInterpreter/ mock interpreters
This unified approach ensures consistent guard behavior across all execution environments.
5. Guard Chain Integration
Guard effects originate from two sources that share the same EffectCommand system:
-
Choreographic Annotations (compile-time): The
choreography!macro generatesEffectCommandsequences from annotations. These represent per-message guard requirements. -
Runtime Guard Chain (send-site): The
GuardChain::standard()evaluates pure guards against aGuardSnapshotat each protocol send site. This enforces invariants like charge-before-send.
Guard Chain Sequence
The runtime guard chain contains CapGuard, FlowGuard, JournalCoupler, and LeakageTracker. These guards enforce authorization and budget constraints:
CapGuardchecks that the active capabilities satisfy the message requirementsFlowGuardchecks that flow budget is available for the context and peerJournalCouplersynchronizes journal updates with protocol executionLeakageTrackerrecords metadata leakage per observer class
Guard evaluation is synchronous over a prepared GuardSnapshot and yields EffectCommand items. An async interpreter executes those commands, keeping guard logic pure while preserving charge-before-send.
graph TD
S[Send] --> A[Annotation Effects];
A --> C[CapGuard];
C --> F[FlowGuard];
F --> J[JournalCoupler];
J --> L[LeakageTracker];
L --> N[Network Send];
This diagram shows the combined guard sequence. Annotation-derived effects execute first, then runtime guards validate and charge budgets before the send.
Combined Execution
Use execute_guarded_choreography() from aura_guards to execute both annotation-derived commands and runtime guards atomically:
#![allow(unused)] fn main() { use aura_guards::{execute_guarded_choreography, GuardChain}; let result = execute_guarded_choreography( &effect_system, &request, annotation_commands, // From choreography macro interpreter, ).await?; }
6. Execution Modes
Aura supports multiple execution environments for the same choreography definitions. Production execution uses admitted VM sessions with real effect handlers. Simulation execution uses deterministic time and fault injection. Test utilities may use narrower runner surfaces when that improves isolation.
Each environment preserves the same protocol structure and admission semantics where applicable. Choreography execution also captures conformance artifacts for native/WASM parity testing. See Test Infrastructure Reference for artifact surfaces and effect classification.
7. Example Protocols
Anti-entropy protocols synchronize CRDT state. They run as choreographies that exchange state deltas. Session types ensure that the exchange pattern follows causal and structural rules.
FROST ceremonies use choreographies to coordinate threshold signing. These ceremonies use the guard chain to enforce authorization rules.
Aura Consensus uses choreographic notation for fast path and fallback flows. Consensus choreographies define execute, witness, and commit messages. Session types ensure evidence propagation and correctness.
#![allow(unused)] fn main() { choreography! { #[namespace = "sync"] protocol AntiEntropy { roles: A, B; A -> B: Delta(data: Vec<u8>); B -> A: Ack(data: Vec<u8>); } } }
This anti-entropy example illustrates a minimal synchronization protocol.
8. Operation Categories and Choreography Use
Not all multi-party operations require full choreographic specification. Aura classifies operations into categories that determine when choreography is necessary.
8.1 When to Use Choreography
Category C (Consensus-Gated) Operations - Full choreography required:
- Guardian rotation ceremonies
- Recovery execution flows
- OTA hard fork activation
- Device revocation
- Adding contacts / creating groups (establishing cryptographic context)
- Adding members to existing groups
These operations require explicit session types because:
- Partial execution is dangerous
- All parties must agree before effects apply
- Strong ordering guarantees are necessary
Example - Invitation Ceremony:
#![allow(unused)] fn main() { choreography! { #[namespace = "invitation"] protocol InvitationCeremony { roles: Sender, Receiver; Sender -> Receiver: Invitation(data: InvitationPayload); Receiver -> Sender: Accept(commitment: Hash32); // Context is now established } } }
8.2 When Choreography is NOT Required
Category A (Optimistic) Operations - No choreography needed:
- Send message (within established context)
- Create channel (within existing relational context)
- Update channel topic
- Block/unblock contact
These use simple CRDT fact emission because:
- Cryptographic context already exists
- Keys derive deterministically from shared state
- Eventual consistency is sufficient
- No coordination required
Example - No choreography:
#![allow(unused)] fn main() { // Just emit a fact - no ceremony needed journal.append(ChannelCheckpoint { context: existing_context_id, channel: new_channel_id, chan_epoch: 0, base_gen: 0, window: 1024, .. }); }
Category B (Deferred) Operations - May use lightweight choreography:
- Change channel permissions
- Remove channel member (may be contested)
- Transfer ownership
These may use a proposal/approval pattern but don't require the full ceremony infrastructure.
8.3 Decision Tree for Protocol Design
Is this operation establishing or modifying cryptographic relationships?
│
├─ YES → Use full choreography (Category C)
│ Define explicit session types and guards
│
└─ NO: Does this affect other users' policies/access?
│
├─ YES: Is strong agreement required?
│ │
│ ├─ YES → Use lightweight choreography (Category B)
│ │ Proposal/approval pattern
│ │
│ └─ NO → Use CRDT facts (Category A)
│ Eventually consistent
│
└─ NO → Use CRDT facts (Category A)
No coordination needed
See Consensus - Operation Categories for detailed categorization.
9. Choreography Inventory
This section lists all choreographies in the codebase with their locations and purposes.
9.1 Core Protocols
| Protocol | Location | Purpose |
|---|---|---|
| AuraConsensus | aura-consensus/src/protocol/choreography.choreo | Fast path and fallback consensus |
| AmpTransport | aura-amp/src/choreography.choreo | Asynchronous message transport |
9.2 Rendezvous Protocols
| Protocol | Location | Purpose |
|---|---|---|
| RendezvousExchange | aura-rendezvous/src/protocol.rendezvous_exchange.choreo | Direct peer discovery |
| RelayedRendezvous | aura-rendezvous/src/protocol.relayed_rendezvous.choreo | Relay-assisted connection |
9.3 Authentication Protocols
| Protocol | Location | Purpose |
|---|---|---|
| GuardianAuthRelational | aura-authentication/src/guardian_auth_relational.choreo | Guardian authentication |
| DkdChoreography | aura-authentication/src/dkd.choreo | Distributed key derivation |
9.4 Recovery Protocols
| Protocol | Location | Purpose |
|---|---|---|
| RecoveryProtocol | aura-recovery/src/recovery_protocol.choreo | Account recovery flow |
| GuardianMembershipChange | aura-recovery/src/guardian_membership.choreo | Guardian add/remove |
| GuardianCeremony | aura-recovery/src/guardian_ceremony.choreo | Guardian key ceremony |
| GuardianSetup | aura-recovery/src/guardian_setup.choreo | Initial guardian setup |
9.5 Invitation Protocols
| Protocol | Location | Purpose |
|---|---|---|
| InvitationExchange | aura-invitation/src/protocol.invitation_exchange.choreo | Contact/channel invitation |
| GuardianInvitation | aura-invitation/src/protocol.guardian_invitation.choreo | Guardian invitation |
| DeviceEnrollment | aura-invitation/src/protocol.device_enrollment.choreo | Device enrollment |
9.6 Sync Protocols
| Protocol | Location | Purpose |
|---|---|---|
| EpochRotationProtocol | aura-sync/src/protocols/epochs.choreo | Epoch rotation sync |
9.7 Runtime Protocols
| Protocol | Location | Purpose |
|---|---|---|
| SessionCoordination | aura-agent/src/handlers/sessions/coordination.choreo | Session creation coordination |
10. Runtime Infrastructure
The runtime provides production choreographic execution through manifest-driven Telltale VM sessions.
10.1 ChoreographicEffects Trait
| Method | Purpose |
|---|---|
send_to_role_bytes | Send message to specific role |
receive_from_role_bytes | Receive message from specific role |
broadcast_bytes | Broadcast to all roles |
start_session | Initialize choreography session |
end_session | Terminate choreography session |
AuraVmEffectHandler is the synchronous host boundary between the VM and Aura runtime services. AuraQueuedVmBridgeHandler provides queued outbound payloads and branch decisions for role-scoped VM sessions.
10.2 Wiring a Choreography
- Store the protocol in a
.choreofile next to the Rust module that loads it. - Use
choreography!(include_str!("..."))to generate the protocol module, VM artifacts, and composition metadata. - Open the session with
open_manifest_vm_session_admitted(...)from the runtime bridge or service layer. - Provide decision sources for
provide_messageandselect_branchthrough the VM host bridge.
10.3 Decision Sourcing
Generated runners call provide_message for outbound payloads and select_branch for choice decisions. These are sourced from runtime state:
| Source Type | Examples |
|---|---|
| Params | Consensus parameters, invitation payloads |
| Journal facts | Local state, authority commitments |
| Service state | Ceremony proposals, channel state |
| UI/Policy | Accept/reject decisions, commit/abort choices |
10.4 Integration Features
The runtime provides guard chain integration (CapGuard → FlowGuard → JournalCoupler), transport effects for message passing, manifest-driven admission, determinism policy enforcement, typed link and delegation checks, and session lifecycle management with metrics.
10.5 Output and Flow Policy Integration Points
Aura binds choreography execution to VM output/flow gates at the runtime boundary.
AuraVmEffectHandler tags VM-observable operations with output-condition predicate hints so OutputConditionPolicy can enforce commit visibility rules. The hardening profile allow-list admits only known predicates (transport send/recv, protocol choice/step, guard acquire/release). Unknown predicates are rejected in CI profiles.
Flow constraints are enforced with FlowPolicy::PredicateExpr(...) derived from Aura role/category constraints. This keeps pre-send flow checks aligned with Aura's information-flow contract while preserving deterministic replay behavior.
Practical integration points:
- Choreography annotations declare intent (
guard_capability,flow_cost,journal_facts,leak). - Macro output emits
EffectCommandsequences. - Guard chain evaluates commands and budgets at send sites.
- VM output/flow policies gate observable commits and cross-role message flow before transport effects execute.
This means choreography-level guard semantics and VM-level hardening are additive, not competing: annotations define required effects; policies constrain which effects are allowed to become observable.
11. Summary
Aura uses choreographic programming to define global protocols. Projection produces local session types. Session types enforce structured communication. Handlers execute protocol steps using effect traits. Extension effects provide authorization, budgeting, and journal updates. Execution modes support testing, simulation, and production. Choreographies define distributed coordination for CRDT sync, FROST signing, and consensus.
Not all multi-party operations need choreography. Operations within established cryptographic contexts use optimistic CRDT facts. Choreography is reserved for Category C operations where partial state would be dangerous.
Transport and Information Flow
This document describes the architecture of transport, guard chains, flow budgets, receipts, and information flow in Aura. It defines the secure channel abstraction and the enforcement mechanisms that regulate message transmission. It explains how context boundaries scope capabilities and budgets.
1. Transport Abstraction
Aura provides a transport layer that delivers encrypted messages between authorities. Each transport connection is represented as a SecureChannel. A secure channel binds a pair of authorities and a context identifier. A secure channel maintains isolation across contexts.
A secure channel exposes a send operation and a receive operation. The channel manages replay protection and handles connection teardown on epoch changes.
#![allow(unused)] fn main() { pub struct SecureChannel { pub context: ContextId, pub peer: AuthorityId, pub channel_id: Uuid, } }
This structure identifies a single secure channel. One channel exists per (ContextId, peer) pair. Channel metadata binds the channel to a specific context epoch.
2. Guard Chain
All transport sends pass through the guard chain defined in Authorization. CapGuard evaluates Biscuit capabilities and sovereign policy. FlowGuard charges the per-context flow budget and produces a receipt. JournalCoupler records the accompanying facts atomically. Each stage must succeed before the next stage executes. Guard evaluation runs synchronously over a prepared GuardSnapshot and returns EffectCommand data. An async interpreter executes those commands so guards never perform I/O directly.
3. Flow Budget and Receipts
Flow budgets limit the amount of data that an authority may send within a context. The flow budget model defines a quota for each (ContextId, peer) pair. A reservation system protects against race conditions.
An authority must reserve budget before sending. A reservation locks a portion of the available budget. The actual charge occurs during the guard chain. If the guard chain succeeds, a receipt is created.
#![allow(unused)] fn main() { /// From aura-core/src/types/flow.rs pub struct Receipt { pub ctx: ContextId, pub src: AuthorityId, pub dst: AuthorityId, pub epoch: Epoch, pub cost: FlowCost, pub nonce: FlowNonce, pub prev: Hash32, pub sig: ReceiptSig, } }
This structure defines a receipt. A receipt binds a cost to a specific context and epoch. The sender signs the receipt. The nonce ensures uniqueness and the prev field chains receipts for auditing. The recipient verifies the signature. Receipts support accountability in multi-hop routing.
4. Information Flow Budgets
Information flow budgets define limits on metadata leakage. Budgets exist for external leakage, neighbor leakage, and group leakage. Each protocol message carries leakage annotations. These annotations specify the cost for each leakage dimension.
Leakage budgets determine if a message can be sent. If the leakage cost exceeds the remaining budget, the message is denied. Enforcement uses padding and batching strategies. Padding hides message size. Batching hides message frequency.
#![allow(unused)] fn main() { pub struct LeakageBudget { pub external: u32, pub neighbor: u32, pub in_group: u32, } }
This structure defines the leakage budget for a message. Leakage costs reduce the corresponding budget on successful send.
5. Context Integration
Capabilities and flow budgets are scoped to a ContextId. Each secure channel associates all guard decisions with its context. A capability is valid only for the context in which it was issued. A flow budget applies only within the same context.
Derived context keys bind communication identities to the current epoch. When the account epoch changes, all context identities must refresh. All secure channels for the context must be renegotiated.
#![allow(unused)] fn main() { pub struct ChannelContext { pub context: ContextId, pub epoch: u64, pub caps: Vec<Capability>, } }
This structure defines the active context state for a channel. All guard chain checks use these values.
6. Failure Modes and Observability
The guard chain defines three categories of failure. A denial failure occurs when capability requirements are not met. A block failure occurs when a flow budget check fails. A commit failure occurs when journal coupling fails.
Denial failures produce no observable behavior. Block failures also produce no observable behavior. Commit failures prevent sending and produce local error logs. None of these failures result in network traffic.
This design ensures that unauthorized or over-budget sends do not produce side channels.
7. Security Properties
Aura enforces no observable behavior without charge. A message cannot be sent unless flow budget is charged first. Capability gated sends ensure that each message satisfies authorization rules. Receipts provide accountability for multi-hop forwarding.
The network layer does not reveal authority structure. Context identifiers do not reveal membership. All metadata is scoped to individual relationships.
8. Secure Channel Lifecycle
Secure channels follow a lifecycle aligned with rendezvous and epoch semantics:
-
Establishment:
- Ranch rendezvous per Rendezvous Architecture to exchange descriptors inside the relational context journal.
- Each descriptor contains transport hints, a handshake PSK derived from the context key, and a
punch_nonce. - Once both parties receive offer/answer envelopes, they perform Noise IKpsk2 using the context-derived keys and establish a QUIC or relay-backed channel bound to
(ContextId, peer).
-
Steady state:
- Guard chain enforces CapGuard → FlowGuard → JournalCoupler for every send.
- FlowBudget receipts created on each hop are inserted into the relational context journal so downstream peers can audit path compliance.
-
Re-keying on epoch change:
- When the account or context epoch changes (as recorded in Authority and Identity / Relational Contexts), the channel detects the mismatch, tears down the existing Noise session, and triggers rendezvous to derive fresh keys.
- Existing receipts are marked invalid for the new epoch, preventing replay.
-
Teardown:
- Channels close explicitly when contexts end or when FlowGuard hits the configured budget limit.
- Receipts emitted during teardown propagate through the relational context journal so guardians or auditors can verify all hops charged their budgets up to the final packet.
By tying establishment and teardown to relational context journals, receipts become part of the same fact set tracked in 115_maintenance.md, ensuring long-term accountability.
8. Privacy-by-Design Patterns
The aura-transport crate demonstrates privacy-by-design principles where privacy mechanisms are integrated into core types rather than added as external concerns. This section extracts key patterns from aura-transport for use across the codebase.
8.1 Core Principles
Privacy-by-Design Integration:
- Privacy mechanisms built into core types (Envelope, FrameHeader, TransportConfig)
- Privacy levels as first-class configuration, not bolt-on features
- Relationship scoping embedded in message routing
- Capability blinding at the envelope level
Minimal Metadata Exposure:
- Frame headers contain only essential routing information
- Capability hints are blinded before transmission
- Selection criteria hide detailed capability requirements
- Peer selection uses privacy-preserving scoring
Context Isolation:
- All messages scoped to RelationshipId or ContextId
- No cross-context message routing
- Connection state partitioned by context
- Context changes trigger re-keying
8.2 Privacy-Aware Envelope Usage
The Envelope type provides three privacy levels:
#![allow(unused)] fn main() { // Clear transmission (no privacy protection) let envelope = Envelope::new(payload); // Relationship-scoped transmission let envelope = Envelope::new_scoped( payload, relationship_id, None, // Optional capability hint ); // Fully blinded transmission let envelope = Envelope::new_blinded( payload, blinded_metadata, ); }
Pattern: Always use new_scoped() for relationship communication. Only use new() for public announcements. Use new_blinded() when metadata exposure must be minimized.
8.3 Privacy-Preserving Peer Selection
Peer selection must not reveal capability requirements or relationship structure:
#![allow(unused)] fn main() { let criteria = PrivacyAwareSelectionCriteria::for_relationship(relationship_id) .require_capability("threshold_signing") // Will be blinded .min_reliability(ReliabilityLevel::High) .prefer_privacy_features(true); let selection = criteria.select_peers(&available_peers); }
Pattern:
- Selection criteria blinded before network transmission
- Selection scores computed without revealing weights
- Rejected candidates not logged or exposed
- Selection reasons use generic categories, not specific capabilities
8.4 Common Privacy Pitfalls
✗ Avoid:
- Logging detailed capability requirements
- Exposing relationship membership in error messages
- Reusing connection state across contexts
- Sending capability names in clear text
- Correlating message sizes with content types
✓ Do:
- Use generic error messages ("authorization failed" not "missing capability: admin")
- Pad messages to fixed sizes when possible
- Rotate connection identifiers on epoch changes
- Blind capability hints before network transmission
- Use privacy-preserving selection scoring
8.5 Testing Privacy Properties
When testing transport code, verify:
- Context Isolation: Messages sent in context A cannot be received in context B
- Metadata Minimization: Only essential headers exposed, no capability details
- Selection Privacy: Peer selection does not leak candidate set or ranking criteria
- Re-keying: Context epoch changes trigger channel teardown and fresh establishment
- No Side Channels: Timing, message size, error messages do not leak sensitive information
See crates/aura-transport/src/types/tests.rs and crates/aura-transport/src/peers/tests.rs for examples.
8.6 Integration with Guard Chain
Transport privacy integrates with the guard chain:
- CapGuard: Evaluates blinded capability hints without exposing requirements
- FlowGuard: Charges budget before transmission (no side channels on failure)
- LeakageTracker: Accounts for metadata exposure in frame headers
- JournalCoupler: Records minimal send facts (no capability details)
The combination ensures that:
- Unauthorized sends produce no network traffic
- Over-budget sends fail silently (no timing side channel)
- Metadata leakage tracked against context budget
- Fact commits reveal only necessary information
9. Sync Status and Delivery Tracking
Category A (optimistic) operations require UI feedback for sync and delivery status. Anti-entropy provides the underlying sync mechanism, but users need visibility into progress. See Operation Categories for the full consistency metadata type definitions.
9.1 Propagation Status
Propagation status tracks journal fact sync via anti-entropy:
#![allow(unused)] fn main() { use aura_core::domain::Propagation; pub enum Propagation { /// Fact committed locally, not yet synced Local, /// Fact synced to some peers Syncing { peers_reached: u16, peers_known: u16 }, /// Fact synced to all known peers Complete, /// Sync failed, will retry Failed { retry_at: PhysicalTime, retry_count: u32, error: String }, } }
Anti-entropy provides callbacks via SyncProgressEvent to track progress:
Started- sync session beganDigestExchanged- digest comparison completePulling/Pushing- fact transfer in progressPeerCompleted- sync finished with one peerAllCompleted- all peers synced
9.2 Acknowledgment Protocol
For facts with ack_tracked = true, the transport layer implements the ack protocol:
Transmission Envelope:
#![allow(unused)] fn main() { pub struct FactEnvelope { pub fact: Fact, pub ack_requested: bool, // Set when fact.ack_tracked = true } }
FactAck Response:
#![allow(unused)] fn main() { pub struct FactAck { pub fact_id: String, pub peer: AuthorityId, pub acked_at: PhysicalTime, } }
Protocol Flow:
- Sender marks fact as
ack_trackedwhen committing - Transport includes
ack_requested: truein envelope - Receiver processes fact and sends
FactAckresponse - Sender records ack in journal's ack table
Acknowledgmentrecords are queryable for delivery status
9.3 Message Delivery Status
Message delivery derives from consistency metadata:
#![allow(unused)] fn main() { use aura_core::domain::{Propagation, Acknowledgment, OptimisticStatus}; // Delivery status is derived, not stored directly let is_sending = matches!(status.propagation, Propagation::Local); let is_sent = matches!(status.propagation, Propagation::Complete); let is_delivered = status.acknowledgment .map(|ack| expected_peers.iter().all(|p| ack.contains(p))) .unwrap_or(false); }
Status progression: Sending (◐) → Sent (✓) → Delivered (✓✓) → Read (✓✓ blue)
Read receipts are semantic (user viewed) and distinct from delivery (device received). They use ChatFact::MessageRead rather than the acknowledgment system.
9.4 UI Status Indicators
Status indicators provide user feedback:
Symbol Meaning Color Animation
─────────────────────────────────────────────────
✓ Confirmed/Sent Green None
✓✓ Delivered Green None
✓✓ Read (blue) Blue None
◐ Syncing/Sending Blue Pulsing
◌ Pending Gray None
⚠ Unconfirmed Yellow None
✗ Failed Red None
For messages:
- Single checkmark (✓) = sent to at least one peer
- Double checkmark (✓✓) = delivered to recipient
- Blue double checkmark = recipient read the message
9.4 Integration with Operation Categories
Sync status applies to Category A operations only:
- Send message → delivery status tracking
- Create channel → sync status to context members
- Update topic → sync status to channel members
- Block contact → local only (no sync needed for privacy)
Category B and C operations have different confirmation models:
- Category B uses proposal/approval state
- Category C uses ceremony completion status
Lifecycle modes (A1/A2/A3) apply within these categories: A1/A2 updates are usable immediately but must be treated as provisional until A3 consensus finalization. Soft-safe A2 should publish convergence certificates and reversion facts so UI and transport can surface any reversion risk during the soft window.
See Consensus - Operation Categories for categorization details.
10. Anti-Entropy Sync Protocol
Anti-entropy implements journal synchronization between peers. The protocol exchanges digests, plans reconciliation, and transfers operations.
10.1 Sync Phases
- Load Local State: Read local
Journal(facts + caps) and the local operation log. - Compute Digest: Compute
JournalDigestfor local state. - Digest Exchange: Send local digest to peer and receive peer digest.
- Reconciliation Planning: Compare digests and choose action (equal, LocalBehind, RemoteBehind, or Diverged).
- Operation Transfer: Pull or push operations in batches.
- Merge + Persist: Convert applied ops to journal delta, merge with local journal, persist once per round.
10.2 Digest Format
#![allow(unused)] fn main() { pub struct JournalDigest { pub operation_count: u64, pub last_epoch: Epoch, pub operation_hash: Hash32, pub fact_hash: Hash32, pub caps_hash: Hash32, } }
The operation_count is the number of operations in the local op log. The last_epoch is the max parent_epoch observed. The operation_hash is computed by streaming op fingerprints in deterministic order. The fact_hash and caps_hash use canonical serialization (DAG-CBOR) then hash.
10.3 Reconciliation Actions
| Digest Comparison | Action |
|---|---|
| Equal | No-op |
| LocalBehind | Request missing ops |
| RemoteBehind | Push ops |
| Diverged | Push + pull |
10.4 Retry Behavior
Anti-entropy can be retried according to AntiEntropyConfig.retry_policy. The default policy is exponential backoff with a bounded max attempt count.
10.5 Failure Semantics
Failures are reported with structured phase context:
SyncPhase::LoadLocalStateSyncPhase::ComputeDigestSyncPhase::DigestExchangeSyncPhase::PlanRequestSyncPhase::ReceiveOperationsSyncPhase::MergeJournalSyncPhase::PersistJournal
This makes failures attributable to a specific phase and peer.
11. Protocol Version Negotiation
All choreographic protocols participate in version negotiation during connection establishment.
11.1 Version Handshake Flow
Initiator Responder
| |
|-- VersionHandshakeRequest -->
| (version, min_version, capabilities, nonce)
| |
|<-- VersionHandshakeResponse -|
| (Accepted/Rejected)
| |
[Use negotiated version or disconnect]
The handler is located at aura-protocol/src/handlers/version_handshake.rs.
11.2 Handshake Outcomes
| Outcome | Response Contents |
|---|---|
| Compatible | negotiated_version (min of both peers), shared capabilities |
| Incompatible | reason, peer version, optional upgrade_url |
11.3 Protocol Capabilities
| Capability | Min Version | Description |
|---|---|---|
ceremony_supersession | 1.0.0 | Ceremony replacement tracking |
version_handshake | 1.0.0 | Protocol version negotiation |
fact_journal | 1.0.0 | Fact-based journal sync |
12. Summary
The transport, guard chain, and information flow architecture enforces strict control over message transmission. Secure channels bind communication to contexts. Guard chains enforce authorization, budget, and journal updates. Flow budgets and receipts regulate data usage. Leakage budgets reduce metadata exposure. Privacy-by-design patterns ensure minimal metadata exposure and context isolation. All operations remain private to the context and reveal no structural information.
Sync status and delivery tracking provide user visibility into Category A operation propagation. Anti-entropy provides the underlying sync mechanism with digest-based reconciliation. Version negotiation ensures protocol compatibility across peers. Delivery receipts enable message read status for enhanced UX.
Aura Messaging Protocol (AMP)
AMP is a secure asynchronous messaging protocol for Aura. It operates within relational contexts and channels. The protocol provides strong post-compromise security and bounded forward secrecy without head-of-line blocking.
1. Scope and Goals
AMP assumes shared state in joint semilattice journals is canonical. Secrets derive locally from shared state combined with authority keys. Ratchet operations remain deterministic and recoverable.
The protocol targets four properties. No head-of-line blocking occurs during message delivery. Strong post-compromise security restores confidentiality after key exposure. Bounded forward secrecy limits exposure within skip windows. Deterministic recovery enables key rederivation from journal state.
1.2 Optimistic Operations Within Contexts
Channels are encrypted substreams within an existing relational context. The relational context (established via invitation ceremony) already provides the shared secret foundation. This means:
- No new key agreement - the relational context already provides the shared secret
- Channel creation is local - just emit a
ChannelCheckpointfact into the existing context journal - Members must already share the context - you can't invite someone to a channel unless you already have a relational context with them
Channel creation and messaging are Category A (optimistic) operations because:
- Both parties already derived the context root via the invitation ceremony
- Channel facts sync via normal journal anti-entropy
- Key derivation is deterministic:
KDF(ContextRoot, ChannelId, epoch) → ChannelBaseKey
The encryption "just works" because the cryptographic foundation was established when the relationship was created. See Consensus - Operation Categories for the full classification.
1.2.1 Bootstrap Messaging (Dealer Key, Provisional)
For newly-created channels where the group key ceremony has not completed yet, AMP supports a bootstrap epoch using a trusted-dealer key (K2/A1). This enables participants to exchange encrypted messages immediately without waiting for consensus-finalized group keys.
Bootstrap flow:
- Dealer generates a bootstrap key (
K_boot) during channel creation. - Dealer emits a distinct protocol fact that records the bootstrap metadata (but not the key):
AmpChannelBootstrap { context, channel, bootstrap_id, dealer, recipients, created_at, expires_at }
- Dealer distributes
K_bootper invitee (via the channel invite payload or other secure delivery). - Participants store
K_bootlocally in secure storage keyed by(context, channel, bootstrap_id). - Messages in epoch 0 use
K_bootfor encryption/decryption.
Security properties:
- Messages are encrypted (not plaintext), but confidentiality is provisional and dealer-trusted.
K_bootnever appears in the journal; only thebootstrap_idhash is recorded.- Late joiners do not receive
K_bootand therefore cannot read bootstrap-epoch messages.
Transition to finalized keys:
- After the group key ceremony completes, the channel bumps to the next epoch.
- Messages in epoch ≥ 1 use the canonical group key derivation.
- Bootstrap messages remain decryptable only to members who stored
K_boot.
1.3 AMP Lifecycle (A1/A2/A3)
AMP channel epoch bumps follow the lifecycle taxonomy:
- A1 (Provisional):
ProposedChannelEpochBumpis emitted immediately. - A2 (Coordinator Soft‑Safe):
ConvergenceCertmarks bounded convergence. - A3 (Consensus‑Finalized):
CommittedChannelEpochBump+CommitFactfinalize the bump.
Fast‑path outputs are usable immediately but are provisional. Durable state is only established once A3 evidence is merged. See Operation Categories for the canonical per-ceremony policy matrix and Consensus for consensus evidence binding.
1.1 Design Requirements
AMP addresses constraints that existing messaging protocols cannot satisfy.
Deterministic recovery requires all ratchet state to be derivable from replicated facts. The journal is the only durable state. Secrets derive from reduced journal state combined with authority keys. No device may maintain ratchet state that cannot be recovered by other devices in the authority.
Multi-device authorities require support for concurrent message sends from different devices within the same authority. External parties cannot observe which device sent a message. All devices must converge to the same ratchet position after merging facts.
Selective consistency requires both eventual consistency for message sends and strong agreement for epoch transitions. The protocol cannot assume a central ordering service.
Authorization integration requires every message send to pass through a guard chain before reaching the network. Authorization failure must prevent key derivation. Budget charges must be atomic with the send.
2. Channel Lifecycle Surface
The AmpChannelEffects trait defines the canonical API for AMP channel lifecycle and messaging. This trait lives in aura-core::effects::amp.
#![allow(unused)] fn main() { #[async_trait] pub trait AmpChannelEffects: Send + Sync { async fn create_channel(&self, params: ChannelCreateParams) -> Result<ChannelId, AmpChannelError>; async fn close_channel(&self, params: ChannelCloseParams) -> Result<(), AmpChannelError>; async fn join_channel(&self, params: ChannelJoinParams) -> Result<(), AmpChannelError>; async fn leave_channel(&self, params: ChannelLeaveParams) -> Result<(), AmpChannelError>; async fn send_message(&self, params: ChannelSendParams) -> Result<AmpCiphertext, AmpChannelError>; } }
The create_channel method writes an AMP checkpoint and policy for a context-scoped channel. The close_channel method records a terminal epoch bump and policy closure. The join_channel and leave_channel methods record membership facts. The send_message method derives current channel state and returns AmpCiphertext containing the header and encrypted payload.
2.1 Implementations
The runtime implementation uses AmpChannelCoordinator in aura-protocol::amp::channel_lifecycle. The simulator implementation uses SimAmpChannels in aura-simulator::amp. The testkit provides MockEffects implementing AmpChannelEffects for deterministic unit tests.
2.2 TUI Wiring
The TUI routes EffectCommand variants through an optional AmpChannelEffects handle. Commands include CreateChannel, CloseChannel, and SendMessage. Events including ChannelCreated, ChannelClosed, and MessageReceived drive view updates.
Demo mode injects the simulator handler into SimulatedBridge. Alice and Carol agents use the same lifecycle as real clients while Bob remains human-controlled.
2.3 Transport Integration (Rendezvous/Noise)
The secure delivery of bootstrap keys and invitations relies on the Rendezvous system (Layer 5). Peers establish transient secure channels using the Noise Protocol (IKpsk2), authenticated by their identity keys (converted from Ed25519 to X25519) and a Pre-Shared Key (PSK) derived from the relational context.
This Noise channel enables the synchronous coordination required to bootstrap the asynchronous AMP ratchet, specifically:
- Bootstrap Key Distribution: Delivering K_boot to new invitees.
- Invitation Exchange: Swapping capabilities and descriptors.
The Noise handshake provides forward secrecy and authentication for this critical setup phase. Once the Noise channel is established and the bootstrap key is exchanged, participants can immediately begin sending encrypted messages using the provisional epoch 0 AMP mechanics. The relational context facts (including the rendezvous receipt and descriptor binding) synchronize via normal anti-entropy, ensuring all parties converge on the same identity credentials and shared secrets.
3. Terminology
3.1 Aura Terms
An authority is an account authority with a commitment tree and FROST keys. A relational context is shared state between authorities identified by ContextId. A journal is a CRDT OR-set of facts with monotone growth and deterministic reduction.
3.2 AMP Terms
A channel is a messaging substream scoped to a relational context. The channel epoch bounds post-compromise security and serves as the KDF base. The ratchet generation is a monotone position derived from reduced journal state. The skip window defines out-of-order tolerance with a default of 1024 generations.
A checkpoint is a journal fact anchoring ratchet windows. The alternating ratchet maintains two overlapping windows at boundaries. A pending bump is a proposed epoch transition awaiting consensus.
4. Commitment Tree Integration
The commitment tree defines authority structure and provides the foundation for key derivation. Each authority maintains an internal tree with branch nodes and leaf nodes. Leaf nodes represent devices holding threshold signing shares. Branch nodes represent subpolicies expressed as m-of-n thresholds.
#![allow(unused)] fn main() { pub struct TreeState { pub epoch: Epoch, pub root_commitment: TreeHash32, pub branches: BTreeMap<NodeIndex, BranchNode>, pub leaves: BTreeMap<LeafId, LeafNode>, } }
The TreeState represents the result of reducing all operations in the OpLog. The epoch field scopes all derived keys. The root_commitment field is a Merkle hash over the ordered tree structure. External parties see only the epoch and root commitment.
Tree operations modify device membership and policies. The AddLeaf operation inserts a new device. The RemoveLeaf operation removes an existing device. The ChangePolicy operation updates threshold requirements. The RotateEpoch operation increments the epoch and invalidates all derived keys.
Each operation appears in the journal as an attested operation signed by the required threshold of devices. The tree supports concurrent updates through deterministic conflict resolution.
graph TD
A[Multiple Concurrent Ops] --> B[Group by Parent Commitment]
B --> C[Select Winner by Hash Order]
C --> D[Apply in Epoch Order]
D --> E[Single TreeState]
The channel base key for a given ChannelId and epoch derives as KDF(TreeRoot, ChannelId, epoch). All devices in the authority compute the same base key.
5. Fact Structures
AMP uses facts inserted into the relational context journal. All facts are monotone. Reduction determines canonical state.
5.1 Channel Checkpoint
#![allow(unused)] fn main() { pub struct ChannelCheckpoint { pub context: ContextId, pub channel: ChannelId, pub chan_epoch: u64, pub base_gen: u64, pub window: u32, pub ck_commitment: Hash32, pub skip_window_override: Option<u32>, } }
Reduction chooses one canonical checkpoint per context, channel, and epoch tuple. The valid ratchet generation set is the union of two windows. Window A spans from base_gen to base_gen + window. Window B spans from base_gen + window + 1 to base_gen + 2 * window. Checkpoints enable deterministic recovery and serve as garbage collection anchors.
5.2 Proposed and Committed Epoch Bumps
#![allow(unused)] fn main() { pub struct ProposedChannelEpochBump { pub context: ContextId, pub channel: ChannelId, pub parent_epoch: u64, pub new_epoch: u64, pub bump_id: Hash32, pub reason: ChannelBumpReason, } pub struct CommittedChannelEpochBump { pub context: ContextId, pub channel: ChannelId, pub parent_epoch: u64, pub new_epoch: u64, pub chosen_bump_id: Hash32, pub consensus_id: Hash32, } }
Proposed bumps represent optimistic ratchet phase transitions. Committed bumps are finalized by Aura Consensus with 2f+1 witnesses. The new epoch always equals parent epoch plus one. Only one committed bump exists per parent epoch.
5.3 Channel Bump Reason
#![allow(unused)] fn main() { pub enum ChannelBumpReason { Routine, SuspiciousActivity, ConfirmedCompromise, } }
The Routine variant indicates cadence-based maintenance. The SuspiciousActivity variant indicates detected anomalies such as AEAD failures or ratchet conflicts. The ConfirmedCompromise variant requires immediate post-compromise security restoration. Both suspicious activity and confirmed compromise bypass routine spacing rules.
6. Derived Channel State
Reduction yields a ChannelEpochState struct containing the canonical epoch and derived state.
#![allow(unused)] fn main() { pub struct ChannelEpochState { pub chan_epoch: u64, pub pending_bump: Option<PendingBump>, pub last_checkpoint_gen: u64, pub current_gen: u64, pub skip_window: u32, } pub struct PendingBump { pub parent_epoch: u64, pub new_epoch: u64, pub bump_id: Hash32, pub reason: ChannelBumpReason, } }
The chan_epoch field holds the current canonical epoch. The pending_bump field holds an in-progress transition if one exists. The last_checkpoint_gen and current_gen fields track ratchet positions. The skip_window field determines out-of-order tolerance.
6.1 Skip Window Computation
The skip window derives from three sources in priority order. The checkpoint skip_window_override takes precedence. The channel policy fact applies if no override exists. The default of 1024 applies otherwise.
6.2 Critical Invariants
Before sending, a device must merge latest facts, reduce channel state, and use updated epoch and generation values. No device may send under stale epochs. Only one bump from epoch e to e+1 may be pending at a time. The new epoch always equals parent epoch plus one.
6.3 Ratchet Generation Semantics
The ratchet_gen value is not a local counter. It derives from reduced journal state. All devices converge to the same ratchet position after merging facts. Generation advances only when send or receive events occur consistent with checkpoint and dual-window rules. This guarantees deterministic recovery and prevents drift across devices.
7. Three-Level Key Architecture
AMP separates key evolution across three levels. Authority epochs provide identity-level post-compromise security. Channel epochs provide channel-level post-compromise security. Ratchet generations provide bounded forward secrecy within each epoch.
7.1 Authority Epochs
Authority epochs rotate the threshold key shares via DKG ceremony. Rotation invalidates all derived context keys. This is the strongest form of post-compromise security.
Authority epoch rotation occurs approximately daily or upon confirmed compromise. The rotation is independent of messaging activity. All relational contexts must re-derive their shared secrets after rotation.
7.2 Channel Epochs
Channel epochs provide post-compromise security at the channel level. Each epoch uses an independent base key derived from the tree root. Epoch rotation invalidates all keys from the previous epoch. The rotation is atomic. All devices observe the same epoch after merging journal facts.
The epoch bump lifecycle has two states. The stable state indicates that epoch e is finalized. The pending state indicates that an optimistic bump has been proposed. No new bump can be proposed while a bump is pending. This single-pending-bump invariant ensures linear epoch progression.
7.3 Ratchet Generations
Ratchet generations provide forward secrecy within each epoch. Message keys derive from the channel base key, generation, and direction.
#![allow(unused)] fn main() { pub struct AmpHeader { pub context: ContextId, pub channel: ChannelId, pub chan_epoch: u64, pub ratchet_gen: u64, } }
The AmpHeader contains the routing and ratchet information. The chan_epoch field identifies which base key to use. The ratchet_gen field identifies which generation key to derive. This header becomes the AEAD additional data.
Ratchet generation is not a local counter. It derives from reduced journal state. Devices compute the current generation by examining send and receive events in the reduced context state. All devices converge to the same ratchet position after merging facts.
8. Channel Epoch Bump Lifecycle
8.1 States
A channel has two states. The stable state indicates the epoch is finalized. The pending state indicates an optimistic bump awaits commitment.
8.2 Spacing Rule for Routine Bumps
AMP enforces a spacing rule for routine bumps. Let base_gen be the anchor generation from the canonical checkpoint. Let current_gen be the current ratchet generation. Let W be the skip window.
A routine bump from epoch e to e+1 requires current_gen - base_gen >= W / 2. With default W of 1024, this threshold equals 512 generations. This spacing ensures structural transitions do not occur too frequently.
Emergency bumps bypass this rule. Examples include multiple AEAD failures, conflicting ratchet commitments, unexpected epoch values, and explicit compromise signals.
8.3 State Transitions
In the stable state, if the spacing rule is satisfied and no bump is pending, the device inserts a ProposedChannelEpochBump and starts consensus. The device enters the pending state and uses the new epoch with alternating windows.
In the pending state, upon observing a committed bump, the device sets the new epoch and clears the pending bump. It may emit a new checkpoint and returns to the stable state.
8.4 Complete Lifecycle Sequence
The following diagram shows the complete double ratchet and epoch bump lifecycle between two devices.
sequenceDiagram
participant A as Device A
participant J as Journal
participant C as Consensus
participant B as Device B
Note over A,B: Channel Creation (Epoch 0)
A->>J: ChannelCheckpoint(epoch=0, base_gen=0, W=1024)
J-->>B: Sync checkpoint fact
B->>B: Reduce state → Epoch 0, gen=0
Note over A,B: Normal Messaging (Ratchet Advancement)
loop Messages within Window A [0..1024]
A->>A: Merge facts, reduce state
A->>A: Derive key(epoch=0, gen=N)
A->>B: AmpHeader(epoch=0, gen=N) + ciphertext
B->>B: Validate gen in [0..2048]
B->>B: Derive key, decrypt
B->>B: Advance receive ratchet
end
Note over A,B: Spacing Rule Satisfied (gen >= 512)
A->>A: Check: current_gen - base_gen >= W/2
A->>J: ProposedChannelEpochBump(0→1, reason=Routine)
A->>C: Start consensus for bump
Note over A,B: Pending State (Dual Windows Active)
rect rgb(240, 240, 255)
Note over A,B: Both epochs valid during transition
A->>B: Messages use epoch=1 (optimistic)
B->>B: Accept epoch=0 OR epoch=1
end
C->>C: Collect 2f+1 witness signatures
C->>J: CommittedChannelEpochBump(0→1)
J-->>A: Sync committed bump
J-->>B: Sync committed bump
Note over A,B: Epoch 1 Finalized
A->>A: Clear pending_bump, set epoch=1
B->>B: Clear pending_bump, set epoch=1
A->>J: ChannelCheckpoint(epoch=1, base_gen=1024)
Note over A,B: Continue in New Epoch
loop Messages in Epoch 1
A->>A: Derive key(epoch=1, gen=M)
A->>B: AmpHeader(epoch=1, gen=M) + ciphertext
end
Note over A,B: Emergency Bump (Compromise Detected)
B->>B: Detect AEAD failures > threshold
B->>J: ProposedChannelEpochBump(1→2, reason=SuspiciousActivity)
Note right of B: Bypasses spacing rule
B->>C: Immediate consensus request
C->>J: CommittedChannelEpochBump(1→2)
Note over A,B: Post-Compromise Security Restored
A->>A: Derive new base key for epoch=2
B->>B: Derive new base key for epoch=2
This diagram illustrates four phases of the protocol. Channel creation establishes the initial checkpoint at epoch 0. Normal messaging advances the ratchet generation within dual windows. Routine bumps occur when spacing rules are satisfied and follow the consensus path. Emergency bumps bypass spacing for immediate post-compromise security.
9. Ratchet Windows
AMP uses an always-dual window model. Every checkpoint defines two consecutive skip windows providing a continuous valid range of 2W generations.
9.1 Window Layout
Given base generation G and skip window W, two windows are defined. Window A spans G to G+W. Window B spans G+W+1 to G+2W. The valid generation set is the union of both windows.
#![allow(unused)] fn main() { // Window layout for checkpoint at generation G with window W let window_a = G..(G + W); let window_b = (G + W + 1)..(G + 2*W); let valid_gen_set = window_a.union(window_b); }
flowchart LR
subgraph "Valid Generation Range"
A["Window A<br/>G to G+W"]
B["Window B<br/>G+W+1 to G+2W"]
end
A --> B
This design eliminates boundary issues. No mode switches are required. The implementation remains simple with robust asynchronous tolerance.
9.2 Window Shifting
When a new checkpoint is issued, the new base generation is chosen far enough ahead per the spacing rule. The dual-window layout guarantees overlap with prior windows. Garbage collection can safely prune older checkpoints when they no longer affect valid generation ranges.
9.3 Asynchronous Delivery
The dual window design solves the asynchronous delivery problem. Messages may arrive out of order by up to W generations. During epoch transitions, messages may be sent under either the old or new epoch. The overlapping windows ensure that all valid messages can be decrypted.
Message derivation uses a KDF chain similar to Signal's construction but with key differences. AMP derives all ratchet state deterministically from replicated journal facts rather than device-local databases. This enables complete recovery across multiple devices without coordination.
10. Sending Messages
10.1 Message Header
#![allow(unused)] fn main() { pub struct AmpHeader { pub context: ContextId, pub channel: ChannelId, pub chan_epoch: u64, pub ratchet_gen: u64, } }
The header contains the context identifier, channel identifier, current epoch, and current generation. These fields form the additional authenticated data for AEAD encryption.
10.2 Reduce-Before-Send Rule
Before sending a message, a device must merge the latest journal facts. It must reduce the channel state to get the current epoch and generation. It must verify that the generation is within the valid window. Only then can it derive the message key and encrypt.
10.3 Send Procedure
Before sending, a device merges new facts and reduces channel state. It asserts that the current generation is within the valid generation set. It asserts the channel epoch is current.
If the spacing rule is satisfied and no bump is pending, the device proposes a new bump. The device derives the message key using a KDF with the channel base key, generation, and direction.
The device creates an AmpHeader, encrypts the payload with AEAD using the message key and header as additional data, and performs the guard chain. Guard evaluation runs over a prepared GuardSnapshot and emits EffectCommand items for async interpretation. The device then advances the local ratchet generation.
11. Receiving Messages
When receiving, a device merges new facts and reduces channel state. It checks that the message epoch is either the current or pending epoch. It checks that the generation is within the valid generation set.
The device rederives the message key using the same KDF parameters. It decrypts the payload using AEAD with the message key and header. It advances the local receive ratchet.
Messages outside valid windows or with unsupported epochs are rejected.
12. Guard Chain Integration
AMP integrates with Aura's guard chain for authorization and flow control. Every message send passes through CapabilityGuard, FlowBudgetGuard, JournalCouplingGuard, and LeakageTrackingGuard before reaching the transport.
Guard evaluation runs over a prepared GuardSnapshot and produces EffectCommand items. An async interpreter executes those commands only after the entire guard chain succeeds. Unauthorized or over-budget sends never touch the network.
#![allow(unused)] fn main() { let outcome = guard_chain.evaluate(&snapshot, &request); if !outcome.is_authorized() { return Err(AuraError::authorization_failed("Guard denied")); } for cmd in outcome.effects { interpreter.execute(cmd).await?; } }
The snapshot captures the current capability frontier and flow budget state. Each guard emits EffectCommand items instead of performing I/O. The interpreter executes the resulting commands in production or simulation. Any failure returns locally without observable effects.
12.1 Flow Budgets
Flow budgets are replicated as spent counters in the journal. The spent counter for a context and peer pair is a monotone fact. The limit computes at runtime from Biscuit capabilities and sovereign policy. Before sending, the flow budget guard checks that spent + cost <= limit.
12.2 Receipts
Receipts provide accountability for multi-hop forwarding. Each relay hop produces a receipt containing the context, source, destination, epoch, cost, and signature. The receipt proves that the relay charged its budget before forwarding.
13. Recovery and Garbage Collection
13.1 Recovery
To recover, a device loads the relational context journal. It reduces to the latest committed epoch, checkpoint, and skip window. It rederives the channel base key from context root key, channel identifier, and epoch. It rederives ratchet state from the checkpoint and window generations.
graph TD
A[Load Journal Facts] --> B[Reduce to TreeState]
B --> C[Reduce to ChannelEpoch]
C --> D[Compute Checkpoint]
D --> E[Derive Base Key]
E --> F[Ready to Message]
The recovery process requires no coordination. The device does not need to contact other participants. It does not need to request missing state. It only needs access to the journal facts. Once reduction completes, the device has the same view as all other participants.
13.2 Garbage Collection
AMP garbage collection maintains protocol safety while reclaiming storage. GC operates on checkpoints, proposed bumps, and committed bumps.
A checkpoint at generation G with window W can be pruned when a newer checkpoint exists at G' where G' exceeds G+2W. The newer checkpoint coverage must not overlap with the old checkpoint coverage. All messages within the old window must be processed or beyond the recovery horizon.
A proposed bump can be pruned when a committed bump supersedes it or when the epoch becomes stale. Committed bumps are retained as consensus evidence until full snapshot compaction.
13.3 Pruning Boundary
The safe pruning boundary for checkpoints is computed as the maximum checkpoint generation minus 2W minus a safety margin. The recommended safety margin is W/2 or 512 generations with default settings.
GC triggers when journal size exceeds threshold, checkpoint count exceeds maximum, manual compaction is requested, or snapshot creation initiates. See Maintenance for integration with the Aura snapshot system.
14. Security Properties
14.1 Forward Secrecy
Forward secrecy is bounded by the skip window size of 1024 within each epoch. Within a single epoch, an attacker who compromises a device learns at most W future message keys. Older epochs use independent base keys. Compromise of epoch e reveals nothing about epoch e-1.
14.2 Post-Compromise Security
Post-compromise security operates at two levels. Channel epoch bumps heal compromise of channel secret state. Context epoch bumps provide stronger PCS by healing compromise of context state. Channel bumps are frequent and cheap. Context bumps are rare and expensive.
The dual window mechanism provides continuous post-compromise healing. When an emergency bump occurs, the new epoch uses a fresh base key. Messages sent under the new epoch cannot be decrypted with the old base key. The overlapping windows ensure that honest messages in flight are not lost.
14.3 No Head-of-Line Blocking
The protocol accepts out-of-order messages up to W generations. Dual windows absorb boundary drift.
14.4 No Ratchet Forks
Consensus finalizes each epoch step. The epoch chain is always linear and monotone.
14.5 Deterministic Recovery
Checkpoints anchor the ratchet. Facts represent all shared state. Keys are always rederived deterministically. This property enables true multi-device messaging. All devices in an authority see the same messages. All devices can send and receive without coordinating who is currently active. The authority appears as a single messaging entity to external parties.
15. Governance
The skip window size W is governable via policy facts. Bump cadence and post-compromise security triggers can be governed by context policies.
16. Failure Modes
AMP drops messages only when they fall outside the cryptographically safe envelope. These failures are intentional safeguards.
16.1 Generation Out of Window
A message is dropped if its generation lies outside the valid generation set. Causes include stale sender state, aged messages post-GC, or invalid generations from attackers.
16.2 Epoch Mismatch
Messages are rejected when the header epoch is inconsistent with reduced epoch state. This includes epochs that are too old, non-linear epochs, or replays from retired epochs.
16.3 AEAD Failure
If AEAD decryption fails, messages are dropped. Repeated failures contribute to suspicious-event classification and may trigger emergency PCS.
16.4 Beyond Recovery Horizon
Messages older than the current checkpoint window cannot be decrypted because older checkpoints have been garbage collected. This is an intentional forward secrecy tradeoff.
16.5 Policy-Enforced Invalidity
After context epoch bumps or high-severity PCS events, messages from previous context epochs are intentionally dropped.
17. Message Delivery Tracking
AMP integrates with Aura's unified consistency metadata system for delivery tracking. For UI presentation of delivery status indicators, see CLI and TUI. For consistency type definitions, see Operation Categories.
17.1 Acknowledgment Tracking
Message delivery uses the generic ack_tracked system:
#![allow(unused)] fn main() { let options = FactOptions { request_acks: true }; runtime_bridge.commit_relational_facts_with_options(&[message_fact], options).await?; }
The transport layer includes ack_requested: true in the transmission envelope. Peers send FactAck responses upon successful processing. Acks are stored in the journal's ack table.
17.2 Read Receipts
Read receipts are distinct from delivery acknowledgments. Delivery indicates the message was received and decrypted. Read indicates the user actually viewed the message.
Read receipts are opt-in per contact and emit a MessageRead chat fact:
#![allow(unused)] fn main() { ChatFact::MessageRead { context_id, channel_id, message_id, reader_id, read_at, } }
17.3 Delivery Policy and GC
Ack tracking is garbage-collected based on DeliveryPolicy:
#![allow(unused)] fn main() { use aura_app::policies::DropWhenFullyAcked; app.register_delivery_policy::<MessageSentSealed>(DropWhenFullyAcked); }
When the policy's should_drop_tracking() returns true, ack records are pruned.
17.4 Privacy Considerations
Delivery acknowledgments leak timing metadata. Applications should batch acknowledgments to reduce timing precision, default read receipts to disabled (opt-in), and allow disabling acks entirely for high-sensitivity contexts with FactOptions { request_acks: false }.
See Also
- Transport and Information Flow for guard chain enforcement
- Consensus for epoch bump finalization
- Journal System for fact semantics and reduction
- Maintenance for snapshot and GC integration
- Authority and Identity for tree operation details
- CLI and TUI for UI delivery status display
Rendezvous Architecture
This document describes the rendezvous architecture in Aura. It explains peer discovery, descriptor propagation, transport selection, channel establishment, and relay-to-direct holepunch upgrades. It aligns with the authority and context model. It scopes all rendezvous behavior to relational contexts.
1. Overview
Rendezvous establishes secure channels between authorities. The RendezvousService exposes prepare_publish_descriptor() and prepare_establish_channel() methods. The service returns guard outcomes that the caller executes through an effect interpreter. Rendezvous operates inside a relational context and uses the context key for encryption. Descriptors appear as facts in the context journal. Propagation uses journal synchronization (aura-sync), not custom flooding.
Rendezvous does not establish global identity. All operations are scoped to a ContextId. A context defines which authorities may see descriptors. Only participating authorities have the keys required to decrypt descriptor payloads.
Rendezvous descriptor exchange can run in provisional (A1) or soft-safe (A2) modes to enable rapid connectivity under poor network conditions, but durable channel epochs and membership changes must be finalized via consensus (A3). Soft-safe flows should emit convergence and reversion facts so participants can reason about reversion risk while channels are warming.
1.1 Secure‑Channel Lifecycle (A1/A2/A3)
- A1: Provisional
Descriptor/ handshake facts allow immediate connectivity. - A2: Coordinator soft‑safe convergence certs indicate bounded divergence.
- A3:
ChannelEstablishedis finalized by consensus and accompanied by aCommitFact.
Rendezvous must treat A1/A2 outputs as provisional until A3 evidence is merged.
See docs/106_consensus.md for commit evidence binding.
2. Architecture
The rendezvous crate follows Aura's fact-based architecture:
- Guard Chain First: All network sends flow through guard evaluation before execution
- Facts Not Flooding: Descriptors are journal facts propagated via
aura-sync, not custom flooding - Standard Receipts: Uses the system
Receipttype with epoch binding and cost tracking - Session-Typed Protocol: Protocols expressed as MPST choreographies with guard annotations
- Unified Transport: Channels established via
SecureChannelwith Noise IKpsk2
2.1 Module Structure
aura-rendezvous/
├── src/
│ ├── lib.rs # Public exports
│ ├── facts.rs # RendezvousFact domain fact type
│ ├── protocol.rs # MPST choreography definition
│ ├── service.rs # RendezvousService (main coordinator)
│ ├── descriptor.rs # Transport selector and builder
│ └── new_channel.rs # SecureChannel, ChannelManager, Handshaker
3. Transport Strategy
The transport layer uses a priority sequence of connection attempts. Direct QUIC is attempted first. QUIC using reflexive addresses via STUN is attempted next. TCP direct is attempted next. WebSocket relay is attempted last. The first successful connection is used.
Aura uses relay-first fallback. Relays use guardians or peers designated to provide relay services. Relay traffic uses end-to-end encryption. Relay capabilities must be valid for the context.
STUN discovery identifies the external address of each participant. Devices query STUN servers periodically. The reflexive address appears in rendezvous descriptors as a transport hint. STUN failure does not prevent rendezvous because relay is always available.
3.1 Holepunching and Upgrade Policy
Aura uses a relay-first, direct-upgrade model for NAT traversal:
- Start on relay as soon as both peers have a valid descriptor path.
- Exchange direct/reflexive candidates from descriptor facts.
- Launch bounded direct upgrade attempts (holepunch) in the background.
- Promote to direct when a recoverable direct path succeeds; otherwise remain on relay.
Retry state is tracked with typed generations (CandidateGeneration, NetworkGeneration) and bounded backoff (AttemptBudget, BackoffWindow) in PeerConnectionActor. Generation changes reset retry budgets, which avoids stale retry loops after interface/NAT changes.
Recoverability is evaluated from local binding/interface provenance, not from reflexive addresses alone. This prevents treating stale external mappings as viable direct paths.
Operationally:
- Relay path is the safety baseline.
- Direct holepunch is an optimization path.
- Network changes can trigger a fresh upgrade cycle without dropping relay connectivity.
4. Data Structures
4.1 Domain Facts
Rendezvous uses domain facts in the relational context journal. Facts are propagated via journal synchronization.
#![allow(unused)] fn main() { /// Rendezvous domain facts stored in context journals pub enum RendezvousFact { /// Transport descriptor advertisement Descriptor(RendezvousDescriptor), /// Channel established acknowledgment ChannelEstablished { initiator: AuthorityId, responder: AuthorityId, channel_id: [u8; 32], epoch: u64, }, /// Descriptor revocation DescriptorRevoked { authority_id: AuthorityId, nonce: [u8; 32], }, } }
4.2 Transport Descriptors
#![allow(unused)] fn main() { /// Transport descriptor for peer discovery pub struct RendezvousDescriptor { /// Authority publishing this descriptor pub authority_id: AuthorityId, /// Context this descriptor is for pub context_id: ContextId, /// Available transport endpoints pub transport_hints: Vec<TransportHint>, /// Handshake PSK commitment (hash of PSK derived from context) pub handshake_psk_commitment: [u8; 32], /// Public key for Noise IK handshake (Ed25519 public key) pub public_key: [u8; 32], /// Validity window start (ms since epoch) pub valid_from: u64, /// Validity window end (ms since epoch) pub valid_until: u64, /// Nonce for uniqueness pub nonce: [u8; 32], /// What the peer wants to be called (optional, for UI purposes) pub nickname_suggestion: Option<String>, } }
4.3 Transport Hints
#![allow(unused)] fn main() { /// Transport endpoint hint pub enum TransportHint { /// Direct QUIC connection QuicDirect { addr: TransportAddress }, /// QUIC via STUN-discovered reflexive address QuicReflexive { addr: TransportAddress, stun_server: TransportAddress }, /// WebSocket relay through a relay authority WebSocketRelay { relay_authority: AuthorityId }, /// TCP direct connection TcpDirect { addr: TransportAddress }, } }
5. MPST Choreographies
Rendezvous protocols are defined as MPST choreographies with guard annotations.
5.1 Direct Exchange Protocol
#![allow(unused)] fn main() { choreography! { #[namespace = "rendezvous_exchange"] protocol RendezvousExchange { roles: Initiator, Responder; // Initiator publishes descriptor (fact insertion, propagates via sync) Initiator[guard_capability = "rendezvous:publish", journal_facts = "descriptor_offered"] -> Responder: DescriptorOffer(RendezvousDescriptor); // Responder publishes response descriptor Responder[guard_capability = "rendezvous:publish", journal_facts = "descriptor_answered"] -> Initiator: DescriptorAnswer(RendezvousDescriptor); // Direct channel establishment Initiator[guard_capability = "rendezvous:connect", -> Responder: HandshakeInit(NoiseHandshake); Responder[guard_capability = "rendezvous:connect", journal_facts = "channel_established"] -> Initiator: HandshakeComplete(NoiseHandshake); } } }
5.2 Relayed Protocol
#![allow(unused)] fn main() { choreography! { #[namespace = "relayed_rendezvous"] protocol RelayedRendezvous { roles: Initiator, Relay, Responder; Initiator[guard_capability = "rendezvous:relay",] -> Relay: RelayRequest(RelayEnvelope); Relay[guard_capability = "relay:forward", leak = "neighbor:1"] -> Responder: RelayForward(RelayEnvelope); Responder[guard_capability = "rendezvous:relay"] -> Relay: RelayResponse(RelayEnvelope); Relay[guard_capability = "relay:forward", leak = "neighbor:1"] -> Initiator: RelayComplete(RelayEnvelope); } } }
6. Descriptor Propagation
Descriptors propagate via journal synchronization. This replaces custom flooding.
- Authority creates a
RendezvousFact::Descriptorfact - Guard chain evaluates the publication request
- On success, fact is inserted into the context journal
- Journal sync (
aura-sync) propagates facts to context participants - Peers query journal for peer descriptors
This model provides:
- Deduplication: Journal sync handles duplicate facts naturally
- Ordering: Facts have causal ordering via journal timestamps
- Authorization: Guard chain validates before insertion
- Consistency: Same propagation mechanism as other domain facts
6.1 aura-sync Integration
The aura-sync crate provides a RendezvousAdapter that bridges peer discovery with rendezvous descriptors:
#![allow(unused)] fn main() { use aura_sync::infrastructure::RendezvousAdapter; // Create adapter linking rendezvous to sync let adapter = RendezvousAdapter::new(&rendezvous_service); // Query peer info from cached descriptors if let Some(peer_info) = adapter.get_peer_info(context_id, peer, now_ms) { if peer_info.has_direct_transport() { // Use direct connection } } // Check which peers need descriptor refresh let stale_peers = adapter.peers_needing_refresh(context_id, now_ms); }
6.2 Four-Layer Discovery Model
Aura uses a progressive disclosure model for peer discovery. The aura-social crate provides SocialTopology which determines the appropriate discovery layer based on the caller's social relationships. The four layers in priority order:
- Direct (priority 0): Target is a known peer with existing relationship. Directly accessible with minimal cost.
- Home (priority 1): Target is unknown but reachable through 0-hop home relays. Home relays can forward messages.
- Neighborhood (priority 2): Target is unknown but reachable through neighborhood traversal. Multiple hops across 1-hop linked homes.
- Rendezvous (priority 3): No social presence - requires global flooding through rendezvous infrastructure.
#![allow(unused)] fn main() { use aura_social::{DiscoveryLayer, SocialTopology}; let topology = SocialTopology::new(local_authority, home, neighborhoods); // Determine discovery strategy for target match topology.discovery_layer(&target) { DiscoveryLayer::Direct => { // Direct connection, minimal cost } DiscoveryLayer::Home => { // Relay through available 0-hop relays let (_, relays) = topology.discovery_context(&target); } DiscoveryLayer::Neighborhood => { // Multi-hop through neighborhood let (layer, peers) = topology.discovery_context(&target); } DiscoveryLayer::Rendezvous => { // Global rendezvous flooding required } } }
The discovery layer determines flow budget costs. Direct has minimal cost. Rendezvous has highest cost due to global propagation. This creates economic incentives to establish social relationships before communication.
7. Protocol Flow
The rendezvous sequence uses the context between two authorities.
sequenceDiagram
participant A as Authority A
participant J as Context Journal
participant B as Authority B
A->>A: Build descriptor
A->>A: Evaluate guards
A->>J: Insert descriptor fact
J-->>B: Sync descriptor fact
B->>B: Query descriptor
B->>A: Select transport
A->>B: Noise IKpsk2 handshake
B->>J: Record ChannelEstablished fact
8. Guard Chain Integration
All rendezvous operations flow through the guard chain.
8.1 Guard Capabilities
#![allow(unused)] fn main() { pub mod guards { pub const CAP_RENDEZVOUS_PUBLISH: &str = "rendezvous:publish"; pub const CAP_RENDEZVOUS_CONNECT: &str = "rendezvous:connect"; pub const CAP_RENDEZVOUS_RELAY: &str = "rendezvous:relay"; } }
8.2 Flow Costs
#![allow(unused)] fn main() { pub const DESCRIPTOR_PUBLISH_COST: u32 = 1; pub const CONNECT_DIRECT_COST: u32 = 2; pub const CONNECT_RELAY_COST: u32 = 3; pub const RELAY_FORWARD_COST: u32 = 1; }
8.3 Guard Evaluation
The service prepares operations and returns GuardOutcome containing effect commands. The caller executes these commands.
#![allow(unused)] fn main() { // 1. Prepare snapshot of current state let snapshot = GuardSnapshot { authority_id: alice, context_id: context, flow_budget_remaining: 100, capabilities: vec!["rendezvous:publish".into()], epoch: 1, }; // 2. Prepare publication (pure, sync) let outcome = service.prepare_publish_descriptor( &snapshot, context, transport_hints, now_ms ); // 3. Check decision and execute effects if outcome.decision.is_allowed() { for cmd in outcome.effects { execute_effect_command(cmd).await?; } } }
9. Secure Channel Establishment
After receiving a valid descriptor, the initiator selects a transport. Both sides run Noise IKpsk2 using a context-derived PSK. Successful handshake yields a SecureChannel.
9.1 Channel Structure
#![allow(unused)] fn main() { pub struct SecureChannel { /// Unique channel identifier channel_id: [u8; 32], /// Context this channel belongs to context_id: ContextId, /// Local authority local: AuthorityId, /// Remote peer remote: AuthorityId, /// Current epoch (for key rotation) epoch: u64, /// Channel state state: ChannelState, /// Agreement mode (A1/A2/A3) for the channel lifecycle agreement_mode: AgreementMode, /// Whether reversion is still possible reversion_risk: bool, /// Whether the channel needs key rotation needs_rotation: bool, /// Bytes sent on this channel (for flow budget tracking) bytes_sent: u64, /// Bytes received on this channel bytes_received: u64, } pub enum ChannelState { Establishing, Active, Rotating, Closed, Error(String), } }
9.2 Channel Manager
The ChannelManager tracks active channels:
#![allow(unused)] fn main() { let mut manager = ChannelManager::new(); // Register a new channel manager.register(channel); // Find channel by context and peer if let Some(ch) = manager.find_by_context_peer(context, peer) { if ch.is_active() { // Use channel } } // Advance epoch and mark channels for rotation manager.advance_epoch(new_epoch); }
9.3 Handshake Flow
The Handshaker state machine handles Noise IKpsk2:
#![allow(unused)] fn main() { // Initiator side let mut initiator = Handshaker::new(HandshakeConfig { local: alice, remote: bob, context_id: context, psk: derived_psk, timeout_ms: 5000, }); let init_msg = initiator.create_init_message(epoch)?; // ... send init_msg to responder ... initiator.process_response(&response_msg)?; let result = initiator.complete(epoch, true)?; let channel = initiator.build_channel(&result)?; }
9.4 Key Rotation
Channels support epoch-based key rotation. When the epoch advances, channels rekey using the new context-derived PSK.
#![allow(unused)] fn main() { impl SecureChannel { pub fn needs_epoch_rotation(&self, current_epoch: u64) -> bool { self.epoch < current_epoch } pub fn rotate(&mut self, new_epoch: u64) -> AuraResult<()> { // Rekey the channel self.state = ChannelState::Rotating; self.epoch = new_epoch; self.needs_rotation = false; self.state = ChannelState::Active; Ok(()) } } }
10. Service Interface
The rendezvous service coordinates descriptor publication and channel establishment.
#![allow(unused)] fn main() { impl RendezvousService { /// Create a new rendezvous service pub fn new(authority_id: AuthorityId, config: RendezvousConfig) -> Self; /// Prepare to publish descriptor to context journal pub fn prepare_publish_descriptor( &self, snapshot: &GuardSnapshot, context_id: ContextId, transport_hints: Vec<TransportHint>, now_ms: u64, ) -> GuardOutcome; /// Prepare to establish channel with peer pub fn prepare_establish_channel( &self, snapshot: &GuardSnapshot, context_id: ContextId, peer: AuthorityId, psk: &[u8; 32], ) -> AuraResult<GuardOutcome>; /// Prepare to handle incoming handshake pub fn prepare_handle_handshake( &self, snapshot: &GuardSnapshot, context_id: ContextId, initiator: AuthorityId, handshake: NoiseHandshake, psk: &[u8; 32], ) -> GuardOutcome; /// Cache a peer's descriptor (from journal sync) pub fn cache_descriptor(&mut self, descriptor: RendezvousDescriptor); /// Get a cached descriptor pub fn get_cached_descriptor( &self, context_id: ContextId, peer: AuthorityId, ) -> Option<&RendezvousDescriptor>; /// Check if our descriptor needs refresh pub fn needs_refresh( &self, context_id: ContextId, now_ms: u64, refresh_window_ms: u64, ) -> bool; } }
11. Effect Commands
The service returns GuardOutcome with effect commands to execute:
#![allow(unused)] fn main() { pub enum EffectCommand { /// Append fact to journal JournalAppend { fact: RendezvousFact }, /// Charge flow budget ChargeFlowBudget { cost: FlowCost }, /// Send handshake init message SendHandshake { peer: AuthorityId, message: HandshakeInit }, /// Send handshake response SendHandshakeResponse { peer: AuthorityId, message: HandshakeComplete }, /// Record operation receipt RecordReceipt { operation: String, peer: AuthorityId }, } }
12. Failure Modes and Privacy
Failures occur during guard evaluation, descriptor validation, or transport establishment. These failures are local. No network packets reveal capability or budget failures.
Context isolation prevents unauthorized authorities from reading descriptors. Transport hints do not reveal authority structure. Relay identifiers reveal only the relay authority. Descriptor contents remain encrypted for transit.
13. Summary
Rendezvous provides encrypted peer discovery and channel establishment scoped to relational contexts. Descriptors propagate through journal synchronization with guard chain enforcement. Secure channels use Noise IKpsk2 and QUIC. All behavior remains private to the context and reveals no structural information. The architecture uses standard Aura primitives: domain facts, guard chains, MPST choreographies, and effect interpretation.
Relational Contexts
This document describes the architecture of relational contexts in Aura. It explains how cross-authority relationships are represented using dedicated context namespaces. It defines the structure of relational facts and the role of Aura Consensus in producing agreed relational state. It also describes privacy boundaries and the interpretation of relational data by participating authorities.
Relational contexts are distinct from authority types. In particular, Neighborhood is modeled as an authority type (see Social Architecture), not as a relational-context type.
1. RelationalContext Abstraction
A relational context is shared state linking two or more authorities. A relational context has its own journal namespace. A relational context does not expose internal authority structure. A relational context contains only the facts that the participating authorities choose to share.
A relational context is identified by a ContextId. Authorities publish relational facts inside the context journal. The context journal is a join semilattice under set union. Reduction produces a deterministic relational state.
#![allow(unused)] fn main() { pub struct ContextId(Uuid); }
This identifier selects the journal namespace for a relational context. It does not encode participant information. It does not reveal the type of relationship. Only the participants know how the context is used.
2. Participants and Fact Store
A relational context has a defined set of participating authorities. This set is not encoded in the ContextId. Participation is expressed by writing relational facts to the context journal. Each fact references the commitments of the participating authorities.
#![allow(unused)] fn main() { /// Relational facts in `aura-journal/src/fact.rs` pub enum RelationalFact { /// Protocol-level facts with core reduction semantics Protocol(ProtocolRelationalFact), /// Domain-level extensibility facts Generic { context_id: ContextId, envelope: FactEnvelope }, } /// Typed fact envelope (from aura-core/src/types/facts.rs) pub struct FactEnvelope { pub type_id: FactTypeId, pub schema_version: u16, pub encoding: FactEncoding, pub payload: Vec<u8>, } }
The Protocol variant wraps core protocol facts that have specialized reduction logic in reduce_context(). These include GuardianBinding, RecoveryGrant, Consensus, AMP channel facts, DKG transcript commits, and lifecycle markers. The Generic variant provides extensibility for domain-specific facts using FactEnvelope which contains a typed payload with schema versioning.
Domain crates implement the DomainFact trait from aura-journal/src/extensibility.rs. They store facts via DomainFact::to_generic() which produces a FactEnvelope. The runtime registers reducers in crates/aura-agent/src/fact_registry.rs so reduce_context() can process Generic facts into RelationalBinding values.
3. Prestate Model
Relational context operations use a prestate model. Aura Consensus verifies that all witnesses see the same authority states. The prestate hash binds the relational fact to current authority states.
#![allow(unused)] fn main() { let prestate_hash = H(C_auth1, C_auth2, C_context); }
This hash represents the commitments of the authorities and the current context. Aura Consensus witnesses check this hash before producing shares. The final commit fact includes a threshold signature over the relational operation.
4. Types of Relational Contexts
Several categories of relational contexts appear in Aura. Each fact type carries a well-defined schema to ensure deterministic reduction.
Neighborhood governance and home-membership state are recorded in neighborhood authority journals, not in relational context journals. Relational contexts are still used for pairwise and small-group cross-authority relationships such as guardian bindings, recovery grants, and application-specific collaboration contexts.
GuardianConfigContext
Stores GuardianBinding facts:
{
"type": "GuardianBinding",
"account_commitment": "Hash32",
"guardian_commitment": "Hash32",
"parameters": {
"recovery_delay_secs": 86400,
"notification_required": true
},
"consensus_commitment": "Hash32",
"consensus_proof": "ThresholdSignature"
}
account_commitment: reduced commitment of the protected authority.guardian_commitment: reduced commitment of the guardian authority.parameters: serialized GuardianParameters (delay, notification policy, etc.).consensus_commitment: commitment hash of the Aura Consensus instance that approved the binding.consensus_proof: aggregated signature from the witness set.
GuardianRecoveryContext
Stores RecoveryGrant facts:
{
"type": "RecoveryGrant",
"account_commitment_old": "Hash32",
"account_commitment_new": "Hash32",
"guardian_commitment": "Hash32",
"operation": {
"kind": "ReplaceTree",
"payload": "Base64(TreeOp)"
},
"consensus_commitment": "Hash32",
"consensus_proof": "ThresholdSignature"
}
account_commitment_old/new: before/after commitments for the account authority.guardian_commitment: guardian authority that approved the grant.operation: serialized recovery operation (matches TreeOp schema).consensus_*: Aura Consensus identifiers tying the grant to witness approvals.
Generic / Application Contexts
Shared group or project contexts can store application-defined facts:
{
"type": "Generic",
"payload": "Base64(opaque application data)",
"bindings": ["Hash32 commitment of participant A", "Hash32 commitment of participant B"],
"labels": ["project:alpha", "role:reviewer"]
}
Generic facts should include enough metadata (bindings, optional labels) for interpreters to apply context-specific rules.
5. Relational Facts
Relational facts express specific cross-authority operations. A GuardianBinding fact defines the guardian authority for an account while a RecoveryGrant fact defines an allowed update to the account state. A Generic fact covers application defined interactions. Consensus-backed facts include the consensus_commitment and aggregated signature so reducers can verify provenance even after witnesses rotate.
SessionDelegation protocol facts record runtime endpoint transfer events. Aura emits these facts when session ownership moves across authorities (for example, guardian handoff or device migration). Each delegation fact includes source authority, destination authority, session id, optional bundle id, and timestamp so reconfiguration decisions remain auditable.
Reduction applies all relational facts to produce relational state. Reduction verifies that authority commitments in each fact match the current reduced state of each authority.
#![allow(unused)] fn main() { /// Reduced relational state from aura-journal/src/reduction.rs pub struct RelationalState { /// Active relational bindings pub bindings: Vec<RelationalBinding>, /// Flow budget state by context pub flow_budgets: BTreeMap<(AuthorityId, AuthorityId, u64), u64>, /// Leakage budget totals for privacy accounting pub leakage_budget: LeakageBudget, /// AMP channel epoch state keyed by channel id pub channel_epochs: BTreeMap<ChannelId, ChannelEpochState>, } pub struct RelationalBinding { pub binding_type: RelationalBindingType, pub context_id: ContextId, pub data: Vec<u8>, } }
This structure represents the reduced relational state. It contains relational bindings, flow budget tracking between authorities, leakage budget totals for privacy accounting, and AMP channel epoch state for message ratcheting. Reduction processes all facts in the context journal to derive this state deterministically.
6. Aura Consensus in Relational Contexts
Some relational operations require strong agreement. Aura Consensus produces these operations. Aura Consensus uses a witness set drawn from participating authorities. Witnesses compute shares after verifying the prestate hash.
Commit facts contain threshold signatures. Each commit fact is inserted into the relational context journal. Reduction interprets the commit fact as a confirmed relational operation.
Aura Consensus binds relational operations to authority state. After consensus completes, the initiator inserts a CommitFact into the relational context journal that includes:
#![allow(unused)] fn main() { pub struct RelationalCommitFact { pub context_id: ContextId, pub consensus_commitment: Hash32, // H(Op, prestate) pub fact: RelationalFact, // GuardianBinding, RecoveryGrant, etc. pub aggregated_signature: ThresholdSignature, pub attesters: BTreeSet<AuthorityId>, } }
Reducers validate aggregated_signature before accepting the embedded RelationalFact. This mirrors the account-level process but scoped to the context namespace.
7. Interpretation of Relational State
Authorities interpret relational state by reading the relational context journal. An account reads GuardianBinding facts to determine its guardian authority. A guardian authority reads the same facts to determine which accounts it protects.
Other relational contexts follow similar patterns. Each context defines its own interpretation rules. No context affects authority internal structure. Context rules remain confined to the specific relationship.
8. Privacy and Isolation
A relational context does not reveal its participants. The ContextId is opaque. Only participating authorities know the relationship. No external party can infer connections between authorities based on context identifiers.
Profile information shared inside a context stays local to that context. Nickname suggestions and contact attributes do not leave the context journal. Each context forms a separate identity boundary. Authorities can maintain many unrelated relationships without cross linking.
9. Implementation Patterns
The implementation in aura-relational provides concrete patterns for working with relational contexts.
Creating and Managing Contexts
#![allow(unused)] fn main() { use aura_core::{AuthorityId, ContextId}; use aura_relational::RelationalContext; // Create a new guardian-account relational context let account_authority = AuthorityId::from_entropy([1u8; 32]); let guardian_authority = AuthorityId::from_entropy([2u8; 32]); let context = RelationalContext::new(vec![account_authority, guardian_authority]); // Or use a specific context ID let context_id = ContextId::from_entropy([3u8; 32]); let context = RelationalContext::with_id( context_id, vec![account_authority, guardian_authority], ); // Check participation assert!(context.is_participant(&account_authority)); assert!(context.is_participant(&guardian_authority)); }
Guardian Binding Pattern
#![allow(unused)] fn main() { use aura_core::relational::{GuardianBinding, GuardianParameters}; use std::time::Duration; let params = GuardianParameters { recovery_delay: Duration::from_secs(86400), notification_required: true, // Use Aura's unified time system (PhysicalTimeEffects/TimeStamp) for expiration. expiration: None, }; let binding = GuardianBinding::new( Hash32::from_bytes(&account_authority.to_bytes()), Hash32::from_bytes(&guardian_authority.to_bytes()), params, ); // Store a binding receipt + full payload via the context's journal-backed API context.add_guardian_binding(account_authority, guardian_authority, binding)?; }
Recovery Grant Pattern
#![allow(unused)] fn main() { use aura_core::relational::{RecoveryGrant, RecoveryOp, ConsensusProof}; // Construct a recovery operation let recovery_op = RecoveryOp::AddDevice { device_public_key: new_device_pubkey.to_bytes(), }; // Create recovery grant (requires consensus proof) let grant = RecoveryGrant { account_old: old_tree_commitment, account_new: new_tree_commitment, guardian: guardian_commitment, operation: recovery_op, consensus_proof: consensus_result.proof, // From Aura Consensus }; // Add to recovery context recovery_context.add_fact(RelationalFact::RecoveryGrant(grant))?; // Check operation type if grant.operation.is_emergency() { // Handle emergency operations immediately execute_emergency_recovery(&grant)?; } }
Query Patterns
#![allow(unused)] fn main() { // Query guardian bindings let bindings = context.guardian_bindings(); for binding in bindings { println!("Guardian: {:?}", binding.guardian_commitment); println!("Recovery delay: {:?}", binding.parameters.recovery_delay); if let Some(expiration) = binding.parameters.expiration { if expiration < Utc::now() { // Binding has expired } } } // Find specific guardian binding if let Some(binding) = context.get_guardian_binding(account_authority) { // Found guardian for this account let guardian_id = binding.guardian_commitment; } // Query recovery grants let grants = context.recovery_grants(); for grant in grants { println!("Operation: {}", grant.operation.description()); println!("From: {:?} -> To: {:?}", grant.account_old, grant.account_new); } }
Generic Binding Pattern
#![allow(unused)] fn main() { use aura_core::relational::{GenericBinding, RelationalFact}; use aura_core::types::facts::{FactEnvelope, FactTypeId, FactEncoding}; // Application-specific binding (e.g., project collaboration) let payload = serde_json::to_vec(&ProjectMetadata { name: "Alpha Project", role: "Reviewer", permissions: vec!["read", "comment"], })?; let envelope = FactEnvelope { type_id: FactTypeId::new("project_collaboration"), schema_version: 1, encoding: FactEncoding::Json, payload, }; let generic = GenericBinding::new(envelope, None); // No consensus proof context.add_fact(RelationalFact::Generic(generic))?; }
Prestate Computation Pattern
#![allow(unused)] fn main() { use aura_core::Prestate; // Collect current authority commitments let authority_commitments = vec![ (account_authority, account_tree_commitment), (guardian_authority, guardian_tree_commitment), ]; // Compute prestate for consensus let prestate = context.compute_prestate(authority_commitments); // Use prestate in consensus protocol let consensus_result = run_consensus( prestate, operation_data, witness_set, ).await?; }
Journal Commitment Pattern
#![allow(unused)] fn main() { // Get deterministic commitment of current relational state let commitment = context.journal.compute_commitment(); // Commitment is deterministic - all replicas compute same value // Used for: // - Prestate computation in consensus // - Context verification // - Anti-entropy sync checkpoints }
Integration with Aura Consensus
#![allow(unused)] fn main() { use aura_core::relational::ConsensusProof; use aura_relational::run_consensus; // 1. Prepare operation requiring consensus let binding = GuardianBindingBuilder::new() .account(account_commitment) .guardian(guardian_commitment) .build()?; // 2. Compute prestate let prestate = context.compute_prestate(authority_commitments); // 3. Run consensus let consensus_proof = run_consensus( prestate, serde_json::to_vec(&binding)?, witness_set, ).await?; // 4. Attach proof to fact let binding_with_proof = GuardianBinding { consensus_proof: Some(consensus_proof), ..binding }; // 5. Add to context context.add_fact(RelationalFact::GuardianBinding(binding_with_proof))?; }
Recovery Operation Selection
#![allow(unused)] fn main() { use aura_core::relational::RecoveryOp; // Select appropriate recovery operation let recovery_op = match recovery_scenario { RecoveryScenario::LostAllDevices => RecoveryOp::ReplaceTree { new_tree_root: new_tree_commitment, }, RecoveryScenario::AddNewDevice => RecoveryOp::AddDevice { device_public_key: device_key.to_bytes(), }, RecoveryScenario::RemoveCompromised => RecoveryOp::RemoveDevice { leaf_index: compromised_device_index, }, RecoveryScenario::ChangeThreshold => RecoveryOp::UpdatePolicy { new_threshold: new_m_of_n.0, }, RecoveryScenario::EmergencyCompromise => RecoveryOp::EmergencyRotation { new_epoch: current_epoch + 1, }, }; // Emergency operations bypass delay if recovery_op.is_emergency() { // No recovery_delay applied } else { // Wait for binding.parameters.recovery_delay } }
Best Practices
Guardian Configuration:
- Use 24-hour minimum recovery delay for security
- Always require notification unless emergency scenario
- Set expiration for temporary guardian relationships
- Rotate guardian bindings periodically
Recovery Grants:
- Always require consensus proof for recovery operations
- Validate prestate commitments before accepting grants
- Log all recovery operations for audit trail
- Emergency operations should be rare and logged prominently
Generic Bindings:
- Document your FactTypeId schemas externally
- Use schema_version field to version your payload format
- Include consensus_proof for critical application bindings
- Keep payload size reasonable for sync performance
Context Management:
- Use opaque ContextId - never encode participant info
- Limit participants to 2-10 authorities for efficiency
- Separate contexts for different relationship types
- Garbage collect expired bindings periodically
Lifecycle Notes:
- A1/A2 facts are usable immediately but are provisional; any durable relational state must be A3 (consensus-finalized).
- Soft-safe operations should emit
ConvergenceCertandReversionFactprotocol facts to make convergence and reversion risk explicit.
10. Contexts vs Channels: Operation Categories
Understanding the distinction between relational contexts and channels is essential for operation categorization.
10.1 Relational Contexts (Category C - Consensus Required)
Creating a relational context establishes a cryptographic relationship between authorities. This is a Category C (consensus-gated) operation because:
- It creates the shared secret foundation for all future communication
- Both parties must agree to establish the relationship
- Partial state (one party thinks relationship exists, other doesn't) is dangerous
Examples:
- Adding a contact (bilateral context between two authorities)
- Creating a group (multi-party context with all members)
- Adding a member to an existing group (extends the cryptographic context)
10.2 Channels Within Contexts (Category A - Optimistic)
Once a relational context exists, channels are Category A (optimistic) operations:
- Channels are just organizational substreams within the context
- No new cryptographic agreement needed - keys derive from context
- Channel facts sync via anti-entropy, eventual consistency is sufficient
Examples within existing context:
- Create channel → emit
ChannelCheckpointfact - Send message → derive key from context, encrypt, send
- Update topic → emit fact to context journal
10.3 The Cost Structure
BILATERAL MULTI-PARTY
(2 members) (3+ members)
─────────── ────────────
Context Creation Invitation ceremony Group ceremony
(Category C - expensive) (Category C - expensive)
Member Addition N/A (already 2) Per-member ceremony
(Category C - expensive)
Channel Creation Optimistic Optimistic
(Category A - cheap) (Category A - cheap)
Messages Optimistic Optimistic
(Category A - cheap) (Category A - cheap)
The expensive part is establishing WHO is in the group. Once that's established, operations WITHIN the group are cheap.
10.4 Multi-Party Context Keys
Groups with >2 members derive keys from all member tree roots:
GroupContext {
context_id: ContextId,
members: [Alice, Bob, Carol],
group_secret: DerivedFromMemberTreeRoots,
epoch: u64,
}
Key Derivation:
1. Each member contributes their tree root commitment
2. Group secret = KDF(sorted_member_roots, context_id)
3. Channel key = KDF(group_secret, channel_id, epoch)
All members derive the SAME group secret from the SAME inputs
10.5 Why Membership Changes Require Ceremony
Group membership changes are Category C because they affect encryption:
-
Forward Secrecy: New members shouldn't read old messages
- Solution: Epoch rotation, new keys for new messages
-
Post-Compromise Security: Removed members shouldn't read new messages
- Solution: Epoch rotation, re-derive group secret without removed member
-
Consistency: All members must agree on who's in the group
- Solution: Ceremony ensures atomic membership view
See Consensus - Operation Categories for the full decision tree.
11. Summary
Relational contexts represent cross-authority relationships in Aura. They provide shared state without revealing authority structure and support guardian configuration, recovery, and application specific collaboration. Aura Consensus ensures strong agreement where needed while deterministic reduction ensures consistent relational state. Privacy boundaries isolate each relationship from all others.
The implementation provides concrete types (RelationalContext, GuardianBinding, RecoveryGrant) with builder patterns, query methods, and consensus integration. All relational facts are stored in a CRDT journal with deterministic commitment computation.
12. Implementation References
- Core Types:
aura-core/src/relational/- RelationalFact, GuardianBinding, RecoveryGrant, ConsensusProof domain types - Journal Facts:
aura-journal/src/fact.rs- Protocol-level RelationalFact with AMP variants - Reduction:
aura-journal/src/reduction.rs- RelationalState, reduce_context() - Context Management:
aura-relational/src/lib.rs- RelationalContext (context-scoped fact journal mirror + helpers) - Consensus Integration:
crates/aura-consensus/src/consensus/relational.rs- consensus implementation - Consensus Adapter:
aura-relational/src/consensus_adapter.rs- thin consensus delegation layer - Prestate Computation:
aura-core/src/domain/consensus.rs- Prestate struct and methods - Protocol Usage:
aura-authentication/src/guardian_auth_relational.rs- Guardian authentication - Recovery Flows:
aura-recovery/src/- Guardian recovery choreographies
See Also
- Operation Categories - Ceremony contract, guardian rotation, device enrollment, and Category B/C operations
- Consensus - Aura Consensus protocol for strong agreement
Database Architecture
This document specifies the architecture for Aura's query system. The journal is the database. Datalog is the query language. Biscuit provides authorization.
1. Core Principles
1.1 Database-as-Journal Equivalence
Aura's fact-based journal functions as the database. There is no separate database layer. The equivalence maps traditional database concepts to Aura components.
Aura treats database state as a composite of the fact journal and the capability frontier. Query execution always combines the reduced fact state with the current capability lattice (the JournalState composite) to enforce authorization and isolate contexts.
| Traditional Database | Aura Component |
|---|---|
| Table | Journal reduction view |
| Row | Fact implementing JoinSemilattice |
| Transaction | Atomic fact append |
| Index | Merkle trees and Bloom filters |
| Query | Datalog program evaluation |
| Replication | CrdtCoordinator with delta sync |
1.2 Authority-First Data Model
Aura's database is partitioned by cryptographic authorities. An AuthorityId owns facts that implement JoinSemilattice. State is derived from those facts.
Data is naturally sharded by authority. Cross-authority operations require explicit choreography. Privacy is the default because no cross-authority visibility exists without permission.
2. Query System
2.1 Query Trait
The Query trait defines typed queries that compile to Datalog:
#![allow(unused)] fn main() { pub trait Query: Send + Sync + Clone + 'static { /// The result type of this query type Result: Clone + Send + Sync + Default + 'static; /// Compile this query to a Datalog program fn to_datalog(&self) -> DatalogProgram; /// Get required Biscuit capabilities for authorization fn required_capabilities(&self) -> Vec<QueryCapability>; /// Get fact predicates for invalidation tracking fn dependencies(&self) -> Vec<FactPredicate>; /// Parse Datalog bindings to typed result fn parse(bindings: DatalogBindings) -> Result<Self::Result, QueryParseError>; /// Unique identifier for caching and subscriptions fn query_id(&self) -> String; } }
This design separates query definition from execution, enabling:
- Portable query definitions across runtimes
- Authorization checking before execution
- Reactive subscriptions via dependency tracking
2.2 Datalog Types
Queries compile to DatalogProgram, the intermediate representation:
#![allow(unused)] fn main() { pub struct DatalogProgram { pub rules: Vec<DatalogRule>, pub facts: Vec<DatalogFact>, pub goal: Option<String>, } pub struct DatalogRule { pub head: DatalogFact, pub body: Vec<DatalogFact>, } pub struct DatalogFact { pub predicate: String, pub args: Vec<DatalogValue>, } pub enum DatalogValue { String(String), Integer(i64), Boolean(bool), Variable(String), Symbol(String), Null, } }
Programs convert to Datalog source via to_datalog_source():
#![allow(unused)] fn main() { let program = DatalogProgram::new(vec![ DatalogRule::new(DatalogFact::new("active_user", vec![DatalogValue::var("name")])) .when(DatalogFact::new("user", vec![DatalogValue::var("name")])) .when(DatalogFact::new("online", vec![DatalogValue::var("name")])) ]) .with_goal("active_user($name)"); let source = program.to_datalog_source(); // Output: // active_user($name) :- user($name), online($name). // ?- active_user($name). }
2.3 Fact Predicates
FactPredicate patterns determine query invalidation:
#![allow(unused)] fn main() { pub struct FactPredicate { /// The predicate name to match pub name: String, /// Optional positional argument patterns (None = wildcard) pub arg_patterns: Vec<Option<String>>, /// Named field constraints for structured facts pub named_constraints: BTreeMap<String, String>, } impl FactPredicate { /// Match any fact with the given name pub fn named(name: impl Into<String>) -> Self; /// Match facts with specific named field constraints pub fn with_args(name: impl Into<String>, args: Vec<(&str, &str)>) -> Self; /// Add a positional argument pattern pub fn with_arg(self, pattern: Option<String>) -> Self; /// Add a named field constraint pub fn with_named_constraint(self, name: impl Into<String>, value: impl Into<String>) -> Self; /// Check if this predicate matches another pub fn matches(&self, other: &FactPredicate) -> bool; /// Check if this predicate matches a fact with positional arguments pub fn matches_fact(&self, fact_name: &str, fact_args: &[String]) -> bool; /// Check if this predicate matches a fact with named fields pub fn matches_named_fact(&self, fact_name: &str, fact_fields: &BTreeMap<String, String>) -> bool; } }
When facts change, subscriptions matching the predicate re-evaluate.
2.4 Query Capabilities
Authorization integrates with Biscuit via QueryCapability:
#![allow(unused)] fn main() { pub struct QueryCapability { pub resource: String, pub action: String, pub constraints: Vec<(String, String)>, } impl QueryCapability { pub fn read(resource: impl Into<String>) -> Self; pub fn list(resource: impl Into<String>) -> Self; pub fn with_constraint(self, key: impl Into<String>, value: impl Into<String>) -> Self; pub fn to_biscuit_check(&self) -> String; } }
Capabilities convert to Biscuit checks:
#![allow(unused)] fn main() { let cap = QueryCapability::read("channels").with_constraint("owner", "alice"); assert_eq!( cap.to_biscuit_check(), "check if right(\"channels\", \"read\"), owner == \"alice\"" ); }
3. Query Effects
3.1 QueryEffects Trait
QueryEffects executes queries against the journal:
#![allow(unused)] fn main() { #[async_trait] pub trait QueryEffects: Send + Sync { /// Execute a one-shot typed query async fn query<Q: Query>(&self, query: &Q) -> Result<Q::Result, QueryError>; /// Execute a raw Datalog program async fn query_raw(&self, program: &DatalogProgram) -> Result<DatalogBindings, QueryError>; /// Subscribe to query updates when facts change fn subscribe<Q: Query>(&self, query: &Q) -> QuerySubscription<Q::Result>; /// Pre-check authorization async fn check_capabilities(&self, capabilities: &[QueryCapability]) -> Result<(), QueryError>; /// Trigger re-evaluation for matching subscriptions async fn invalidate(&self, predicate: &FactPredicate); } }
3.2 Query Execution Flow
flowchart TD
A[Query::to_datalog] --> B[QueryEffects::query];
B --> C{Check capabilities};
C -->|Pass| D[Load journal facts];
C -->|Fail| E[QueryError::AuthorizationFailed];
D --> F[Execute Datalog];
F --> G[Query::parse];
G --> H[Typed result];
3.3 Reactive Subscriptions
QuerySubscription wraps SignalStream for live updates:
#![allow(unused)] fn main() { pub struct QuerySubscription<T: Clone + Send + 'static> { stream: SignalStream<T>, query_id: String, } impl<T: Clone + Send + 'static> QuerySubscription<T> { pub fn query_id(&self) -> &str; pub fn try_recv(&mut self) -> Option<T>; pub async fn recv(&mut self) -> Result<T, QueryError>; } }
Usage pattern:
#![allow(unused)] fn main() { let mut subscription = effects.subscribe(&ChannelsQuery::default()); while let Ok(channels) = subscription.recv().await { println!("Channels updated: {} total", channels.len()); } }
3.4 Query Isolation
QueryIsolation specifies consistency requirements for queries:
#![allow(unused)] fn main() { pub enum QueryIsolation { /// See all facts including uncommitted (CRDT state) - fastest ReadUncommitted, /// Only see facts with consensus commit ReadCommitted { wait_for: Vec<ConsensusId> }, /// Snapshot at specific prestate (time-travel query) Snapshot { prestate_hash: Hash32 }, /// Wait for all pending consensus in scope ReadLatest { scope: ResourceScope }, } }
Usage:
#![allow(unused)] fn main() { // Fast query - may see uncommitted facts let result = effects.query(&ChannelsQuery::default()).await?; // Wait for specific consensus before querying let result = effects.query_with_isolation( &ChannelsQuery::default(), QueryIsolation::ReadCommitted { wait_for: vec![consensus_id] }, ).await?; }
3.5 Query Statistics
QueryStats provides execution metrics:
#![allow(unused)] fn main() { pub struct QueryStats { pub execution_time: Duration, pub facts_scanned: u32, pub facts_matched: u32, pub cache_hit: bool, pub isolation_used: QueryIsolation, pub consensus_wait_time: Option<Duration>, /// Consistency metadata for matched facts pub consistency: ConsistencyMap, } }
Usage:
#![allow(unused)] fn main() { let (channels, stats) = effects.query_with_stats(&ChannelsQuery::default()).await?; println!("Query took {:?}, scanned {} facts", stats.execution_time, stats.facts_scanned); }
3.6 Query Errors
#![allow(unused)] fn main() { pub enum QueryError { AuthorizationFailed { reason: String }, MissingCapability { capability: String }, ExecutionError { reason: String }, ParseError(QueryParseError), SubscriptionNotFound { query_id: String }, JournalError { reason: String }, HandlerUnavailable, Internal { reason: String }, ConsensusTimeout { consensus_id: ConsensusId }, SnapshotNotAvailable { prestate_hash: Hash32 }, IsolationNotSupported { reason: String }, } }
4. Concrete Query Examples
4.1 ChannelsQuery
#![allow(unused)] fn main() { #[derive(Clone, Default)] pub struct ChannelsQuery { pub channel_type: Option<String>, } impl Query for ChannelsQuery { type Result = Vec<Channel>; fn to_datalog(&self) -> DatalogProgram { DatalogProgram::new(vec![ DatalogRule::new(DatalogFact::new( "channel", vec![ DatalogValue::var("id"), DatalogValue::var("name"), DatalogValue::var("type"), ], )) .when(DatalogFact::new( "channel_fact", vec![ DatalogValue::var("id"), DatalogValue::var("name"), DatalogValue::var("type"), ], )), ]) .with_goal("channel($id, $name, $type)") } fn required_capabilities(&self) -> Vec<QueryCapability> { vec![QueryCapability::list("channels")] } fn dependencies(&self) -> Vec<FactPredicate> { vec![FactPredicate::named("channel_fact")] } fn parse(bindings: DatalogBindings) -> Result<Self::Result, QueryParseError> { bindings.rows.iter().map(|row| { Ok(Channel { id: row.get_string("id") .ok_or(QueryParseError::MissingField { field: "id".into() })? .to_string(), name: row.get_string("name") .ok_or(QueryParseError::MissingField { field: "name".into() })? .to_string(), channel_type: row.get_string("type").map(String::from), }) }).collect() } } }
4.2 MessagesQuery
#![allow(unused)] fn main() { #[derive(Clone)] pub struct MessagesQuery { pub channel_id: String, pub limit: Option<usize>, } impl Query for MessagesQuery { type Result = Vec<Message>; fn to_datalog(&self) -> DatalogProgram { DatalogProgram::new(vec![ DatalogRule::new(DatalogFact::new( "message", vec![ DatalogValue::var("id"), DatalogValue::var("content"), DatalogValue::var("sender"), DatalogValue::var("timestamp"), ], )) .when(DatalogFact::new( "message_fact", vec![ DatalogValue::String(self.channel_id.clone()), DatalogValue::var("id"), DatalogValue::var("content"), DatalogValue::var("sender"), DatalogValue::var("timestamp"), ], )), ]) .with_goal("message($id, $content, $sender, $timestamp)") } fn required_capabilities(&self) -> Vec<QueryCapability> { vec![QueryCapability::read("messages") .with_constraint("channel", &self.channel_id)] } fn dependencies(&self) -> Vec<FactPredicate> { vec![FactPredicate::with_args( "message_fact", vec![("channel_id", &self.channel_id)], )] } fn parse(bindings: DatalogBindings) -> Result<Self::Result, QueryParseError> { // Parse message rows... } } }
5. Indexing Layer
5.1 IndexedJournalEffects
The IndexedJournalEffects trait provides efficient indexed lookups:
#![allow(unused)] fn main() { pub trait IndexedJournalEffects: Send + Sync { /// Subscribe to journal fact updates as they are added fn watch_facts(&self) -> Box<dyn FactStreamReceiver>; /// Get all facts with the given predicate/key async fn facts_by_predicate(&self, predicate: &str) -> Result<Vec<IndexedFact>, AuraError>; /// Get all facts created by the given authority async fn facts_by_authority(&self, authority: &AuthorityId) -> Result<Vec<IndexedFact>, AuraError>; /// Get all facts within the given time range (inclusive) async fn facts_in_range(&self, start: TimeStamp, end: TimeStamp) -> Result<Vec<IndexedFact>, AuraError>; /// Return all indexed facts (append-only view) async fn all_facts(&self) -> Result<Vec<IndexedFact>, AuraError>; /// Fast membership test using Bloom filter fn might_contain(&self, predicate: &str, value: &FactValue) -> bool; /// Get the Merkle root commitment for the current index state async fn merkle_root(&self) -> Result<[u8; 32], AuraError>; /// Verify a fact against the Merkle tree async fn verify_fact_inclusion(&self, fact: &IndexedFact) -> Result<bool, AuraError>; /// Get the Bloom filter for fast membership tests async fn get_bloom_filter(&self) -> Result<BloomFilter, AuraError>; /// Get statistics about the index async fn index_stats(&self) -> Result<IndexStats, AuraError>; } }
The might_contain method uses Bloom filters for fast negative answers with O(1) lookup and less than 1% false positive rate.
5.2 Index Structure
#![allow(unused)] fn main() { pub struct AuthorityIndex { merkle_tree: MerkleTree<FactHash>, predicate_filters: BTreeMap<String, BloomFilter>, by_predicate: BTreeMap<String, Vec<FactId>>, by_authority: BTreeMap<AuthorityId, Vec<FactId>>, by_time: BTreeMap<TimeStamp, Vec<FactId>>, } }
- Merkle trees: Integrity verification
- Bloom filters: Fast membership tests (<1% false positive rate)
- B-trees: Ordered lookups (O(log n))
Indexes update on fact commit. Performance target: <10ms for 10k facts.
6. Biscuit Integration
6.1 AuraQuery Wrapper
AuraQuery wraps Biscuit's authorizer for query execution:
#![allow(unused)] fn main() { pub struct AuraQuery { authorizer: biscuit_auth::Authorizer, } impl AuraQuery { pub fn add_journal_facts(&mut self, facts: &[Fact]) -> Result<()> { for fact in facts { self.authorizer.add_fact(fact.to_biscuit_fact()?)?; } Ok(()) } pub fn query(&self, rule: &str) -> Result<Vec<biscuit_auth::Fact>> { self.authorizer.query(rule) } } }
6.2 Guard Chain Integration
Database operations flow through the guard chain:
flowchart LR
A[Query Request] --> B[CapGuard];
B --> C[FlowGuard];
C --> D[JournalCoupler];
D --> E[QueryEffects];
- CapGuard: Evaluates Biscuit token authorization
- FlowGuard: Charges budget for query cost
- JournalCoupler: Logs query execution
- QueryEffects: Executes the query
Each guard must succeed before the next executes.
7. Transaction Model
7.1 Coordination Matrix
Database operations use two orthogonal dimensions:
| Single Authority | Cross-Authority | |
|---|---|---|
| Monotone | Direct fact insertion (0 RTT) | CRDT merge via anti-entropy (0 RTT + sync) |
| Consensus | Single-authority consensus (1-2 RTT) | Multi-authority consensus (2-3 RTT) |
7.2 Examples
- Monotone + Single: Append message to own channel (
journal.insert_fact()) - Monotone + Cross-Authority: Guardian adds trust fact (
journal.insert_relational_fact()) - Consensus + Single: Remove device from account (
consensus_single_shot()) - Consensus + Cross-Authority: Recovery grant with guardian approval (
federated_consensus())
Aura Consensus is not linearizable by default. Each consensus instance independently agrees on a single operation and prestate. To sequence operations, use session types (see MPST and Choreography).
Agreement modes are orthogonal to the coordination matrix: A1 (provisional) and A2 (soft-safe) may provide immediate usability, but any durable shared database state must be A3 (consensus-finalized) with prestate binding. Soft-safe windows should be bounded with convergence certificates and explicit reversion facts.
BFT-DKG integration: When key material is required (K3), the database must bind
operations to a consensus‑finalized DkgTranscriptCommit. This ensures the
transaction prestate and the cryptographic prestate are aligned.
7.3 Mutation Receipts
MutationReceipt indicates how a mutation was coordinated:
#![allow(unused)] fn main() { pub enum MutationReceipt { /// Monotone operation completed immediately via CRDT merge Immediate { fact_ids: Vec<FactId>, timestamp: PhysicalTime, }, /// Non-monotone operation submitted to consensus Consensus { consensus_id: ConsensusId, prestate_hash: Hash32, submit_latency: Duration, }, } }
Usage pattern:
#![allow(unused)] fn main() { let receipt = effects.mutate(operation).await?; match receipt { MutationReceipt::Immediate { fact_ids, .. } => { // Facts are immediately visible println!("{} facts created", fact_ids.len()); } MutationReceipt::Consensus { consensus_id, .. } => { // Wait for consensus completion before querying let result = effects.query_with_isolation( &MyQuery::default(), QueryIsolation::read_committed(consensus_id), ).await?; } } }
8. Temporal Database Model
Aura uses a Datomic-inspired immutable database model where all changes are represented as append-only facts with temporal metadata.
8.1 Core Concepts
Facts are never deleted. They are either:
- Asserted: Added to a scope
- Retracted: Marked as no longer valid (but remain queryable in history)
- Epoch-bumped: Bulk invalidation of facts in a scope
- Checkpointed: Snapshotted for temporal queries
Facts are organized in hierarchical scopes:
#![allow(unused)] fn main() { // Scope path examples "authority:abc123" // Authority-level "authority:abc123/chat" // Named sub-scope "authority:abc123/chat/channel:xyz" // Typed sub-scope }
Facts progress through finality levels:
#![allow(unused)] fn main() { pub enum Finality { Local, // Written locally only Replicated { ack_count: u16 }, // Acknowledged by N peers Checkpointed, // In a durable checkpoint Consensus { proof: ConsensusId }, // Confirmed via consensus Anchored { anchor: AnchorProof }, // External chain anchor } }
8.2 Fact Operations
#![allow(unused)] fn main() { pub enum FactOp { /// Assert a new fact Assert { content: FactContent, scope: Option<ScopeId> }, /// Mark a specific fact as retracted Retract { target: FactId, reason: RetractReason }, /// Invalidate all facts in scope before new epoch EpochBump { scope: ScopeId, new_epoch: Epoch, checkpoint: Option<Hash32> }, /// Create a queryable snapshot Checkpoint { scope: ScopeId, state_hash: Hash32, supersedes: Vec<FactId> }, } }
Monotonic vs non-monotonic:
AssertandCheckpointare monotonic (no coordination required)RetractandEpochBumpare non-monotonic (may require consensus)
8.3 FactEffects Trait
The write interface for the temporal database:
#![allow(unused)] fn main() { #[async_trait] pub trait FactEffects: Send + Sync { /// Apply a single fact operation async fn apply_op(&self, op: FactOp, scope: &ScopeId) -> Result<FactReceipt, FactError>; /// Apply a transaction atomically async fn apply_transaction(&self, tx: Transaction) -> Result<TransactionReceipt, FactError>; /// Wait for finality level async fn wait_for_finality(&self, fact_id: FactId, target: Finality) -> Result<Finality, FactError>; /// Configure scope finality requirements async fn configure_scope(&self, config: ScopeFinalityConfig) -> Result<(), FactError>; /// Query with temporal constraints async fn query_temporal(&self, scope: &ScopeId, temporal: TemporalQuery) -> Result<Vec<TemporalFact>, FactError>; } }
8.4 Transactions
For atomic operations, facts can be grouped:
#![allow(unused)] fn main() { let tx = Transaction::new(ScopeId::authority("abc")) .with_op(FactOp::assert(content1)) .with_op(FactOp::assert(content2)) .with_op(FactOp::retract(old_id, RetractReason::Superseded { by: new_id })) .with_finality(Finality::Checkpointed); let receipt = effects.apply_transaction(tx).await?; }
Hybrid transaction model:
- Simple monotonic operations: Direct
apply_op()(no transaction overhead) - Atomic operations: Explicit
Transactiongrouping when needed
8.5 Temporal Queries
Query facts with respect to time:
#![allow(unused)] fn main() { pub enum TemporalQuery { /// Database state at a point in time AsOf(TemporalPoint), /// Changes since a point in time (delta) Since(TemporalPoint), /// Full history over a time range History { from: TemporalPoint, to: TemporalPoint }, } pub enum TemporalPoint { Physical(PhysicalTime), // Wall-clock time Order(OrderTime), // Opaque order token AfterTransaction(TransactionId), // After specific transaction AtEpoch { scope: ScopeId, epoch: Epoch }, // At scope epoch Now, // Current state } }
Usage:
#![allow(unused)] fn main() { // Query current state let facts = effects.query_temporal(&scope, TemporalQuery::current()).await?; // Time-travel query let historical = effects.query_temporal( &scope, TemporalQuery::as_of(TemporalPoint::Physical(past_time)), ).await?; // Get changes since last sync let delta = effects.query_temporal( &scope, TemporalQuery::since(TemporalPoint::AfterTransaction(last_tx_id)), ).await?; }
8.6 Finality Configuration
Per-scope finality with operation override:
#![allow(unused)] fn main() { let config = ScopeFinalityConfig::new(ScopeId::parse("authority:abc/payments")?) .with_default(Finality::Checkpointed) // Default for this scope .with_minimum(Finality::replicated(2)) // Minimum required .with_cascade(true) // Inherit to child scopes .with_override(ContentFinalityOverride::new( "high_value_transfer", Finality::consensus(ConsensusId([0; 32])), // Require consensus for transfers )); effects.configure_scope(config).await?; }
9. Consistency Metadata
Query results include consistency metadata that tracks the agreement, propagation, and acknowledgment status of each fact.
9.1 ConsistencyMap
The ConsistencyMap type provides per-item consistency status in query results:
#![allow(unused)] fn main() { pub struct ConsistencyMap { entries: HashMap<String, Consistency>, } impl ConsistencyMap { pub fn get(&self, id: &str) -> Option<&Consistency>; pub fn is_finalized(&self, id: &str) -> bool; pub fn acked_by(&self, id: &str) -> Option<&[AckRecord]>; } }
9.2 Querying with Consistency
Use query_with_consistency() to get both results and consistency metadata:
#![allow(unused)] fn main() { let (messages, consistency) = handler.query_with_consistency(&MessagesQuery::default()).await?; for msg in &messages { let status = if consistency.is_finalized(&msg.id) { "finalized" } else { "pending" }; println!("{}: {}", msg.content, status); } }
9.3 QueryStats with Consistency
QueryStats now includes a ConsistencyMap for tracking consistency of scanned facts:
#![allow(unused)] fn main() { let (result, stats) = handler.query_with_stats(&query).await?; if stats.consistency.any_finalized() { println!("Some results are finalized"); } }
9.4 Consistency Dimensions
Each Consistency entry tracks three orthogonal dimensions:
| Dimension | Type | Purpose |
|---|---|---|
| Agreement | Agreement | A1/A2/A3 finalization level |
| Propagation | Propagation | Gossip/sync status to peers |
| Acknowledgment | Acknowledgment | Per-peer delivery confirmation |
See Operation Categories for full details on these types and their usage.
10. Implementation Location
| Component | Location |
|---|---|
Query trait | aura-core/src/query.rs |
QueryEffects trait | aura-core/src/effects/query.rs |
FactEffects trait | aura-core/src/effects/fact.rs |
QueryIsolation | aura-core/src/query.rs |
QueryStats | aura-core/src/query.rs |
MutationReceipt | aura-core/src/query.rs |
ConsensusId, FactId | aura-core/src/query.rs |
ConsistencyMap | aura-core/src/domain/consistency.rs |
Agreement, Propagation | aura-core/src/domain/ |
| Temporal types | aura-core/src/domain/temporal.rs |
ScopeId, Finality | aura-core/src/domain/temporal.rs |
FactOp, Transaction | aura-core/src/domain/temporal.rs |
TemporalQuery, TemporalPoint | aura-core/src/domain/temporal.rs |
AuraQuery wrapper | aura-effects/src/database/query.rs |
QueryHandler | aura-effects/src/query/handler.rs |
| Concrete queries | aura-app/src/queries/ |
IndexedJournalEffects | aura-core/src/effects/indexed.rs |
See Also
- Journal System - Fact storage, reduction, and flow budgets
- Operation Categories - Agreement, propagation, acknowledgment, and ceremony contract
- Authorization - Biscuit token evaluation
- Effect System - Effect implementation patterns
Social Architecture
This document defines Aura's social organization model using a digital urban metaphor. The system layers social organization, privacy, consent, and governance into three tiers: messages, homes, and neighborhoods.
1. Overview
1.1 Design Goals
The model produces human-scaled social structures with natural scarcity based on physical analogs. Organic community dynamics emerge from bottom-up governance. The design aligns with Aura's consent-based privacy guarantees and capability-based authorization.
1.2 Three-Tier Structure
Messages are communication contexts. Direct messages are private relational contexts. Home messages are semi-public messaging for home members and participants.
Homes are semi-public communities capped by storage constraints. Each home has a 10 MB total allocation. Members and participants allocate storage to participate.
Neighborhoods are authority-level collections of homes connected via 1-hop links and access policies. Homes allocate storage to neighborhood infrastructure.
1.3 Terminology
An authority (AuthorityId) is the cryptographic identity that holds capabilities and participates in consensus. A nickname is a local mapping from an authority to a human-understandable name. Each device maintains its own nickname mappings. There is no global username registry.
A nickname suggestion (nickname_suggestion) is metadata an authority optionally shares when connecting with someone. Users configure a default suggestion sent to all new connections. Users can share different suggestions with different people or opt out entirely.
1.4 Unified Naming Pattern
The codebase uses a consistent naming pattern across entities (contacts, devices, discovered peers). The EffectiveName trait in aura-app/src/views/naming.rs defines the resolution order:
- Local nickname (user-assigned override) if non-empty
- Shared nickname_suggestion (what entity wants to be called) if non-empty
- Fallback identifier (truncated authority/device ID)
This pattern ensures consistent display names across all UI surfaces while respecting both local preferences and shared suggestions.
2. Message Types
2.1 Direct Messages
Direct messages are small private relational contexts built on AMP. There is no hop-based expansion across homes. All participants must be explicitly added. New members do not receive historical message sync.
2.2 Home Messages
Home messages are semi-public messaging for home members and participants. They use the same AMP infrastructure as direct messages. When a new participant joins, current members send a window of recent messages.
Membership and participation are tied to home policy. Leaving the home revokes access. Multiple channels may exist per home for different purposes.
#![allow(unused)] fn main() { pub struct HomeMessage { home_id: HomeId, channel: String, content: Vec<u8>, author: AuthorityId, timestamp: TimeStamp, } }
The message structure identifies the home, channel, content, author, and timestamp. Historical sync is configurable, typically the last 500 messages.
3. Home Architecture
3.1 Home Structure
A home is an authority-scoped context with its own journal. The total storage allocation is 10 MB. Capability templates define limited, partial, full, participant, and moderator patterns. Local governance is encoded via policy facts.
#![allow(unused)] fn main() { pub struct Home { /// Unique identifier for this home pub home_id: HomeId, /// Total storage limit in bytes pub storage_limit: u64, /// Maximum number of participants pub max_participants: u8, /// Maximum number of neighborhoods this home can join pub neighborhood_limit: u8, /// Current participants (authority IDs) pub participants: Vec<AuthorityId>, /// Current moderators with their capabilities pub moderators: Vec<(AuthorityId, ModeratorCapabilities)>, /// Current storage budget tracking pub storage_budget: HomeStorageBudget, } }
The home structure contains the identifier, storage limit, configuration limits, participant list, moderator designation list with capabilities, and storage budget tracking.
3.2 Membership and Participation
Home participation derives from possessing capability bundles, meeting entry requirements defined by policy, and allocating participant-specific storage. In v1, each user belongs to exactly one home.
Joining a home follows a defined sequence. The authority requests capability. Home governance approves using local policy via Biscuit evaluation and consensus. The authority accepts the capability bundle and allocates storage. Historical home messages sync from current members.
3.3 Moderator Designation
Moderators are designated via governance decisions in the home. A moderator must also be a member. Moderator capability bundles include moderation, pin and unpin operations, and governance facilitation. Moderator designation is auditable because capability issuance is visible via relational facts.
4. Neighborhood Architecture
4.1 Neighborhood Structure
A neighborhood is an authority type linking multiple homes. It contains a combined pinned infrastructure pool equal to the number of homes times 1 MB. A 1-hop link graph connects homes. Access-level and inter-home policy logic define movement rules.
#![allow(unused)] fn main() { pub struct Neighborhood { /// Unique identifier for this neighborhood pub neighborhood_id: NeighborhoodId, /// Member homes pub member_homes: Vec<HomeId>, /// 1-hop links between homes pub one_hop_links: Vec<(HomeId, HomeId)>, } }
The neighborhood structure contains the identifier, member homes, and 1-hop link edges.
4.2 Home Membership
Homes allocate 1 MB of their budget per neighborhood joined. In v1, each home may join a maximum of 4 neighborhoods. This limits 1-hop graph complexity and effect delegation routing.
5. Position and Traversal
5.1 Discovery Layers
Discovery is represented through the DiscoveryLayer enum, which indicates the best strategy to reach a target based on social relationships:
#![allow(unused)] fn main() { pub enum DiscoveryLayer { /// No relationship with target - must use rendezvous/flooding discovery Rendezvous, /// We have neighborhood presence and can use traversal Neighborhood, /// Target is reachable via home-level relay Home, /// Target is personally known - we have a direct relationship Direct, } }
The discovery layer determines routing strategy and flow costs. Direct has minimal cost. Home relays through same-home participants. Neighborhood uses multi-hop traversal. Rendezvous requires global flooding and has the highest cost.
5.2 Movement Rules
Movement is possible when a Biscuit capability authorizes entry, neighborhood policy allows movement along a 1-hop link, and home policy or invitations allow deeper access levels. Movement does not replicate pinned data. Visitors operate on ephemeral local state.
Traversal does not reveal global identity. Only contextual identities within encountered homes are visible.
6. Storage Constraints
6.1 Block-Level Allocation
Homes have a fixed size of 10 MB total. Allocation depends on neighborhood participation.
| Neighborhoods | Allocation | Participant Storage | Shared Storage |
|---|---|---|---|
| 1 | 1.0 MB | 1.6 MB | 7.4 MB |
| 2 | 2.0 MB | 1.6 MB | 6.4 MB |
| 3 | 3.0 MB | 1.6 MB | 5.4 MB |
| 4 | 4.0 MB | 1.6 MB | 4.4 MB |
More neighborhood connections mean less local storage for home culture. This creates meaningful trade-offs.
6.2 Flow Budget Integration
Storage constraints are enforced via the flow budget system.
#![allow(unused)] fn main() { pub struct HomeFlowBudget { /// Home ID (typed identifier) pub home_id: HomeId, /// Current number of participants pub participant_count: u8, /// Storage used by participants (spent counter as fact) pub participant_storage_spent: u64, /// Number of neighborhoods joined pub neighborhood_count: u8, /// Total neighborhood allocations pub neighborhood_allocations: u64, /// Storage used by pinned content (spent counter as fact) pub pinned_storage_spent: u64, } }
The spent counters are persisted as journal facts. The count fields track current membership. Limits are derived at runtime from home policy and Biscuit capabilities. Participant storage limit is 1.6 MB for 8 participants at 200 KB each.
7. Fact Schema
7.1 Home Facts
Home facts enable Datalog queries.
home(home_id, created_at, storage_limit).
home_config(home_id, max_participants, neighborhood_limit).
participant(authority_id, home_id, joined_at, storage_allocated).
moderator(authority_id, home_id, designated_by, designated_at, capabilities).
pinned(content_hash, home_id, pinned_by, pinned_at, size_bytes).
These facts express home existence, configuration, participation, moderator designation, and pin state.
7.2 Neighborhood Facts
Neighborhood facts express neighborhood existence, home membership, 1-hop links, and access permissions.
neighborhood(neighborhood_id, created_at).
home_member(home_id, neighborhood_id, joined_at, allocated_storage).
one_hop_link(home_a, home_b, neighborhood_id).
access_allowed(from_home, to_home, capability_requirement).
7.3 Query Examples
Queries use Biscuit Datalog.
participants_of(Home) <- participant(Auth, Home, _, _).
visitable(Target) <-
participant(Me, Current, _, _),
one_hop_link(Current, Target, _),
access_allowed(Current, Target, Cap),
has_capability(Me, Cap).
The first query finds all participants of a home. The second finds homes a user can visit from their current position via 1-hop links.
8. IRC-Style Commands
8.1 User Commands
User commands are available to all participants.
| Command | Description | Capability |
|---|---|---|
/msg <user> <text> | Send private message | send_dm |
/me <action> | Send action | send_message |
/nick <name> | Update contact suggestion | update_contact |
/who | List participants | view_members |
/leave | Leave current context | leave_context |
8.2 Moderator Commands
Moderator commands require moderator capabilities, and moderators must be members.
| Command | Description | Capability |
|---|---|---|
/kick <user> | Remove from home | moderate:kick |
/ban <user> | Ban from home | moderate:ban |
/mute <user> | Silence user | moderate:mute |
/pin <msg> | Pin message | pin_content |
8.3 Command Execution
Commands execute through the guard chain.
flowchart LR
A[Parse Command] --> B[CapGuard];
B --> C[FlowGuard];
C --> D[JournalCoupler];
D --> E[TransportEffects];
The command is parsed into a structured type. CapGuard checks capability requirements. FlowGuard charges the moderation action budget. JournalCoupler commits the action fact. TransportEffects notifies affected parties.
9. Governance
9.1 Home Governance
Homes govern themselves through capability issuance, consensus-based decisions, member moderation designations, and moderation. Home governance uses Aura Consensus for irreversible or collective decisions.
9.2 Neighborhood Governance
Neighborhoods govern home admission, 1-hop graph maintenance, access rules, and shared civic norms. High-stakes actions use Aura Consensus.
10. Privacy Model
10.1 Contextual Identity
Identity in Aura is contextual and relational. Joining a home reveals a home-scoped identity. Leaving a home causes that contextual identity to disappear. Profile data shared inside a context stays local to that context.
10.2 Consent Model
Disclosure is consensual. The device issues a join request. Home governance approves using local policy. The authority accepts the capability bundle. This sequence ensures all participation is explicit.
11. V1 Constraints
For the initial release, the model is simplified with three constraints.
Each user is a member of exactly one home. This eliminates multi-membership complexity and allows core infrastructure to stabilize.
Each home has a maximum of 8 participants. This human-scale limit enables strong community bonds and manageable governance.
Each home may join a maximum of 4 neighborhoods. This limits 1-hop graph complexity and effect delegation routing overhead.
12. Infrastructure Roles
Homes and neighborhoods provide infrastructure services beyond social organization. The aura-social crate implements these roles through materialized views and relay selection.
12.1 Home Infrastructure
Homes provide data availability and relay services for members and participants:
- Data Replication: Home members and participants replicate pinned data across available devices. The
HomeAvailabilitytype coordinates replication factor and failover. - Message Relay: 0-hop relays serve as first-hop relays for unknown destinations. Social topology queries return currently reachable relays.
- Storage Coordination: The
StorageServiceenforces storage budgets per participant and tracks usage facts.
12.2 Neighborhood Infrastructure
Neighborhoods enable multi-hop routing and cross-home coordination:
- Descriptor Propagation: Neighborhood 1-hop links define descriptor propagation paths. Connected homes exchange routing information.
- Access Capabilities:
TraversalAllowedFactgrants movement between homes. Access-level limits constrain routing overhead. - Multi-Hop Relay: When home-level relay fails, neighborhood traversal provides alternate paths.
12.3 Progressive Discovery Layers
The aura-social crate implements a four-layer discovery model:
| Layer | Priority | Resources Required | Flow Cost |
|---|---|---|---|
| Direct | 0 | Known peer relationship | Minimal |
| Home | 1 | 0-hop relays available | Low |
| Neighborhood | 2 | 1-hop/2-hop routes | Medium |
| Rendezvous | 3 | Global flooding | High |
Discovery layer selection uses SocialTopology::discovery_layer():
#![allow(unused)] fn main() { let topology = SocialTopology::new(local_authority, home, neighborhoods); let layer = topology.discovery_layer(&target); }
Lower priority layers are preferred when available. This creates economic incentives to establish social relationships before communication.
12.4 Relay Selection
The SocialTopology provides relay candidate generation:
#![allow(unused)] fn main() { let topology = SocialTopology::new(local_authority, home, neighborhoods); let candidates = topology.build_relay_candidates(&destination, |peer| is_reachable(peer)); }
Candidates are returned in priority order: same-home relays first, then 1-hop and 2-hop neighborhood relays, then guardians. Reachability checks filter unreachable peers. Each candidate includes relay-relationship metadata indicating same-home, neighborhood-hop, or guardian routing.
See Also
Database Architecture describes fact storage and queries. Transport and Information Flow covers AMP messaging. Authorization describes capability evaluation. Rendezvous Architecture details the four-layer discovery model integration.
Distributed Maintenance Architecture
This document describes distributed maintenance in Aura. It explains snapshots, garbage collection, cache invalidation, OTA upgrades, admin replacement, epoch handling, and backup procedures. All maintenance operations align with the authority and relational context model. All maintenance operations insert facts into appropriate journals. All replicas converge through join-semilattice rules.
1. Maintenance Facts
Maintenance uses facts stored in an authority journal. Facts represent monotone knowledge. Maintenance logic evaluates local predicates over accumulated facts. These predicates implement constraints such as GC eligibility or upgrade readiness. The authoritative schema lives in crates/aura-maintenance/src/facts.rs.
#![allow(unused)] fn main() { pub enum MaintenanceFact { SnapshotProposed(SnapshotProposed), SnapshotCompleted(SnapshotCompleted), CacheInvalidated(CacheInvalidated), AdminReplacement(AdminReplacement), ReleaseDistribution(ReleaseDistributionFact), ReleasePolicy(ReleasePolicyFact), UpgradeExecution(UpgradeExecutionFact), } }
This fact model defines snapshot, cache, release-distribution, policy-publication, and scoped upgrade-execution events. Each fact is immutable and merges by set union. Devices reduce maintenance facts with deterministic rules.
2. Snapshots and Garbage Collection
Snapshots bound storage size. A snapshot proposal announces a target epoch and a digest of the journal prefix. Devices verify the digest. If valid, they contribute signatures. A threshold signature completes the snapshot.
Snapshot completion inserts a SnapshotCompleted fact. Devices prune facts whose epochs fall below the snapshot epoch. Devices prune blobs whose retractions precede the snapshot. This pruning does not affect correctness because the snapshot represents a complete prefix.
DKG transcript blobs follow the same garbage collection fence: once a snapshot is finalized, transcripts with epochs older than the snapshot retention window may be deleted. This keeps long-lived key ceremonies from accumulating unbounded storage while preserving the ability to replay from the latest snapshot.
#![allow(unused)] fn main() { pub struct Snapshot { pub epoch: Epoch, pub commitment: TreeHash32, pub roster: Vec<LeafId>, pub policies: BTreeMap<NodeIndex, Policy>, pub state_cid: Option<TreeHash32>, pub timestamp: u64, pub version: u8, } }
This structure defines the snapshot type from aura_core::tree. Devices fetch the blob when restoring state. Devices hydrate journal state and replay the tail of post-snapshot facts.
3. Cache Invalidation
State mutations publish CacheInvalidated facts. A cache invalidation fact contains cache keys and an epoch floor. Devices maintain local maps from keys to epoch floors. A cache entry is valid only when the current epoch exceeds its floor.
Cache invalidation is local. No CRDT cache is replicated. Devices compute validity using meet predicates on epoch constraints.
#![allow(unused)] fn main() { pub struct CacheKey(pub String); }
This structure identifies a cached entry. Devices invalidate cached data when they observe newer invalidation facts.
4. OTA Upgrades
OTA in Aura separates two concerns:
- global and eventual release distribution
- local or scope-bound staging, activation, cutover, and rollback
Aura does not model "the whole network is now in cutover" as a valid primitive. Release propagation is multi-directional and eventual. Hard cutover is meaningful only inside a scope that actually has agreement or a legitimate fence.
4.1 Release Identity and Provenance
#![allow(unused)] fn main() { pub struct AuraReleaseProvenance { pub source_repo_url: String, pub source_bundle_hash: Hash32, pub build_recipe_hash: Hash32, pub output_hash: Hash32, pub nix_flake_hash: Hash32, pub nix_flake_lock_hash: Hash32, } pub struct AuraReleaseManifest { pub series_id: AuraReleaseSeriesId, pub release_id: AuraReleaseId, pub version: SemanticVersion, pub provenance: AuraReleaseProvenance, pub artifacts: Vec<AuraArtifactDescriptor>, pub compatibility: AuraCompatibilityManifest, pub suggested_activation_time_unix_ms: Option<u64>, } }
AuraReleaseId is derived from the release series and the full provenance. source_repo_url participates in that derivation, so the declared upstream repository location is part of canonical release identity. Builder authorities may publish deterministic build certificates over the same provenance. TEE attestation is optional hardening, not the source of release identity.
4.2 Policy Surfaces
OTA policy is not one switch. Aura distinguishes:
- discovery policy: what release authorities, builders, and contexts a device is willing to learn from
- sharing policy: what manifests, artifacts, certificates, or recommendations it is willing to forward or pin
- activation policy: what trust, compatibility, health, approval, and fence conditions must hold before local activation
Discovering a release does not imply forwarding it. Forwarding it does not imply activating it.
4.3 Activation Scopes and State
Activation is modeled per scope, not globally.
#![allow(unused)] fn main() { pub enum AuraActivationScope { DeviceLocal { device_id: DeviceId }, AuthorityLocal { authority_id: AuthorityId }, RelationalContext { context_id: ContextId }, ManagedQuorum { context_id: ContextId, participants: BTreeSet<AuthorityId>, }, } pub enum ReleaseResidency { LegacyOnly, Coexisting, TargetOnly, } pub enum TransitionState { Idle, AwaitingCutover, CuttingOver, RollingBack, } }
ReleaseResidency describes which release set may currently run in the scope. TransitionState describes whether the scope is stable, waiting on evidence, actively switching, or rolling back.
4.4 Cutover and Rollback
Scoped activation uses journal facts plus local policy evaluation. A scope may move toward cutover only when the relevant evidence is present:
- manifest and certificate verification
- compatibility classification
- staged artifacts
- local trust policy satisfaction
- optional local-policy respect for
suggested_activation_time_unix_ms - threshold approval, if that scope actually supports threshold approval
- epoch fence, if that scope actually owns the relevant fence
- health gate checks
Hard-fork behavior is explicit. After local cutover, incompatible new sessions are rejected. In-flight incompatible sessions must drain, abort, or delegate according to policy. If post-cutover validation fails, rollback is deterministic and recorded in UpgradeExecutionFact.
Managed quorum cutover requires explicit approval from the participant set bound into AuraActivationScope::ManagedQuorum. Staged revoked releases are canceled before cutover. Active revoked releases follow the local rollback preference. Automatic rollback queues the revert path immediately. Manual rollback leaves the scope failed until an operator approves rollback.
4.5 Updater / Launcher Boundary
Aura does not rely on in-place self-replacement of the running runtime. Layer 6 owns an updater/launcher control plane that:
- stages manifests, artifacts, and certificates
- emits explicit activate/rollback commands
- records scoped upgrade state
- restores the previous release deterministically when rollback is required
5. Admin Replacement
Admin replacement uses a maintenance fact. The fact records the old admin, new admin, and activation epoch. Devices use this fact to ignore operations from retired administrators.
#![allow(unused)] fn main() { pub struct AdminReplacement { pub authority_id: AuthorityId, pub old_admin: AuthorityId, pub new_admin: AuthorityId, pub activation_epoch: Epoch, } }
This structure defines an admin replacement. Devices enforce this rule locally. The replacement fact is monotone and resolves disputes using journal evidence.
6. Epoch Handling
Maintenance logic uses identity epochs for consistency. A maintenance session uses a tuple containing the identity epoch and snapshot epoch. A session aborts if the identity epoch advances. Devices retry under the new epoch.
Snapshot completion sets the snapshot epoch equal to the identity epoch. Garbage collection rules use the snapshot epoch to prune data safely. Upgrade fences use the same epoch model to enforce activation.
#![allow(unused)] fn main() { use aura_maintenance::MaintenanceEpoch; pub struct MaintenanceEpoch { pub identity_epoch: Epoch, pub snapshot_epoch: Epoch, } }
This structure captures epoch state for maintenance workflows. Devices use this structure for guard checks.
7. Backup and Restore
Backup uses the latest snapshot and recent journal facts. Devices export an encrypted archive containing the snapshot blob and journal tail. Restore verifies the snapshot signature, hydrates state, and replays the journal tail.
Backups use existing storage and verification effects. No separate protocol exists. Backup correctness follows from snapshot correctness.
8. Automatic Synchronization
Automatic synchronization implements periodic journal replication between devices. The synchronization service coordinates peer discovery, session management, and fact exchange. All synchronization uses the journal primitives described in Journal.
8.1 Peer Discovery and Selection
Devices discover sync peers through the rendezvous system described in Rendezvous. The peer manager maintains metadata for each discovered peer. This metadata includes connection state, trust level, sync success rate, and active session count.
#![allow(unused)] fn main() { pub struct PeerMetadata { pub device_id: DeviceId, pub status: PeerStatus, pub discovered_at: PhysicalTime, pub last_status_change: PhysicalTime, pub successful_syncs: u64, pub failed_syncs: u64, pub average_latency_ms: u64, pub last_seen: PhysicalTime, pub last_successful_sync: PhysicalTime, pub trust_level: u8, pub has_sync_capability: bool, pub active_sessions: usize, } }
This structure tracks peer state for selection decisions. All timestamp fields use PhysicalTime from the unified time system. The peer manager calculates a score for each peer using weighted factors. Trust level contributes 50 percent. Success rate contributes 30 percent. Load factor contributes 20 percent. Higher scores indicate better candidates for synchronization.
Devices select peers when their score exceeds a threshold. Devices limit concurrent sessions per peer. This prevents resource exhaustion. Devices skip peers that have reached their session limit.
8.2 Session Management
The session manager tracks active synchronization sessions. Each session has a unique identifier and references a peer device. Sessions enforce rate limits and concurrency bounds.
#![allow(unused)] fn main() { pub struct SessionManager<T> { sessions: HashMap<SessionId, SessionState<T>>, config: SessionConfig, metrics: Option<MetricsCollector>, last_cleanup: PhysicalTime, session_counter: u64, } }
This structure maintains session state. Sessions are indexed by SessionId rather than DeviceId. Configuration is provided via SessionConfig. All timestamp fields use PhysicalTime from the unified time system. Devices close sessions after fact exchange completes. Devices abort sessions when the identity epoch advances. Session cleanup releases resources for new synchronization rounds.
8.3 Rate Limiting and Metrics
The synchronization service enforces rate limits per peer and globally. Rate limiting prevents network saturation. Metrics track sync latency, throughput, and error rates.
Devices record metrics for each sync operation. These metrics include fact count, byte count, and duration. Devices aggregate metrics to monitor service health. Degraded peers receive lower priority in future rounds.
8.4 Integration with Journal Effects
Automatic synchronization uses JournalEffects to read and write facts. The service queries local journals for recent facts. The service sends these facts to peers. Peers merge incoming facts using join-semilattice rules.
All fact validation rules apply during automatic sync. Devices reject invalid facts. Devices do not rollback valid facts already merged. This maintains journal monotonicity.
9. Migration Infrastructure
The MigrationCoordinator in aura-agent/src/runtime/migration.rs orchestrates data migrations between protocol versions.
9.1 Migration Trait
#![allow(unused)] fn main() { #[async_trait] pub trait Migration: Send + Sync { fn source_version(&self) -> SemanticVersion; fn target_version(&self) -> SemanticVersion; fn name(&self) -> &str; async fn validate(&self, ctx: &MigrationContext) -> Result<(), MigrationError>; async fn execute(&self, ctx: &MigrationContext) -> Result<(), MigrationError>; async fn rollback(&self, ctx: &MigrationContext) -> Result<bool, MigrationError> { Ok(false) // Default: rollback not supported } } }
Each migration specifies source and target versions, a name for logging, validate/execute methods, and an optional rollback method. The default rollback implementation returns Ok(false) to indicate rollback is not supported.
9.2 Coordinator API
| Method | Purpose |
|---|---|
needs_migration(from) | Check if upgrade is needed |
get_migration_path(from, to) | Find ordered migration sequence |
migrate(from, to) | Execute migrations with validation |
validate_migration(from, to) | Dry-run validation only |
9.3 Migration Guarantees
Migrations are ordered by target version. Each migration runs at most once (idempotent via version tracking). Failed migrations leave the system in a consistent state. Progress is recorded in the journal for auditability.
10. Evolution
Maintenance evolves in phases. Current OTA work focuses on release identity/provenance, scoped activation, deterministic rollback, and Aura-native distribution. Future phases may add richer replicated cache CRDTs, stronger builder attestation, staged rollout tooling, and automatic snapshot triggers.
Future phases build on the same journal schema. Maintenance semantics remain compatible with older releases.
11. Summary
Distributed maintenance uses journal facts to coordinate snapshots, cache invalidation, release distribution, scoped upgrades, and admin replacement. All operations use join-semilattice semantics. All reductions are deterministic. Devices prune storage only after observing snapshot completion. OTA release propagation is eventual, while activation is always local or scope-bound. The system remains consistent across offline and online operation.
CLI and Terminal User Interface
This document describes the aura-terminal user interface layer. It covers the non-interactive CLI commands and the iocraft-based TUI. It also describes how both frontends share aura-app through AppCore and the reactive signal system.
Demo mode is compiled only with --features development.
Goals and constraints
The CLI and the TUI are thin frontends over AppCore. They should not become alternate application runtimes. They should also avoid owning long-lived domain state.
Both frontends must respect the guard chain, journaling, and effect system boundaries described in Aura System Architecture. The CLI optimizes for scriptability and stable output. The TUI optimizes for deterministic navigation with reactive domain data.
Concepts
- Command. A user-invoked operation such as
aura statusoraura chat send. - Handler. A CLI implementation function that uses
HandlerContextand returnsCliOutput. - Screen. A routed view that renders domain data and local UI state.
- Modal. A blocking overlay that captures focus. Modals are queued and only one can be visible at a time.
- Toast. A transient notification. Toasts are queued and only one can be visible at a time.
- Signal. A reactive stream of domain values from
aura-app. - Intent. A journaled application command dispatched through
AppCore.dispatch(legacy in the TUI). Most TUI actions are runtime-backed workflows viaIoContext. - Role/access labels. UI copy should use canonical terms from Theoretical Model:
Member,Participant,Moderator, and access levelsFull/Partial/Limited.
Running
Run the CLI and the TUI from the development shell.
CLI
The CLI entry point is crates/aura-terminal/src/main.rs. Commands are parsed by bpaf and dispatched to CliHandler and the handler modules.
nix develop
aura status -c ./aura.toml
aura chat list
This enters the Nix development environment and runs two example commands. CLI commands return structured CliOutput and then render it to stdout and stderr.
TUI production
The production TUI command is aura tui.
nix develop
aura tui
This launches the production TUI and enters fullscreen mode.
TUI demo
To run the demo, compile with --features development and pass --demo.
nix develop
cargo run -p aura-terminal --features development -- aura tui --demo
This launches a deterministic demo environment with simulated peers.
Useful flags and environment variables
CLI:
aura -venables verbose output and logsaura -c CONFIGselects a global config file
TUI:
aura tui --data-dir DIRsets the Aura data directory. It falls back toAURA_PATH.aura tui --device-id DEVICEselects the device identifier for the session.AURA_TUI_ALLOW_STDIO=1disables fullscreen stderr redirection. It is intended for debugging.AURA_TUI_LOG_PATHoverrides the TUI log file location.
Architecture overview
The CLI and the TUI share the same backend boundary. Both construct an AppCore value and use it as the primary interface to domain workflows and views. Both also rely on aura-agent for effect handlers and runtime services.
The user interface split is:
crates/aura-appprovides portable domain logic, reactive state, and signals throughAppCore.crates/aura-terminal/src/cli/defines bpaf parsers and CLI argument types.crates/aura-terminal/src/handlers/implements CLI commands and shared terminal glue.crates/aura-terminal/src/tui/implements iocraft UI code and deterministic navigation.
Relationship to aura-app
AppCore is the shared boundary for both frontends. It owns reactive state and provides stable APIs for dispatching intents and reading derived views. Frontends should use the aura_app::ui facade (especially aura_app::ui::signals and aura_app::ui::workflows) as the public API surface.
The frontends use AppCore in two ways:
- Trigger work by calling
AppCore.dispatch(intent)or by calling effect-backed handlers that ultimately produce journaled facts. - Read state by reading views or by subscribing to signals for push-based updates.
This split keeps domain semantics centralized. It also makes it possible to reuse the same workflows in multiple user interfaces.
Time system in UI
UI code must never read OS clocks (for example, SystemTime::now() or Instant::now()). All wall-clock needs must flow through algebraic effects (PhysicalTimeEffects via the handler/effect system). Demo mode and relative-time UI (e.g., “Synced Xm ago”) must be driven by runtime time so simulations remain deterministic.
Aura time domains are: PhysicalClock (wall time), LogicalClock (causality), OrderClock (privacy-preserving ordering), and Range (validity windows). When attested time is required, use ProvenancedTime/TimeComparison rather than embedding OS timestamps in UI state.
Ordering across domains must be explicit: use TimeStamp::compare(policy) (never compare raw ms) when you need deterministic ordering across mixed time domains.
Shared infrastructure in aura-terminal
CLI commands are parsed in crates/aura-terminal/src/cli/commands.rs and routed through crates/aura-terminal/src/main.rs. Implementations live under crates/aura-terminal/src/handlers/ and use HandlerContext from crates/aura-terminal/src/handlers/handler_context.rs. Handler functions typically return CliOutput for testable rendering.
The TUI launcher lives in crates/aura-terminal/src/handlers/tui.rs. It sets up tracing and constructs the IoContext and callback registry. The fullscreen stdio policy is defined in crates/aura-terminal/src/handlers/tui_stdio.rs.
CLI execution model
The CLI is request and response. Each command parses arguments, runs one handler, and exits. Long-running commands such as daemon modes should still use the same effect boundaries.
flowchart TD A[CLI args] --> P[bpaf parser] P --> H[CliHandler] H --> HC[HandlerContext] HC --> E[Effects and services] E --> AC[AppCore state] AC --> O[CliOutput render]
This diagram shows the main CLI path from parsing to rendering. Some handlers also read derived state from AppCore after effects complete.
Reactive data model
The reactive system follows a fact-based architecture where typed facts are the source of truth for UI state.
Signals as single source of truth
ReactiveEffects signals (CHAT_SIGNAL, CONTACTS_SIGNAL, RECOVERY_SIGNAL, etc.) are the canonical source for all UI state. They are updated by the ReactiveScheduler processing typed facts from the journal.
flowchart LR F[Typed Facts] -->|commit| J[Journal] J -->|publish| S[ReactiveScheduler] S -->|emit| Sig[Signals] Sig --> TUI[TUI Screens] Sig --> CLI[CLI Commands]
The data flow is:
- Typed facts (
aura_journal::fact::Fact) are committed to the journal - ReactiveScheduler processes committed facts via registered
ReactiveViewimplementations - SignalViews (e.g.,
ContactsSignalView,ChatSignalView) update their internal state and emit snapshots to signals - UI components subscribe to signals and render the current state
This architecture ensures a single source of truth and eliminates dual-write bugs. Code that needs to update UI state must commit facts (production) or emit directly to signals (demo/test).
ViewState is internal
AppCore contains an internal ViewState used for legacy compatibility and non-signal use cases. ViewState changes do not propagate to signals. The signal forwarding infrastructure was removed in favor of scheduler-driven updates.
For compile-time safety, there are no public methods on AppCore to mutate ViewState for UI-affecting state. Code that needs to update what the UI displays must:
- Production: Commit facts via
RuntimeBridge.commit_relational_facts(). Facts flow through the scheduler to signals. - Demo/Test: Emit directly to signals via
ReactiveEffects::emit(). This is explicit and type-safe.
This design prevents the "dual-write" bug class where code updates ViewState expecting UI changes, but signals remain unchanged.
The CLI usually reads state at a point in time. It can still use signals for watch-like commands or daemon commands. When a command needs continuous updates, it should subscribe to the relevant signals and render incremental output.
Reading and subscribing to signals
Signals are accessed through AppCore's ReactiveEffects implementation. Read the current value with read() and subscribe for updates with subscribe().
#![allow(unused)] fn main() { // Read current state from signal let contacts = { let core = app_core.read().await; core.read(&*CONTACTS_SIGNAL).await.unwrap_or_default() }; // Subscribe for ongoing updates let mut stream = { let core = app_core.read().await; core.subscribe(&*CONTACTS_SIGNAL) }; while let Ok(state) = stream.recv().await { render_contacts(&state); } }
For initial render, read the current signal value first to avoid a blank frame. Then subscribe for updates. This pattern is used heavily by TUI screens.
Subscriptions and ownership
Long-lived subscriptions that drive global TUI elements live in crates/aura-terminal/src/tui/screens/app/subscriptions.rs. Screen-local subscriptions should live with the screen module.
Subscriptions should be owned by the component that renders the data. A subscription should not mutate TuiState unless it is updating navigation, focus, or overlay state.
Connection status (peer count)
The footer "connected peers" count is a UI convenience signal. It must represent how many of your contacts are online, not a seeded or configured peer list.
- Source:
CONNECTION_STATUS_SIGNAL(emitted byaura_app::ui::workflows::system::refresh_account()). - Contact set: read from
CONTACTS_SIGNAL(signal truth), not fromViewStatesnapshots. - Online check:
RuntimeBridge::is_peer_online(contact_id)(best-effort). Demo uses a shared in-memory transport. Production can use real transport channel health.
Deterministic UI model
The TUI separates domain state from UI state. Domain state is push-based and comes from aura-app signals. UI state is deterministic and lives in TuiState.
Navigation, focus, input buffers, modal queues, and toast queues are updated by a pure transition function. The entry point is crates/aura-terminal/src/tui/state/mod.rs. The runtime executes TuiCommand values in crates/aura-terminal/src/tui/runtime.rs.
Dispatch bridge
The TUI dispatch path uses IoContext in crates/aura-terminal/src/tui/context/io_context.rs.
Today the TUI uses a runtime-backed dispatch model: IoContext routes EffectCommand to DispatchHelper and OperationalHandler. Most domain-affecting behavior occurs through aura-app workflows that call RuntimeBridge (which commits facts and drives signals).
flowchart TD UI[User input] --> SM[TuiState transition] SM -->|TuiCommand::Dispatch| IO[IoContext dispatch] IO --> DH[DispatchHelper] DH --> OP[OperationalHandler] OP --> W[Workflows / RuntimeBridge] W --> J[Commit facts] J --> S[Emit signals] S --> UI2[Screens subscribe]
This diagram shows the primary TUI dispatch path. Operational commands may also emit operational signals such as SYNC_STATUS_SIGNAL, CONNECTION_STATUS_SIGNAL, and ERROR_SIGNAL.
Screens, modals, and callbacks
The root iocraft component is in crates/aura-terminal/src/tui/screens/app/shell.rs. Global modals live in crates/aura-terminal/src/tui/screens/app/modal_overlays.rs. Long-lived signal subscriptions for the shell live in crates/aura-terminal/src/tui/screens/app/subscriptions.rs.
Modals and toasts are routed through explicit queues in TuiState. The modal enum is QueuedModal in crates/aura-terminal/src/tui/state/modal_queue.rs. Avoid per-modal visible flags.
Invitation codes are managed from the Contacts workflow (modals), not via a dedicated routed Invitations screen.
Callbacks are registered in crates/aura-terminal/src/tui/callbacks/. Asynchronous results are surfaced through UiUpdate in crates/aura-terminal/src/tui/updates.rs. Prefer subscribing to domain signals when a signal already exists.
Fullscreen I/O policy
Writing to stderr while iocraft is in fullscreen can corrupt the terminal buffer. The TUI redirects stderr away from the terminal while fullscreen is active. Tracing is written to a log file.
The policy is enforced with type-level stdio tokens in crates/aura-terminal/src/handlers/tui_stdio.rs. The token used before fullscreen is consumed while iocraft is running. This prevents accidental println! and eprintln! calls in the fullscreen scope.
This policy aligns with Privacy and Information Flow and Effect System.
Errors and user feedback
Domain and dispatch failures are emitted through aura_app::ui::signals::ERROR_SIGNAL. The app shell subscribes to this signal and renders errors as queued toasts. When the account setup modal is active, errors are routed into the modal instead of creating a toast.
UI-only failures use UiUpdate::OperationFailed. This is used primarily for account file operations that occur before AppCore dispatch.
CLI commands should return errors through TerminalResult and render them through CliOutput. Avoid printing error text directly from deep helper functions. Prefer returning structured error types.
Invariants and common pitfalls
The state machine owns navigation, focus, and overlay visibility. Screen components should render TuiState and should not mutate it directly. They should send events and let the state machine decide transitions.
The domain owns the reactive state. Avoid caching domain data in TuiState. Prefer subscribing to aura-app signals and deriving view props inside the screen component.
Single source of truth invariants
- Signals are the source of truth for UI state, not ViewState.
- Facts drive signals in production. Commit facts via RuntimeBridge.
- Direct emission is only for demo/test scenarios via
ReactiveEffects::emit(). - No ViewState mutation for UI state. AppCore has no public methods to mutate ViewState for UI-affecting state.
Common pitfalls
- Calling
println!andeprintln!while fullscreen is active - Storing domain state in
TuiStateinstead of subscribing to signals - Adding per-modal
visibleflags instead of usingQueuedModaland the modal queue - Using
UiUpdateas a general event bus instead of subscribing to signals - Expecting ViewState changes to appear in the UI. ViewState does not propagate to signals.
- Emitting directly to domain signals in production code. Use fact commits instead.
Testing strategy
The CLI should be tested with handler unit tests and structured output assertions. Prefer pure formatting helpers and CliOutput snapshots over stdout capture.
The deterministic boundary for the TUI is the state machine. Prefer unit tests that call transition() directly for navigation and modal behavior. For headless terminal event tests, use TuiRuntime<T> from crates/aura-terminal/src/tui/runtime.rs with a mock TerminalEffects handler.
Code map
crates/aura-terminal/src/
main.rs
cli/
commands.rs
handlers/
mod.rs
handler_context.rs
tui.rs
tui_stdio.rs
tui/
context/
screens/
app/
shell.rs
modal_overlays.rs
subscriptions.rs
state/
runtime.rs
hooks.rs
effects/
components/
This map shows the primary module boundaries for the CLI and the TUI. CLI logic should live under handlers/ and cli/. TUI view logic should live under tui/.
Demo mode
Demo mode is under crates/aura-terminal/src/demo/. It compiles only with --features development. Production builds should not require demo-only types or props.
Demo architecture
Demo mode uses the same fact-based pipeline as production where possible:
- Guardian bindings: Committed as
RelationalFact::GuardianBindingfacts throughRuntimeBridge.commit_relational_facts(). These flow through the scheduler to updateCONTACTS_SIGNAL. - Chat messages: Emitted directly to
CHAT_SIGNALviaReactiveEffects::emit(). Sealed message facts would require cryptographic infrastructure not available in demo. - Recovery approvals: Emitted directly to
RECOVERY_SIGNAL. Production would use consensus-basedRecoveryGrantfacts.
The DemoSignalCoordinator in crates/aura-terminal/src/demo/signal_coordinator.rs handles bidirectional event routing between the TUI and simulated agents (Alice and Carol).
Demo shortcuts
Demo mode supports convenience shortcuts. Invite code entry supports Ctrl+a and Ctrl+l when demo codes are present.
Recovery scenario walkthrough
The CLI can run a complete guardian recovery demo through the simulator. This scenario shows Bob onboarding with guardians, losing his device, and recovering with help from Alice and Carol.
Run the recovery demo from the repository root:
cargo run -p aura-terminal -- scenarios run --directory scenarios/integration --pattern cli_recovery_demo
This command uses the CLI scenario runner plus the simulator to execute the guardian setup and recovery choreography. Logs are written to work/scenario_logs/cli_recovery_demo.log.
The scenario executes in eight phases:
alice_carol_setupcreates Alice and Carol authorities using Ed25519 single-signer mode for initial key generation.bob_onboardingcreates Bob authority (Ed25519 initially), sends guardian requests, and configures threshold 2 (switching to FROST).group_chat_setuphas Alice create a group chat and invite Bob and Carol. Context keys are derived for the chat.group_messagingsends normal chat messages among all three participants. History is persisted.bob_account_losssimulates total device loss for Bob. He cannot access his authority.recovery_initiationhas Bob initiate recovery. Alice and Carol validate and approve the request. Guardian approval threshold (2) is met.account_restorationruns threshold key recovery. Bob's chat history is synchronized back.post_recovery_messaginghas Bob send messages again and see full history. The group remains functional.
New accounts use standard Ed25519 signatures because FROST requires at least 2 signers. Once Bob adds guardians and configures threshold 2, subsequent signing operations use the full FROST threshold protocol. See Cryptographic Architecture for details on signing modes.
Testing
Run tests inside the development shell.
Standard
just test-crate aura-terminal
This runs the aura-terminal test suite in the standard project workflow.
Offline
For offline testing, use the workspace offline mode.
CARGO_NET_OFFLINE=true cargo test -p aura-terminal --tests --offline
This runs the full aura-terminal test suite without network access.
Message status indicators
Chat messages display delivery status and finalization indicators in the message bubble header. Status is derived from the unified consistency metadata types in aura_core::domain.
Status indicator legend
Symbol Meaning Color Animation Source
────────────────────────────────────────────────────────────────
◐ Syncing/Sending Blue Pulsing Propagation::Local
◌ Pending Gray None Agreement::Provisional
✓ Sent Green None Propagation::Complete
✓✓ Delivered Green None Acknowledgment.acked_by includes recipient
✓✓ Read Blue None ChatFact::MessageRead
⚠ Unconfirmed Yellow None Agreement::SoftSafe (pending A3)
✗ Failed Red None Propagation::Failed
◆ Finalized Muted None Agreement::Finalized
Delivery status lifecycle
Messages progress through the following states:
| Status | Icon | Meaning |
|---|---|---|
| Sending | ◐ | Message being transmitted (Propagation::Local) |
| Sent | ✓ | Synced to all known peers (Propagation::Complete) |
| Delivered | ✓✓ | Recipient acked via transport protocol |
| Read | ✓✓ (blue) | Recipient viewed message (ChatFact::MessageRead) |
| Failed | ✗ | Sync failed (Propagation::Failed with retry) |
Delivery status is derived from OptimisticStatus consistency metadata:
#![allow(unused)] fn main() { use aura_core::domain::{Propagation, Acknowledgment, OptimisticStatus}; fn delivery_icon(status: &OptimisticStatus, expected_peers: &[AuthorityId]) -> &'static str { match &status.propagation { Propagation::Local => "◐", Propagation::Syncing { .. } => "◐", Propagation::Failed { .. } => "✗", Propagation::Complete => { // Check if all expected peers have acked let delivered = status.acknowledgment .as_ref() .map(|ack| expected_peers.iter().all(|p| ack.contains(p))) .unwrap_or(false); if delivered { "✓✓" } else { "✓" } } } } }
The transport layer implements ack tracking via ack_tracked on facts. When ack_tracked = true, recipients send FactAck responses that are recorded in the journal's ack table. Read receipts are semantic (user viewed) and use ChatFact::MessageRead.
Agreement and finalization
Agreement level affects display:
| Agreement | Display | Meaning |
|---|---|---|
| Provisional (A1) | Normal | Usable but may change |
| SoftSafe (A2) | ⚠ badge | Coordinator-safe with convergence cert |
| Finalized (A3) | ◆ badge | Consensus-finalized, durable |
The finalization indicator (◆) appears when a message achieves A3 consensus (2f+1 witnesses). This indicates the message is durably committed and cannot be rolled back.
Implementation notes
Status indicators are rendered in MessageBubble (crates/aura-terminal/src/tui/components/message_bubble.rs). The consistency metadata flows from ChatState through the CHAT_SIGNAL to the TUI. The ChatState includes OptimisticStatus for each message, which contains:
agreement: Current A1/A2/A3 levelpropagation: Sync status (Local, Syncing, Complete, Failed)acknowledgment: Which peers have acked (for delivery tracking)
See Operation Categories for the full consistency metadata type definitions and AMP Protocol for the acknowledgment flow.
See also
- Aura System Architecture
- Privacy and Information Flow
- Effect System
- Runtime
- AMP Protocol for message delivery acknowledgments
- Operation Categories for status tracking patterns
Test Infrastructure Reference
This document describes the architecture of aura-testkit, the test infrastructure crate that provides fixtures, mock handlers, and verification utilities for testing Aura protocols.
Overview
The aura-testkit crate occupies Layer 8 in the Aura architecture. It provides reusable test infrastructure without containing production code. All test utilities follow effect system guidelines to ensure deterministic execution.
The crate serves three purposes. It provides stateful effect handlers for controllable test environments. It offers fixture builders for consistent test setup. It includes verification utilities for property testing and differential testing.
Stateful Effect Handlers
Stateful effect handlers maintain internal state across calls. They enable deterministic testing by controlling time, randomness, and storage. These handlers implement the same traits as production handlers but store state for inspection and manipulation.
Handler Categories
The stateful_effects module provides handlers for each effect category.
| Handler | Effect Trait | Purpose |
|---|---|---|
SimulatedTimeHandler | PhysicalTimeEffects | Controllable simulated time |
MockRandomHandler | RandomCoreEffects | Seeded deterministic randomness |
MemoryStorageHandler | StorageEffects | In-memory storage with inspection |
MockJournalHandler | JournalEffects | Journal with fact tracking |
MockCryptoHandler | CryptoCoreEffects | Crypto with key inspection |
MockConsoleHandler | ConsoleEffects | Captured console output |
Time Handler
The SimulatedTimeHandler provides controllable time for tests.
#![allow(unused)] fn main() { use aura_testkit::stateful_effects::SimulatedTimeHandler; use aura_core::effects::PhysicalTimeEffects; let time = SimulatedTimeHandler::new(); let now = time.physical_time().await?; time.advance_time(5000); let later = time.physical_time().await?; }
This handler starts at the current system time by default. Use SimulatedTimeHandler::new_at_epoch() for tests starting at Unix epoch, or SimulatedTimeHandler::new_with_time(start_ms) for a specific start time. Tests can verify time-dependent behavior without wall-clock delays.
Random Handler
The MockRandomHandler provides seeded randomness for reproducible tests.
#![allow(unused)] fn main() { use aura_testkit::stateful_effects::MockRandomHandler; let random = MockRandomHandler::with_seed(42); let bytes = random.random_bytes(32).await; }
Given the same seed, this handler produces identical sequences across runs. This enables deterministic property testing and failure reproduction.
Fixture System
The fixture system provides consistent test environment setup. Fixtures encapsulate common configuration patterns and reduce boilerplate.
TestFixture
The TestFixture type provides a complete test environment.
#![allow(unused)] fn main() { use aura_testkit::infrastructure::harness::{TestFixture, TestConfig}; let fixture = TestFixture::new().await?; let device_id = fixture.device_id(); let context = fixture.context(); }
A fixture creates deterministic identifiers, initializes effect handlers, and provides access to test context. The default configuration suits most unit tests.
TestConfig
Custom configurations enable specialized test scenarios.
#![allow(unused)] fn main() { let config = TestConfig { name: "threshold_test".to_string(), deterministic_time: true, capture_effects: true, timeout: Some(Duration::from_secs(60)), }; let fixture = TestFixture::with_config(config).await?; }
The deterministic_time flag enables StatefulTimeHandler. The capture_effects flag records effect calls for later inspection.
Builder Utilities
Builder functions create test data with deterministic inputs. They live in the builders module.
Account Builders
#![allow(unused)] fn main() { use aura_testkit::builders::test_account_with_seed; let account = test_account_with_seed(42).await; }
This creates an account with deterministic keys derived from the seed. Multiple calls with the same seed produce identical accounts.
Key Builders
#![allow(unused)] fn main() { use aura_testkit::builders::test_key_pair; let (signing_key, verifying_key) = test_key_pair(1337); }
Key pairs derive from the provided seed. This enables testing signature verification with known keys.
Identifier Generation
Tests must use deterministic identifiers to ensure reproducibility.
#![allow(unused)] fn main() { use aura_core::identifiers::AuthorityId; let auth1 = AuthorityId::from_entropy([1u8; 32]); let auth2 = AuthorityId::from_entropy([2u8; 32]); }
Never use Uuid::new_v4() or similar entropy-consuming methods in tests. Incrementing byte patterns create distinct but reproducible identifiers.
Verification Utilities
The verification module provides utilities for property testing and differential testing.
Proptest Strategies
The strategies module defines proptest strategies for Aura types.
#![allow(unused)] fn main() { use aura_testkit::verification::strategies::{arb_device_id, arb_account_id, arb_key_pair}; use proptest::prelude::*; proptest! { #[test] fn device_id_deterministic(id in arb_device_id()) { assert_ne!(id.to_string(), ""); } #[test] fn key_pair_valid((sk, vk) in arb_key_pair()) { assert_eq!(sk.verifying_key(), vk); } } }
Available strategies include arb_device_id, arb_account_id, arb_session_id, and arb_key_pair. These generate valid, deterministic instances for property testing.
Lean Oracle
The lean_oracle module provides integration with Lean theorem proofs.
#![allow(unused)] fn main() { use aura_testkit::verification::lean_oracle::LeanOracle; let oracle = LeanOracle::new()?; let result = oracle.verify_journal_merge(&state1, &state2)?; }
The oracle invokes compiled Lean code to verify properties. This enables differential testing against proven implementations.
Capability Soundness
The capability_soundness module provides formal verification for capability system properties.
#![allow(unused)] fn main() { use aura_testkit::verification::capability_soundness::{ CapabilitySoundnessVerifier, SoundnessProperty, CapabilityState }; let mut verifier = CapabilitySoundnessVerifier::with_defaults(); let result = verifier.verify_property( SoundnessProperty::NonInterference, initial_state ).await?; assert!(result.holds); }
The verifier checks five soundness properties: NonInterference, Monotonicity, TemporalConsistency, ContextIsolation, and AuthorizationSoundness. Use verify_all_properties to check all properties at once.
Consensus Testing
The consensus module provides infrastructure for consensus protocol testing.
ITF Loader
The itf_loader module loads ITF traces for replay testing.
#![allow(unused)] fn main() { use aura_testkit::consensus::itf_loader::ITFLoader; let trace = ITFLoader::load("artifacts/traces/consensus_happy_path.itf.json")?; for state in trace.states { // Verify state against implementation } }
ITF traces come from Quint model checking. The loader parses them into Rust types for conformance testing.
Reference Implementation
The reference module provides a minimal consensus implementation for differential testing.
#![allow(unused)] fn main() { use aura_testkit::consensus::reference::ReferenceConsensus; let reference = ReferenceConsensus::new(config); let expected = reference.process_vote(vote)?; let actual = production_consensus.process_vote(vote)?; assert_eq!(expected.outcome, actual.outcome); }
The reference implementation prioritizes clarity over performance. It serves as a specification against which production code is tested.
Mock Runtime Bridge
The MockRuntimeBridge simulates the runtime environment for TUI testing.
#![allow(unused)] fn main() { use aura_testkit::mock_runtime_bridge::MockRuntimeBridge; let bridge = MockRuntimeBridge::new(); bridge.inject_chat_update(chat_state); bridge.inject_contact_update(contacts); }
This bridge injects signals that would normally come from the reactive pipeline. It enables testing TUI state machines without a full runtime.
Conformance Framework
The conformance module provides artifact validation for native/WASM parity testing.
Artifact Format
The AuraConformanceArtifactV1 captures execution state for comparison. The type is defined in aura_core::conformance:
#![allow(unused)] fn main() { use aura_core::{AuraConformanceArtifactV1, AuraConformanceRunMetadataV1, ConformanceSurfaceName}; let mut artifact = AuraConformanceArtifactV1::new(AuraConformanceRunMetadataV1 { target: "native".to_string(), profile: "native_coop".to_string(), scenario: "test_scenario".to_string(), seed: Some(42), commit: None, async_host_transcript_entries: None, async_host_transcript_digest_hex: None, }); // Insert required surfaces artifact.insert_surface(ConformanceSurfaceName::Observable, observable_surface); artifact.insert_surface(ConformanceSurfaceName::SchedulerStep, scheduler_surface); artifact.insert_surface(ConformanceSurfaceName::Effect, effect_surface); artifact.validate_required_surfaces()?; }
Every conformance artifact must capture three surfaces:
| Surface | Purpose | Content |
|---|---|---|
Observable | Protocol-visible outputs | Normalized message contents |
SchedulerStep | Logical progression | Step index, session state, role progression |
Effect | Effect envelope trace | Sequence of effect calls with arguments |
Missing surfaces cause validation failure.
Metadata aids debugging but does not affect comparison:
#![allow(unused)] fn main() { pub struct AuraConformanceRunMetadataV1 { pub target: String, pub profile: String, pub scenario: String, pub seed: Option<u64>, pub commit: Option<String>, pub async_host_transcript_entries: Option<usize>, pub async_host_transcript_digest_hex: Option<String>, } }
Effect Envelope Classification
Each effect kind has a comparison class that determines how differences are evaluated:
| Effect Kind | Class | Comparison Rule |
|---|---|---|
send_decision | commutative | Order-insensitive under normalization |
invoke_step | commutative | Scheduler interleavings normalized |
handle_recv | strict | Byte-exact match required |
handle_choose | strict | Branch choice must match |
handle_acquire | strict | Guard semantics must match |
handle_release | strict | Guard semantics must match |
topology_event | algebraic | Reduced via topology-normal form |
The strict class requires exact matches. The commutative class normalizes order before comparison. The algebraic class applies domain-specific reduction before comparison.
New effect kinds must be classified before use:
#![allow(unused)] fn main() { use aura_core::conformance::AURA_EFFECT_ENVELOPE_CLASSIFICATIONS; AURA_EFFECT_ENVELOPE_CLASSIFICATIONS.insert( "new_effect_kind", ComparisonClass::Strict, ); aura_core::assert_effect_kinds_classified(&effect_trace)?; }
Unclassified effect kinds cause conformance checks to fail.
Module Structure
aura-testkit/
├── src/
│ ├── builders/ # Test data builders
│ ├── configuration/ # Test configuration
│ ├── consensus/ # Consensus testing utilities
│ ├── conformance.rs # Conformance artifact support
│ ├── differential.rs # Differential testing
│ ├── fixtures/ # Reusable test fixtures
│ ├── foundation.rs # Core test utilities
│ ├── handlers/ # Mock effect handlers
│ ├── infrastructure/ # Test harness infrastructure
│ ├── mock_effects.rs # Simple mock implementations
│ ├── stateful_effects/ # Stateful effect handlers
│ └── verification/ # Property testing utilities
├── tests/ # Integration tests
└── benches/ # Performance benchmarks
Related Documentation
See Testing Guide for how to write tests using this infrastructure. See Effect System for effect trait definitions.
Simulation Infrastructure Reference
This document describes the architecture of aura-simulator, the simulation crate that provides deterministic protocol testing through effect handler composition, fault injection, and scenario execution.
Overview
The aura-simulator crate occupies Layer 6 in the Aura architecture. It enables testing distributed protocols under controlled conditions. The simulator uses a handler-based architecture rather than a monolithic simulation engine.
The crate provides four capabilities. It offers specialized effect handlers for simulation control. It includes a middleware system for fault injection. It supports TOML-based scenario definitions. It integrates with Quint for model-based testing.
Handler-Based Architecture
The simulator composes effect handlers rather than wrapping them in a central engine. Each simulated participant uses its own handler instances. This approach aligns with Aura's stateless effect architecture.
graph TD
A[Protocol Code] --> B[Effect Traits]
B --> C[SimulationTimeHandler]
B --> D[SimulationFaultHandler]
B --> E[Other Handlers]
C --> F[Simulated State]
D --> F
E --> F
Handlers implement effect traits from aura-core. Protocol code calls effect methods without knowing whether handlers are production or simulation instances.
Simulation Handlers
SimulationTimeHandler
This handler provides deterministic time control.
#![allow(unused)] fn main() { use aura_simulator::handlers::SimulationTimeHandler; use aura_core::effects::PhysicalTimeEffects; use std::time::Duration; let mut time = SimulationTimeHandler::new(); time.jump_to_time(Duration::from_secs(10)); let now = time.physical_time().await?; }
Time starts at zero and advances only through explicit jump_to_time calls or sleep_ms invocations. The sleep_ms method returns immediately after advancing simulated time by the scaled duration. This enables testing timeout behavior without wall-clock delays. Use set_acceleration to adjust time scaling.
SimulationFaultHandler
This handler injects faults into protocol execution.
#![allow(unused)] fn main() { use aura_simulator::handlers::SimulationFaultHandler; use aura_core::{AuraFault, AuraFaultKind, FaultEdge}; use std::time::Duration; let faults = SimulationFaultHandler::new(42); // seed for determinism // Inject a message delay fault let delay_fault = AuraFault { fault: AuraFaultKind::MessageDelay { delay_ms: 200 }, edge: FaultEdge::new("*", "*"), }; faults.inject_fault(delay_fault, Some(Duration::from_secs(60)))?; // Inject a message drop fault let drop_fault = AuraFault { fault: AuraFaultKind::MessageDrop { probability: 0.1 }, edge: FaultEdge::new("*", "*"), }; faults.inject_fault(drop_fault, None)?; // permanent }
Fault types include MessageDelay, MessageDrop, MessageCorruption, NodeCrash, NetworkPartition, FlowBudgetExhaustion, and JournalCorruption. Faults can be temporary (with duration) or permanent. The handler implements ChaosEffects for async fault injection.
SimulationScenarioHandler
This handler manages scenario-driven testing.
#![allow(unused)] fn main() { use aura_simulator::handlers::{ SimulationScenarioHandler, ScenarioDefinition, TriggerCondition, InjectionAction, }; let mut scenarios = SimulationScenarioHandler::new(); scenarios.add_scenario(ScenarioDefinition { name: "partition".to_string(), trigger: TriggerCondition::AfterTime(Duration::from_secs(5)), action: InjectionAction::PartitionNetwork { group_a: vec![device1, device2], group_b: vec![device3], }, }); }
Scenarios define triggered actions based on time or protocol state. They enable testing recovery from transient failures.
SimulationEffectComposer
This type composes handlers into complete simulation environments.
#![allow(unused)] fn main() { use aura_simulator::handlers::SimulationEffectComposer; use aura_core::DeviceId; let device_id = DeviceId::new_from_entropy([1u8; 32]); let composer = SimulationEffectComposer::for_testing(device_id).await?; let env = composer .with_time_control() .with_fault_injection() .build()?; }
The composer provides a builder pattern for handler configuration. It produces an effect system instance suitable for simulation.
TOML Scenario System
TOML scenarios provide human-readable integration tests with fault injection.
File Format
Scenario files live in the scenarios/ directory.
[metadata]
name = "dkd_basic_derivation"
description = "Basic P2P deterministic key derivation"
version = "1.0"
[[phases]]
name = "setup"
actions = [
{ type = "create_participant", id = "alice" },
{ type = "create_participant", id = "bob" },
]
[[phases]]
name = "derivation"
actions = [
{ type = "run_choreography", choreography = "p2p_dkd", participants = ["alice", "bob"] },
]
[[phases]]
name = "verification"
actions = [
{ type = "verify_property", property = "derived_keys_match" },
]
[[properties]]
name = "derived_keys_match"
property_type = "safety"
expression = "alice.derived_key == bob.derived_key"
Each scenario has metadata, ordered phases, and property definitions. Phases contain action sequences that execute in order.
Action Types
| Action | Parameters | Description |
|---|---|---|
create_participant | id | Create a simulated participant |
run_choreography | choreography, participants | Execute a choreographic protocol |
verify_property | property | Check a named property |
simulate_data_loss | participant, percentage | Delete random stored data |
apply_network_condition | condition, duration | Apply network fault |
advance_time | duration | Advance simulated time |
Execution
The SimulationScenarioHandler executes TOML scenarios.
#![allow(unused)] fn main() { use aura_simulator::handlers::SimulationScenarioHandler; let handler = SimulationScenarioHandler::new(); let result = handler.execute_file("scenarios/core_protocols/dkd_basic.toml").await?; assert!(result.all_properties_passed()); }
Execution proceeds phase by phase. Failures stop execution and report the failing action.
Configuration System
The simulator uses configuration types from aura_simulator::types.
SimulatorConfig
#![allow(unused)] fn main() { use aura_simulator::types::{SimulatorConfig, NetworkConfig}; use aura_core::DeviceId; let config = SimulatorConfig { seed: 42, deterministic: true, time_scale: 1.0, network: Some(NetworkConfig { latency: std::time::Duration::from_millis(50), packet_loss_rate: 0.02, bandwidth_bps: Some(1_000_000), }), ..Default::default() }; }
NetworkConfig
Network configuration controls simulated network conditions.
| Field | Type | Description |
|---|---|---|
latency | Duration | Base network latency |
packet_loss_rate | f64 | Probability of dropping messages |
bandwidth_bps | Option<u64> | Bytes per second limit |
SimulatorContext
The SimulatorContext tracks execution state during simulation runs.
#![allow(unused)] fn main() { use aura_simulator::types::SimulatorContext; let context = SimulatorContext::new("scenario_id".into(), "run_123".into()) .with_seed(42) .with_participants(3, 2) .with_debug(true); println!("Tick: {}", context.tick); println!("Timestamp: {:?}", context.timestamp); }
The context advances through advance_tick and advance_time methods.
Async Host Boundary
The AsyncSimulatorHostBridge provides an async request/resume interface for Telltale integration.
Design
#![allow(unused)] fn main() { use aura_simulator::{AsyncHostRequest, AsyncSimulatorHostBridge}; let mut host = AsyncSimulatorHostBridge::new(42); host.submit(AsyncHostRequest::VerifyAllProperties); let entry = host.resume_next().await?; }
The bridge maintains deterministic ordering through FIFO processing and monotone sequence IDs.
Determinism Constraints
The async host boundary enforces several constraints. Requests process in submission order. Each request receives a unique monotone sequence ID. No wall-clock time affects host decisions. Transcript entries enable replay comparison.
Transcript Artifacts
#![allow(unused)] fn main() { use aura_simulator::AsyncHostTranscriptEntry; let entry = host.resume_next().await?; assert_eq!(entry.sequence, 0); assert!(entry.request.is_verify_properties()); }
Transcript entries record request/response pairs. They enable sync-versus-async host parity testing.
Factory Abstraction
The SimulationEnvironmentFactory trait decouples simulation from effect system internals.
#![allow(unused)] fn main() { use aura_core::effects::{SimulationEnvironmentFactory, SimulationEnvironmentConfig}; let config = SimulationEnvironmentConfig { seed: 42, authority_id, device_id: Some(device_id), test_mode: true, }; let effects = factory.create_simulation_environment(config).await?; }
This abstraction enables stable simulation code across effect system changes. Only the factory implementation requires updates when internals change.
Quint Integration
The quint module provides integration with Quint formal specifications. See Formal Verification Reference for complete details.
Telltale Parity Integration
The simulator exposes telltale parity as an artifact-level boundary. The boundary lives in aura_simulator::telltale_parity. Default simulator execution does not run telltale VM directly.
Entry Points
Use TelltaleParityInput and TelltaleParityRunner when both baseline and candidate artifacts are already in memory. Use run_telltale_parity_file_lane with TelltaleParityFileRun for file-driven CI workflows.
The file lane accepts baseline and candidate artifact paths, a comparison profile, and an output report path. It emits one stable JSON report artifact.
Canonical Surface Mapping
Telltale parity lanes use one canonical mapping:
| Telltale Event Family | Aura Surface | Normalization |
|---|---|---|
observable | observable | identity |
scheduler_step | scheduler_step | tick normalization |
effect | effect | envelope classification |
Reports use schema aura.telltale-parity.report.v1.
Expected Outputs
The simulator telltale parity lane writes:
artifacts/telltale-parity/report.json
The report includes:
- comparison classification (
strictorenvelope_bounded) - first mismatch surface
- first mismatch step index
- full differential comparison payload
ITF Trace Format
ITF (Informal Trace Format) traces come from Quint model checking. Each trace captures a sequence of states and transitions.
{
"#meta": {
"format": "ITF",
"source": "quint",
"version": "1.0"
},
"vars": ["phase", "participants", "messages"],
"states": [
{
"#meta": { "index": 0 },
"phase": "Setup",
"participants": [],
"messages": []
},
{
"#meta": { "index": 1, "action": "addParticipant" },
"phase": "Setup",
"participants": ["alice"],
"messages": []
}
]
}
Each state represents a model state. Transitions between states correspond to actions.
ITF traces capture non-deterministic choices for replay:
{
"#meta": {
"index": 3,
"action": "selectLeader",
"nondet_picks": { "leader": "bob" }
}
}
The nondet_picks field records choices made by Quint. Replay uses these values to seed RandomEffects.
ITFLoader
#![allow(unused)] fn main() { use aura_simulator::quint::itf_loader::ITFLoader; let trace = ITFLoader::load("trace.itf.json")?; for (i, state) in trace.states.iter().enumerate() { let action = state.meta.action.as_deref(); let picks = &state.meta.nondet_picks; } }
The loader validates trace format and extracts typed state.
GenerativeSimulator
#![allow(unused)] fn main() { use aura_simulator::quint::generative_simulator::GenerativeSimulator; let simulator = GenerativeSimulator::new(config)?; let result = simulator.replay_trace(&trace).await?; }
The generative simulator replays ITF traces through real effect handlers.
Module Structure
aura-simulator/
├── src/
│ ├── handlers/ # Simulation effect handlers
│ │ ├── time_control.rs
│ │ ├── fault_simulation.rs
│ │ ├── scenario.rs
│ │ └── effect_composer.rs
│ ├── middleware/ # Effect interception
│ ├── quint/ # Quint integration
│ │ ├── itf_loader.rs
│ │ ├── action_registry.rs
│ │ ├── state_mapper.rs
│ │ └── generative_simulator.rs
│ ├── scenarios/ # Scenario execution
│ ├── async_host.rs # Async host boundary
│ └── testkit_bridge.rs # Testkit integration
├── tests/ # Integration tests
└── examples/ # Usage examples
Related Documentation
See Simulation Guide for how to write simulations. See Testing Guide for conformance testing. See Formal Verification Reference for Quint integration details.
Formal Verification Reference
This document describes the formal verification infrastructure that provides mathematical guarantees for Aura protocols through Quint model checking, Lean theorem proving, and Telltale session type verification.
Overview
Aura uses three complementary verification systems. Quint provides executable state machine specifications with model checking. Lean provides mathematical theorem proofs. Telltale provides session type guarantees for choreographic protocols.
The systems form a trust chain. Quint specifications define correct behavior. Lean proofs verify mathematical properties. Telltale ensures protocol implementations match session type specifications.
Verification Boundary
Aura separates domain proof ownership from runtime parity checks.
| Verification Surface | Primary Tools | Guarantee Class | Ownership |
|---|---|---|---|
| Consensus and CRDT domain properties | Quint + Lean | model and theorem correctness | verification/quint/ and verification/lean/ |
| Runtime execution conformance | Telltale parity + conformance artifacts | implementation parity under declared envelopes | aura-agent, aura-simulator, aura-testkit |
| Bridge consistency | aura-quint bridge pipeline | cross-validation between model checks and certificates | aura-quint |
Telltale runtime parity does not replace domain theorem work. It validates runtime behavior against admitted profiles and artifact envelopes.
Assurance Summary
This architecture provides five assurance classes.
-
Boundary assurance. Domain theorem claims and runtime parity claims are separated. This reduces proof-surface ambiguity.
-
Runtime parity assurance. Telltale parity lanes compare runtime artifacts with deterministic profiles. This provides replayable evidence for conformance under declared envelopes.
-
Bridge consistency assurance. Bridge pipelines check model-check outcomes against certificate outcomes. This detects drift between proof artifacts and executable checks.
-
CI gate assurance. Parity and bridge lanes run as CI gates. This prevents silent regression of conformance checks.
-
Coverage drift assurance. Coverage documentation is validated against repository state by script checks. This prevents long-term drift between claims and implementation.
Limits remain explicit. Parity success is not a replacement for new Quint or Lean domain proofs. Parity checks are coverage-bounded by scenarios, seeds, and artifact surfaces.
Quint Architecture
Quint specifications live in verification/quint/. They define protocol state machines and verify properties through model checking with Apalache.
Directory Structure
verification/quint/
├── core.qnt # Shared runtime utilities
├── authorization.qnt # Guard chain security
├── recovery.qnt # Guardian recovery
├── consensus/ # Consensus protocol specs
│ ├── core.qnt
│ ├── liveness.qnt
│ └── adversary.qnt
├── journal/ # Journal CRDT specs
│ ├── core.qnt
│ ├── counter.qnt
│ └── anti_entropy.qnt
├── keys/ # Key management specs
│ └── dkg.qnt
├── sessions/ # Session management specs
│ ├── core.qnt
│ └── groups.qnt
├── harness/ # Simulator harnesses
├── tui/ # TUI state machine
└── traces/ # Generated ITF traces
Each specification focuses on a single protocol or subsystem.
Specification Pattern
Specifications follow a consistent structure.
module protocol_example {
// Type definitions
type Phase = Setup | Active | Completed | Failed
type State = { phase: Phase, data: Data }
// State variables
var state: State
// Initial state
action init = {
state' = { phase: Setup, data: emptyData }
}
// State transitions
action transition(input: Input): bool = all {
state.phase != Completed,
state.phase != Failed,
state' = computeNextState(state, input)
}
// Invariants
val safetyInvariant = state.phase != Failed or hasRecoveryPath(state)
}
Actions define state transitions. Invariants define properties that must hold in all reachable states.
Harness Modules
Harness modules provide standardized entry points for simulation.
module harness_example {
import protocol_example.*
action register(id: Id): bool = init
action step(input: Input): bool = transition(input)
action complete(): bool = state.phase == Completed
}
Harnesses enable Quint simulation and ITF trace generation.
Available Specifications
| Specification | Purpose | Key Invariants |
|---|---|---|
consensus/core.qnt | Fast-path consensus | UniqueCommitPerInstance, CommitRequiresThreshold |
consensus/liveness.qnt | Liveness properties | ProgressUnderSynchrony, RetryBound |
consensus/adversary.qnt | Byzantine tolerance | ByzantineThreshold, EquivocationDetected |
journal/core.qnt | Journal CRDT | NonceUnique, FactsOrdered |
journal/anti_entropy.qnt | Sync protocol | FactsMonotonic, EventualConvergence |
authorization.qnt | Guard chain | NoCapabilityWidening, ChargeBeforeSend |
Lean Architecture
Lean proofs live in verification/lean/. They provide mathematical verification of safety properties.
Directory Structure
verification/lean/
├── lakefile.lean # Build configuration
├── Aura/
│ ├── Assumptions.lean # Cryptographic axioms
│ ├── Types.lean # Core type definitions
│ ├── Types/
│ │ ├── ByteArray32.lean
│ │ └── OrderTime.lean
│ ├── Proofs/
│ │ ├── Consensus/
│ │ │ ├── Agreement.lean
│ │ │ ├── Validity.lean
│ │ │ ├── Equivocation.lean
│ │ │ ├── Liveness.lean
│ │ │ ├── Evidence.lean
│ │ │ ├── Adversary.lean
│ │ │ └── Frost.lean
│ │ ├── Journal.lean
│ │ ├── FlowBudget.lean
│ │ ├── GuardChain.lean
│ │ ├── KeyDerivation.lean
│ │ ├── TimeSystem.lean
│ │ └── ContextIsolation.lean
│ └── Runner.lean # CLI for differential testing
Axioms
Cryptographic assumptions appear as axioms in Assumptions.lean.
axiom frost_threshold_unforgeability :
∀ (k n : Nat) (shares : List Share),
k ≤ shares.length →
shares.length ≤ n →
validShares shares →
unforgeable (aggregate shares)
axiom hash_collision_resistance :
∀ (a b : ByteArray), hash a = hash b → a = b
Proofs that depend on these assumptions are sound under standard cryptographic hardness assumptions.
The consensus proofs also depend on domain-level axioms for signature binding. These axioms establish that valid signatures bind to unique results. See verification/lean/Aura/Assumptions.lean for the full axiom reduction analysis.
Claims Bundles
Related theorems group into claims bundles.
structure ValidityClaims where
commit_has_threshold : ∀ c, isCommit c → hasThreshold c
validity : ∀ c, isCommit c → validPrestate c
distinct_signers : ∀ c, isCommit c → distinctSigners c.shares
def validityClaims : ValidityClaims := {
commit_has_threshold := Validity.commit_has_threshold
validity := Validity.validity
distinct_signers := Validity.distinct_signers
}
Bundles provide easy access to related proofs.
Proof Status
| Module | Status | Notes |
|---|---|---|
Validity | Complete | All theorems proven |
Equivocation | Complete | Detection soundness/completeness |
Evidence | Complete | CRDT properties |
Frost | Complete | Aggregation properties |
Agreement | Uses axiom | Depends on FROST uniqueness |
Liveness | Axioms | Timing assumptions |
Journal | Complete | CRDT semilattice properties |
aura-quint Crate
The aura-quint crate provides Rust integration with Quint specifications.
QuintRunner
The runner executes Quint verification and parses results.
#![allow(unused)] fn main() { use aura_quint::runner::{QuintRunner, RunnerConfig}; use aura_quint::PropertySpec; let config = RunnerConfig { default_timeout: Duration::from_secs(60), max_steps: 1000, generate_counterexamples: true, ..Default::default() }; let mut runner = QuintRunner::with_config(config)?; let spec = PropertySpec::invariant("UniqueCommitPerInstance"); let result = runner.verify_property(&spec).await?; }
The runner provides verify_property for invariant checking and simulate for trace-based testing. It caches results and can generate counterexamples.
Property Evaluator
The evaluator checks properties against Rust state.
#![allow(unused)] fn main() { use aura_quint::evaluator::PropertyEvaluator; let evaluator = PropertyEvaluator::new(); let result = evaluator.evaluate("chargeBeforeSend", &state)?; }
Properties translate from Quint syntax to Rust predicates.
Property Categories
The evaluator classifies properties by keyword patterns.
| Category | Keywords | Examples |
|---|---|---|
| Authorization | grant, permit, guard | guardChainOrder |
| Budget | budget, charge, spent | chargeBeforeSend |
| Integrity | attenuation, signature | attenuationOnlyNarrows |
| Liveness | eventually, progress | eventualConvergence |
| Safety | never, always, invariant | uniqueCommit |
Categories help organize verification coverage reports.
Quint Integration in aura-simulator
The simulator provides deeper Quint integration for model-based testing.
ITFLoader
#![allow(unused)] fn main() { use aura_simulator::quint::itf_loader::ITFLoader; let trace = ITFLoader::load("trace.itf.json")?; }
The loader parses ITF traces into typed Rust structures.
QuintMappable Trait
Types that map between Quint and Rust implement QuintMappable.
#![allow(unused)] fn main() { use aura_core::effects::quint::QuintMappable; impl QuintMappable for ConsensusState { fn from_quint(value: &QuintValue) -> Result<Self> { // Parse Quint JSON into Rust type } fn to_quint(&self) -> QuintValue { // Convert Rust type to Quint JSON } } }
This trait enables bidirectional state mapping.
ActionRegistry
The registry maps Quint action names to Rust handlers.
#![allow(unused)] fn main() { use aura_simulator::quint::action_registry::{ActionRegistry, ActionHandler}; let mut registry = ActionRegistry::new(); registry.register("initContext", Box::new(InitContextHandler)); registry.register("submitVote", Box::new(SubmitVoteHandler)); let result = registry.execute("initContext", ¶ms, &effects).await?; }
Handlers implement Quint actions using real effect handlers.
StateMapper
The mapper converts between Aura and Quint state representations.
#![allow(unused)] fn main() { use aura_simulator::quint::state_mapper::StateMapper; let mapper = StateMapper::default(); let quint_state = mapper.aura_to_quint(&aura_state)?; let updated_aura = mapper.quint_to_aura(&quint_state)?; }
Bidirectional mapping enables state synchronization during trace replay.
GenerativeSimulator
The simulator replays ITF traces through real effect handlers.
#![allow(unused)] fn main() { use aura_simulator::quint::generative_simulator::{ GenerativeSimulator, GenerativeSimConfig, }; let config = GenerativeSimConfig { max_steps: 1000, check_invariants_every: 10, seed: Some(42), }; let simulator = GenerativeSimulator::new(config)?; let result = simulator.replay_trace(&trace).await?; }
Replay validates that implementations match Quint specifications.
Telltale Formal Guarantees
Telltale provides session type verification for choreographic protocols.
Session Type Projections
Choreographies project to local session types for each participant.
#![allow(unused)] fn main() { #[choreography] async fn two_party_exchange<A, B>( #[role] alice: A, #[role] bob: B, ) { alice.send(bob, message)?; let response = bob.recv(alice)?; } }
The macro generates session types that ensure protocol compliance.
Leakage Tracking
The LeakageTracker monitors information flow during protocol execution.
#![allow(unused)] fn main() { use aura_mpst::LeakageTracker; let tracker = LeakageTracker::new(budget); tracker.record_send(recipient, message_size)?; let remaining = tracker.remaining_budget(); }
Choreography annotations specify leakage costs. The tracker enforces budgets at runtime.
Guard Annotations
Guards integrate with session types through annotations.
#![allow(unused)] fn main() { #[guard_capability("send_message")] #[flow_cost(100)] #[journal_facts("MessageSent")] async fn send_step() { // Implementation } }
Annotations generate guard chain invocations. The Telltale compiler verifies annotation consistency.
Related Documentation
- Verification Guide — Practical workflows, commands, and Quint-Lean correspondence tables
- Verification Coverage — Current metrics, file inventories, and CI gates
- Simulation Infrastructure Reference — ITF trace format and generative simulation
Runtime
Overview
The Aura runtime assembles effect handlers into working systems. It manages lifecycle, executes the guard chain, schedules reactive updates, and exposes services through AuraAgent. The AppCore provides a unified interface for all frontends.
This document covers runtime composition and execution. See Effect System for trait definitions and handler design.
Lifecycle Management
Aura defines a lightweight LifecycleManager for initialization and shutdown. It coordinates session cleanup timeouts and shutdown behavior. If richer lifecycle orchestration becomes necessary, it should be introduced via a dedicated design pass.
Long-lived runtimes must periodically prune caches and stale in-memory state. Aura handles this in Layer 6 via background maintenance tasks scheduled by the RuntimeTaskRegistry. Domain crates expose cleanup APIs but do not self-schedule. The agent runtime wires these up during startup.
#![allow(unused)] fn main() { // In aura-agent runtime builder system.start_maintenance_tasks(); // Internally, maintenance tasks call: // - sync_service.maintenance_cleanup(...) // - ceremony_tracker.cleanup_timed_out() }
This approach keeps time-based policy in the runtime layer and preserves deterministic testing. The simulator controls time directly. Layer 4 and 5 crates remain decoupled from runtime concerns.
Guard Chain Execution
The runtime enforces guard chain sequencing defined in Authorization. Each projected choreography message expands to three phases. First, snapshot preparation gathers capability frontier, budget headroom, and metadata. Second, pure guard evaluation runs synchronously over the snapshot. Third, command interpretation executes the resulting effect commands.
flowchart LR
A[Snapshot prep] -->|async| B[Guard eval]
B -->|sync| C[Interpreter]
C -->|async| D[Transport]
This diagram shows the guard execution flow. Snapshot preparation is async. Guard evaluation is pure and synchronous. Command interpretation is async and performs actual I/O.
GuardSnapshot
The runtime prepares a GuardSnapshot immediately before entering the guard chain. It contains every stable datum a guard may inspect while remaining read-only.
#![allow(unused)] fn main() { pub struct GuardSnapshot { pub now: TimeStamp, pub caps: Cap, pub budgets: FlowBudgetView, pub metadata: MetadataView, pub rng_seed: [u8; 32], } }
Guards evaluate synchronously against this snapshot and the incoming request. They cannot mutate state or perform I/O. This keeps guard evaluation deterministic, replayable, and WASM-compatible.
EffectCommands
Guards do not execute side effects directly. Instead, they return EffectCommand items for the interpreter to run. Each command is a minimal description of work.
#![allow(unused)] fn main() { pub enum EffectCommand { ChargeBudget { context: ContextId, authority: AuthorityId, peer: AuthorityId, amount: FlowCost, }, AppendJournal { entry: JournalEntry }, RecordLeakage { bits: u32 }, StoreMetadata { key: String, value: String }, SendEnvelope { to: NetworkAddress, peer_id: Option<uuid::Uuid>, envelope: Vec<u8> }, GenerateNonce { bytes: usize }, } }
Commands describe what happened rather than how. Interpreters can batch, cache, or reorder commands as long as the semantics remain intact. This vocabulary keeps the guard interface simple.
EffectInterpreter
The EffectInterpreter trait encapsulates async execution of commands. Production runtimes hook it to aura-effects handlers. The simulator hooks deterministic interpreters that record events instead of hitting the network.
#![allow(unused)] fn main() { #[async_trait] pub trait EffectInterpreter: Send + Sync { async fn execute(&self, cmd: EffectCommand) -> Result<EffectResult>; fn interpreter_type(&self) -> &'static str; } }
ProductionEffectInterpreter performs real I/O for storage, transport, and journal. SimulationEffectInterpreter records deterministic events and consumes simulated time. This design lets the guard chain enforce authorization, flow budgets, and journal coupling without leaking implementation details.
Reactive Scheduling
The ReactiveScheduler in aura-agent processes journal facts and drives UI signal updates. It receives facts from multiple sources including journal commits, network receipts, and timers. It batches them in a 5ms window and drives all signal updates.
Intent → Fact Commit → FactPredicate → Query Invalidation → Signal Emit → UI Update
This flow shows how facts propagate to UI. Services emit facts rather than directly mutating view state. The scheduler processes fact batches and updates registered signal views. This eliminates dual-write bugs where different signal sources could desync.
Signal Views
Domain signals are driven by signal views in the reactive scheduler. ChatSignalView, ContactsSignalView, and InvitationsSignalView process facts and emit full state snapshots to their respective signals.
#![allow(unused)] fn main() { // Define application signals pub static CHAT_SIGNAL: LazyLock<Signal<ChatState>> = LazyLock::new(|| Signal::new("app:chat")); // Bind signal to query at initialization pub async fn register_app_signals_with_queries<R: ReactiveEffects>( handler: &R, ) -> Result<(), ReactiveError> { handler.register_query(&*CHAT_SIGNAL, ChatQuery::default()).await?; Ok(()) } }
This example shows signal definition and query binding. Signals are defined as static lazy values. They are bound to queries during initialization. When facts change, queries invalidate and signals update automatically.
Fact Processing
The scheduler integrates with the effect system through fact sinks. Facts flow from journal commits through the scheduler to signal views.
#![allow(unused)] fn main() { // In RuntimeSystem (aura-agent) effect_system.attach_fact_sink(pipeline.fact_sender()); // The scheduler processes fact batches and updates signal views. }
Terminal screens subscribe and automatically receive updates. This enables push-based UI updates without polling.
UnifiedHandler
The UnifiedHandler composes Query and Reactive effects into a single cohesive handler. It holds a QueryHandler, a shared ReactiveHandler, and an optional capability context.
The commit_fact method adds a fact and invalidates affected queries. The query method checks capabilities and executes the query. BoundSignal<Q> pairs a signal with its source query for registration and invalidation tracking.
Service Pattern
Domain crates define stateless handlers that take effect references per call. The agent layer wraps these with services that manage shared access.
Handler Layer
Handlers in domain crates are stateless and return GuardOutcome values. They produce pure plans describing effect commands rather than performing I/O directly.
#![allow(unused)] fn main() { // aura-chat/src/fact_service.rs pub struct ChatFactService; impl ChatFactService { pub fn new() -> Self { Self } pub fn prepare_create_channel( &self, snapshot: &GuardSnapshot, channel_id: ChannelId, name: String, ) -> GuardOutcome { GuardOutcome::authorized(vec![ EffectCommand::AppendJournal { entry: /* ... */ }, ]) } } }
This example shows a domain handler returning a guard outcome. The handler performs pure evaluation over the snapshot. It does not execute I/O or hold state. This keeps domain crates testable without tokio dependencies.
Service Layer
Services in aura-agent wrap handlers, run guard evaluation, and interpret commands.
#![allow(unused)] fn main() { // aura-agent/src/handlers/chat_service.rs pub struct ChatServiceApi { handler: ChatFactService, effects: Arc<AuraEffectSystem>, } impl ChatServiceApi { pub fn new(effects: Arc<AuraEffectSystem>) -> Self { Self { handler: ChatFactService::new(), effects, } } pub async fn create_group( &self, name: &str, creator_id: AuthorityId, ) -> AgentResult<ChatGroup> { let snapshot = self.effects.prepare_snapshot().await?; let outcome = self.handler.prepare_create_channel( &snapshot, ChannelId::new(), name.to_string(), ); self.effects.interpret(outcome).await } } }
Services gather snapshots, call handlers, and interpret outcomes. They share the effect system via Arc<AuraEffectSystem>. Error normalization converts domain errors to AgentError.
Agent API
The agent exposes services through accessor methods.
#![allow(unused)] fn main() { impl AuraAgent { pub fn chat(&self) -> ChatServiceApi { ChatServiceApi::new(self.runtime.effects()) } pub fn sessions(&self) -> &SessionServiceApi { ... } pub fn auth(&self) -> &AuthServiceApi { ... } pub fn invitations(&self) -> &InvitationServiceApi { ... } pub fn recovery(&self) -> &RecoveryServiceApi { ... } } }
This pattern provides clean service access. Services are created on demand with no lazy-init overhead. The ServiceRegistry initializes all services during agent startup and holds Arc references to each.
Session Management
The runtime manages the lifecycle of distributed protocols. Choreographies define protocol logic. Sessions represent single stateful executions of choreographies. The runtime uses the effect system to create, manage, and execute sessions.
Session Interface
The SessionManagementEffects trait provides the abstract interface for all session operations.
#![allow(unused)] fn main() { pub trait SessionManagementEffects: Send + Sync { async fn create_choreographic_session( &self, session_type: SessionType, participants: Vec<ParticipantInfo>, ) -> Result<SessionId>; async fn send_choreographic_message( &self, session_id: SessionId, message: Vec<u8>, ) -> Result<()>; } }
This trait abstracts session management into an effect. Application logic remains decoupled from the underlying implementation. Sessions can use in-memory or persistent state.
Session State
Concrete implementations act as the engine for the session system. Each session has a SessionId for unique identification, a SessionStatus indicating the current phase, a SessionEpoch for coordinating state changes, and a list of participants.
The creation and lifecycle of sessions are themselves managed as choreographic protocols. The SessionLifecycleChoreography in aura-protocol ensures consistency across all participants.
Telltale Integration
Aura executes production choreography sessions through the Telltale VM in Layer 6. Production startup is manifest-driven. Generated CompositionManifest metadata defines the protocol id, required capabilities, determinism profile reference, link constraints, and delegation constraints for each choreography. AuraChoreoEngine runs admitted VM sessions and exposes deterministic trace, replay, and envelope-validation APIs.
Runtime ownership is fragment-scoped. One admitted VM fragment has one local owner at a time. A choreography without link metadata is one fragment. A choreography with link metadata yields one fragment per linked bundle. Ownership claims, transfer, and release flow through AuraEffectSystem and ReconfigurationManager.
The synchronous callback boundary is VmBridgeEffects. AuraVmEffectHandler and AuraQueuedVmBridgeHandler use it for session-local payload queues, blocked receive snapshots, branch choices, and scheduler signals. Async transport, guard-chain execution, journal coupling, and storage remain outside VM callbacks in vm_host_bridge and service loops.
Dynamic reconfiguration follows the same rule. Runtime code must go through ReconfigurationManager for link and delegation so bundle evidence, capability admission, and coherence checks are enforced before any transfer occurs.
VM Profiles
Telltale VM execution is configured through explicit runtime profiles. Use build_vm_config(hardening, parity) instead of ad-hoc VMConfig mutation.
AuraVmHardeningProfile controls safety posture. The Dev profile enables host contract assertions and full trace capture. The Ci profile enforces strict output predicate allow-lists and sequence replay mode. The Prod profile preserves safety checks with bounded overhead.
AuraVmParityProfile controls deterministic cross-target lanes. NativeCooperative provides the native parity baseline. WasmCooperative provides the WASM parity lane. Both use cooperative scheduling and strict effect determinism.
Determinism and scheduler policy are protocol-driven rather than ad hoc. Admission resolves the protocol class, applies the configured determinism tier and replay mode, validates the selected VM profile, and chooses scheduler controls from weighted measure plus guard-capacity signals. Production code should not mutate these settings directly after admission.
Operational envelopes are part of admission, not a host convention. Cooperative fragments run on the canonical VM. Replay-deterministic and envelope-bounded fragments select the threaded runtime only through the policy and admission path. Envelope-bounded fragments must provide admissible envelope artifacts or admission fails closed.
Mixed workloads are allowed. Cooperative and threaded fragments may coexist in the same runtime. The contract is per fragment. Envelope validation, diff capture, and fallback to canonical execution are also per fragment.
Boundary Review Checklist
Use this checklist for any change to the Aura and Telltale boundary:
- Confirm that VM callbacks perform only synchronous
VmBridgeEffectsoperations. - Confirm that async transport, journal, storage, and guard work stay outside the callback path.
- Confirm that each admitted fragment has exactly one local owner at a time.
- Confirm that
delegateandlinkflows useReconfigurationManagerand fragment ownership APIs. - Confirm that threaded or envelope-bounded execution is selected only through policy and admission.
- Confirm that boundary changes add or update replay, conformance, and fault-path tests.
Fact Registry
The FactRegistry provides domain-specific fact type registration and reduction for reactive scheduling. It lives in aura-journal and is integrated via AuraEffectSystem::fact_registry().
Registration
Domain crates register their fact types during effect system assembly.
#![allow(unused)] fn main() { registry.register( "chat", ChatFact::type_id(), |facts| ChatFact::reduce(facts), ); }
This example shows aura-chat registering its fact type. Registered domains include Chat for message threading, Invitation for device invitations, Contact for relationship management, and Moderation for home and mute facts.
Scheduling Integration
The reactive scheduler uses the registry to process domain facts. When facts arrive, the scheduler looks up the registered reducer. It applies the reducer to compute derived state and notifies reactive subscribers of state changes.
Production code obtains the registry via effect_system.fact_registry(). Tests may use build_fact_registry() for isolation. The registry assembly stays in Layer 6 rather than Layer 1.
Delivery Policy
The DeliveryPolicy trait enables domain crates to define custom acknowledgment retention and garbage collection behavior. This keeps policy decisions in the appropriate layer while the journal provides generic ack storage.
Domain crates implement the trait to control acknowledgment lifecycle. Key methods include min_retention and max_retention for time bounds, requires_ack to check tracking needs, and can_gc to determine GC eligibility.
#![allow(unused)] fn main() { let chat_policy = ChatDeliveryPolicy { min_retention: Duration::from_secs(30 * 24 * 60 * 60), max_retention: Duration::from_secs(90 * 24 * 60 * 60), }; effect_system.register_delivery_policy("chat", Arc::new(chat_policy)); }
This example shows domain-specific retention configuration. Chat messages retain acks for 30 to 90 days. Different domains can have vastly different retention needs. The maintenance system uses registered policies during GC passes.
AppCore
The AppCore in aura-app provides a unified portable interface for all frontend platforms. It is headless and runtime-agnostic. It can run without a runtime bridge for offline or demo modes. It can be wired to a concrete runtime via the RuntimeBridge trait.
Architecture
AppCore sits between frontends and a runtime bridge.
flowchart TB
subgraph Frontends
A[TUI]
B[CLI]
C[iOS]
D[Web]
end
A --> E[AppCore]
B --> E
C --> E
D --> E
E --> F[RuntimeBridge]
F --> G[AuraAgent]
This diagram shows the AppCore architecture. Frontends import UI-facing types from aura-app. They may additionally depend on a runtime crate to obtain a concrete RuntimeBridge. This keeps aura-app portable while allowing multiple runtime backends.
Construction Modes
AppCore supports two construction modes.
#![allow(unused)] fn main() { // Demo/Offline mode let app = AppCore::new(config)?; // Production mode let agent = AgentBuilder::new() .with_config(agent_config) .with_authority(authority_id) .build_production() .await?; let app = AppCore::with_runtime(config, agent.as_runtime_bridge())?; }
Demo mode enables offline development and testing. Production mode provides full effect system capabilities. The runtime bridge exposes async capabilities for sync, signing, and protocols.
Reactive Flow
All state changes flow through the reactive pipeline. Services emit facts rather than directly mutating view state. UI subscribes to signals using signal.for_each(). This preserves push semantics and avoids polling.
Local Intent ───┐
│
Service Result ─┼──► Fact ──► Journal ──► Reduce ──► ViewState
│ │
Remote Sync ────┘ ↓
Signal<T> ──► UI
This flow shows the push-based reactive model. Facts from any source flow through the journal. Reduction computes view state. Signals push updates to UI subscribers.
Runtime Access
When AppCore has a runtime, it provides access to runtime-backed operations.
#![allow(unused)] fn main() { if app.has_runtime() { let runtime = app.runtime().unwrap(); let status = runtime.get_sync_status().await; } }
The runtime bridge exposes async capabilities while keeping aura-app decoupled from any specific runtime implementation. Frontends import app-facing types from aura-app and runtime types from aura-agent directly.
Hello World Guide
This guide gets you running with Aura in 15 minutes. You will build a simple ping-pong protocol, deploy it locally, and interact with it using the CLI.
Setup
Aura uses Nix for reproducible builds. Install Nix with flakes support.
Enter the development environment:
nix develop
This command activates all required tools and dependencies. The environment includes Rust, development tools, and build scripts.
Build the project:
just build
The build compiles all Aura components and generates the CLI binary. This takes a few minutes on the first run.
Creating an Agent
Aura provides platform-specific builder presets for creating agents. The CLI preset is the simplest path for terminal applications.
#![allow(unused)] fn main() { use aura_agent::AgentBuilder; // CLI preset - simplest path for terminal applications let agent = AgentBuilder::cli() .data_dir("~/.aura") .testing_mode() .build() .await?; }
The CLI preset provides sensible defaults for command-line tools. It uses file-based storage, real cryptographic operations, and TCP transport.
For custom environments that need explicit control over effect handlers, use AgentBuilder::custom() with typestate enforcement. This requires providing all five core effects (crypto, storage, time, random, console) before build() is available.
Platform-specific presets are available for iOS (AgentBuilder::ios()), Android (AgentBuilder::android()), and Web/WASM (AgentBuilder::web()). These require feature flags to enable. See Effects and Handlers Guide for detailed builder examples.
See Project Structure for details on the 8-layer architecture and effect handler organization.
Hello World Protocol
Create a simple ping-pong choreography. This protocol demonstrates basic message exchange between two devices.
#![allow(unused)] fn main() { use aura_macros::choreography; use aura_core::effects::{ConsoleEffects, NetworkEffects, TimeEffects}; use aura_core::time::PhysicalTime; use serde::{Serialize, Deserialize}; /// Sealed supertrait for ping-pong effects pub trait PingPongEffects: ConsoleEffects + NetworkEffects + TimeEffects {} impl<T> PingPongEffects for T where T: ConsoleEffects + NetworkEffects + TimeEffects {} #[derive(Debug, Clone, Serialize, Deserialize)] pub struct Ping { pub message: String, pub timestamp: PhysicalTime, } #[derive(Debug, Clone, Serialize, Deserialize)] pub struct Pong { pub response: String, pub timestamp: PhysicalTime, } choreography! { #[namespace = "hello_world"] protocol HelloWorld { roles: Alice, Bob; Alice[guard_capability = "send_ping", flow_cost = 10] -> Bob: SendPing(Ping); Bob[guard_capability = "send_pong", flow_cost = 10, journal_facts = "pong_sent"] -> Alice: SendPong(Pong); } } }
The choreography defines a global protocol. Alice sends a ping to Bob. Bob responds with a pong. Guard capabilities control access and flow costs manage rate limiting.
Implement the Alice session:
#![allow(unused)] fn main() { pub async fn execute_alice_session<E: PingPongEffects>( effects: &E, ping_message: String, bob_device: aura_core::DeviceId, ) -> Result<Pong, HelloWorldError> { let ping = Ping { message: ping_message, timestamp: effects.current_timestamp().await, }; let ping_bytes = serde_json::to_vec(&ping)?; effects.send_to_peer(bob_device.into(), ping_bytes).await?; let (peer_id, pong_bytes) = effects.receive().await?; let pong: Pong = serde_json::from_slice(&pong_bytes)?; Ok(pong) } }
Alice serializes the ping message and sends it to Bob. She then waits for Bob's response and deserializes the pong message. See Effect System for details on effect-based execution.
Local Deployment
Initialize a local Aura account:
just quickstart init
This command creates a 2-of-3 threshold account configuration. The account uses three virtual devices with a threshold of two signatures for operations.
Check account status:
just quickstart status
The status command shows account health, device connectivity, and threshold configuration. All virtual devices should show as connected.
Run quickstart smoke checks:
just quickstart smoke
This command runs a local end-to-end smoke flow (init, status, and threshold-signature checks) across multiple virtual devices.
CLI Interaction
The Aura CLI provides commands for account management and protocol testing. These commands demonstrate core functionality.
View account information:
aura status --verbose
This shows detailed account state including journal facts, capability sets, and trust relationships. The journal contains all distributed state updates.
Run a threshold signature test:
aura threshold-test --message "hello world" --threshold 2
The threshold test coordinates signature generation across virtual devices. Two devices must participate to create a valid signature.
View recent protocol activity:
aura journal-query --limit 10
This command shows recent journal entries created by protocol execution. Each entry represents a state change with cryptographic verification.
Testing Your Protocol
Create a test script for the hello world protocol:
#![allow(unused)] fn main() { use aura_macros::aura_test; use aura_testkit::*; use aura_agent::runtime::AuraEffectSystem; use aura_agent::AgentConfig; #[aura_test] async fn test_hello_world_protocol() -> aura_core::AuraResult<()> { // Create test fixture with automatic tracing let fixture = create_test_fixture().await?; // Create deterministic test effect systems let alice_effects = AuraEffectSystem::simulation_for_named_test_with_salt( &AgentConfig::default(), "test_hello_world_protocol", 0, )?; let bob_effects = AuraEffectSystem::simulation_for_named_test_with_salt( &AgentConfig::default(), "test_hello_world_protocol", 1, )?; // Get device IDs for routing let alice_device = fixture.create_device_id(); let bob_device = fixture.create_device_id(); let ping_message = "Hello Bob!".to_string(); // Run protocol sessions concurrently let (alice_result, bob_result) = tokio::join!( execute_alice_session(&alice_effects, ping_message.clone(), bob_device), execute_bob_session(&bob_effects, ping_message.clone()) ); assert!(alice_result.is_ok(), "Alice session failed"); assert!(bob_result.is_ok(), "Bob session failed"); let pong = alice_result?; assert!(pong.response.contains(&ping_message)); Ok(()) } }
This test creates deterministic, seeded effect systems for Alice and Bob using simulation_for_named_test_with_salt(...). The identity + salt pair makes failures reproducible. For comprehensive testing approaches, see Testing Guide.
Run the test:
cargo test test_hello_world_protocol
The test validates protocol correctness without requiring network infrastructure. Mock handlers provide deterministic behavior for testing.
Understanding System Invariants
As you develop protocols, be aware of Aura's system invariants - properties that must always hold true:
- Charge-Before-Send: All messages pass through the guard chain, which evaluates over a prepared
GuardSnapshotand emitsEffectCommanditems that the interpreter executes before any transport send - CRDT Convergence: Identical facts always produce identical state
- Context Isolation: Information stays within relational context boundaries
- Secure Channel Lifecycle: Channels are epoch-bound and follow strict state transitions
See Project Structure for details. When developing, ensure your protocols respect these invariants to maintain system integrity.
Next Steps
You now have a working Aura development environment. The hello world protocol demonstrates basic choreographic programming concepts.
Continue with Effects and Handlers Guide to learn about effect systems, platform implementation, and handler patterns. Learn choreographic programming in Choreography Guide. For session type theory, see MPST and Choreography.
Explore testing and simulation in Testing Guide and Simulation Guide.
Effects and Handlers Guide
This guide covers how to work with Aura's algebraic effect system. Use it when you need to extend the system at its boundaries: adding handlers, implementing platform support, or creating new effect traits.
For the full effect system specification, see Effect System.
1. Code Location
A critical distinction guides where code belongs in the architecture.
Single-party operations go in aura-effects. These are stateless, context-free handlers that take input and produce output without maintaining state or coordinating with other handlers.
Examples:
sign(key, msg) -> Signature- one device, one cryptographic operationstore_chunk(id, data) -> Ok(())- one device, one writeRealCryptoHandler- self-contained cryptographic operations
Multi-party coordination goes in aura-protocol. These orchestrate multiple handlers together with stateful, context-specific operations.
Examples:
execute_anti_entropy(...)- orchestrates sync across multiple partiesCrdtCoordinator- manages state of multiple CRDT handlersGuardChain- coordinates authorization checks across sequential operations
Rule of thumb: If removing one effect handler requires changing the logic of how other handlers are called (not just removing calls), it's orchestration and belongs in Layer 4.
Decision Matrix
| Pattern | Characteristics | Location |
|---|---|---|
| Single effect trait method | Stateless, single operation | aura-effects |
| Multiple effects/handlers | Stateful, multi-handler | aura-protocol |
| Multi-party coordination | Distributed state, orchestration | aura-protocol |
| Domain types and semantics | Pure logic, no handlers | Domain crate |
| Complete reusable protocol | End-to-end, no UI | Feature crate |
| Handler/protocol assembly | Runtime composition | aura-agent |
| User-facing application | Has main() entry point | aura-terminal |
Boundary Questions
- Stateless or stateful? Stateless goes in
aura-effects. Stateful goes inaura-protocol. - One party or multiple? Single-party goes in
aura-effects. Multi-party goes inaura-protocol. - Context-free or context-specific? Context-free goes in
aura-effects. Context-specific goes inaura-protocol.
2. Effect Handler Pattern
Effect handlers are stateless. Each handler implements one or more effect traits from aura-core. It receives input, performs a single operation, and returns output. No state is maintained between calls.
Production handlers (like RealCryptoHandler) use real libraries. Mock handlers (like MockCryptoHandler in aura-testkit) use deterministic implementations for testing.
See Cryptographic Architecture for cryptographic handler requirements.
Implementing a Handler
Step 1: Define the trait in aura-core
#![allow(unused)] fn main() { #[async_trait] pub trait MyEffects: Send + Sync { async fn my_operation(&self, input: Input) -> Result<Output, EffectError>; } }
Step 2: Implement the production handler in aura-effects
#![allow(unused)] fn main() { pub struct RealMyHandler; #[async_trait] impl MyEffects for RealMyHandler { async fn my_operation(&self, input: Input) -> Result<Output, EffectError> { // Implementation using real libraries } } }
Step 3: Implement the mock handler in aura-testkit
#![allow(unused)] fn main() { pub struct MockMyHandler { seed: u64, } #[async_trait] impl MyEffects for MockMyHandler { async fn my_operation(&self, input: Input) -> Result<Output, EffectError> { // Deterministic implementation for testing } } }
Adding a Cryptographic Primitive
- Define the type in
aura-corecrypto module - Implement
aura-coretraits for the type's semantics - Add a single-operation handler in
aura-effectsthat implements the primitive - Use the handler in feature crates or protocols through the effect system
3. Platform Implementation
Use the AgentBuilder API to assemble the runtime with appropriate effect handlers for each platform.
Builder Strategies
| Strategy | Use Case | Compile-Time Safety |
|---|---|---|
| Platform preset | Standard platforms (CLI, iOS, Android, Web) | Configuration validation |
| Custom preset | Full control over all effects | Typestate enforcement |
| Effect overrides | Preset with specific customizations | Mixed |
Platform Presets
#![allow(unused)] fn main() { // CLI let agent = AgentBuilder::cli() .data_dir("~/.aura") .build() .await?; // iOS (requires --features ios) let agent = AgentBuilder::ios() .app_group("group.com.example.aura") .keychain_access_group("com.example.aura") .build() .await?; // Android (requires --features android) let agent = AgentBuilder::android() .application_id("com.example.aura") .use_strongbox(true) .build() .await?; // Web/WASM (requires --features web) let agent = AgentBuilder::web() .storage_prefix("aura_") .build() .await?; }
Custom Preset with Typestate
#![allow(unused)] fn main() { let agent = AgentBuilder::custom() .with_crypto(Arc::new(RealCryptoHandler::new())) .with_storage(Arc::new(FilesystemStorageHandler::new("~/.aura".into()))) .with_time(Arc::new(PhysicalTimeHandler::new())) .with_random(Arc::new(RealRandomHandler::new())) .with_console(Arc::new(RealConsoleHandler::new())) .build() .await?; }
All five required effects must be provided or the code won't compile.
Required Effects
| Effect | Purpose | Trait |
|---|---|---|
| Crypto | Signing, verification, encryption | CryptoEffects |
| Storage | Persistent data storage | StorageEffects |
| Time | Wall-clock timestamps | PhysicalTimeEffects |
| Random | Cryptographically secure randomness | RandomEffects |
| Console | Logging and output | ConsoleEffects |
Optional Effects
| Effect | Default Behavior |
|---|---|
TransportEffects | TCP transport |
LogicalClockEffects | Derived from storage |
OrderClockEffects | Derived from random |
ReactiveEffects | Default reactive handler |
JournalEffects | Derived from storage + crypto |
BiometricEffects | Fallback no-op handler |
Platform Implementation Checklist
- Identify platform-specific APIs for crypto, storage, time, random, console
- Implement the five core effect traits
- Create a preset builder (optional)
- Add feature flags for platform-specific dependencies
- Write integration tests using mock handlers
- Document platform-specific security considerations
- Consider transport requirements (WebSocket, BLE, etc.)
4. Testing Handlers
Test handlers using mock implementations from aura-testkit.
#![allow(unused)] fn main() { use aura_testkit::*; #[aura_test] async fn test_my_handler() -> aura_core::AuraResult<()> { let fixture = create_test_fixture().await?; // Use fixture.effects() to get mock effect system let result = my_operation(&fixture.effects()).await?; assert!(result.is_valid()); Ok(()) } }
Key principles:
- Never use real system calls in tests (no
SystemTime::now(),thread_rng(), etc.) - Use deterministic seeds for reproducibility
- Test both success and error paths
See Testing Guide for comprehensive testing patterns.
5. Effect System Architecture
For deeper understanding of the effect system architecture, see:
- Effect System - Full specification
- Cryptographic Architecture - Crypto handler requirements
- System Architecture - Layer boundaries
Key Concepts
The effect system uses three layers:
- Foundation effects (
aura-core): Crypto, storage, time, random, console, transport - Infrastructure effects (
aura-effects): Production handlers implementing foundation traits - Composite effects: Built by composing foundation effects (e.g.,
TreeEffects= storage + crypto)
All impure operations (time, randomness, filesystem, network) must flow through effect traits. Direct calls break simulation determinism and WASM compatibility.
Run just check-arch to validate effect trait placement and layer boundaries.
Choreography Development Guide
This guide covers how to build distributed protocols using Aura's choreographic programming system. Use it when you need to coordinate multiple parties with session types, CRDTs, and multi-phase workflows.
For theoretical foundations, see MPST and Choreography. For operation categorization, see Operation Categories.
1. When to Use Choreography
Use choreographic protocols when:
- Multiple parties must coordinate (threshold signing, consensus, sync)
- Session guarantees matter (no deadlock, no message mismatch)
- You need formal verification of protocol correctness
Do not use choreography for:
- Single-party operations (use effect handlers)
- Simple request-response (use direct transport)
2. Protocol Development Pipeline
This pipeline applies to all Layer 4/5 choreographies and all Category C ceremonies.
Phase 1: Classification and Facts
Classify the operation using Operation Categories:
- Category A: Local operations, no coordination
- Category B: Optimistic CRDT operations, eventual consistency
- Category C: Ceremonies requiring threshold agreement
Define fact types with schema versioning:
#![allow(unused)] fn main() { use aura_macros::ceremony_facts; #[ceremony_facts] pub enum InvitationFact { CeremonyInitiated { ceremony_id: CeremonyId, agreement_mode: Option<AgreementMode>, trace_id: Option<String>, timestamp_ms: u64, }, CeremonyCommitted { ceremony_id: CeremonyId, relationship_id: String, agreement_mode: Option<AgreementMode>, trace_id: Option<String>, timestamp_ms: u64, }, CeremonyAborted { ceremony_id: CeremonyId, reason: String, trace_id: Option<String>, timestamp_ms: u64, }, } }
The macro provides canonical ceremony_id() and ceremony_timestamp_ms() accessors.
Phase 2: Choreography Specification
Write the choreography in a .choreo file:
#![allow(unused)] fn main() { use aura_macros::choreography; choreography! { #[namespace = "secure_request"] protocol SecureRequest { roles: Client, Server; Client[guard_capability = "send_request", flow_cost = 50] -> Server: SendRequest(RequestData); Server[guard_capability = "send_response", flow_cost = 30, journal_facts = "response_sent"] -> Client: SendResponse(ResponseData); } } }
Annotation syntax: Role[guard_capability = "...", flow_cost = N, journal_facts = "..."] -> Target: Message
Select the narrowest TimeStamp domain for each time field. See Effect System for time domains.
Phase 3: Runtime Wiring
Create the protocol implementation:
#![allow(unused)] fn main() { use aura_agent::runtime::open_manifest_vm_session_admitted; let (mut engine, handler, vm_sid) = open_manifest_vm_session_admitted( &my_protocol::COMPOSITION_MANIFEST, "Initiator", &my_protocol::global_type(), &my_protocol::local_types(), scheduler_signals, ).await?; let status = engine.run_to_completion(vm_sid)?; }
This wiring opens an admitted VM session from generated choreography metadata. The runtime source of truth is the composition manifest, not an ad hoc adapter. Register the service with the runtime and integrate it with the guard chain. Category C operations must follow the ceremony contract.
Production services should treat the admitted unit as a VM fragment. If the manifest declares link bundles, each linked bundle becomes its own ownership unit. Runtime transfer must use ReconfigurationManager. Do not bypass fragment ownership through service-local state.
The runtime also derives execution mode from admitted policy. Cooperative protocols stay on the canonical VM path. Replay-deterministic and envelope-bounded protocols select the threaded path only through the admission and hardening surface. Service code should not construct ad hoc threaded runtimes.
Phase 4: Status and Testing
Implement CeremonyStatus for Category C or protocol-specific status views:
#![allow(unused)] fn main() { pub fn ceremony_status(facts: &[InvitationFact]) -> CeremonyStatus { // Reduce facts to current status } }
Definition of Done:
- Operation category declared (A/B/C)
- Facts defined with reducer and schema version
- Choreography specified with roles/messages documented
- Runtime wiring added (role runners + registration)
- Fragment ownership uses manifest admission and runtime ownership APIs
-
delegateandlinkflows useReconfigurationManager - Threaded or envelope-bounded execution uses admitted policy only
- Category C uses ceremony runner and emits standard facts
- Status output implemented
- Shared-bus integration test added
- Simulation test added
- Choreography parity/replay tests added (Category C)
See crates/aura-consensus/src/protocol/ for canonical examples.
3. CRDT Integration
CRDTs handle state consistency in choreographic protocols. See Journal for CRDT theory.
CRDT Coordinator
Use CrdtCoordinator to manage CRDT state in protocols:
#![allow(unused)] fn main() { use aura_protocol::effects::crdt::CrdtCoordinator; // State-based CRDT let coordinator = CrdtCoordinator::with_cv_state(authority_id, initial_journal); // Delta CRDT with compaction threshold let coordinator = CrdtCoordinator::with_delta_threshold(authority_id, 100); // Meet-semilattice for constraints let coordinator = CrdtCoordinator::with_mv_state(authority_id, capability_set); }
Protocol Integration
Protocols consume and return coordinators with updated state:
#![allow(unused)] fn main() { use aura_sync::choreography::anti_entropy::execute_as_requester; let (result, updated_coordinator) = execute_anti_entropy( authority_id, config, is_requester, &effect_system, coordinator, ).await?; let synchronized_state = updated_coordinator.cv_handler().get_state(); }
4. Protocol Composition
Complex applications require composing multiple protocols.
Sequential Composition
Chain protocols for multi-phase workflows:
#![allow(unused)] fn main() { pub async fn execute_authentication_flow( &self, target_device: aura_core::DeviceId, ) -> Result<AuthenticationResult, ProtocolError> { // Phase 1: Identity exchange let identity_result = self.execute_identity_exchange(target_device).await?; // Phase 2: Capability negotiation let capability_result = self.execute_capability_negotiation( target_device, &identity_result ).await?; // Phase 3: Session establishment let session_result = self.execute_session_establishment( target_device, &capability_result ).await?; Ok(AuthenticationResult { identity: identity_result, capabilities: capability_result, session: session_result, }) } }
Each phase uses results from previous phases. Failed phases abort the entire workflow.
Parallel Composition
Execute independent protocols concurrently:
#![allow(unused)] fn main() { pub async fn execute_distributed_computation( &self, worker_devices: Vec<aura_core::DeviceId>, ) -> Result<ComputationResult, ProtocolError> { // Launch parallel worker protocols let worker_futures = worker_devices.iter().map(|device| { self.execute_worker_protocol(*device) }); // Wait for all workers with timeout let worker_results = tokio::time::timeout( self.config.worker_timeout, futures::future::try_join_all(worker_futures) ).await??; // Aggregate results self.aggregate_worker_results(worker_results).await } }
Effect Program Composition
Compose protocols through effect programs:
#![allow(unused)] fn main() { let composed_protocol = Program::new() .ext(ValidateCapability { capability: "coordinate".into(), role: Coordinator }) .then(anti_entropy_program) .then(threshold_ceremony_program) .ext(LogEvent { event: "protocols_complete".into() }) .end(); }
5. Error Handling and Resilience
Timeout and Retry
Implement robust timeout handling with exponential backoff:
#![allow(unused)] fn main() { pub async fn execute_with_resilience<T>( &self, protocol_fn: impl Fn() -> BoxFuture<'_, Result<T, ProtocolError>>, operation_name: &str, ) -> Result<T, ProtocolError> { let mut attempt = 0; while attempt < self.config.max_attempts { match tokio::time::timeout( self.config.operation_timeout, protocol_fn() ).await { Ok(Ok(result)) => return Ok(result), Ok(Err(e)) if !e.is_retryable() => return Err(e), _ => { // Exponential backoff with jitter let delay = self.config.base_delay * 2_u32.pow(attempt); tokio::time::sleep(self.add_jitter(delay)).await; attempt += 1; } } } Err(ProtocolError::MaxRetriesExceeded) } }
Compensation and Rollback
For multi-phase protocols, implement compensation for partial failures:
#![allow(unused)] fn main() { pub async fn execute_compensating_transaction( &self, operations: Vec<Operation>, ) -> Result<TransactionResult, TransactionError> { let mut completed = Vec::new(); for operation in &operations { match self.execute_operation(operation).await { Ok(result) => { completed.push((operation.clone(), result)); } Err(e) => { // Compensate in reverse order self.execute_compensation(&completed).await?; return Err(TransactionError::OperationFailed { operation: operation.clone(), cause: e, }); } } } Ok(TransactionResult { completed }) } }
Circuit Breakers
Prevent cascading failures with circuit breakers:
#![allow(unused)] fn main() { pub enum CircuitState { Closed { failure_count: usize }, Open { opened_at: Instant }, HalfOpen { test_requests: usize }, } pub async fn execute_with_circuit_breaker<T>( &self, protocol_fn: impl Fn() -> BoxFuture<'_, Result<T, ProtocolError>>, ) -> Result<T, ProtocolError> { let should_execute = match &*self.circuit_state.lock() { CircuitState::Closed { failure_count } => *failure_count < self.config.failure_threshold, CircuitState::Open { opened_at } => opened_at.elapsed() >= self.config.recovery_timeout, CircuitState::HalfOpen { test_requests } => *test_requests < self.config.test_threshold, }; if !should_execute { return Err(ProtocolError::CircuitBreakerOpen); } match protocol_fn().await { Ok(result) => { self.record_success(); Ok(result) } Err(e) => { self.record_failure(); Err(e) } } } }
6. Guard Chain Integration
The guard chain enforces authorization, flow budgets, and journal commits. See Authorization for the full specification.
Guard Chain Pattern
Guards are pure: evaluation runs synchronously over a prepared GuardSnapshot:
#![allow(unused)] fn main() { // Phase 1: Authorization via Biscuit + policy (async, cached) let token = effects.verify_biscuit(&request.token).await?; // Phase 2: Prepare snapshot and evaluate guards (sync) let snapshot = GuardSnapshot { capabilities: token.capabilities(), flow_budget: current_budget, ..Default::default() }; let commands = guard_chain.evaluate(&snapshot, &request)?; // Phase 3: Execute commands (async) for command in commands { interpreter.execute(command).await?; } }
No transport observable occurs until the interpreter executes commands in order.
Security Annotations
Choreography annotations compile to guard chain commands:
guard_capability: Creates capability check before sendflow_cost: Charges flow budgetjournal_facts: Records facts after successful sendleak: Records leakage budget charge
7. Domain Service Pattern
Domain crates define stateless handlers; the agent layer wraps them with services.
Domain Handler
#![allow(unused)] fn main() { // In domain crate (e.g., aura-chat/src/service.rs) pub struct ChatHandler; impl ChatHandler { pub async fn send_message<E>( &self, effects: &E, channel_id: ChannelId, content: String, ) -> Result<MessageId> where E: StorageEffects + RandomEffects + PhysicalTimeEffects { let message_id = effects.random_uuid().await; // ... domain logic Ok(message_id) } } }
Agent Service Wrapper
#![allow(unused)] fn main() { // In aura-agent/src/handlers/chat_service.rs pub struct ChatService { handler: ChatHandler, effects: Arc<RwLock<AuraEffectSystem>>, } impl ChatService { pub async fn send_message( &self, channel_id: ChannelId, content: String, ) -> AgentResult<MessageId> { let effects = self.effects.read().await; self.handler.send_message(&*effects, channel_id, content) .await .map_err(Into::into) } } }
Benefits: Domain crate stays pure (no tokio/RwLock), testable with mock effects, consistent pattern across crates.
8. Testing Choreographies
Unit Testing Guard Logic
#![allow(unused)] fn main() { #[test] fn test_cap_guard_denies_unauthorized() { let snapshot = GuardSnapshot { capabilities: vec![], flow_budget: FlowBudget { limit: 100, spent: 0, epoch: 0 }, ..Default::default() }; let result = CapGuard::evaluate(&snapshot, &SendRequest::default()); assert!(result.is_err()); } }
Integration Testing Protocols
#![allow(unused)] fn main() { #[aura_test] async fn test_sync_protocol() -> aura_core::AuraResult<()> { let fixture = create_test_fixture().await?; let coordinator = CrdtCoordinator::with_cv_state( fixture.authority_id(), fixture.initial_journal(), ); let (result, _) = execute_anti_entropy( fixture.authority_id(), SyncConfig::default(), true, // is_requester &fixture.effects(), coordinator, ).await?; assert!(result.is_success()); Ok(()) } }
Simulation Testing
See Simulation Guide for fault injection and adversarial testing.
Related Documentation
- MPST and Choreography - Session type theory
- Operation Categories - Category A/B/C classification
- Authorization - Guard chain specification
- Journal - CRDT and fact semantics
- Testing Guide - Test patterns
- Simulation Guide - Fault injection testing
Testing Guide
This guide covers how to write tests for Aura protocols using the testing infrastructure. It includes unit testing, integration testing, property-based testing, conformance testing, and runtime harness validation.
For infrastructure details, see Test Infrastructure Reference.
1. Core Philosophy
Aura tests follow four principles:
- Effect-based: Tests use effect traits, never direct impure functions
- Real handlers: Tests run actual protocol logic through real handlers
- Deterministic: Tests produce reproducible results
- Comprehensive: Tests validate both happy paths and error conditions
Harness Policy
Aura's runtime harness is the primary end-to-end validation lane. Default harness runs exercise the real Aura runtime with real TUI and webfront ends. The goal is to catch integration failures in the actual product, not just prove a model.
Quint and other verification tools generate models, traces, and invariants. They are not a replacement for real frontends.
aura-app owns the shared semantic scenario and UI contracts.
aura-harness consumes those contracts and drives real frontends.
aura-simulator is the separate alternate runtime substrate.
Use this lane matrix when selecting harness mode.
| Lane | Backend | Command |
|---|---|---|
| Local deterministic | mock | just harness-run -- --config configs/harness/local-loopback.toml --scenario scenarios/harness/local-discovery-smoke.toml |
| Patchbay relay realism | patchbay | just harness-run -- --config configs/harness/local-loopback.toml --scenario scenarios/harness/scenario2-social-topology-e2e.toml --network-backend patchbay |
| Patchbay-vm relay realism | patchbay-vm | just harness-run -- --config configs/harness/local-loopback.toml --scenario scenarios/harness/scenario2-social-topology-e2e.toml --network-backend patchbay-vm |
| Browser | Playwright | just harness-run-browser scenarios/harness/local-discovery-smoke.toml |
All shared flows should use typed scenario primitives and structured snapshot waits.
aura-app::ui_contract is the canonical module for shared flow support.
It defines SharedFlowId, SHARED_FLOW_SUPPORT, SHARED_FLOW_SCENARIO_COVERAGE,
UiSnapshot, compare_ui_snapshots_for_parity, OperationInstanceId, and
RuntimeEventSnapshot.
Use semantic readiness and state assertions before using fallback text matching.
Direct usage of SystemTime::now(), thread_rng(), File::open(), or Uuid::new_v4() is forbidden. These operations must flow through effect traits instead.
2. The #[aura_test] Macro
The macro provides async test setup with tracing and timeout:
#![allow(unused)] fn main() { use aura_macros::aura_test; use aura_testkit::*; #[aura_test] async fn test_basic_operation() -> aura_core::AuraResult<()> { let fixture = create_test_fixture().await?; let result = some_operation(&fixture).await?; assert!(result.is_valid()); Ok(()) } }
The macro wraps the test body with tracing initialization and a 30-second timeout. Create fixtures explicitly.
3. Test Fixtures
Fixtures provide consistent test environments with deterministic configuration.
Creating Fixtures
#![allow(unused)] fn main() { use aura_testkit::infrastructure::harness::TestFixture; let fixture = TestFixture::new().await?; let device_id = fixture.device_id(); let context = fixture.context(); }
Custom Configuration
#![allow(unused)] fn main() { use aura_testkit::infrastructure::harness::{TestFixture, TestConfig}; let config = TestConfig { name: "threshold_test".to_string(), deterministic_time: true, capture_effects: true, timeout: Some(Duration::from_secs(60)), }; let fixture = TestFixture::with_config(config).await?; }
Deterministic Identifiers
Use deterministic identifier generation:
#![allow(unused)] fn main() { use aura_core::identifiers::AuthorityId; let auth1 = AuthorityId::from_entropy([1u8; 32]); let auth2 = AuthorityId::from_entropy([2u8; 32]); }
Incrementing byte patterns create distinct but reproducible identifiers.
4. Unit Tests
Unit tests validate individual functions or components:
#![allow(unused)] fn main() { #[aura_test] async fn test_single_function() -> aura_core::AuraResult<()> { let fixture = create_test_fixture().await?; let input = TestInput::new(42); let output = process_input(&fixture, input).await?; assert_eq!(output.value, 84); Ok(()) } }
Unit tests should be fast and focused, testing one behavior per function.
5. Integration Tests
Integration tests validate complete workflows:
#![allow(unused)] fn main() { use aura_agent::runtime::AuraEffectSystem; use aura_agent::AgentConfig; #[aura_test] async fn test_threshold_workflow() -> aura_core::AuraResult<()> { let fixture = create_test_fixture().await?; let device_ids: Vec<_> = (0..5) .map(|i| DeviceId::new_from_entropy([i as u8 + 1; 32])) .collect(); let effect_systems: Result<Vec<_>, _> = (0..5) .map(|i| { AuraEffectSystem::simulation_for_named_test_with_salt( &AgentConfig::default(), "test_threshold_workflow", i as u64, ) }) .collect(); let result = execute_protocol(&effect_systems?, &device_ids).await?; assert!(result.is_complete()); Ok(()) } }
Use simulation_for_test* helpers for all tests. For multi-instance tests from one callsite, use simulation_for_named_test_with_salt(...) and keep the identity and salt stable. This allows failures to be replayed deterministically.
6. Property-Based Testing
Property tests validate invariants across diverse inputs using proptest.
Synchronous Properties
#![allow(unused)] fn main() { use proptest::prelude::*; fn arbitrary_message() -> impl Strategy<Value = Vec<u8>> { prop::collection::vec(any::<u8>(), 1..=1024) } proptest! { #[test] fn message_roundtrip(message in arbitrary_message()) { let encoded = encode(&message); let decoded = decode(&encoded).unwrap(); assert_eq!(message, decoded); } } }
Async Properties
#![allow(unused)] fn main() { proptest! { #[test] fn async_property(data in arbitrary_message()) { tokio::runtime::Runtime::new().unwrap().block_on(async { let fixture = create_test_fixture().await.unwrap(); let result = async_operation(&fixture, data).await; assert!(result.is_ok()); }); } } }
7. GuardSnapshot Pattern
The guard chain separates pure evaluation from async execution, enabling testing without async runtime.
Testing Pure Guard Logic
#![allow(unused)] fn main() { #[test] fn test_cap_guard_denies_unauthorized() { let snapshot = GuardSnapshot { capabilities: vec![], flow_budget: FlowBudget { limit: 100, spent: 0, epoch: 0 }, ..Default::default() }; let result = CapGuard::evaluate(&snapshot, &SendRequest::default()); assert!(result.is_err()); } }
Testing Flow Budget
#![allow(unused)] fn main() { #[test] fn test_flow_guard_blocks_over_budget() { let snapshot = GuardSnapshot { flow_budget: FlowBudget { limit: 100, spent: 95, epoch: 0 }, ..Default::default() }; let request = SendRequest { cost: 10, ..Default::default() }; let result = FlowGuard::evaluate(&snapshot, &request); assert!(matches!(result.unwrap_err(), GuardError::BudgetExceeded)); } }
8. TUI and CLI Testing
TUI State Machine Tests
#![allow(unused)] fn main() { mod support; use support::TestTui; use aura_terminal::tui::screens::Screen; #[test] fn test_screen_navigation() { let mut tui = TestTui::new(); tui.assert_screen(Screen::Block); tui.send_char('2'); tui.assert_screen(Screen::Neighborhood); } }
CLI Handler Testing
#![allow(unused)] fn main() { use aura_terminal::handlers::{CliOutput, HandlerContext}; #[tokio::test] async fn test_status_handler() { let ctx = create_test_handler_context().await; let output = status::handle_status(&ctx).await.unwrap(); let lines = output.stdout_lines(); assert!(lines.iter().any(|l| l.contains("Authority"))); } }
Quint Trace Usage
Quint traces are model artifacts. Export them through the shared semantic
scenario contract and execute real TUI/web flows through aura-harness
rather than replaying Quint traces directly against the TUI implementation.
9. Conformance Testing
Conformance tests validate that implementations produce identical results across environments.
Conformance Lanes
CI runs two lanes.
Strict lane (native vs WASM cooperative):
just ci-conformance-strict
Differential lane (native threaded vs cooperative):
just ci-conformance-diff
Run both:
just ci-conformance
Mismatch Taxonomy
| Type | Description | Fix |
|---|---|---|
strict | Byte-level difference | Remove hidden state or ordering-sensitive side effects |
envelope_bounded | Outside declared envelopes | Add or correct envelope classification |
surface_missing | Required surface not present | Emit observable, scheduler_step, and effect |
Reproducing Failures
AURA_CONFORMANCE_SCENARIO=scenario_name \
AURA_CONFORMANCE_SEED=42 \
cargo test -p aura-agent \
--features choreo-backend-telltale-vm \
--test telltale_vm_parity test_name \
-- --nocapture
10. Runtime Harness
The runtime harness executes real Aura instances in PTYs for end-to-end validation.
Harness Overview
The harness is the single executor for real frontend scenarios. Scripted mode uses the shared semantic scenario contract. Agent mode uses LLM-driven execution toward goals.
Shared flows should be authored semantically once, then executed through the harness using either the TUI or browser driver. Do not create a second frontend execution path for MBT or simulator replay.
Core shared scenarios should use semantic actions and state-based assertions.
Avoid raw selector steps, raw press_key steps, and label-based browser clicks
except in dedicated low-level driver tests.
Run Config
schema_version = 1
[run]
name = "local-loopback-smoke"
pty_rows = 40
pty_cols = 120
seed = 4242
[[instances]]
id = "alice"
mode = "local"
data_dir = "artifacts/harness/state/local-loopback/alice"
device_id = "alice-dev-01"
bind_address = "127.0.0.1:41001"
Scenario File
id = "discovery-smoke"
goal = "Validate semantic harness observation against a real TUI"
[[steps]]
id = "launch"
action = "launch_actors"
timeout_ms = 5000
[[steps]]
id = "nav-chat"
actor = "alice"
action = "navigate"
screen_id = "chat"
timeout_ms = 2000
[[steps]]
id = "chat-ready"
actor = "alice"
action = "readiness_is"
readiness = "ready"
timeout_ms = 2000
Running the Harness
# Lint before running
just harness-lint -- --config configs/harness/local-loopback.toml \
--scenario scenarios/harness/semantic-observation-tui-smoke.toml
# Execute
just harness-run -- --config configs/harness/local-loopback.toml \
--scenario scenarios/harness/semantic-observation-tui-smoke.toml
# Replay for deterministic reproduction
just harness-replay -- --bundle artifacts/harness/local-loopback-smoke/replay_bundle.json
Interactive Mode
Use tool_repl for manual validation:
cargo run -p aura-harness --bin tool_repl -- \
--config configs/harness/local-loopback.toml
Send JSON requests:
{"id":1,"method":"screen","params":{"instance_id":"alice"}}
{"id":2,"method":"send_keys","params":{"instance_id":"alice","keys":"3n"}}
{"id":3,"method":"wait_for","params":{"instance_id":"alice","pattern":"Create","timeout_ms":4000}}
Harness CI
just ci-harness-build
just ci-harness-contract
just ci-harness-replay
just ci-shared-flow-policy
just ci-shared-flow-policy validates the shared-flow contract end to end. It checks that aura-app shared-flow support declarations are internally consistent, that every fully shared flow has explicit parity-scenario coverage, and that required shell and modal ids still exist. It confirms browser control and field mappings still line up with the shared contract and that core shared scenarios have not drifted back to raw mechanics.
Use just ci-ui-parity-contract for the narrower parity gate. That lane validates shared screen/module mappings, shared-flow scenario coverage, and parity-manifest consistency without running a full scenario matrix.
11. Test Organization
Organize tests by category:
#![allow(unused)] fn main() { #[cfg(test)] mod tests { mod unit { #[aura_test] async fn test_single_function() -> aura_core::AuraResult<()> { Ok(()) } } mod integration { #[aura_test] async fn test_full_workflow() -> aura_core::AuraResult<()> { Ok(()) } } mod properties { proptest! { #[test] fn invariant_holds(input in any::<u64>()) { assert!(input == input); } } } } }
Running Tests
# All tests
just test
# Specific crate
just test-crate aura-agent
# With output
cargo test --workspace -- --nocapture
# TUI state machine tests
cargo test --package aura-terminal --test unit_state_machine
12. Best Practices
Test one behavior per function and name tests descriptively. Use fixtures for common setup. Prefer real handlers over mocks.
Test error conditions explicitly. Avoid testing implementation details. Focus on observable behavior.
Keep tests fast. Parallelize independent tests.
13. Holepunch Backends and Artifact Triage
Use the harness --network-backend option to select execution mode:
# Deterministic local backend
cargo run -p aura-harness --bin aura-harness -- \
run --config configs/harness/local-loopback.toml \
--network-backend mock
# Native Linux Patchbay (requires Linux + userns/capabilities)
cargo run -p aura-harness --bin aura-harness -- \
run --config configs/harness/local-loopback.toml \
--network-backend patchbay
# Cross-platform VM runner (macOS/Linux)
cargo run -p aura-harness --bin aura-harness -- \
run --config configs/harness/local-loopback.toml \
--network-backend patchbay-vm
Harness writes backend resolution details to:
artifacts/harness/<run>/network_backend_preflight.json
Patchbay is the authoritative NAT-realism backend for holepunch validation.
Use native patchbay on Linux CI and Linux developers when capabilities are available.
Use patchbay-vm on macOS and as Linux fallback to run the same scenarios in a Linux VM.
Keep deterministic non-network logic in mock backend tests to preserve fast feedback.
Implementation follows three tiers.
Tier 1 covers deterministic and property tests in aura-testkit for retry and path-selection invariants.
Tier 2 covers Patchbay integration scenarios in aura-harness for PR gating.
Tier 3 covers Patchbay stress and flake detection suites on scheduled CI.
When a scenario fails, triage artifacts in this order.
- Check
network_backend_preflight.jsonto confirm selected backend and fallback reason. - Check
startup_summary.jsonandscenario_report.jsonfor run context and failing step. - Check
events.jsonand backend timeline artifacts for event ordering. - Check namespace and network dumps and pcap files for packet and routing diagnosis.
- Check agent logs for authority-local failures and retry state transitions.
For harness-specific state debugging, treat timeout_diagnostics.json as the first failure bundle.
It includes semantic state snapshots, render readiness, and runtime event history.
14. Browser Harness Workflow (WASM + Playwright)
Use this flow to run harness scenarios in browser mode:
# 1) Check wasm/frontend compilation
just web-check
# 2) Install/update Playwright driver deps
cd crates/aura-harness/playwright-driver
npm ci
npm run install-browsers
npm test
cd ../..
# 3) Serve the web app
just web-serve
In a second shell:
# Lint browser run/scenario config
just harness-lint-browser scenarios/harness/semantic-observation-browser-smoke.toml
# Run browser scenarios
just harness-run-browser scenarios/harness/semantic-observation-browser-smoke.toml
# Replay the latest browser run bundle
just harness-replay-browser
Browser harness artifacts are written under:
artifacts/harness/browser/
When debugging browser failures, check web-serve.log for bundle and runtime startup issues. Check preflight_report.json for browser prerequisites including Node, Playwright, and app URL. Check timeout_diagnostics.json for authoritative and normalized snapshots and per-instance log tails. Playwright screenshots and traces are stored under each instance data_dir in playwright-artifacts/.
timeout_diagnostics.json is now the primary authoritative failure bundle. In addition to UiSnapshot, it should be treated as the first source for runtime event history through runtime_events. It contains operation lifecycle and instance ids. It provides render and readiness diagnostics along with browser and TUI backend log tails.
For browser runs, the harness observes the semantic state contract first and uses DOM/text fallbacks only for diagnostics. If semantic state and rendered UI diverge, treat that as a product or frontend contract bug rather than papering over it with text-based assertions.
Frontend Shell Roadmap
aura-ui is the shared Dioxus UI core. It supports web-first delivery today and future multi-target shells.
aura-web(current): browser shell and harness bridge- Desktop shell (future): desktop-specific shell reusing
aura-ui - Mobile shell (future): mobile-specific shell reusing
aura-ui
Related Documentation
- Test Infrastructure Reference - Infrastructure details
- Simulation Guide - Fault injection testing
- Verification Guide - Formal methods
Simulation Guide
This guide covers how to use Aura's simulation infrastructure for testing distributed protocols under controlled conditions.
When to Use Simulation
Simulation suits scenarios that unit tests cannot address. Use simulation for fault injection testing. Use it for multi-participant protocol testing. Use it for time-dependent behavior validation.
Do not use simulation for simple unit tests. Direct effect handler testing is faster and simpler for single-component validation. Do not treat simulation as the default end-to-end correctness oracle for user-facing flows. Aura's primary feedback loop remains the real-runtime harness running against the real software stack.
See Simulation Infrastructure Reference for the complete architecture documentation.
Simulation vs Harness
Simulation is a first-class alternate runtime substrate, not the primary one. Use the real-runtime harness by default when validating product behavior through the TUI or webapp. Use simulation when you need deterministic virtual time, controlled network faults, scheduler control, or MBT and trace replay under constrained distributed conditions. Promote high-value simulation findings back into real-runtime harness coverage when the flow is user-visible or integration-sensitive.
This separation keeps responsibilities clean. The harness executes real frontends and gathers semantic observations. The simulator provides controlled runtime conditions when explicitly selected. Quint and related verification tooling generate traces and invariants.
The shared semantic contracts for UI state and scenario execution live in
aura-app. Simulation should integrate through those contracts rather than
introducing a parallel frontend-driving format.
Two Simulation Systems
Aura provides two complementary simulation systems.
TOML scenarios suit human-written integration tests. They provide readable scenario definitions with named phases and fault injection. Use them for end-to-end protocol testing.
Quint actions suit model-based testing. They enable generative state space exploration driven by formal specifications. Use them for conformance testing against Quint models.
| Use Case | System |
|---|---|
| End-to-end integration | TOML scenarios |
| Named fault injection | TOML scenarios |
| Conformance testing | Quint actions |
| State space exploration | Quint actions |
When you need user-facing coverage, promote the scenario into the real-runtime harness lane after it is stable in simulation. Treat simulation as a substrate for controlled runtime conditions, not as the final UI executor.
TOML Scenario Authoring
Creating Scenarios
Scenario files live in the scenarios/ directory.
[metadata]
name = "recovery_basic"
description = "Basic guardian recovery flow"
[[phases]]
name = "setup"
actions = [
{ type = "create_participant", id = "owner" },
{ type = "create_participant", id = "guardian1" },
{ type = "create_participant", id = "guardian2" },
]
[[phases]]
name = "recovery"
actions = [
{ type = "run_choreography", choreography = "guardian_recovery", participants = ["owner", "guardian1", "guardian2"] },
]
[[properties]]
name = "owner_recovered"
property_type = "safety"
Each scenario has metadata, ordered phases, and property definitions.
Defining Phases
Phases execute in order. Each phase contains a list of actions.
[[phases]]
name = "fault_injection"
actions = [
{ type = "apply_network_condition", condition = "partition", duration = "5s" },
{ type = "advance_time", duration = "10s" },
]
Actions within a phase execute sequentially. Use multiple phases to organize complex scenarios.
Adding Fault Injection
Fault injection actions simulate adverse conditions.
[[phases]]
name = "chaos"
actions = [
{ type = "simulate_data_loss", participant = "guardian1", percentage = 50 },
{ type = "apply_network_condition", condition = "high_latency", duration = "30s" },
]
Available conditions include partition, high_latency, packet_loss, and byzantine.
Running Scenarios
cargo run --package aura-terminal -- scenario run scenarios/recovery_basic.toml
The scenario handler parses and executes the TOML file. Results report property pass/fail status.
Working with Handlers
Basic Handler Composition
#![allow(unused)] fn main() { use aura_simulator::handlers::SimulationEffectComposer; use aura_core::DeviceId; let device_id = DeviceId::new_from_entropy([1u8; 32]); let composer = SimulationEffectComposer::for_testing(device_id).await?; let env = composer .with_time_control() .with_fault_injection() .build()?; }
The composer builds a complete effect environment from handler components.
Time Control
#![allow(unused)] fn main() { use aura_simulator::handlers::SimulationTimeHandler; let mut time = SimulationTimeHandler::new(); let start = time.physical_time().await?; time.jump_to_time(Duration::from_secs(60)); let later = time.physical_time().await?; }
Simulated time advances only through explicit calls (jump_to_time) or sleep operations. This enables testing timeout behavior without delays.
Fault Injection
#![allow(unused)] fn main() { use aura_simulator::handlers::SimulationFaultHandler; use aura_core::{AuraFault, AuraFaultKind, FaultEdge}; let faults = SimulationFaultHandler::new(42); faults.inject_fault( AuraFault::new(AuraFaultKind::MessageDelay { edge: FaultEdge::new("alice", "bob"), min: Duration::from_millis(100), max: Duration::from_millis(500), }), None, )?; faults.inject_fault( AuraFault::new(AuraFaultKind::MessageDrop { edge: FaultEdge::new("alice", "bob"), probability: 0.1, }), None, )?; }
AuraFault is the canonical simulator fault model. Legacy scenario fault forms should be converted to AuraFault before injection or replay.
Triggered Scenarios
#![allow(unused)] fn main() { use aura_simulator::handlers::{ SimulationScenarioHandler, ScenarioDefinition, TriggerCondition, InjectionAction, }; use aura_core::{AuraFault, AuraFaultKind}; let handler = SimulationScenarioHandler::new(42); handler.register_scenario(ScenarioDefinition { id: "late_partition".to_string(), name: "Late Partition".to_string(), trigger: TriggerCondition::AfterTime(Duration::from_secs(30)), actions: vec![InjectionAction::TriggerFault { fault: AuraFault::new(AuraFaultKind::NetworkPartition { partition: vec![vec!["device1".into(), "device2".into()], vec!["device3".into()]], duration: Some(Duration::from_secs(15)), }), }, duration: Some(Duration::from_secs(45)), priority: 10, }); }
Triggered scenarios inject faults at specific times or protocol states.
Integrating Feature Crates
Layer 5 feature crates (sync, recovery, chat, etc.) integrate with simulation through the effect system. This section covers patterns for wiring feature crates into simulation environments.
Required Effects
Feature crates are generic over effect traits. Common requirements include:
| Effect Trait | Purpose |
|---|---|
NetworkEffects | Transport and peer communication |
JournalEffects | Fact retrieval and commits |
CryptoEffects | Hashing and signature verification |
PhysicalTimeEffects | Timeouts and scheduling |
RandomEffects | Nonce generation |
Pass the effect system from aura-simulator or aura-testkit for deterministic testing. In production, use aura-agent's runtime effects.
Configuration for Simulation
Feature crates typically provide testing configurations that minimize timeouts and remove jitter:
#![allow(unused)] fn main() { use aura_sync::SyncConfig; // Production: conservative timeouts, adaptive scheduling let prod_config = SyncConfig::for_production(); // Testing: fast timeouts, no jitter, predictable behavior let test_config = SyncConfig::for_testing(); // Validate before use test_config.validate()?; }
Environment variables (prefixed per-crate, e.g., AURA_SYNC_*) allow per-process tuning without code changes.
Guard Chain Integration
Feature crates assume the guard chain is active when running protocols:
CapGuard → FlowGuard → LeakageTracker → JournalCoupler → Transport
For simulation, capability checks rely on Biscuit tokens evaluated by AuthorizationEffects. Guard evaluators must be provided by the runtime before sync operations. Validation occurs before sending or applying any protocol data.
Observability in Simulation
Connect feature crates to MetricsCollector for simulation diagnostics:
#![allow(unused)] fn main() { use aura_core::metrics::MetricsCollector; let metrics = MetricsCollector::new(); // Protocol timings, retries, and failure reasons flow to metrics // Log transport and authorization failures for debugging }
Safety Requirements
Feature crates must follow effect system rules. All I/O and timing must flow through effects with no direct runtime calls. Validate Biscuit tokens before accepting peer data. Enforce flow budgets and leakage constraints at transport boundaries.
Verify compliance before simulation:
just ci-effects
Debugging Simulations
Deterministic Configuration
Always use deterministic settings for reproducible debugging.
#![allow(unused)] fn main() { let config = SimulatorConfig { device_id: DeviceId::new_from_entropy([1u8; 32]), network: NetworkConfig::default(), enable_fault_injection: false, deterministic_time: true, }; }
Deterministic identifiers and time enable exact failure reproduction.
Effect System Compliance
Verify protocol code follows effect guidelines before simulation.
just check-arch
The architecture checker flags direct time, randomness, or I/O usage. Non-compliant code breaks simulation determinism.
Monitoring State
#![allow(unused)] fn main() { let metrics = middleware.get_metrics(); println!("Messages: {}", metrics.messages_sent); println!("Faults: {}", metrics.faults_injected); println!("Duration: {:?}", metrics.simulation_duration); }
Middleware metrics help identify unexpected behavior.
Common Issues
Flaky simulation results indicate non-determinism. Check for direct system calls. Check for uncontrolled concurrency. Check for ordering assumptions.
Slow simulations indicate inefficient fault configuration. Reduce fault rates for initial debugging. Increase rates for stress testing.
Online Property Monitoring
Aura simulator supports per-tick property monitoring through aura_simulator::AuraProperty, aura_simulator::AuraPropertyMonitor, and aura_simulator::default_property_suite(...).
The monitor checks properties on each simulation tick using PropertyStateSnapshot input. This includes events, buffer sizes, local-type depths, flow budgets, and optional session, coroutine, and journal snapshots.
#![allow(unused)] fn main() { use aura_simulator::{ AuraPropertyMonitor, ProtocolPropertyClass, ProtocolPropertySuiteIds, PropertyMonitoringConfig, SimulationScenarioConfig, }; let monitoring = PropertyMonitoringConfig::new( ProtocolPropertyClass::Consensus, ProtocolPropertySuiteIds { session, context }, ) .with_check_interval(1) .with_snapshot_provider(|tick| build_snapshot_for_tick(tick)); let config = SimulationScenarioConfig { property_monitoring: Some(monitoring), ..SimulationScenarioConfig::default() }; let results = env.run_scenario("consensus".into(), "with property checks".into(), config).await?; assert!(results.property_violations.is_empty()); }
Default suites are available for consensus, sync, chat, and recovery protocol classes. Scenario results include properties_checked and property_violations for CI reporting.
Quint Integration
Quint actions enable model-based testing. See Verification and MBT Guide for complete workflows.
When to Use Quint
Use Quint actions for conformance testing against formal specifications. Use them for generative state exploration. Do not use them for simple integration tests.
Basic Trace Replay
#![allow(unused)] fn main() { use aura_simulator::quint::itf_loader::ITFLoader; use aura_simulator::quint::generative_simulator::GenerativeSimulator; let trace = ITFLoader::load("trace.itf.json")?; let simulator = GenerativeSimulator::new(config)?; let result = simulator.replay_trace(&trace).await?; assert!(result.all_properties_passed()); }
Trace replay validates implementation against Quint model behavior.
Conformance Workflow
Simulation feeds native/WASM conformance testing. See Testing Guide for conformance lanes and corpus policy.
Generating Corpus
quint run --out-itf=trace.itf.json verification/quint/consensus/core.qnt
ITF traces from Quint become conformance test inputs.
Running Conformance
just ci-conformance
Conformance lanes compare execution across platforms using simulation-controlled environments.
Checkpoints, Contracts, and Shared Replay
Phase 0 hardening uses three simulator workflows as CI gates.
Checkpoint Snapshot Workflow
SimulationScenarioHandler supports portable checkpoint snapshots. The export_checkpoint_snapshot(label) function exports a serializable ScenarioCheckpointSnapshot. The import_checkpoint_snapshot(snapshot) function restores a checkpoint into a fresh simulator instance.
This enables baseline checkpoint persistence for representative choreography suites. It also supports restore-and-continue regression tests. Upgrade smoke tests can resume from pre-upgrade snapshots.
Use checkpoints when validating runtime upgrades or migration safety, not only end-to-end success.
Scenario Contract Workflow
Conformance CI includes scenario contracts for consensus, sync, recovery, and reconfiguration. Each bundle is validated over a seed corpus. Validation checks terminal status (AllDone), required labels observed in trace, and minimum observable event count.
Contract results are written as JSON artifacts. CI fails on any violation with structured output.
Shared Replay Workflow
Replay-heavy parity lanes should use shared replay APIs. Use run_replay_shared(...) and run_concurrent_replay_shared(...) for this purpose.
These APIs reduce duplicate replay state across lanes and keep replay artifacts compatible with canonical trace fragments. Conformance lanes also emit deterministic replay metrics artifacts so regressions in replay footprint are visible during CI review.
For fault-aware replays, persist entries + faults bundles and re-inject faults before replay. Use aura_testkit::ReplayTrace::load_file(...) to load traces. Use ReplayTrace::replay_faults(...) and aura_simulator::AsyncSimulatorHostBridge::replay_expected_with_faults(...) to replay with faults.
Differential Replay Workflow
Use aura_simulator::DifferentialTester to compare baseline and candidate conformance artifacts. Two profiles are available. The strict profile requires byte-identical surfaces. The envelope_bounded profile uses Aura law-aware comparison with commutative and algebraic envelopes.
For parity debugging, run:
just ci-choreo-parity
aura replay --trace-file artifacts/choreo-parity/native_replay/<scenario>__seed_<seed>.json
The replay command validates required conformance surfaces and verifies stored step/run digests against recomputed values.
Best Practices
Start with simple scenarios and add faults incrementally. Use deterministic seeds. Capture metrics for analysis.
Prefer TOML scenarios for human-readable tests. Prefer Quint actions for specification conformance. Combine both for comprehensive coverage.
Related Documentation
- Simulation Infrastructure Reference - Handler APIs
- Verification and MBT Guide - Quint workflows
- Testing Guide - Conformance testing
Verification and MBT Guide
This guide covers how to use formal verification and model-based testing to validate Aura protocols. It focuses on practical workflows with Quint, Lean, and generative testing.
Quint, simulator, and harness have distinct responsibilities. Quint defines models, traces, and invariants. The aura-simulator crate is a selectable deterministic runtime substrate. The aura-harness crate is the single executor for real TUI and web frontend flows. Shared semantic UI and scenario contracts live in aura-app.
aura-app is also the home of the shared-flow support and parity contract used by the real-runtime harness. This includes SharedFlowId, SHARED_FLOW_SUPPORT, SHARED_FLOW_SCENARIO_COVERAGE, UiSnapshot, semantic parity comparison helpers, and typed runtime event diagnostics.
When to Verify
Verification suits protocols with complex state machines or security-critical properties. Use Quint model checking for exhaustive state exploration. Use Lean proofs for mathematical guarantees. Use generative testing to validate implementations against models.
Unit tests suffice for simple, well-understood behavior. Do not over-invest in verification for straightforward code.
See Formal Verification Reference for the complete architecture documentation.
Writing Quint Specifications
Getting Started
Create a new specification in verification/quint/.
module protocol_example {
type State = { phase: str, value: int }
var state: State
action init = {
state' = { phase: "setup", value: 0 }
}
action increment(amount: int): bool = all {
amount > 0,
state' = { ...state, value: state.value + amount }
}
val nonNegative = state.value >= 0
}
Run quint typecheck to validate syntax. Run quint run to simulate execution.
Authority Model
Specifications should use AuthorityId for identity, not DeviceId. Model relational semantics without device-level details.
type AuthorityId = str
type Participant = { authority: AuthorityId, role: Role }
This aligns specifications with Aura's authority-centric design.
State Machine Design
Define clear phases with explicit transitions.
type Phase = Setup | Active | Completed | Failed
action transition(target: Phase): bool = all {
state.phase != Completed,
state.phase != Failed,
validTransition(state.phase, target),
state' = { ...state, phase: target }
}
Disallow transitions from terminal states. Validate transition legality explicitly.
Invariant Design
Define invariants before actions. Clear invariants guide action design.
val safetyInvariant = or {
state.phase != Failed,
hasRecoveryPath(state)
}
val progressInvariant = state.step < MAX_STEPS
Invariants should be checkable at every state. Avoid invariants that require execution history.
Harness Modules
Create harness modules for simulation and trace generation.
module harness_example {
import protocol_example.*
action register(): bool = init
action step(amount: int): bool = increment(amount)
action done(): bool = state.phase == Completed
}
Harnesses provide standardized entry points for tooling. They should emit semantic traces and invariants, not frontend-specific scripts or key sequences.
Model Checking Workflow
Type Checking
quint typecheck verification/quint/protocol_example.qnt
Type checking validates syntax and catches type errors. Run it before any other operation.
Simulation
quint run --main=harness_example verification/quint/protocol_example.qnt
Simulation executes random traces. It finds bugs quickly but does not provide exhaustive coverage.
Invariant Checking
quint run --invariant=safetyInvariant verification/quint/protocol_example.qnt
Invariant checking verifies properties hold across simulated traces.
Model Checking with Apalache
quint verify --max-steps=10 --invariant=safetyInvariant verification/quint/protocol_example.qnt
Apalache performs exhaustive model checking. It proves invariants hold for all reachable states up to the step bound.
Interpreting Violations
Violations produce counterexample traces. The trace shows the state sequence leading to the violated invariant.
[State 0] phase: Setup, value: 0
[State 1] phase: Active, value: 5
[State 2] phase: Active, value: -3 <- VIOLATION: nonNegative
Use counterexamples to identify specification bugs or missing preconditions.
Generative Testing Workflow
Generative testing validates Rust implementations against Quint models.
The Trust Chain
Quint Specification
│
▼ generates
ITF Traces
│
▼ replayed through
Rust Effect Handlers
│
▼ produces
Property Verdicts
Each link adds verification value. Specifications validate design. Traces validate reachability. Replay validates implementation.
Generating Semantic Traces
just quint-semantic-trace spec=verification/quint/harness/flows.qnt \
out=verification/quint/traces/harness_flows.itf.json
ITF traces capture semantic state sequences and non-deterministic choices.
These traces are model artifacts. Real TUI and web execution belongs to the
harness, which consumes the shared semantic scenario contract. Shared web/TUI parity assertions also run against the same UiSnapshot contract rather than renderer text or DOM structure.
Do not add direct Quint-to-TUI or Quint-to-browser execution paths. Quint should hand off semantic traces to the shared contract layer, then let the harness or simulator consume them through their own adapters.
For shared end-to-end flows, the harness contract is semantic and state-based. Do not introduce frontend-specific Quint replay formats that encode raw keypress sequences, browser selectors, or label-based button targeting. Those belong in driver adapters and diagnostics, not in the semantic trace contract.
Direct Conformance Testing
The recommended approach compares Rust behavior to Quint expected states.
#![allow(unused)] fn main() { use aura_simulator::quint::itf_loader::ITFLoader; #[test] fn test_matches_quint() { let trace = ITFLoader::load("trace.itf.json").unwrap(); for states in trace.states.windows(2) { let pre = State::from_quint(&states[0]).unwrap(); let action = states[1].meta.action.as_deref().unwrap(); let actual = apply_action(&pre, action).unwrap(); let expected = State::from_quint(&states[1]).unwrap(); assert_eq!(actual, expected); } } }
This tests production code directly. Quint serves as the single source of truth.
Generative Exploration
For state space exploration with Rust-driven non-determinism, use the action registry.
#![allow(unused)] fn main() { use aura_simulator::quint::action_registry::ActionRegistry; let mut registry = ActionRegistry::new(); registry.register("increment", Box::new(IncrementHandler)); let result = registry.execute("increment", ¶ms, &effects).await?; }
This approach requires handlers that re-implement Quint logic. Prefer direct conformance testing for new protocols.
ITF Trace Handling
Loading Traces
#![allow(unused)] fn main() { use aura_simulator::quint::itf_loader::ITFLoader; let trace = ITFLoader::load("trace.itf.json")?; for state in &trace.states { let index = state.meta.index; let action = state.meta.action.as_deref(); let picks = &state.meta.nondet_picks; } }
The loader parses ITF JSON into typed Rust structures.
Non-Deterministic Choices
ITF traces capture non-deterministic choices for reproducible replay.
{
"#meta": { "index": 3, "nondet_picks": { "leader": "alice" } }
}
The simulator injects these choices into RandomEffects to ensure deterministic replay.
State Mapping
Types implementing QuintMappable convert between Quint and Rust representations.
#![allow(unused)] fn main() { use aura_core::effects::quint::QuintMappable; let rust_state = State::from_quint(&quint_value)?; let quint_value = rust_state.to_quint(); }
Bidirectional mapping enables state comparison during replay.
Feeding Conformance Corpus
MBT traces should feed conformance testing.
Deriving Seeds
AURA_CONFORMANCE_ITF_TRACE=trace.itf.json cargo test conformance
ITF traces become inputs for native/WASM conformance lanes.
Coupling Model to Corpus
When Quint models change, regenerate traces and update the conformance corpus. This couples model evolution to test coverage.
quint run --out-itf=artifacts/traces/new_trace.itf.json verification/quint/updated_spec.qnt
just ci-conformance
See Testing Guide for corpus policy details.
The main repository policy gate for shared-flow drift is:
just ci-shared-flow-policy
This complements Quint and simulator checks by enforcing that shared real-runtime
scenarios still use the semantic contract and that the shared-flow support map in
aura-app remains consistent with the harness/frontend surfaces.
Telltale Verification Workflow in Aura
Use this workflow for choreography and simulator-level verification that depends on Telltale-derived checks.
1) Choreography compatibility gate (CI/tooling)
Run:
nix develop --command scripts/check/protocol-compat.sh --self-test
nix develop --command just ci-protocol-compat
This validates that known-compatible fixture pairs pass async subtyping checks. It confirms known-breaking fixture pairs fail as expected. It ensures changed .choreo files stay backward-compatible unless intentionally breaking.
Fixtures live in crates/aura-testkit/fixtures/protocol_compat/.
2) Macro-time coherence gate
Run:
nix develop --command cargo test -p aura-macros
This enforces compile-time coherence validation for choreographies, including negative compile-fail coverage.
3) Simulator invariant monitoring under injected faults
Run:
nix develop --command cargo test -p aura-simulator --test fault_invariant_monitor
This verifies that injected faults produce monitor-visible invariant violations (for example, NoFaults violations), and that a gate configured to require zero violations fails accordingly.
4) telltale-lean-bridge integration status
As of March 5, 2026, Aura includes telltale-lean-bridge as a workspace dependency and exposes it through aura-quint.
This adds direct access to upstream Lean runner and equivalence utilities from the Telltale project. It provides explicit schema and version linkage with upstream bridge contracts for cross-tool consistency. It also creates a cleaner path for future migration of local bridge helpers to upstream bridge APIs.
The aura-quint crate re-exports the upstream crate as upstream_telltale_lean_bridge. Call aura_quint::upstream_telltale_lean_bridge_schema_version() to get the upstream schema version. CI lanes remain just ci-lean-quint-bridge and just ci-simulator-telltale-parity.
5) aura-testkit Lean verification API migration (March 5, 2026)
As of March 5, 2026, legacy Lean verification compatibility types were removed from
aura_testkit::verification and aura_testkit::verification::lean_oracle.
Use the canonical full-fidelity types and methods.
| Legacy Type | Canonical Type |
|---|---|
Fact | LeanFact |
ComparePolicy | LeanComparePolicy |
TimeStamp | LeanCompareTimeStamp (compare payloads) or LeanTimeStamp (journal facts) |
Ordering | LeanTimestampOrdering |
FlowChargeInput/FlowChargeResult | LeanFlowChargeInput/LeanFlowChargeResult |
TimestampCompareInput/TimestampCompareResult | LeanTimestampCompareInput/LeanTimestampCompareResult |
| Legacy Method | Canonical Method |
|---|---|
verify_merge | verify_journal_merge |
verify_reduce | verify_journal_reduce |
verify_charge | verify_flow_charge |
verify_compare | verify_timestamp_compare |
Import Lean verification payload types from aura_testkit::verification which re-exports from lean_types. Construct structured journals with LeanJournal and LeanNamespace. Update tests to compare LeanTimestampOrdering values directly.
Lean-Quint Bridge
The bridge connects Quint model checking with Telltale and Lean proof artifacts. It enables exporting Quint session models to a stable interchange format, importing Telltale and Lean properties back into Quint harnesses, and running cross-validation to detect divergence early in CI.
Operator Workflow
Run the bridge lane:
just ci-lean-quint-bridge
Inspect outputs at artifacts/lean-quint-bridge/bridge.log, artifacts/lean-quint-bridge/bridge_discrepancy_report.json, and artifacts/lean-quint-bridge/report.json.
Run the simulator telltale parity lane:
just ci-simulator-telltale-parity
Inspect output at artifacts/telltale-parity/report.json.
Data Contract
aura-quint defines a versioned interchange schema for bridge workflows.
| Type | Purpose |
|---|---|
BridgeBundleV1 | Top-level bundle with schema_version = "aura.lean-quint-bridge.v1" |
SessionTypeInterchangeV1 | Session graph exchange |
PropertyInterchangeV1 | Quint, Telltale, and Lean property exchange |
ProofCertificateV1 | Proof or model-check evidence |
Use this schema as the canonical data contract when exporting Quint sessions to Telltale formats or importing Telltale and Lean properties into Quint harnesses.
Export Workflow
Export moves session models from Quint to Telltale format.
- Parse Quint JSON IR with
parse_quint_modules(...) - Build the bridge bundle with
export_quint_to_telltale_bundle(...) - Validate structural correctness with
validate_export_bundle(...)
Import Workflow
Import brings Telltale and Lean properties back into Quint harnesses.
- Select importable properties with
parse_telltale_properties(...) - Generate Quint invariant module text with
generate_quint_invariant_module(...) - Map certificates into Quint assertion comments with
map_certificates_to_quint_assertions(...)
Cross-Validation Workflow
Cross-validation detects proof and model divergence. Use run_cross_validation(...) from aura-quint to execute Quint checks through a QuintModelCheckExecutor, compare outcomes to bridge proof certificates, and emit a CrossValidationReport with explicit discrepancy entries.
Run cross-validation in CI:
just ci-lean-quint-bridge
This command produces artifacts under artifacts/lean-quint-bridge/ including bridge.log and report.json.
Handling Discrepancies
When cross-validation reports discrepancies, follow these steps. First confirm the property identity mapping (property_id) between model and proof pipelines. Then re-run the failing property in Quint and capture the trace or counterexample. Next re-check proof certificate assumptions against the current protocol model. Do not merge until the mismatch is resolved or explicitly justified.
For telltale parity mismatches, read comparison_classification, first_mismatch_surface, and first_mismatch_step_index first. Re-run the failing lane with the same scenario and seed. Confirm that required surfaces (observable, scheduler_step, effect) were captured before examining envelope differences.
Lean Proof Development
Adding Theorems
Create or extend modules in verification/lean/Aura/Proofs/.
theorem new_property : ∀ s : State, isValid s → preservesInvariant s := by
intro s h
simp [isValid, preservesInvariant] at *
exact h
Use Lean 4 tactic mode for proofs.
Using Claims Bundles
Access related theorems through claims bundles.
import Aura.Proofs.Consensus
#check Aura.Consensus.Validity.validityClaims.commit_has_threshold
Bundles organize proofs by domain.
Working with Axioms
Cryptographic assumptions appear in Assumptions.lean.
axiom frost_threshold_unforgeability : ...
Proofs depending on axioms are sound under standard hardness assumptions. Document axiom dependencies clearly.
Building Proofs
cd verification/lean
lake build
The build succeeds only if all proofs complete without sorry.
Checking Status
just lean-status
This reports per-module proof status including incomplete proofs.
Running Verification
Quint Commands
quint typecheck spec.qnt # Type check
quint run --main=harness spec.qnt # Simulate
quint run --invariant=inv spec.qnt # Check invariant
quint verify --max-steps=10 spec.qnt # Model check
Lean Commands
just verify-lean # Build proofs
just lean-status # Check status
just test-differential # Rust vs Lean tests
Full Verification
just verify-all
This runs Quint model checking, Lean proof building, and conformance tests.
Best Practices
Start with invariants. Define properties before implementing actions. Clear invariants guide design.
Use unique variant names. Quint requires globally unique sum type variants. Prefix with domain names.
Test harnesses separately. Verify harness modules parse before integrating with the simulator.
Start with short traces. Debug action mappings with 3-5 step traces before exhaustive exploration.
Isolate properties. Test one property at a time during development. Combine for coverage testing.
Adding or Updating Invariants
When adding or modifying invariants, follow this workflow to maintain traceability across docs, tests, and proofs.
- Add or update the invariant under
## Invariantsin the crate'sARCHITECTURE.md. - Add a detailed specification section in the same file with invariant name, enforcement locus, failure mode, and verification hooks.
- Use canonical
InvariantXxxnaming for traceability across docs, tests, and proofs. - Add or update tests and simulator scenarios that detect violations.
- Update the traceability matrix in Project Structure if the invariant is cross-crate or contract-level.
Formal and model checks should reference the same canonical names listed in the traceability matrix.
Quint-Lean Correspondence
This section maps Quint model invariants to Lean theorem proofs, providing traceability between model checking and formal proofs.
Types Correspondence
| Quint Type | Lean Type | Rust Type |
|---|---|---|
ConsensusId | Aura.Domain.Consensus.Types.ConsensusId | consensus::types::ConsensusId |
ResultId | Aura.Domain.Consensus.Types.ResultId | consensus::types::ResultId |
PrestateHash | Aura.Domain.Consensus.Types.PrestateHash | consensus::types::PrestateHash |
AuthorityId | Aura.Domain.Consensus.Types.AuthorityId | core::AuthorityId |
ShareData | Aura.Domain.Consensus.Types.ShareData | consensus::types::SignatureShare |
ThresholdSignature | Aura.Domain.Consensus.Types.ThresholdSignature | consensus::types::ThresholdSignature |
CommitFact | Aura.Domain.Consensus.Types.CommitFact | consensus::types::CommitFact |
WitnessVote | Aura.Domain.Consensus.Types.WitnessVote | consensus::types::WitnessVote |
Evidence | Aura.Domain.Consensus.Types.Evidence | consensus::types::Evidence |
Invariant-Theorem Correspondence
Agreement Properties
| Quint Invariant | Lean Theorem | Status |
|---|---|---|
InvariantUniqueCommitPerInstance | Aura.Proofs.Consensus.Agreement.agreement | proven |
InvariantUniqueCommitPerInstance | Aura.Proofs.Consensus.Agreement.unique_commit | proven |
| - | Aura.Proofs.Consensus.Agreement.commit_determinism | proven |
Validity Properties
| Quint Invariant | Lean Theorem | Status |
|---|---|---|
InvariantCommitRequiresThreshold | Aura.Proofs.Consensus.Validity.commit_has_threshold | proven |
InvariantSignatureBindsToCommitFact | Aura.Proofs.Consensus.Validity.validity | proven |
| - | Aura.Proofs.Consensus.Validity.distinct_signers | proven |
| - | Aura.Proofs.Consensus.Validity.prestate_binding_unique | proven |
| - | Aura.Proofs.Consensus.Validity.honest_participation | proven |
| - | Aura.Proofs.Consensus.Validity.threshold_unforgeability | axiom |
FROST Integration Properties
| Quint Invariant | Lean Theorem | Status |
|---|---|---|
InvariantSignatureThreshold | Aura.Proofs.Consensus.Frost.aggregation_threshold | proven |
| - | Aura.Proofs.Consensus.Frost.share_session_consistency | proven |
| - | Aura.Proofs.Consensus.Frost.share_result_consistency | proven |
| - | Aura.Proofs.Consensus.Frost.distinct_signers | proven |
| - | Aura.Proofs.Consensus.Frost.share_binding | proven |
Evidence CRDT Properties
| Quint Invariant | Lean Theorem | Status |
|---|---|---|
| - | Aura.Proofs.Consensus.Evidence.merge_comm_votes | proven |
| - | Aura.Proofs.Consensus.Evidence.merge_assoc_votes | proven |
| - | Aura.Proofs.Consensus.Evidence.merge_idem | proven |
| - | Aura.Proofs.Consensus.Evidence.merge_preserves_commit | proven |
| - | Aura.Proofs.Consensus.Evidence.commit_monotonic | proven |
Equivocation Detection Properties
| Quint Invariant | Lean Theorem | Status |
|---|---|---|
InvariantEquivocationDetected | Aura.Proofs.Consensus.Equivocation.detection_soundness | proven |
InvariantEquivocationDetected | Aura.Proofs.Consensus.Equivocation.detection_completeness | proven |
InvariantEquivocatorsExcluded | Aura.Proofs.Consensus.Equivocation.exclusion_correctness | proven |
InvariantHonestMajorityCanCommit | Aura.Proofs.Consensus.Equivocation.honest_never_detected | proven |
| - | Aura.Proofs.Consensus.Equivocation.verified_proof_sound | proven |
Byzantine Tolerance (Adversary Module)
| Quint Invariant | Lean Theorem | Status |
|---|---|---|
InvariantByzantineThreshold | Aura.Proofs.Consensus.Adversary.adversaryClaims.byzantine_cannot_forge | claim |
InvariantEquivocationDetected | Aura.Proofs.Consensus.Adversary.adversaryClaims.equivocation_detectable | claim |
InvariantHonestMajorityCanCommit | Aura.Proofs.Consensus.Adversary.adversaryClaims.honest_majority_sufficient | claim |
InvariantEquivocatorsExcluded | Aura.Proofs.Consensus.Adversary.adversaryClaims.equivocators_excluded | claim |
InvariantCompromisedNoncesExcluded | - | Quint only |
Liveness Properties
| Quint Property | Lean Support | Notes |
|---|---|---|
InvariantProgressUnderSynchrony | Aura.Proofs.Consensus.Liveness.livenessClaims.terminationUnderSynchrony | axiom |
InvariantByzantineTolerance | byzantine_threshold | axiom |
FastPathProgressCheck | Aura.Proofs.Consensus.Liveness.livenessClaims.fastPathBound | axiom |
SlowPathProgressCheck | Aura.Proofs.Consensus.Liveness.livenessClaims.fallbackBound | axiom |
NoDeadlock | Aura.Proofs.Consensus.Liveness.livenessClaims.noDeadlock | axiom |
InvariantRetryBound | - | Quint model checking only |
Module Correspondence
| Lean Module | Quint File | What It Proves |
|---|---|---|
Proofs.ContextIsolation | authorization.qnt, leakage.qnt | Context separation and bridge authorization |
Proofs.Consensus.Agreement | consensus/core.qnt | Agreement safety (unique commits) |
Proofs.Consensus.Evidence | consensus/core.qnt | CRDT semilattice properties |
Proofs.Consensus.Frost | consensus/frost.qnt | Threshold signature correctness |
Proofs.Consensus.Liveness | consensus/liveness.qnt | Synchrony model axioms |
Proofs.Consensus.Adversary | consensus/adversary.qnt | Byzantine tolerance bounds |
Proofs.Consensus.Equivocation | consensus/adversary.qnt | Detection soundness/completeness |
Related Documentation
See Formal Verification Reference for architecture details. See Simulation Guide for trace replay. See Testing Guide for conformance testing. See Project Structure for the invariant index and traceability matrix.
System Internals Guide
This guide covers deep system patterns for contributors working on Aura core. Use it when you need to understand guard chain internals, service layer patterns, core types, and reactive scheduling.
1. Guard Chain Internals
The guard chain coordinates authorization, flow budgets, and journal effects in strict sequence. See Authorization for the full specification.
Three-Phase Pattern
Guards are pure: evaluation runs synchronously over a prepared GuardSnapshot and yields EffectCommand items that an async interpreter executes.
#![allow(unused)] fn main() { // Phase 1: Authorization via Biscuit + policy (async, cached) let token = effects.verify_biscuit(&request.token).await?; let capabilities = token.capabilities(); // Phase 2: Prepare snapshot and evaluate guards (sync) let snapshot = GuardSnapshot { capabilities, flow_budget: current_budget, leakage_budget: current_leakage, ..Default::default() }; let commands = guard_chain.evaluate(&snapshot, &request)?; // Phase 3: Execute commands (async) for command in commands { match command { EffectCommand::ChargeBudget { cost } => { budget_handler.charge(cost).await?; } EffectCommand::RecordLeakage { budget } => { leakage_handler.record(budget).await?; } EffectCommand::CommitJournal { facts } => { journal_handler.commit(facts).await?; } EffectCommand::SendTransport { message } => { transport_handler.send(message).await?; } } } }
No transport observable occurs until the interpreter executes commands in order.
Guard Chain Sequence
The guards execute in this order:
- CapabilityGuard: Validates Biscuit token capabilities
- FlowBudgetGuard: Checks and charges flow budget
- LeakageTracker: Records privacy leakage
- JournalCoupler: Commits facts to journal
- TransportEffects: Sends messages
Security Patterns
Privacy Budget Enforcement:
#![allow(unused)] fn main() { // Secure default: denies undefined budgets let tracker = LeakageTracker::new(); // Backward compatibility: allows undefined budgets let tracker = LeakageTracker::legacy_permissive(); // Configurable default let tracker = LeakageTracker::with_undefined_policy(DefaultBudget(1000)); }
2. Service Layer Patterns
Domain crates define stateless handlers that take effect references per-call. The agent layer wraps these with services that manage RwLock access.
Domain Handler (Layer 2-5)
#![allow(unused)] fn main() { // In domain crate (e.g., aura-chat/src/service.rs) pub struct ChatFactService; impl ChatFactService { pub fn new() -> Self { Self } pub async fn send_message<E>( &self, effects: &E, channel_id: ChannelId, content: String, ) -> Result<MessageId> where E: StorageEffects + RandomEffects + PhysicalTimeEffects { let message_id = effects.random_uuid().await; let timestamp = effects.physical_time().await?; // ... domain logic using effects Ok(message_id) } } }
Agent Service Wrapper (Layer 6)
#![allow(unused)] fn main() { // In aura-agent/src/handlers/chat_service.rs pub struct ChatService { handler: ChatFactService, effects: Arc<RwLock<AuraEffectSystem>>, } impl ChatService { pub fn new(effects: Arc<RwLock<AuraEffectSystem>>) -> Self { Self { handler: ChatFactService::new(), effects } } pub async fn send_message( &self, channel_id: ChannelId, content: String, ) -> AgentResult<MessageId> { let effects = self.effects.read().await; self.handler.send_message(&*effects, channel_id, content) .await .map_err(Into::into) } } }
Agent API Exposure
#![allow(unused)] fn main() { // In aura-agent/src/core/api.rs impl AuraAgent { pub fn chat_service(&self) -> ChatService { ChatService::new(self.runtime.effects()) } } }
Benefits:
- Domain crate stays pure (no tokio/RwLock)
- Testable with mock effects
- Consistent pattern across crates
3. Type Reference
ProtocolType
Canonical definition in aura-core. All crates re-export this definition.
#![allow(unused)] fn main() { pub enum ProtocolType { Dkd, // Deterministic Key Derivation Counter, // Counter reservation protocol Resharing, // Key resharing for threshold updates Locking, // Resource locking protocol Recovery, // Account recovery protocol Compaction, // Ledger compaction protocol } }
SessionStatus
Lifecycle order:
Initializing- Session initializingActive- Session executingWaiting- Waiting for participant responsesCompleted- Completed successfullyFailed- Failed with errorExpired- Expired due to timeoutTimedOut- Timed out during executionCancelled- Was cancelled
TimeStamp Domains
TimeStamp in aura-core is the only time type for new facts and public APIs.
| Domain | Effect Trait | Primary Use |
|---|---|---|
PhysicalClock | PhysicalTimeEffects | Wall time: cooldowns, receipts, liveness |
LogicalClock | LogicalClockEffects | Causal ordering: CRDT merge, happens-before |
OrderClock | OrderClockEffects | Deterministic ordering without timing leakage |
Range | PhysicalTimeEffects + policy | Validity windows with bounded skew |
ProvenancedTime | TimeComparison | Attested timestamps for consensus |
Guidelines:
- Use effect traits for all time reads (no
SystemTime::now()) - Use the narrowest domain that satisfies the requirement
- Compare mixed domains with
TimeStamp::compare(policy) - Persist
TimeStampvalues directly in facts
Anti-patterns:
- Mixing clock domains in one sort path without explicit policy
- Using
PhysicalClockfor privacy-sensitive ordering - Using UUID or insertion order as time proxy
- Exposing
SystemTimeor chrono types in interfaces
Capability System Layering
The capability system uses multiple layers:
- Canonical types in
aura-core: Lightweight references - Authorization layer (
aura-authorization): Policy enforcement - Storage layer (
aura-store): Capability-based access control
Clear conversion paths enable inter-layer communication.
4. Reactive Scheduling
The ReactiveScheduler in aura-agent/src/reactive/ processes journal facts and emits application signals.
Signal System Overview
#![allow(unused)] fn main() { // Application signals pub const CHAT_SIGNAL: &str = "chat"; pub const CONTACTS_SIGNAL: &str = "contacts"; pub const CHANNELS_SIGNAL: &str = "channels"; pub const RECOVERY_SIGNAL: &str = "recovery"; }
The scheduler:
- Subscribes to journal fact streams
- Reduces facts to view state
- Emits signals when state changes
- TUI/CLI components subscribe to signals
TUI Reactive State
The TUI uses futures-signals for fine-grained reactive state management.
Note: The reactive architecture pattern below represents the target design for TUI state management. Implementation status varies by view.
Signal Types
#![allow(unused)] fn main() { use futures_signals::signal::Mutable; use futures_signals::signal_vec::MutableVec; // Single reactive value let count = Mutable::new(0); count.set(5); let value = count.get_cloned(); // Reactive collection let items = MutableVec::new(); items.lock_mut().push_cloned("item1"); }
View Pattern
#![allow(unused)] fn main() { pub struct ChatView { channels: MutableVec<Channel>, messages: MutableVec<Message>, selected_channel: Mutable<Option<String>>, } impl ChatView { // Synchronous delta application pub fn apply_delta(&self, delta: ChatDelta) { match delta { ChatDelta::ChannelAdded { channel } => { self.channels.lock_mut().push_cloned(channel); // Signals automatically notify subscribers } ChatDelta::MessageReceived { channel_id, message } => { if self.selected_channel.get_cloned() == Some(channel_id) { self.messages.lock_mut().push_cloned(message); } } } } } }
Best Practices
- Delta application should be synchronous (not async)
- Use
.get_cloned()for reading,.set()for mutations - Never hold lock guards across await points
- Use derived signals for computed values
5. Policy Compliance
Application code must follow policies defined in Project Structure.
Impure Function Usage
All time, randomness, filesystem, and network operations must flow through effect traits.
Forbidden:
#![allow(unused)] fn main() { // Direct system calls break simulation and WASM let now = SystemTime::now(); let random = thread_rng().gen(); let file = File::open("path")?; }
Required:
#![allow(unused)] fn main() { // Use effect traits let now = effects.physical_time().await?; let random = effects.random_bytes(32).await?; let data = effects.read_storage("key").await?; }
Serialization
- Wire protocols and facts: DAG-CBOR via
aura_core::util::serialization - User-facing configs: JSON allowed
- Debug output: JSON allowed
Architectural Validation
Run just check-arch before submitting changes. The checker validates:
- Layer boundaries
- Effect trait placement
- Impure function routing
- Guard chain integrity
6. Architecture Compliance Checklist
- Layer dependencies flow downward only
-
Effect traits defined in
aura-coreonly -
Infrastructure effects implemented in
aura-effects - Application effects in domain crates
- No direct impure function usage outside effect implementations
-
All async functions propagate
EffectContext -
Production handlers are stateless, test handlers in
aura-testkit - Guard chain sequence respected
Related Documentation
- Effect System - Effect specification
- Runtime - Runtime specification
- Authorization - Guard chain specification
- System Architecture - Layer boundaries
- CLI and TUI - Terminal specification
Distributed Maintenance Guide
This guide covers practical workflows for the Maintenance and OTA (Over-the-Air) update system in Aura. Use it for snapshots, cache invalidation, and distributed upgrades.
For the maintenance architecture specification, see Distributed Maintenance Architecture.
Maintenance Philosophy
The maintenance system is built on three key principles:
- Coordinated Operations - Threshold approval is used where the chosen maintenance scope actually has a quorum or authority set that can approve
- Epoch Fencing - Hard fork upgrades may use identity epochs for safe coordination, but only inside scopes that own that fence
- Journal-Based Facts - All maintenance events are replicated through the journal CRDT
The system supports snapshots for garbage collection, cache management, and both soft and hard fork upgrades.
Maintenance Events
The maintenance service publishes events to the journal as facts. These events are replicated across all replicas and interpreted deterministically.
Event Types
The system defines several event families:
SnapshotProposed marks the beginning of a snapshot operation. It contains the proposal identifier, proposer authority, target epoch, and state digest of the candidate snapshot.
SnapshotCompleted records a successful snapshot. It includes the accepted proposal identifier, finalized snapshot payload, participating authorities, and threshold signature attesting to the snapshot.
CacheInvalidated signals cache invalidation. It specifies which cache keys must be refreshed and the earliest identity epoch the cache entry remains valid for.
ReleaseDistribution facts announce release declarations, build certificates, and artifact availability.
ReleasePolicy facts announce discovery, sharing, and activation policy publications.
UpgradeExecution facts announce scoped staging, residency changes, transition changes, cutover results, partition outcomes, and rollback execution.
AdminReplacement announces an administrator change. This allows users to fork away from a malicious admin by tracking previous and new administrators with activation epoch.
Snapshot Protocol
The snapshot protocol coordinates garbage collection with threshold approval. It implements writer fencing to ensure consistent snapshot capture across all devices.
Snapshot Workflow
The snapshot process follows five steps:
- Propose snapshot at target epoch with state digest
- Activate writer fence to block concurrent writes
- Capture state and verify digest
- Collect M-of-N threshold approvals
- Commit snapshot and clean obsolete facts
Basic Snapshot Operation
#![allow(unused)] fn main() { use aura_sync::services::{MaintenanceService, MaintenanceServiceConfig}; use aura_core::{Epoch, Hash32}; async fn propose_snapshot( service: &MaintenanceService, authority_id: aura_core::AuthorityId, target_epoch: Epoch, state_digest: Hash32, ) -> Result<(), Box<dyn std::error::Error>> { // Propose snapshot at target epoch service .propose_snapshot(authority_id, target_epoch, state_digest) .await?; // Writer fence is now active - all concurrent writes blocked // Collect approvals from M-of-N authorities // Once threshold reached, commit service.commit_snapshot().await?; Ok(()) } }
Snapshots provide deterministic checkpoints of authority state at specific epochs. This enables garbage collection of obsolete facts while maintaining verifiable state recovery.
Snapshot Proposal
#![allow(unused)] fn main() { use aura_sync::services::MaintenanceService; use aura_core::{AuthorityId, Epoch, Hash32}; async fn snapshot_workflow( service: &MaintenanceService, authority_id: AuthorityId, ) -> Result<(), Box<dyn std::error::Error>> { // Determine target epoch and compute state digest let target_epoch = 100; let current_state = service.get_current_state().await?; let state_digest = current_state.compute_digest(); // Propose snapshot service .propose_snapshot(authority_id, Epoch::new(target_epoch), state_digest) .await?; // Wait for other authorities to activate fence // and collect approvals Ok(()) } }
Proposals include the proposer authority identifier, unique proposal ID, target epoch, and canonical state digest. All participants must agree on the digest before committing.
Writer Fence
The writer fence blocks all writes during snapshot capture. This prevents concurrent modifications that could invalidate the snapshot digest.
#![allow(unused)] fn main() { use aura_sync::services::MaintenanceService; async fn capture_with_fence( service: &MaintenanceService, ) -> Result<(), Box<dyn std::error::Error>> { // Writer fence is automatically activated by snapshot proposal // All write operations will be blocked or queued // Capture state atomically let snapshot = service.capture_snapshot().await?; // Once snapshot is committed, fence is released service.commit_snapshot().await?; Ok(()) } }
Fence enforcement is implicit in snapshot proposal. The protocol guarantees no conflicting writes occur during snapshot capture.
Approval Collection
#![allow(unused)] fn main() { use aura_sync::services::MaintenanceService; async fn collect_approvals( service: &MaintenanceService, threshold: usize, ) -> Result<(), Box<dyn std::error::Error>> { // Get pending snapshot proposals let proposals = service.pending_snapshots().await?; for proposal in proposals { // Verify state digest let current_state = service.get_current_state().await?; let digest = current_state.compute_digest(); if digest == proposal.state_digest { // Approve the snapshot service.approve_snapshot(&proposal.proposal_id).await?; } } Ok(()) } }
Each device verifies the state digest independently and approves if correct. The system collects approvals until threshold is reached.
Snapshot Commitment
#![allow(unused)] fn main() { use aura_sync::services::MaintenanceService; async fn finalize_snapshot( service: &MaintenanceService, ) -> Result<(), Box<dyn std::error::Error>> { // Commit snapshot once threshold reached service.commit_snapshot().await?; // This records SnapshotCompleted fact to journal // All devices deterministically reduce this fact // Obsolete facts before this epoch can be garbage collected Ok(()) } }
Commitment publishes SnapshotCompleted fact to the journal. All devices deterministically reduce this fact to the same relational state. Facts older than the snapshot epoch can then be safely discarded.
OTA (Over-the-Air) Upgrade Protocol
Aura OTA is split into two layers:
- release distribution: manifests, artifacts, and build certificates spread through Aura storage and anti-entropy
- scoped activation: each device, authority, context, or managed quorum decides when a staged release may activate
There is no network-wide authoritative cutover phase for the whole Aura network. Hard cutover is valid only inside a scope that actually has agreement or a legitimate fence.
Upgrade Types
Soft Fork upgrades are compatibility-preserving. Old and new code can interoperate while one scope is in ReleaseResidency::Coexisting.
Hard Fork upgrades are scope-bound incompatibility transitions. They reject incompatible new sessions after local cutover and require explicit in-flight handling for old sessions. A hard fork may use threshold approval or epoch fencing, but only if the chosen scope actually owns that mechanism.
Basic Upgrade Operation
#![allow(unused)] fn main() { use aura_maintenance::AuraReleaseActivationPolicy; use aura_sync::services::{ActivationCandidate, OtaPolicyEvaluator}; async fn evaluate_activation( policy: &AuraReleaseActivationPolicy, candidate: &ActivationCandidate<'_>, ) { // Activation is local or scope-bound. // Discovery and sharing do not imply activation. // The evaluator checks trust, compatibility, staged artifacts, // health gates, threshold approval, and scope-owned fences. let _decision = OtaPolicyEvaluator::new().evaluate_activation(policy, candidate); } }
If policy enables it, suggested_activation_time_unix_ms acts only as a local "not before" hint against the local clock. It is advisory metadata, not a global synchronization fence.
Soft Fork Workflow
#![allow(unused)] fn main() { use aura_agent::runtime::services::OtaManager; use aura_maintenance::{AuraActivationScope, AuraCompatibilityClass}; use aura_sync::services::InFlightIncompatibilityAction; async fn soft_fork_workflow( manager: &OtaManager, scope: AuraActivationScope, ) -> Result<(), Box<dyn std::error::Error>> { let plan = manager .begin_scoped_cutover(&scope, InFlightIncompatibilityAction::Drain, false) .await?; assert!(!plan.partition_required); Ok(()) } }
Soft forks do not require a globally shared instant. Each scope moves from legacy-only residency to coexistence and then to target-only residency based on its own evidence and policy.
Hard Fork Workflow
#![allow(unused)] fn main() { use aura_agent::runtime::services::OtaManager; use aura_maintenance::AuraActivationScope; use aura_sync::services::InFlightIncompatibilityAction; async fn execute_managed_quorum_cutover( manager: &OtaManager, scope: AuraActivationScope, ) -> Result<(), Box<dyn std::error::Error>> { let plan = manager .begin_scoped_cutover(&scope, InFlightIncompatibilityAction::Delegate, true) .await?; assert!(plan.partition_required || plan.in_flight == InFlightIncompatibilityAction::Delegate); Ok(()) } }
For hard forks, the operator must define:
- the activation scope
- the compatibility class
- how incompatible in-flight sessions are handled: drain, abort, or delegate
- whether threshold approval or an epoch fence is actually available in that scope
If post-cutover checks fail, rollback is explicit and deterministic.
Managed Quorum Approval Runbook
Use managed quorum cutover only when the scope has an explicit participant set. Record approval from every participant in the quorum before starting cutover. Reject approval from authorities that are not members of that scope.
If one participant has not approved, keep the scope waiting for cutover evidence. Do not begin launcher activation for that scope. Resolve membership or policy disagreement before retrying.
Failed Rollout Runbook
Check the failure classification before acting. AuraUpgradeFailureClass::HealthGateFailed means the new release started and failed local verification. AuraUpgradeFailureClass::LauncherActivationFailed means the launcher handoff failed before healthy activation.
If policy uses AuraRollbackPreference::Automatic, allow the queued rollback to execute and confirm the scope returns to legacy-only residency with an idle transition state. If policy uses AuraRollbackPreference::ManualApproval, keep the scope failed and require operator approval before rollback.
Revoked Release Runbook
Treat a revoked staged release differently from a revoked active release. If the target release is only staged, cancel the staged scope and remove it from activation consideration. Do not proceed to cutover for that scope.
If the revoked release is already active, follow the configured rollback preference. Automatic rollback should queue a rollback to the prior staged release. Manual rollback should leave the scope failed until an operator approves the revert path.
Partition Response Runbook
If SessionCompatibilityPlan.partition_required is true, assume incompatible peers may separate cleanly rather than interoperate. Stop admitting incompatible new sessions in that scope. Drain, abort, or delegate in-flight sessions according to the recorded incompatibility action.
Record partition observations with the associated failure classification and scope. This keeps rollback and peer-partition handling auditable.
Cache Management
The maintenance system coordinates cache invalidation across all devices. Cache invalidation facts specify which keys must be refreshed and the epoch where the cache entry is no longer valid.
Cache Invalidation
#![allow(unused)] fn main() { use aura_sync::services::MaintenanceService; use aura_core::Epoch; async fn invalidate_cache( service: &MaintenanceService, keys: Vec<String>, epoch_floor: Epoch, ) -> Result<(), Box<dyn std::error::Error>> { // Publish cache invalidation fact service.invalidate_cache_keys(keys, epoch_floor).await?; // All devices receive fact through journal // Each device deterministically invalidates matching keys Ok(()) } }
Cache invalidation facts are replicated through the journal. All devices apply the invalidation deterministically based on epoch and key matching.
Cache Query Patterns
#![allow(unused)] fn main() { use aura_sync::services::MaintenanceService; use aura_core::Epoch; async fn query_cache( service: &MaintenanceService, key: &str, epoch: Epoch, ) -> Result<Option<Vec<u8>>, Box<dyn std::error::Error>> { // Check if cache entry is valid at epoch if service.is_cache_valid(key, epoch).await? { service.get_cached(key).await } else { // Cache invalidated - fetch fresh let value = service.fetch_fresh(key).await?; service.cache_set(key, value.clone(), epoch).await?; Ok(Some(value)) } } }
Cache validity is epoch-based. Entries are valid up to their invalidation epoch. After that, fresh data must be fetched and cached at the new epoch.
Configuration and Best Practices
Service Configuration
#![allow(unused)] fn main() { use aura_sync::services::{MaintenanceService, MaintenanceServiceConfig}; use aura_sync::protocols::{SnapshotConfig, OTAConfig}; use std::time::Duration; fn create_maintenance_service() -> Result<MaintenanceService, Box<dyn std::error::Error>> { let config = MaintenanceServiceConfig { snapshot: SnapshotConfig { proposal_timeout: Duration::from_secs(300), approval_timeout: Duration::from_secs(600), max_proposals: 10, }, ota: OTAConfig { readiness_timeout: Duration::from_secs(3600), max_pending: 5, soft_fork_auto_activate: true, }, cache: Default::default(), }; Ok(MaintenanceService::new(config)?) } }
Configuration controls timeouts, limits, and behavior. Snapshot timeouts should be shorter than OTA staging and activation timeouts since snapshots are more frequent. OTA policies should separately configure discovery, sharing, and activation rather than bundling them into one setting.
Snapshot Best Practices
Keep snapshots frequent but not excessive. Snapshot every 100-500 epochs depending on journal size. Too frequent snapshots create overhead. Too infrequent snapshots reduce garbage collection effectiveness.
Always verify state digest before approving. Use canonical serialization for digest computation. Record all snapshots in the journal for audit trail.
Upgrade Best Practices
Plan upgrades carefully. Soft forks can be deployed flexibly inside one scope. Hard forks require a clear activation scope, an explicit compatibility class, and an operator decision about whether in-flight incompatible sessions drain, abort, or delegate.
Always test upgrades in simulation before deployment. Use threshold approval or epoch fences only in scopes that actually own those mechanisms. Treat suggested_activation_time_unix_ms as advisory rollout metadata, not as a coordination primitive.
Include rollback procedures for hard forks. Document migration paths for state format changes and keep launcher activation/rollback steps separate from the running runtime.
Cache Best Practices
Invalidate cache conservatively. Over-invalidation reduces performance. Under-invalidation risks stale data.
Use epoch floors to scope invalidation. Invalidate only keys that actually changed at that epoch.
Monitor cache hit rates. Low hit rates indicate invalidation is too aggressive.
Monitoring and Debugging
Snapshot Status
#![allow(unused)] fn main() { use aura_sync::services::MaintenanceService; async fn monitor_snapshots( service: &MaintenanceService, ) -> Result<(), Box<dyn std::error::Error>> { let status = service.snapshot_status().await?; println!("Last snapshot: epoch {}", status.last_snapshot_epoch); println!("Pending proposals: {}", status.pending_proposals); println!("Writer fence active: {}", status.fence_active); Ok(()) } }
Check snapshot status regularly to ensure the snapshot cycle is healthy. Long intervals between snapshots may indicate approval delays.
Upgrade Status
#![allow(unused)] fn main() { use aura_sync::services::MaintenanceService; async fn monitor_upgrades( service: &MaintenanceService, ) -> Result<(), Box<dyn std::error::Error>> { let status = service.upgrade_status().await?; for upgrade in status.active_upgrades { println!("Upgrade: version {}", upgrade.version); println!(" Ready devices: {}/{}", upgrade.ready_count, upgrade.total_devices); println!(" Threshold: {}", upgrade.threshold); } Ok(()) } }
Monitor upgrade progress to catch devices that are not ready. Missing devices may need manual intervention.
Integration with Choreography
The maintenance system integrates with the choreography runtime for threshold approval ceremonies. Snapshot and upgrade proposals are published to the journal where choreography protocols can coordinate approval.
The maintenance service publishes events as facts. Choreography protocols subscribe to these facts and coordinate the necessary approvals through their own message flows.
Summary
The Maintenance and OTA system provides coordinated maintenance operations with threshold approval and epoch fencing where those mechanisms actually exist. Snapshots enable garbage collection with writer fencing. OTA release distribution is eventual, while activation is local or scope-bound. Cache invalidation is replicated through the journal for consistency.
Use snapshots regularly for garbage collection. Plan upgrades carefully with sufficient notice for hard forks. Test all upgrades in simulation. Monitor snapshot and upgrade cycles to ensure system health.
Implementation References
- Maintenance Service:
aura-sync/src/services/maintenance.rs - Snapshot Protocol:
aura-sync/src/protocols/snapshots.rs - OTA Protocol:
aura-sync/src/protocols/ota.rs - Cache Management:
aura-sync/src/infrastructure/cache_manager.rs - Integration Examples:
aura-agent/src/handlers/maintenance.rs
UX Flow Coverage Report
This document tracks end-to-end UX coverage for Aura's runtime harness scenarios across TUI and web surfaces.
Coverage Boundary Statement
UX flow coverage validates user-visible behavior and interaction wiring through runtime harness scenarios. It does not replace protocol conformance, theorem proofs, or differential parity lanes. Use this report for UI/product flow traceability and regression targeting.
Summary Metrics
| Metric | Count |
|---|---|
| Harness UX Scenarios | 13 |
| Parity-Critical Scenarios (TUI + Web) | 11 |
| Mixed-Runtime Scenarios (TUI + Web distinct keys) | 2 |
| Auxiliary Coverage Scenarios | 8 |
| Core UX Flow Domains | 11 |
Canonical UX Scenario Set
| Scenario | File | Primary Flow |
|---|---|---|
| Scenario 1 | scenarios/harness/scenario1-invitation-chat-e2e.toml | Invitation acceptance + shared channel + bidirectional chat |
| Scenario 2 | scenarios/harness/scenario2-social-topology-e2e.toml | Social topology and neighborhood operations |
| Scenario 3 | scenarios/harness/scenario3-irc-slash-commands-e2e.toml | Slash command lifecycle and moderation commands |
| Scenario 4 | scenarios/harness/scenario4-global-nav-and-help-e2e.toml | Global navigation and help modal behavior |
| Scenario 5 | scenarios/harness/scenario5-chat-modal-and-retry-e2e.toml | Chat wizard/modals and retry actions |
| Scenario 6 | scenarios/harness/scenario6-contacts-lan-and-contact-lifecycle-e2e.toml | Contacts, LAN scan, contact removal |
| Scenario 7 | scenarios/harness/scenario7-neighborhood-keypath-parity-e2e.toml | Neighborhood keypath parity and detail navigation |
| Scenario 8 | scenarios/harness/scenario8-settings-devices-authority-e2e.toml | Settings: profile, devices, authority panels |
| Scenario 9 | scenarios/harness/scenario9-guardian-and-mfa-ceremonies-e2e.toml | Guardian and MFA ceremony flows |
| Scenario 10 | scenarios/harness/scenario10-recovery-and-notifications-e2e.toml | Recovery request and notifications surfaces |
| Scenario 11 | scenarios/harness/scenario11-demo-full-tui-flow-e2e.toml | Full end-to-end demo-grade TUI flow |
| Scenario 12 | scenarios/harness/scenario12-mixed-device-enrollment-removal-e2e.toml | Mixed TUI/Web device enrollment + removal |
| Scenario 13 | scenarios/harness/scenario13-mixed-contact-channel-message-e2e.toml | Mixed TUI/Web contact invite + channel messaging |
UX Flow Matrix
| Flow Domain | Main Coverage | Secondary Coverage | Runtime Context |
|---|---|---|---|
| Invitation create/accept | Scenario 1 | Scenarios 2, 5, 6, 9, 11, 13 | TUI + Web |
| Contact lifecycle | Scenario 6 | Scenarios 1, 2, 5, 9, 13 | TUI + Web |
| Chat channel + messaging | Scenario 1 | Scenarios 3, 5, 11, 13 | TUI + Web |
| Slash commands and moderation | Scenario 3 | moderation-and-modal-coverage.toml, moderator-assign.toml | TUI-heavy |
| Global navigation/help | Scenario 4 | Scenario 11 | TUI + Web |
| Neighborhood/home operations | Scenario 2 | Scenarios 7, 11, home-roles.toml | TUI + Web |
| Settings panels | Scenario 8 | Scenarios 9, 10, 12 | TUI + Web |
| Device add/remove | Scenario 12 | Scenario 8 | Mixed runtime |
| Guardian/MFA ceremonies | Scenario 9 | Scenario 10 | TUI + Web |
| Recovery + notifications | Scenario 10 | Scenario 8 | TUI + Web |
| Mixed-device and mixed-user interoperability | Scenarios 12 and 13 | cross-authority-contact.toml | Mixed runtime |
Auxiliary Scenario Coverage
These scenarios are maintained as focused supplements and smoke checks:
| Scenario File | Focus |
|---|---|
local-discovery-smoke.toml | Local discovery smoke coverage |
mixed-topology-smoke.toml | Mixed-topology connectivity smoke |
mixed-topology-agent.toml | Agent-level mixed topology behavior |
moderation-and-modal-coverage.toml | Moderation + modal interaction sweep |
moderator-assign.toml | Moderator assignment and kick operations |
access-override.toml | Access override modal flow |
shared-storage.toml | Shared-storage user flow |
cross-authority-contact.toml | Cross-authority contact + neighborhood path |
Coverage Expectations
PR Gate Expectations
- Changes to global navigation, settings, chat, contacts, neighborhood, or ceremonies should have at least one impacted canonical scenario updated or re-validated.
- Changes that affect both TUI and web behavior should be validated against parity-critical scenarios (1-11) in both runtimes.
- Changes to mixed-instance behavior should include scenario 12 and/or 13 coverage.
CI Enforcement
Fast CI runs scripts/check/ux-flow-coverage.sh (just ci-ux-flow-coverage) to enforce that flow-relevant source changes are paired with scenario updates or an update to this report.
Residual Risk Areas
| Area | Current Risk | Mitigation Direction |
|---|---|---|
| Long-tail modal sequencing | Medium | Add focused scenario fragments for rare wizard branch paths |
| Toast timing/race windows | Medium | Prefer persistent-state assertions over toast-only checks |
| Cross-topology regressions | Medium | Keep mixed-topology smoke scenarios in scheduled lanes |
References
Verification Coverage Report
This document provides an overview of the formal verification, model checking, and conformance testing infrastructure in Aura.
Verification Boundary Statement
Aura keeps consensus and CRDT domain proof ownership in Quint models and Lean theorems. Telltale parity lanes validate runtime conformance behavior from replay artifacts. Telltale parity success does not count as new domain theorem coverage. See Formal Verification Reference for the assurance classification and limits.
Summary Metrics
| Metric | Count |
|---|---|
| Quint Specifications | 41 |
| Quint Invariants | 191 |
| Quint Temporal Properties | 11 |
| Quint Type Definitions | 362 |
| Lean Source Files | 38 |
| Lean Theorems | 118 |
| Conformance Fixtures | 4 |
| ITF Trace Harnesses | 8 |
| Testkit Tests | 113 |
| Bridge Modules | 4 |
| CI Verification Gates | 11 |
| Telltale Parity Modules | 1 |
| Bridge Pipeline Fixtures | 3 |
Verification Layers
Layer 1: Quint Specifications
Formal protocol specifications in verification/quint/ organized by subsystem.
| Subsystem | Files | Contents |
|---|---|---|
| Root | 11 | core, authorization, recovery, invitation, interaction, leakage, sbb, time_system, transport, epochs, cli_recovery_demo |
| consensus/ | 4 | core, liveness, adversary, frost |
| journal/ | 3 | core, counter, anti_entropy |
| keys/ | 3 | dkg, dkd, resharing |
| sessions/ | 4 | core, choreography, groups, locking |
| amp/ | 1 | channel |
| liveness/ | 3 | connectivity, timing, properties |
| harness/ | 8 | amp_channel, counter, dkg, flows, groups, locking, recovery, resharing |
| tui/ | 4 | demo_recovery, flows, signals, state |
Harness modules generate ITF traces on demand via just quint-generate-traces. Traces are not checked into the repository. CI runs just ci-conformance-itf to generate traces and replay them through Rust handlers.
Key Specifications
| Specification | Purpose | Key Properties |
|---|---|---|
consensus/core.qnt | Fast-path consensus protocol | UniqueCommitPerInstance, CommitRequiresThreshold, EquivocatorsExcluded |
consensus/liveness.qnt | Progress guarantees | ProgressUnderSynchrony, RetryBound, CommitRequiresHonestParticipation |
consensus/adversary.qnt | Byzantine tolerance | ByzantineThreshold, EquivocationDetected, HonestMajorityCanCommit |
consensus/frost.qnt | Threshold signatures | Share aggregation, commitment validation |
journal/core.qnt | CRDT journal semantics | NonceUnique, FactsOrdered, NonceMergeCommutative, LamportMonotonic |
journal/anti_entropy.qnt | Sync protocol | FactsMonotonic, EventualConvergence, VectorClockConsistent |
authorization.qnt | Guard chain security | NoCapabilityWidening, ChargeBeforeSend |
time_system.qnt | Timestamp ordering | TimeStamp domain semantics and comparison |
Layer 2: Rust Integration
Files implementing Quint-Rust correspondence and model-based testing.
Core Integration (aura-core)
effects/quint.rs-QuintMappabletrait for bidirectional type mappingeffects/mod.rs- Effect trait definitions with Quint correspondence
Quint Crate (aura-quint)
runner.rs-QuintRunnerwith property caching and verification statisticsproperties.rs-PropertySpec,PropertySuite, and property categorizationevaluator.rs-QuintEvaluatorsubprocess wrapper for Quint CLIhandler.rs- Effect handler integration
Lean-Quint Bridge (aura-quint)
Cross-validation modules for Lean↔Quint correspondence:
| Module | Purpose |
|---|---|
bridge_export.rs | Export Quint state to Lean-readable format |
bridge_import.rs | Import Lean outputs back to Quint structures |
bridge_format.rs | Shared serialization format definitions |
bridge_validate.rs | Cross-validation assertions and checks |
Simulator Integration (aura-simulator/src/quint/)
17 modules implementing generative simulation:
| Module | Purpose |
|---|---|
action_registry.rs | Maps Quint action names to Rust handlers |
state_mapper.rs | Bidirectional state conversion (Rust <-> Quint JSON) |
generative_simulator.rs | Orchestrates ITF trace replay with property checking |
itf_loader.rs | Parses ITF traces from Quint model checking |
itf_fuzzer.rs | Model-based fuzzing with coverage analysis |
trace_converter.rs | Converts between trace formats |
simulation_evaluator.rs | Evaluates properties during simulation |
properties.rs | Property extraction and classification |
domain_handlers.rs | Domain-specific action handlers |
amp_channel_handlers.rs | AMP reliable channel handlers |
byzantine_mapper.rs | Byzantine fault strategy mapping |
chaos_generator.rs | Chaos/fault scenario generation |
aura_state_extractors.rs | Aura-specific state extraction |
cli_runner.rs | CLI integration for Quint verification |
ast_parser.rs | Quint AST parsing for analysis |
mod.rs | Module exports and re-exports |
types.rs | Shared type definitions |
Differential Verification (aura-simulator)
differential_tester.rs- Cross-implementation parity testing between Quint models and Rust handlerstelltale_parity.rs- Telltale-backed parity boundary, canonical surface mapping, and report artifact generation
Consensus Verification (aura-consensus)
core/verification/mod.rs- Verification module facadecore/verification/quint_mapping.rs- Consensus type mappings
Layer 3: Lean Proofs
Lean 4 mathematical proofs in verification/lean/ providing formal guarantees.
Type Modules (10 files)
| Module | Content |
|---|---|
Types/ByteArray32.lean | 32-byte hash representation (6 theorems) |
Types/OrderTime.lean | Opaque ordering tokens (4 theorems) |
Types/TimeStamp.lean | 4-variant time enum |
Types/FactContent.lean | Structured fact types |
Types/ProtocolFacts.lean | Protocol-specific fact types |
Types/AttestedOp.lean | Attested operation types |
Types/TreeOp.lean | Tree operation types |
Types/Namespace.lean | Authority/Context namespaces |
Types/Identifiers.lean | Identifier types |
Types.lean | Type module aggregation |
Domain Modules (9 files)
| Module | Purpose |
|---|---|
Domain/Consensus/Types.lean | Consensus message types (8 definitions) |
Domain/Consensus/Frost.lean | FROST signature types |
Domain/Journal/Types.lean | Fact and Journal structures |
Domain/Journal/Operations.lean | merge, reduce, factsEquiv (1 theorem) |
Domain/FlowBudget.lean | Budget types and charging |
Domain/GuardChain.lean | Guard types and evaluation |
Domain/TimeSystem.lean | Timestamp comparison |
Domain/KeyDerivation.lean | Key derivation types |
Domain/ContextIsolation.lean | Context isolation model |
Proof Modules (14 files, 118 theorems)
Consensus Proofs:
| Module | Theorems | Content |
|---|---|---|
Proofs/Consensus/Agreement.lean | 3 | No two honest parties commit different values |
Proofs/Consensus/Validity.lean | 7 | Only valid proposals can be committed |
Proofs/Consensus/Equivocation.lean | 5 | Detection soundness and completeness |
Proofs/Consensus/Evidence.lean | 8 | Evidence CRDT semilattice properties |
Proofs/Consensus/Frost.lean | 12 | FROST share aggregation safety |
Proofs/Consensus/Liveness.lean | 3 | Progress under timing assumptions (axioms) |
Proofs/Consensus/Adversary.lean | 7 | Byzantine model bounds |
Proofs/Consensus/Summary.lean | - | Master consensus claims bundle |
Infrastructure Proofs:
| Module | Theorems | Content |
|---|---|---|
Proofs/Journal.lean | 14 | CRDT semilattice (commutativity, associativity, idempotence) |
Proofs/FlowBudget.lean | 5 | Charging correctness |
Proofs/GuardChain.lean | 7 | Guard evaluation determinism |
Proofs/TimeSystem.lean | 8 | Timestamp ordering properties |
Proofs/KeyDerivation.lean | 3 | PRF isolation proofs |
Proofs/ContextIsolation.lean | 16 | Context separation and bridge authorization |
Entry Points (4 files)
Aura.lean- Top-level documentationAura/Proofs.lean- Main reviewer entry with all Claims bundlesAura/Assumptions.lean- Cryptographic axioms (FROST unforgeability, hash collision resistance, PRF security)Aura/Runner.lean- CLI for differential testing
Layer 4: Conformance Testing
Deterministic parity validation infrastructure in aura-testkit.
Conformance Fixtures
| Fixture | Purpose |
|---|---|
consensus.json | Consensus protocol conformance |
sync.json | Synchronization protocol conformance |
recovery.json | Guardian recovery conformance |
invitation.json | Invitation protocol conformance |
Conformance Modules
| Module | Purpose |
|---|---|
conformance.rs | Artifact loading, replay, and verification |
conformance_diff.rs | Law-aware comparison with envelope classifications |
Effect Envelope Classifications
| Class | Effect Kinds | Comparison Rule |
|---|---|---|
strict | handle_recv, handle_choose, handle_acquire, handle_release | Byte-exact match required |
commutative | send_decision, invoke_step | Order-insensitive under normalization |
algebraic | topology_event | Reduced via domain-normal form |
Verified Invariants
Consensus Invariants
| Invariant | Location |
|---|---|
InvariantUniqueCommitPerInstance | consensus/core.qnt |
InvariantCommitRequiresThreshold | consensus/core.qnt |
InvariantCommittedHasCommitFact | consensus/core.qnt |
InvariantEquivocatorsExcluded | consensus/core.qnt |
InvariantProposalsFromWitnesses | consensus/core.qnt |
InvariantProgressUnderSynchrony | consensus/liveness.qnt |
InvariantRetryBound | consensus/liveness.qnt |
InvariantCommitRequiresHonestParticipation | consensus/liveness.qnt |
InvariantQuorumPossible | consensus/liveness.qnt |
InvariantByzantineThreshold | consensus/adversary.qnt |
InvariantEquivocationDetected | consensus/adversary.qnt |
InvariantCompromisedNoncesExcluded | consensus/adversary.qnt |
InvariantHonestMajorityCanCommit | consensus/adversary.qnt |
Journal Invariants
| Invariant | Location |
|---|---|
InvariantNonceUnique | journal/core.qnt |
InvariantFactsOrdered | journal/core.qnt |
InvariantFactsMatchNamespace | journal/core.qnt |
InvariantLifecycleCompletedImpliesStable | journal/core.qnt |
InvariantNonceMergeCommutative | journal/core.qnt |
InvariantLamportMonotonic | journal/core.qnt |
InvariantReduceDeterministic | journal/core.qnt |
InvariantPhaseRegistered | journal/counter.qnt |
InvariantCountersRegistered | journal/counter.qnt |
InvariantLifecycleStatusDefined | journal/counter.qnt |
InvariantOutcomeWhenCompleted | journal/counter.qnt |
InvariantFactsMonotonic | journal/anti_entropy.qnt |
InvariantFactsSubsetOfGlobal | journal/anti_entropy.qnt |
InvariantVectorClockConsistent | journal/anti_entropy.qnt |
InvariantEventualConvergence | journal/anti_entropy.qnt |
InvariantDeltasFromSource | journal/anti_entropy.qnt |
InvariantCompletedSessionsConverged | journal/anti_entropy.qnt |
Temporal Properties
| Property | Location |
|---|---|
livenessEventualCommit | consensus/core.qnt |
safetyImmutableCommit | consensus/core.qnt |
authorizationSoundness | authorization.qnt |
budgetMonotonicity | authorization.qnt |
flowBudgetFairness | authorization.qnt |
canAlwaysExit | tui/state.qnt |
modalEventuallyCloses | tui/state.qnt |
insertModeEventuallyExits | tui/state.qnt |
InvariantLeakageBounded | leakage.qnt |
InvariantObserverHierarchyMaintained | leakage.qnt |
InvariantBudgetsPositive | leakage.qnt |
CI Verification Gates
Automated verification lanes wired into CI pipelines.
Core Verification
| Gate | Command | Purpose |
|---|---|---|
| Property Monitor | just ci-property-monitor | Runtime property assertion monitoring |
| Simulator Telltale Parity | just ci-simulator-telltale-parity | Artifact-driven telltale vs Aura simulator differential comparison |
| Choreography Parity | just ci-choreo-parity | Session type projection consistency |
| Quint Typecheck | just ci-quint-typecheck | Quint specification type safety |
Conformance Gates
| Gate | Command | Purpose |
|---|---|---|
| Conformance Policy | just ci-conformance-policy | Policy rule validation |
| Conformance Contracts | just ci-conformance-contracts | Contract satisfaction checks |
| Golden Fixtures | conformance_golden_fixtures | Deterministic replay against known-good traces |
Formal Methods
| Gate | Command | Purpose |
|---|---|---|
| Lean Build | just ci-lean-build | Compile Lean proofs |
| Lean Completeness | just ci-lean-check-sorry | Check for incomplete proofs (sorry) |
| Lean-Quint Bridge | just ci-lean-quint-bridge | Cross-validation between Lean and Quint |
| Kani BMC | just ci-kani | Bounded model checking for unsafe code |
CI Artifacts
Conformance artifacts upload to CI for failure triage:
artifacts/conformance/
├── native_coop/
│ └── scenario_seed_artifact.json
├── wasm_coop/
│ └── scenario_seed_artifact.json
└── diff_report.json
The diff report highlights specific mismatches for investigation.
Telltale parity and bridge lanes emit additional artifacts:
artifacts/telltale-parity/
└── report.json
artifacts/lean-quint-bridge/
├── bridge.log
├── bridge_discrepancy_report.json
└── report.json
artifacts/telltale-parity/report.json uses schema aura.telltale-parity.report.v1.
artifacts/lean-quint-bridge/bridge_discrepancy_report.json uses schema aura.lean-quint-bridge.discrepancy.v1.
Bridge Pipeline Fixtures
aura-quint bridge pipeline checks use deterministic fixture inputs:
| Fixture | Purpose |
|---|---|
positive_bundle.json | Expected consistent cross-validation outcome |
negative_bundle.json | Expected discrepancy detection outcome |
quint_ir_fixture.json | Export/import pipeline coverage for Quint IR |
Fixtures live in crates/aura-quint/tests/fixtures/bridge/.
Related Documentation
- Formal Verification Reference - Architecture and specification patterns
- Verification and MBT Guide - Practical verification workflows
- Simulation Infrastructure Reference - Generative simulation details
- Testing Guide - Testing infrastructure and conformance testing
Project Structure
This document provides the authoritative reference for Aura's crate organization, dependencies, and development policies.
The primary specifications live in docs/ (e.g., consensus in docs/106_consensus.md, ceremony lifecycles in docs/107_operation_categories.md). The work/ directory is non-authoritative scratch and may be removed.
8-Layer Architecture
Aura's codebase is organized into 8 clean architectural layers. Each layer builds on the layers below without circular dependencies.
┌────────────────────────────────────────────────────┐
│ Layer 8: Testing & Development Tools │
│ • aura-testkit • aura-quint • aura-harness │
├────────────────────────────────────────────────────┤
│ Layer 7: User Interface │
│ • aura-terminal │
├────────────────────────────────────────────────────┤
│ Layer 6: Runtime Composition │
│ • aura-agent • aura-simulator • aura-app │
├────────────────────────────────────────────────────┤
│ Layer 5: Feature/Protocol Implementation │
│ • aura-authentication • aura-chat │
│ • aura-invitation • aura-recovery │
│ • aura-relational • aura-rendezvous │
│ • aura-sync • aura-social │
├────────────────────────────────────────────────────┤
│ Layer 4: Orchestration │
│ • aura-protocol • aura-guards │
│ • aura-consensus • aura-amp │
│ • aura-anti-entropy │
├────────────────────────────────────────────────────┤
│ Layer 3: Implementation │
│ • aura-effects • aura-composition │
├────────────────────────────────────────────────────┤
│ Layer 2: Specification │
│ Domain Crates: │
│ • aura-journal • aura-authorization │
│ • aura-signature • aura-store │
│ • aura-transport • aura-maintenance │
│ Choreography: │
│ • aura-mpst • aura-macros │
├────────────────────────────────────────────────────┤
│ Layer 1: Foundation │
│ • aura-core │
└────────────────────────────────────────────────────┘
Layer 1: Foundation — aura-core
Purpose: Single source of truth for all domain concepts and interfaces.
Contains:
- Effect traits for core infrastructure, authentication, storage, network, cryptography, privacy, configuration, and testing
- Domain types:
AuthorityId,ContextId,SessionId,FlowBudget,ObserverClass,Capability - Cryptographic utilities: key derivation, FROST types, merkle trees, Ed25519 helpers
- Semantic traits:
JoinSemilattice,MeetSemilattice,CvState,MvState - Error types:
AuraError, error codes, and guard metadata - Configuration system with validation and multiple formats
- Causal context types for CRDT ordering
- AMP channel lifecycle effect surface:
aura-core::effects::amp::AmpChannelEffects(implemented by runtime, simulator, and testkit mocks).
Key principle: Interfaces only, no implementations or business logic.
Exceptions:
-
Extension traits providing convenience methods are allowed (e.g.,
LeakageChoreographyExt,SimulationEffects,AuthorityRelationalEffects). These blanket implementations extend existing effect traits with domain-specific convenience methods while maintaining interface-only semantics. -
Arc
blanket implementations for effect traits are required in aura-core due to Rust's orphan rules. These are not "runtime instantiations" - they are purely mechanical delegations that enableArc<AuraEffectSystem>to satisfy trait bounds. Example:#![allow(unused)] fn main() { impl<T: CryptoEffects + ?Sized> CryptoEffects for std::sync::Arc<T> { async fn ed25519_sign(&self, msg: &[u8], key: &[u8]) -> Result<Vec<u8>, CryptoError> { (**self).ed25519_sign(msg, key).await // Pure delegation } } }Why this is architecturally sound:
Arcis a language-level primitive (likeVec,Box, or&), not a "runtime" in the architectural sense. These implementations add no behavior or state - they simply say "if T can do X, then Arccan too by asking T." Without these, any handler wrapped in Arcwould fail to satisfy effect trait bounds, breaking the entire dependency injection pattern.
Architectural Compliance: aura-core maintains strict interface-only semantics. Test utilities like MockEffects are provided in aura-testkit (Layer 8) where they architecturally belong.
Dependencies: None (foundation crate).
Commitment Tree Types and Functions
Location: aura-core/src/tree/
Contains:
- Core tree types:
TreeOp,AttestedOp,Policy,LeafNode,BranchNode,BranchSigningKey,TreeCommitment,Epoch - Commitment functions:
commit_branch(),commit_leaf(),policy_hash(),compute_root_commitment() - Policy meet-semilattice implementation for threshold refinement
- Snapshot types:
Snapshot,Cut,ProposalId,Partial - Verification module:
verify_attested_op()(cryptographic),check_attested_op()(state consistency),compute_binding_message()
Why Layer 1?
Commitment tree types MUST remain in aura-core because:
- Effect traits require them:
TreeEffectsandSyncEffectsinaura-core/src/effects/use these types in their signatures - FROST primitives depend on them:
aura-core/src/crypto/tree_signing.rsimplements threshold signing over tree operations - Authority abstraction needs them:
aura-core/src/authority.rsusesPolicy,AttestedOp, andTreeOpKind - Foundational cryptographic structures: Commitment trees are merkle trees with threshold policies - core cryptographic primitives, not domain logic
Layer 2 separation (aura-journal) contains:
- Tree state machine: Full
TreeStatewith branches, leaves, topology, and path validation - Reduction logic: Deterministic state derivation from
OpLog<AttestedOp> - Domain validation: Business rules for tree operations (e.g., policy monotonicity, leaf lifecycle)
- Application logic:
apply_verified(), compaction, garbage collection - Re-exports:
pub use aura_core::tree::*for convenience viaaura_journal::commitment_tree
Key architectural distinction:
- Layer 1 (
aura-core): Tree types and cryptographic commitment functions (pure primitives) - Layer 2 (
aura-journal): Tree state machine, CRDT semantics, and validation rules (domain implementation)
This separation allows effect traits in Layer 1 to reference tree types without creating circular dependencies, while keeping the stateful CRDT logic in the appropriate domain crate.
Layer 2: Specification — Domain Crates and Choreography
Purpose: Define domain semantics and protocol specifications.
Layer 2 Architecture Diff (Invariants)
Layer 2 is the specification layer: pure domain semantics with zero runtime coupling.
Must hold:
- No handler composition, runtime assembly, or UI dependencies.
- Domain facts are versioned and encoded via canonical DAG-CBOR.
- Fact reducers register through
FactRegistry; no direct wiring inaura-journal. - Authorization scopes use
aura-coreResourceScopeand typed operations. - No in-memory production state; stateful test handlers live in
aura-testkit.
Forbidden:
- Direct OS access (time, fs, network) outside effect traits.
- Tokio/async-std usage in domain protocols.
- State-bearing singletons or process-wide caches.
Domain Crates
| Crate | Domain | Responsibility |
|---|---|---|
aura-journal | Fact-based journal | CRDT semantics, tree state machine, reduction logic, validation (re-exports tree types from aura-core) |
aura-authorization | Trust and authorization | Capability refinement, Biscuit token helpers |
aura-signature | Identity semantics | Signature verification, device lifecycle |
aura-store | Storage domain | Storage types, capabilities, domain logic |
aura-transport | Transport semantics | P2P communication abstractions |
aura-maintenance | Maintenance facts | Snapshot, cache invalidation, OTA activation, admin replacement facts + reducer |
Key characteristics: Implement domain logic without effect handlers or coordination.
Extensible Fact Types (aura-journal)
The journal provides generic fact infrastructure that higher-level crates extend with domain-specific fact types. This follows the Open/Closed Principle: the journal is open for extension but closed for modification.
Protocol-Level vs Domain-Level Facts
The RelationalFact enum in aura-journal/src/fact.rs contains two categories:
Protocol-Level Facts (stay in aura-journal):
These are core protocol constructs with complex reduction logic in reduce_context(). They have interdependencies and specialized state derivation that cannot be delegated to simple domain reducers:
| Fact | Purpose | Why Protocol-Level |
|---|---|---|
Protocol(GuardianBinding) | Guardian relationship | Core recovery protocol |
Protocol(RecoveryGrant) | Recovery capability | Core recovery protocol |
Protocol(Consensus) | Aura Consensus results | Core agreement mechanism |
Protocol(AmpChannelCheckpoint) | Ratchet window anchoring | Complex epoch state computation |
Protocol(AmpProposedChannelEpochBump) | Optimistic epoch transitions | Spacing rules, bump selection |
Protocol(AmpCommittedChannelEpochBump) | Finalized epoch transitions | Epoch chain validation |
Protocol(AmpChannelPolicy) | Channel policy overrides | Skip window derivation |
Domain-Level Facts (via Generic + FactRegistry):
Application-specific facts use RelationalFact::Generic and are reduced by registered FactReducer implementations.
| Domain Crate | Fact Type | Purpose |
|---|---|---|
aura-chat | ChatFact | Channels, messages |
aura-invitation | InvitationFact | Invitation lifecycle |
aura-relational | ContactFact | Contact management |
aura-social/moderation | Block*Fact | Block, mute, ban, kick |
Design Pattern:
-
aura-journalprovides:DomainFacttrait for fact type identity and serializationFactReducertrait for domain-specific reduction logicFactRegistryfor runtime fact type registrationRelationalFact::Genericas the extensibility mechanism
-
Domain crates implement:
- Their own typed fact enums (e.g.,
ChatFact,InvitationFact) DomainFacttrait withto_generic()for storageFactReducerfor reduction toRelationalBinding
- Their own typed fact enums (e.g.,
-
aura-agent/src/fact_registry.rsregisters all domain reducers:#![allow(unused)] fn main() { pub fn build_fact_registry() -> FactRegistry { let mut registry = FactRegistry::new(); registry.register::<ChatFact>(CHAT_FACT_TYPE_ID, Box::new(ChatFactReducer)); registry.register::<InvitationFact>(INVITATION_FACT_TYPE_ID, Box::new(InvitationFactReducer)); registry.register::<ContactFact>(CONTACT_FACT_TYPE_ID, Box::new(ContactFactReducer)); register_moderation_facts(&mut registry); registry } }
Why This Architecture:
- Open/Closed Principle: New domain facts don't require modifying
aura-journal - Domain Isolation: Each crate owns its fact semantics
- Protocol Integrity: Core protocol facts with complex reduction stay in
aura-journal - Testability: Domain facts can be tested independently
- Type Safety: Compile-time guarantees within each domain
Core Fact Types in aura-journal:
Only facts fundamental to journal operation remain as direct enum variants:
AttestedOp: Commitment tree operations (cryptographic primitives)Snapshot: Journal compaction checkpointsRendezvousReceipt: Cross-authority coordination receipts- Protocol-level
RelationalFactvariants listed above
Fact Implementation Patterns by Layer
Aura uses two distinct fact patterns based on architectural layer to prevent circular dependencies:
Layer 2 Pattern (Domain Crates: aura-maintenance, aura-authorization, aura-signature, aura-store, aura-transport):
These crates use the aura-core::types::facts pattern with NO dependency on aura-journal:
#![allow(unused)] fn main() { use aura_core::types::facts::{FactTypeId, FactError, FactEnvelope, FactDeltaReducer}; pub static MY_FACT_TYPE_ID: FactTypeId = FactTypeId::new("my_domain"); pub const MY_FACT_SCHEMA_VERSION: u16 = 1; impl MyFact { pub fn try_encode(&self) -> Result<Vec<u8>, FactError> { aura_core::types::facts::try_encode_fact( &MY_FACT_TYPE_ID, MY_FACT_SCHEMA_VERSION, self, ) } pub fn to_envelope(&self) -> Result<FactEnvelope, FactError> { // Create envelope manually for Generic wrapping } } impl FactDeltaReducer<MyFact, MyFactDelta> for MyFactReducer { fn apply(&self, fact: &MyFact) -> MyFactDelta { /* ... */ } } }
Layer 4/5 Pattern (Feature Crates: aura-chat, aura-invitation, aura-relational, aura-recovery, aura-social):
These crates use the aura-journal::extensibility::DomainFact pattern and register with FactRegistry:
#![allow(unused)] fn main() { use aura_journal::extensibility::{DomainFact, FactReducer}; use aura_macros::DomainFact; #[derive(DomainFact)] #[domain_fact(type_id = "my_domain", schema_version = 1, context_fn = "context_id")] pub enum MyFact { /* ... */ } impl FactReducer for MyFactReducer { fn handles_type(&self) -> &'static str { /* ... */ } fn reduce_envelope(...) -> Option<RelationalBinding> { /* ... */ } } }
Why Two Patterns?
- Layer 2 → Layer 2 dependencies create circular risk:
aura-journalis itself a Layer 2 crate. If other Layer 2 crates depend onaura-journalfor theDomainFacttrait, we risk circular dependencies. - Layer 4/5 can safely depend on Layer 2: Higher layers depend on lower layers by design, so feature crates can use the
DomainFacttrait fromaura-journal. - Registration location differs: Layer 2 facts are wrapped manually in
RelationalFact::Generic. Layer 4/5 facts register withFactRegistryinaura-agent/src/fact_registry.rs.
For a quick decision tree on pattern selection, see CLAUDE.md under "Agent Decision Aids".
Choreography Specification
aura-mpst: Aura-facing compatibility crate over Telltale. Re-exports choreography/runtime surfaces and Aura extension traits used by generated protocols, VM artifacts, and testing utilities.
aura-macros: Compile-time choreography frontend. Parses Aura annotations (guard_capability, flow_cost, journal_facts, leak) and emits Telltale-backed generated modules plus Aura effect-bridge helpers.
Layer 3: Implementation — aura-effects and aura-composition
Purpose: Effect implementation and handler composition.
aura-effects — Stateless Effect Handlers
Purpose: Stateless, single-party effect implementations. Architectural Decision: aura-effects is the designated singular point of interaction with non-deterministic operating system services (entropy, wall-clock time, network I/O, file system). This design choice makes the architectural boundary explicit and centralizes impure operations.
Contains:
- Production handlers:
RealCryptoHandler,TcpTransportHandler,FilesystemStorageHandler,PhysicalTimeHandler - OS integration adapters that delegate to system services
- Pure functions that transform inputs to outputs without state
What doesn't go here:
- Handler composition or registries
- Multi-handler coordination
- Stateful implementations
- Mock/test handlers
Key characteristics: Each handler should be independently testable and reusable. No handler should know about other handlers. This enables clean dependency injection and modular testing.
Dependencies: aura-core and external libraries.
Note: Mock and test handlers are located in aura-testkit (Layer 8) to maintain clean separation between production and testing concerns.
aura-composition — Handler Composition
Purpose: Assemble individual handlers into cohesive effect systems.
Contains:
- Effect registry and builder patterns
- Handler composition utilities
- Effect system configuration
- Handler lifecycle management (start/stop/configure)
- Reactive infrastructure:
Dynamic<T>FRP primitives for composing view updates over effect changes
What doesn't go here:
- Individual handler implementations
- Multi-party protocol logic
- Runtime-specific concerns
- Application lifecycle
Key characteristics: Feature crates need to compose handlers without pulling in full runtime infrastructure. This is about "how do I assemble handlers?" not "how do I coordinate distributed protocols?"
Dependencies: aura-core, aura-effects.
Layer 4: Orchestration — aura-protocol + subcrates
Purpose: Multi-party coordination and distributed protocol orchestration.
Contains:
- Guard chain coordination (
CapGuard → FlowGuard → JournalCoupler) inaura-guards - Multi-party protocol orchestration (consensus in
aura-consensus, anti-entropy inaura-anti-entropy) - Quorum-driven DKG orchestration and transcript handling in
aura-consensus/src/dkg/ - Cross-handler coordination logic (
TransportCoordinator,StorageCoordinator, etc.) - Distributed state management
- Stateful coordinators for multi-party protocols
What doesn't go here:
- Effect trait definitions (all traits belong in
aura-core) - Handler composition infrastructure (belongs in
aura-composition) - Single-party effect implementations (belongs in
aura-effects) - Test/mock handlers (belong in
aura-testkit) - Runtime assembly (belongs in
aura-agent) - Application-specific business logic (belongs in domain crates)
Key characteristics: This layer coordinates multiple handlers working together across network boundaries. It implements the "choreography conductor" pattern, ensuring distributed protocols execute correctly with proper authorization, flow control, and state consistency. All handlers here manage multi-party coordination, not single-party operations.
Dependencies: aura-core, aura-effects, aura-composition, aura-mpst, domain crates, and Layer 4 subcrates (aura-guards, aura-consensus, aura-amp, aura-anti-entropy). Performance-critical protocol operations may require carefully documented exceptions for direct cryptographic library usage.
Layer 5: Feature/Protocol Implementation
Purpose: Complete end-to-end protocol implementations.
Crates:
| Crate | Protocol | Purpose |
|---|---|---|
aura-authentication | Authentication | Device, threshold, and guardian auth flows |
aura-chat | Chat | Chat domain facts + view reducers; local chat prototype |
aura-invitation | Invitations | Peer onboarding and relational facts |
aura-recovery | Guardian recovery | Recovery grants and dispute escalation |
aura-relational | Cross-authority relationships | RelationalContext protocols (domain types in aura-core) |
aura-rendezvous | Peer discovery | Context-scoped rendezvous and routing |
aura-social | Social topology | Block/neighborhood materialized views, relay selection, progressive discovery layers, role/access semantics (Member/Participant, Full/Partial/Limited) |
aura-sync | Synchronization | Journal sync and anti-entropy protocols |
Key characteristics: Reusable building blocks with no UI or binary entry points.
Notes:
- Layer 5 crates now include
ARCHITECTURE.mddescribing facts, invariants, and operation categories. OPERATION_CATEGORIESconstants in each Layer 5 crate map operations to A/B/C classes.- Runtime-owned caches (e.g., invitation/rendezvous descriptors) live in Layer 6 handlers, not in Layer 5 services.
- Layer 5 facts use versioned binary encoding (bincode) with JSON fallback for debug and compatibility.
- FactKey helper types are required for reducers/views to keep binding key derivation consistent.
- Ceremony facts carry optional
trace_idvalues to support cross-protocol traceability.
Dependencies: aura-core, aura-effects, aura-composition, aura-mpst, plus Layer 4 orchestration crates (aura-protocol, aura-guards, aura-consensus, aura-amp, aura-anti-entropy).
Layer 6: Runtime Composition — aura-agent, aura-simulator, and aura-app
Purpose: Assemble complete running systems for production deployment.
aura-agent: Production runtime for deployment with application lifecycle management, runtime-specific configuration, production deployment concerns, and system integration.
aura-app: Portable headless application core providing the business logic and state management layer for all platforms. Exposes a platform-agnostic API consumed by terminal, iOS, Android, and web frontends. Contains intent processing, view derivation, and platform feature flags (native, ios, android, web-js, web-dominator).
aura-simulator: Deterministic simulation runtime with virtual time, transport shims, failure injection, and generative testing via Quint integration (see aura-simulator/src/quint/ for generative simulation bridge).
Contains:
- Application lifecycle management (startup, shutdown, signals)
- Runtime-specific configuration and policies
- Production deployment concerns
- System integration and monitoring hooks
- Reactive event loop:
ReactiveScheduler(Tokio task) that orchestrates fact ingestion, journal updates, and view propagation
What doesn't go here:
- Effect handler implementations
- Handler composition utilities
- Protocol coordination logic
- CLI or UI concerns
Key characteristics: This is about "how do I deploy and run this as a production system?" It's the bridge between composed handlers/protocols and actual running applications.
Dependencies: All domain crates, aura-effects, aura-composition, and Layer 4 orchestration crates (aura-protocol, aura-guards, aura-consensus, aura-amp, aura-anti-entropy).
Layer 7: User Interface — aura-terminal
Purpose: User-facing applications with main entry points.
aura-terminal: Terminal-based interface combining CLI commands and an interactive TUI (Terminal User Interface). Provides account and device management, recovery status visualization, chat interfaces, and scenario execution. Consumes aura-app for all business logic and state management.
Key characteristic: Contains main() entry point that users run directly. Binary is named aura.
Dependencies: aura-app, aura-agent, aura-core, aura-recovery, and Layer 4 orchestration crates (aura-protocol, aura-guards, aura-consensus, aura-amp, aura-anti-entropy).
Layer 8: Testing and Development Tools
Purpose: Cross-cutting test utilities, formal verification bridges, and generative testing infrastructure.
aura-testkit: Comprehensive testing infrastructure including:
- Shared test fixtures and scenario builders
- Property test helpers and deterministic utilities
- Mock effect handlers:
MockCryptoHandler,SimulatedTimeHandler,MemoryStorageHandler, etc. - Stateful test handlers that maintain controllable state for deterministic testing
aura-quint: Formal verification bridge to Quint model checker including:
- Native Quint subprocess interface for parsing and type checking
- Property specification management with classification (authorization, budget, integrity)
- Verification runner with caching and counterexample generation
- Effect trait implementations for property evaluation during simulation
aura-harness: Multi-instance runtime harness for orchestrating test scenarios including:
- Coordinator and executor for managing multiple Aura instances
- Scenario definition and replay capabilities
- Artifact synchronization and determinism validation
- Screen normalization and VT100 terminal emulation for TUI testing
- Shared semantic contracts from
aura-appfor scenario actions and UI snapshots - Backend selection for
mock,patchbay, andpatchbay-vmholepunch lanes - Resource guards and capability checking
Key characteristics: Mock handlers in aura-testkit are allowed to be stateful (using Arc<Mutex<>>, etc.) since they need controllable, deterministic state for testing. This maintains the stateless principle for production handlers in aura-effects while enabling comprehensive testing.
Deterministic seed policy: Test construction of AuraEffectSystem must use seeded helper constructors (simulation_for_test*). Do not call legacy testing* or raw simulation* constructors from test code. For multiple instances from the same callsite, use simulation_for_named_test_with_salt(...) so each seed is unique and replayable.
Dependencies: aura-core (for aura-harness); aura-agent, aura-composition, aura-journal, aura-transport, aura-core, aura-protocol, aura-guards, aura-consensus, aura-amp, aura-anti-entropy (for aura-testkit and aura-quint).
Workspace Structure
crates/
├── aura-agent Runtime composition and agent lifecycle
├── aura-app Portable headless application core (multi-platform)
├── aura-authentication Authentication protocols
├── aura-anti-entropy Anti-entropy sync and reconciliation
├── aura-amp Authenticated messaging protocol (AMP)
├── aura-chat Chat facts + local prototype service
├── aura-composition Handler composition and effect system assembly
├── aura-consensus Consensus protocol implementation
├── aura-core Foundation types and effect traits
├── aura-effects Effect handler implementations
├── aura-guards Guard chain enforcement
├── aura-harness Multi-instance runtime harness
├── aura-invitation Invitation choreographies
├── aura-journal Fact-based journal domain
├── aura-macros Choreography DSL compiler
├── aura-maintenance Maintenance facts and reducers
├── aura-mpst Session types and choreography specs
├── aura-protocol Orchestration and coordination
├── aura-quint Quint formal verification
├── aura-recovery Guardian recovery protocols
├── aura-relational Cross-authority relationships
├── aura-rendezvous Peer discovery and routing
├── aura-simulator Deterministic simulation engine
├── aura-social Social topology and progressive disclosure
├── aura-store Storage domain types
├── aura-sync Synchronization protocols
├── aura-terminal Terminal UI (CLI + TUI)
├── aura-testkit Testing utilities and fixtures
├── aura-transport P2P communication layer
├── aura-signature Identity verification
└── aura-authorization Web-of-trust authorization
Dependency Graph
graph TD
%% Layer 1: Foundation
core[aura-core]
%% Layer 2: Specification
signature[aura-signature]
journal[aura-journal]
authorization[aura-authorization]
store[aura-store]
transport[aura-transport]
mpst[aura-mpst]
macros[aura-macros]
maintenance[aura-maintenance]
%% Layer 3: Implementation
effects[aura-effects]
composition[aura-composition]
%% Layer 4: Orchestration
guards[aura-guards]
anti_entropy[aura-anti-entropy]
consensus[aura-consensus]
amp[aura-amp]
protocol[aura-protocol]
%% Layer 5: Feature
social[aura-social]
chat[aura-chat]
relational[aura-relational]
auth[aura-authentication]
rendezvous[aura-rendezvous]
invitation[aura-invitation]
recovery[aura-recovery]
sync[aura-sync]
%% Layer 6: Runtime
app[aura-app]
agent[aura-agent]
simulator[aura-simulator]
%% Layer 7: Application
terminal[aura-terminal]
%% Layer 8: Testing
testkit[aura-testkit]
quint[aura-quint]
harness[aura-harness]
%% Layer 2 dependencies
signature --> core
journal --> core
authorization --> core
store --> core
transport --> core
mpst --> core
macros --> core
maintenance --> core
%% Layer 3 dependencies
effects --> core
composition --> core
composition --> effects
composition --> mpst
%% Layer 4 dependencies
guards --> core
guards --> authorization
guards --> mpst
guards --> journal
anti_entropy --> core
anti_entropy --> guards
anti_entropy --> journal
consensus --> core
consensus --> macros
consensus --> journal
consensus --> mpst
consensus --> guards
amp --> core
amp --> effects
amp --> journal
amp --> transport
amp --> consensus
amp --> guards
protocol --> core
protocol --> effects
protocol --> composition
protocol --> journal
protocol --> guards
protocol --> consensus
protocol --> amp
protocol --> anti_entropy
protocol --> authorization
protocol --> transport
protocol --> mpst
protocol --> store
%% Layer 5 dependencies
social --> core
social --> journal
chat --> core
chat --> journal
chat --> composition
chat --> guards
relational --> core
relational --> journal
relational --> consensus
relational --> effects
auth --> core
auth --> effects
auth --> journal
auth --> protocol
auth --> guards
auth --> relational
auth --> signature
auth --> authorization
rendezvous --> core
rendezvous --> journal
rendezvous --> guards
rendezvous --> social
invitation --> core
invitation --> effects
invitation --> guards
invitation --> authorization
invitation --> auth
invitation --> journal
invitation --> composition
recovery --> core
recovery --> journal
recovery --> composition
recovery --> signature
recovery --> auth
recovery --> authorization
recovery --> effects
recovery --> protocol
recovery --> relational
sync --> core
sync --> protocol
sync --> guards
sync --> journal
sync --> authorization
sync --> maintenance
sync --> rendezvous
sync --> effects
sync --> anti_entropy
%% Layer 6 dependencies
app --> core
app --> effects
app --> journal
app --> relational
app --> chat
app --> social
app --> maintenance
app --> protocol
app --> recovery
agent --> core
agent --> app
agent --> effects
agent --> composition
agent --> protocol
agent --> guards
agent --> consensus
agent --> journal
agent --> relational
agent --> chat
agent --> auth
agent --> invitation
agent --> rendezvous
agent --> social
agent --> sync
agent --> maintenance
agent --> transport
agent --> recovery
agent --> authorization
agent --> signature
agent --> store
simulator --> core
simulator --> agent
simulator --> effects
simulator --> journal
simulator --> amp
simulator --> consensus
simulator --> protocol
simulator --> testkit
simulator --> sync
simulator --> quint
simulator --> guards
%% Layer 7 dependencies
terminal --> app
terminal --> core
terminal --> agent
terminal --> protocol
terminal --> recovery
terminal --> invitation
terminal --> auth
terminal --> sync
terminal --> effects
terminal --> authorization
terminal --> maintenance
terminal --> chat
terminal --> journal
terminal --> relational
%% Layer 8 dependencies
testkit --> core
testkit --> effects
testkit --> mpst
testkit --> journal
testkit --> relational
testkit --> social
testkit --> transport
testkit --> authorization
testkit --> consensus
testkit --> anti_entropy
testkit --> amp
testkit --> protocol
testkit --> app
quint --> core
quint --> effects
harness --> core
%% Styling
classDef foundation fill:#e1f5fe
classDef spec fill:#f3e5f5
classDef impl fill:#e8f5e9
classDef orch fill:#fff3e0
classDef feature fill:#fce4ec
classDef runtime fill:#f1f8e9
classDef application fill:#e0f2f1
classDef test fill:#ede7f6
class core foundation
class signature,journal,authorization,store,transport,mpst,macros,maintenance spec
class effects,composition impl
class guards,anti_entropy,consensus,amp,protocol orch
class social,chat,relational,auth,rendezvous,invitation,recovery,sync feature
class app,agent,simulator runtime
class terminal application
class testkit,quint,harness test
Effect Trait Classification
Not all effect traits are created equal. Aura organizes effect traits into three categories that determine where their implementations should live:
Fundamental Principle: All effect trait definitions belong in aura-core (Layer 1) to maintain a single source of truth for interfaces. This includes infrastructure effects (OS integration), application effects (domain-specific), and protocol coordination effects (multi-party orchestration).
Infrastructure Effects (Implemented in aura-effects)
Infrastructure effects are truly foundational capabilities that every Aura system needs. These traits define OS-level operations that are universal across all Aura use cases.
Characteristics:
- OS integration (file system, network, cryptographic primitives)
- No Aura-specific semantics
- Reusable across any distributed system
- Required for basic system operation
Examples:
CryptoEffects: Ed25519 signing, key generation, hashingNetworkEffects: TCP connections, message sending/receivingStorageEffects: File read/write, directory operationsPhysicalTimeEffects,LogicalClockEffects,OrderClockEffects: Unified time systemRandomEffects: Cryptographically secure random generationConfigurationEffects: Configuration file parsingConsoleEffects: Terminal input/outputLeakageEffects: Cross-cutting metadata leakage tracking (composable infrastructure)ReactiveEffects: Type-safe signal-based state management for UI and inter-component communication
Implementation Location: These traits have stateless handlers in aura-effects that delegate to OS services.
Application Effects (Implemented in Domain Crates)
Application effects encode Aura-specific abstractions and business logic. These traits capture domain concepts that are meaningful only within Aura's architecture.
Characteristics:
- Aura-specific semantics and domain knowledge
- Built on top of infrastructure effects
- Implement business logic and domain rules
- May have multiple implementations for different contexts
Examples:
JournalEffects: Fact-based journal operations, specific to Aura's CRDT design (aura-journal)AuthorityEffects: Authority-specific operations, central to Aura's identity modelFlowBudgetEffects: Privacy budget management, unique to Aura's information flow control (aura-authorization)AuthorizationEffects: Biscuit token evaluation, tied to Aura's capability system (aura-authorization)RelationalEffects: Cross-authority relationship managementGuardianEffects: Recovery protocol operations
Protocol Coordination Effects (new category):
ChoreographicEffects: Multi-party protocol coordinationEffectApiEffects: Event sourcing and audit for protocolsSyncEffects: Anti-entropy synchronization operations
Implementation Location: Application effects are implemented in their respective domain crates (aura-journal, aura-authorization, etc.). Protocol coordination effects are implemented in Layer 4 orchestration crates (aura-protocol, aura-guards, aura-consensus, aura-amp, aura-anti-entropy) as they manage multi-party state.
Why Not in aura-effects?: Moving these to aura-effects would create circular dependencies. Domain crates need to implement these effects using their own domain logic, but aura-effects cannot depend on domain crates due to the layered architecture.
Implementation Pattern: Domain crates implement application effects by creating domain-specific handler structs that compose infrastructure effects for OS operations while encoding Aura-specific business logic.
#![allow(unused)] fn main() { // Example: aura-journal implements JournalEffects pub struct JournalHandler<C: CryptoEffects, S: StorageEffects> { crypto: C, storage: S, // Domain-specific state } impl<C: CryptoEffects, S: StorageEffects> JournalEffects for JournalHandler<C, S> { async fn append_fact(&self, fact: Fact) -> Result<(), AuraError> { // 1. Domain validation using Aura-specific rules self.validate_fact_semantics(&fact)?; // 2. Cryptographic operations via infrastructure effects let signature = self.crypto.sign(&fact.hash()).await?; // 3. Storage operations via infrastructure effects let entry = JournalEntry { fact, signature }; self.storage.write_chunk(&entry.id(), &entry.encode()).await?; // 4. Domain-specific post-processing self.update_fact_indices(&fact).await?; Ok(()) } } }
Common Effect Placement Mistakes
Here are examples of incorrect effect placement and how to fix them:
#![allow(unused)] fn main() { // WRONG: Domain handler using OS operations directly // File: aura-journal/src/effects.rs impl JournalEffects for BadJournalHandler { async fn read_facts(&self, namespace: Namespace) -> Vec<Fact> { // VIOLATION: Direct file system access in domain handler let data = std::fs::read("journal.dat")?; serde_json::from_slice(&data)? } } // CORRECT: Inject StorageEffects for OS operations impl<S: StorageEffects> JournalEffects for GoodJournalHandler<S> { async fn read_facts(&self, namespace: Namespace) -> Vec<Fact> { // Use injected storage effects let data = self.storage.read_chunk(&namespace.to_path()).await?; self.deserialize_facts(data) } } }
#![allow(unused)] fn main() { // WRONG: Application effect implementation in aura-effects // File: aura-effects/src/journal_handler.rs pub struct JournalHandler { } impl JournalEffects for JournalHandler { // VIOLATION: Domain logic in infrastructure crate async fn validate_fact(&self, fact: &Fact) -> bool { match fact { Fact::TreeOp(op) => self.validate_tree_semantics(op), Fact::Commit(c) => self.validate_commit_rules(c), } } } // CORRECT: Application effects belong in domain crates // File: aura-journal/src/effects.rs impl<C, S> JournalEffects for JournalHandler<C, S> { // Domain validation logic belongs here } }
#![allow(unused)] fn main() { // WRONG: Infrastructure effect in domain crate // File: aura-journal/src/network_handler.rs pub struct CustomNetworkHandler { } impl NetworkEffects for CustomNetworkHandler { // VIOLATION: OS-level networking in domain crate async fn connect(&self, addr: &str) -> TcpStream { TcpStream::connect(addr).await? } } // CORRECT: Use existing NetworkEffects from aura-effects impl<N: NetworkEffects> MyDomainHandler<N> { async fn send_fact(&self, fact: Fact) -> Result<()> { // Compose with injected network effects self.network.send(fact.encode()).await } } }
Key principles for domain effect implementations:
- Domain logic first: Encode business rules and validation specific to the domain
- Infrastructure composition: Use infrastructure effects for OS operations, never direct syscalls
- Clean separation: Domain handlers should not contain OS integration code
- Testability: Mock infrastructure effects for unit testing domain logic
Fallback Handlers and the Null Object Pattern
Infrastructure effects sometimes require fallback implementations for platforms or environments where the underlying capability is unavailable (e.g., biometric hardware on servers, secure enclaves in CI, HSMs in development).
When fallback handlers are appropriate:
- The effect trait represents optional hardware/OS capabilities
- Code must run on platforms without the capability
- Graceful degradation is preferable to compile-time feature flags everywhere
Naming conventions:
- Good:
FallbackBiometricHandler,NoOpSecureEnclaveHandler,UnsupportedHsmHandler - Avoid:
RealBiometricHandler(misleading - implies real implementation)
Fallback handler behavior:
- Return
falsefor capability checks (is_available(),supports_feature()) - Return descriptive errors for operations (
Err(NotSupported)) - Never panic or silently succeed when the capability is unavailable
For a checklist on removing stub handlers, see CLAUDE.md under "Agent Decision Aids".
Why this matters: A fallback handler is not dead code if its trait is actively used. It's the Null Object Pattern providing safe defaults. The architectural violation is a misleading name, not the existence of the fallback.
Composite Effects (Convenience Extensions)
Composite effects provide convenience methods that combine multiple lower-level operations. These are typically extension traits that add domain-specific convenience to infrastructure effects.
Characteristics:
- Convenience wrappers around other effects
- Domain-specific combinations of operations
- Often implemented as blanket implementations
- Improve developer ergonomics
Examples:
TreeEffects: CombinesCryptoEffectsandStorageEffectsfor merkle tree operationsSimulationEffects: Testing-specific combinations for deterministic simulationLeakageChoreographyExt: Combines leakage tracking with choreography operations
Implementation Location: Usually implemented as extension traits in aura-core or as blanket implementations in domain crates.
Effect Classification
For quick decision aids (decision matrix, decision tree), see CLAUDE.md under "Agent Decision Aids".
Examples:
- CryptoEffects → Infrastructure (OS crypto, no Aura semantics, reusable)
- JournalEffects → Application (Aura facts, domain validation, not reusable)
- NetworkEffects → Infrastructure (TCP/UDP, no domain logic, reusable)
- FlowBudgetEffects → Application (Aura privacy model, domain rules)
This classification ensures that:
- Infrastructure effects have reliable, stateless implementations available in
aura-effects - Application effects can evolve with their domain logic in domain crates
- Composite effects provide ergonomic interfaces without architectural violations
- The dependency graph remains acyclic
- Domain knowledge stays in domain crates, OS knowledge stays in infrastructure
- Clean composition enables testing domain logic independently of OS integration
Architecture Principles
No Circular Dependencies
Each layer builds on lower layers without reaching back down. This enables independent testing, reusability, and clear responsibility boundaries.
The layered architecture means that Layer 1 has no dependencies on any other Aura crate. Layer 2 depends only on Layer 1. Layer 3 depends on Layers 1 and 2. This pattern continues through all 8 layers.
Code Location Policy
The 8-layer architecture enforces strict placement rules. Violating these rules creates circular dependencies or breaks architectural invariants.
For a quick reference table of layer rules, see CLAUDE.md under "Agent Decision Aids".
For practical guidance on effects and handlers, see Effects Guide. For choreography development, see Choreography Guide.
Pure Mathematical Utilities
Some effect traits in aura-core (e.g., BloomEffects) represent pure mathematical operations without OS integration. These follow the standard trait/handler pattern for consistency, but are technically not "effects" in the algebraic sense (no side effects).
This is acceptable technical debt - the pattern consistency outweighs the semantic impurity. Future refactoring may move pure math to methods on types in aura-core.
Architectural Compliance Checking
The project includes an automated architectural compliance checker to enforce these layering principles:
Command: just check-arch
Script: scripts/check/arch.sh
What it validates:
- Layer boundary violations (no upward dependencies)
- Dependency direction (Lx→Ly where y≤x only)
- Effect trait classification and placement
- Domain effect implementation patterns
- Stateless handler requirements in
aura-effects(noArc<Mutex>,Arc<RwLock>) - Mock handler location in
aura-testkit - Guard chain integrity (no bypass of CapGuard → FlowGuard → JournalCoupler)
- Impure function routing through effects (
SystemTime::now,thread_rng, etc.) - Physical time guardrails (
tokio::time::sleepconfinement) - Handler composition patterns (no direct instantiation)
- Placeholder/TODO detection
- Invariants documentation schema validation
The checker reports violations that must be fixed and warnings for review. Run it before submitting changes to ensure architectural compliance.
Feature Flags
Aura uses a minimal set of deliberate feature flags organized into three tiers.
Tier 1: Workspace-Wide Features
| Feature | Crate | Purpose |
|---|---|---|
simulation | aura-core, aura-effects | Enables simulation/testing effect traits and handlers. Required by aura-simulator, aura-quint, aura-testkit. |
proptest | aura-core | Property-based testing support via the proptest crate. |
Tier 2: Platform Features (aura-app)
| Feature | Purpose |
|---|---|
native | Rust consumers (aura-terminal, tests). Enables futures-signals API. |
ios | iOS via UniFFI → Swift bindings. |
android | Android via UniFFI → Kotlin bindings. |
wasm | Web via wasm-bindgen → JavaScript/TypeScript. |
web-dominator | Pure Rust WASM apps using futures-signals + dominator. |
Development features: instrumented (tracing), debug-serialize (JSON debug output), host (binding stub).
Tier 3: Crate-Specific Features
| Crate | Feature | Purpose |
|---|---|---|
aura-terminal | terminal | TUI mode (default). |
aura-terminal | development | Includes simulator, testkit, debug features. |
aura-testkit | full-effect-system | Optional aura-agent integration for higher-layer tests. |
aura-testkit | lean | Lean oracle for differential testing against formal models. |
aura-agent | dev-console | Optional development console server. |
aura-agent | real-android-keystore | Real Android Keystore implementation. |
aura-mpst | debug | Choreography debugging output. |
aura-macros | proc-macro | Proc-macro compilation (default). |
Feature Usage Examples
# Standard development (default features)
cargo build
# With simulation support
cargo build -p aura-core --features simulation
# Terminal with development tools
cargo build -p aura-terminal --features development
# Lean differential testing (requires Lean toolchain)
just lean-oracle-build
cargo test -p aura-testkit --features lean --test lean_differential
# iOS build
cargo build -p aura-app --features ios
Effect System and Impure Function Guidelines
Core Principle: Deterministic Simulation
Aura's effect system ensures fully deterministic simulation by requiring all impure operations (time, randomness, filesystem, network) to flow through effect traits. This enables:
- Predictable testing: Mock all external dependencies for unit tests
- WASM compatibility: No blocking operations or OS thread assumptions
- Cross-platform support: Same code runs in browsers and native environments
- Simulation fidelity: Virtual time and controlled randomness for property testing
Impure Function Classification
FORBIDDEN: Direct impure function usage
#![allow(unused)] fn main() { // VIOLATION: Direct system calls let now = SystemTime::now(); let random = thread_rng().gen::<u64>(); let file = File::open("data.txt")?; let socket = TcpStream::connect("127.0.0.1:8080").await?; // VIOLATION: Global state static CACHE: Mutex<HashMap<String, String>> = Mutex::new(HashMap::new()); }
REQUIRED: Effect trait usage
#![allow(unused)] fn main() { // CORRECT: Via effect traits with explicit context async fn my_operation<T: TimeEffects + RandomEffects + StorageEffects>( ctx: &EffectContext, effects: &T, ) -> Result<ProcessedData> { let timestamp = effects.current_time().await; let nonce = effects.random_bytes(32).await?; let data = effects.read_chunk(&chunk_id).await?; // ... business logic with pure functions Ok(ProcessedData { timestamp, nonce, data }) } }
Legitimate Effect Injection Sites
The architectural compliance checker ONLY allows direct impure function usage in these specific locations:
1. Effect Handler Implementations (aura-effects)
#![allow(unused)] fn main() { // ALLOWED: Production effect implementations impl PhysicalTimeEffects for PhysicalTimeHandler { async fn physical_time(&self) -> Result<PhysicalTime, TimeError> { // OK: This IS the effect implementation let now = SystemTime::now().duration_since(UNIX_EPOCH)?; Ok(PhysicalTime::from_ms(now.as_millis() as u64)) } } impl RandomCoreEffects for RealRandomHandler { async fn random_bytes(&self, len: usize) -> Vec<u8> { let mut bytes = vec![0u8; len]; rand::thread_rng().fill_bytes(&mut bytes); // OK: Legitimate OS randomness source bytes } } }
2. Runtime Effect Assembly (runtime/effects.rs)
#![allow(unused)] fn main() { // ALLOWED: Effect system bootstrapping pub fn create_production_effects() -> AuraEffectSystem { AuraEffectSystemBuilder::new() .with_handler(Arc::new(PhysicalTimeHandler::new())) .with_handler(Arc::new(RealRandomHandler::new())) // OK: Assembly point .build() } }
3. Pure Functions (aura-core::hash)
#![allow(unused)] fn main() { // ALLOWED: Deterministic, pure operations pub fn hash(data: &[u8]) -> [u8; 32] { blake3::hash(data).into() // OK: Pure function, no external state } }
Exemption Rationale
Why these exemptions are architecturally sound:
- Effect implementations MUST access the actual system - that's their purpose
- Runtime assembly is the controlled injection point where production vs. mock effects are chosen
- Pure functions are deterministic regardless of when/where they're called
Why broad exemptions are dangerous:
- Crate-level exemptions (
aura-agent,aura-protocol,aura-guards,aura-consensus,aura-amp,aura-anti-entropy) would allow business logic to bypass effects - This breaks simulation determinism and WASM compatibility
- Makes testing unreliable by introducing hidden external dependencies
Effect System Usage Patterns
Correct: Infrastructure Effects in aura-effects
#![allow(unused)] fn main() { // File: crates/aura-effects/src/transport/tcp.rs pub struct TcpTransportHandler { config: TransportConfig, } impl TcpTransportHandler { pub async fn connect(&self, addr: TransportSocketAddr) -> TransportResult<TransportConnection> { let stream = TcpStream::connect(*addr.as_socket_addr()).await?; // OK: Implementation // ... connection setup Ok(connection) } } }
Correct: Domain Effects in Domain Crates
#![allow(unused)] fn main() { // File: crates/aura-journal/src/effects.rs pub struct JournalHandler<C: CryptoEffects, S: StorageEffects> { crypto: C, storage: S, } impl<C: CryptoEffects, S: StorageEffects> JournalEffects for JournalHandler<C, S> { async fn append_fact(&self, ctx: &EffectContext, fact: Fact) -> Result<()> { // Domain validation (pure) self.validate_fact_semantics(&fact)?; // Infrastructure effects for impure operations let signature = self.crypto.sign(&fact.hash()).await?; self.storage.write_chunk(&entry.id(), &entry.encode()).await?; Ok(()) } } }
Violation: Direct impure access in domain logic
#![allow(unused)] fn main() { // File: crates/aura-core/src/crypto/tree_signing.rs pub async fn start_frost_ceremony() -> Result<()> { let start_time = SystemTime::now(); // VIOLATION: Should use TimeEffects let session_id = Uuid::new_v4(); // VIOLATION: Should use RandomEffects // This breaks deterministic simulation! ceremony_with_timing(start_time, session_id).await } }
Context Propagation Requirements
All async operations must propagate EffectContext:
#![allow(unused)] fn main() { // CORRECT: Explicit context propagation async fn process_request<T: AllEffects>( ctx: &EffectContext, // Required for tracing/correlation effects: &T, request: Request, ) -> Result<Response> { let start = effects.current_time().await; // Context flows through the call chain let result = process_business_logic(ctx, effects, request.data).await?; let duration = effects.current_time().await.duration_since(start)?; tracing::info!( request_id = %ctx.request_id, duration_ms = duration.as_millis(), "Request processed" ); Ok(result) } }
Mock Testing Pattern
Tests use controllable mock effects:
#![allow(unused)] fn main() { // File: tests/integration/frost_test.rs #[tokio::test] async fn test_frost_ceremony_timing() { // Controllable time for deterministic tests let mock_time = SimulatedTimeHandler::new(); mock_time.set_time(PhysicalTime::from_ms(1000_000)); let effects = TestEffectSystem::new() .with_time(mock_time) .with_random(MockRandomHandler::deterministic()) .build(); let ctx = EffectContext::test(); // Test runs deterministically regardless of wall-clock time let result = start_frost_ceremony(&ctx, &effects).await; assert!(result.is_ok()); } }
WASM Compatibility Guidelines
Forbidden in all crates (except effect implementations):
std::thread- No OS threads in WASMstd::fs- No filesystem in browsersSystemTime::now()- Time must be injectedrand::thread_rng()- Randomness must be controllable- Blocking operations - Everything must be async
Required patterns:
- Async/await for all I/O operations
- Effect trait injection for all impure operations
- Explicit context propagation through call chains
- Builder patterns for initialization with async setup
Compliance Checking
The just check-arch command validates these principles by:
- Scanning for direct impure usage: Detects
SystemTime::now,thread_rng(),std::fs::, etc. - Enforcing precise exemptions: Only allows usage in
impl.*Effects,runtime/effects.rs - Context propagation validation: Warns about async functions without
EffectContext - Global state detection: Catches
lazy_static,Mutex<static>anti-patterns
Run before every commit to maintain architectural compliance and simulation determinism.
Serialization Policy
Aura uses DAG-CBOR as the canonical serialization format for:
- Wire protocols: Network messages between peers
- Facts: CRDT state, journal entries, attestations
- Cryptographic commitments: Content-addressable hashes (determinism required)
Canonical Module
All serialization should use aura_core::util::serialization:
#![allow(unused)] fn main() { use aura_core::util::serialization::{to_vec, from_slice, hash_canonical}; use aura_core::util::serialization::{VersionedMessage, SemanticVersion}; // Serialize to DAG-CBOR let bytes = to_vec(&value)?; // Deserialize from DAG-CBOR let value: MyType = from_slice(&bytes)?; // Content-addressable hash (deterministic) let hash = hash_canonical(&value)?; }
Why DAG-CBOR?
- Deterministic canonical encoding: Required for FROST threshold signatures where all parties must produce identical bytes
- Content-addressable: IPLD compatibility for content hashing and Merkle trees
- Forward/backward compatible: Semantic versioning support via
VersionedMessage<T> - Efficient binary encoding: Better than JSON, comparable to bincode
Allowed Alternatives
| Format | Use Case | Example |
|---|---|---|
serde_json | User-facing config files | .aura/config.json |
serde_json | Debug output and logging | tracing spans |
serde_json | Dynamic metadata | HashMap<String, Value> |
Versioned Facts Pattern
All fact types should use the versioned serialization pattern:
#![allow(unused)] fn main() { use aura_core::util::serialization::{to_vec, from_slice, SerializationError}; const CURRENT_VERSION: u32 = 1; impl MyFact { pub fn to_bytes(&self) -> Result<Vec<u8>, SerializationError> { to_vec(self) } pub fn from_bytes(bytes: &[u8]) -> Result<Self, SerializationError> { // Try DAG-CBOR first, fall back to JSON for compatibility from_slice(bytes).or_else(|_| { serde_json::from_slice(bytes) .map_err(|e| SerializationError::Deserialization(e.to_string())) }) } } }
Enforcement
The just check-arch --serialization command validates:
- Wire protocol files use canonical serialization
- Facts files use versioned serialization
Invariant Traceability
This section indexes invariants across Aura and maps them to enforcement loci. Invariant specifications live in crate ARCHITECTURE.md files. Contracts in Theoretical Model, Privacy and Information Flow Contract, and Distributed Systems Contract define the cross-crate safety model.
Canonical Naming
Use InvariantXxx names in proofs and tests. Use prose aliases for readability when needed. When both forms appear, introduce the alias once and then reference the canonical name.
Examples:
Charge-Before-Sendmaps toInvariantSentMessagesHaveFactsandInvariantFlowBudgetNonNegative.Context Isolationmaps toInvariantContextIsolation.Secure Channel Lifecyclemaps toInvariantReceiptValidityWindowandInvariantCrossEpochReplayPrevention.
Use shared terminology from Theoretical Model:
- Role terms:
Member,Participant,Moderator - Access terms:
AccessLevelwithFull,Partial,Limited - Storage/pinning terms:
Shared Storage,allocation, andpinnedfacts
Core Invariant Index
| Alias | Canonical Name(s) | Primary Enforcement | Related Contracts |
|---|---|---|---|
| Charge-Before-Send | InvariantSentMessagesHaveFacts, InvariantFlowBudgetNonNegative | crates/aura-guards/ARCHITECTURE.md | Privacy and Information Flow Contract, Distributed Systems Contract |
| CRDT Convergence | InvariantCRDTConvergence | crates/aura-journal/ARCHITECTURE.md | Theoretical Model, Distributed Systems Contract |
| Context Isolation | InvariantContextIsolation | crates/aura-core/ARCHITECTURE.md | Theoretical Model, Privacy and Information Flow Contract, Distributed Systems Contract |
| Secure Channel Lifecycle | InvariantSecureChannelLifecycle, InvariantReceiptValidityWindow, InvariantCrossEpochReplayPrevention | crates/aura-rendezvous/ARCHITECTURE.md | Privacy and Information Flow Contract, Distributed Systems Contract |
| Authority Tree Topology and Commitment Coherence | InvariantAuthorityTreeTopologyCommitmentCoherence | crates/aura-journal/ARCHITECTURE.md | Theoretical Model, Distributed Systems Contract |
Distributed Contract Invariants
The distributed and privacy contracts define additional canonical names used by proofs and conformance tests:
InvariantUniqueCommitPerInstanceInvariantCommitRequiresThresholdInvariantEquivocatorsExcludedInvariantNonceUniqueInvariantSequenceMonotonicInvariantReceiptValidityWindowInvariantCrossEpochReplayPreventionInvariantVectorClockConsistentInvariantHonestMajorityCanCommitInvariantCompromisedNoncesExcluded
When a crate enforces one of these invariants, record the same canonical name in that crate's ARCHITECTURE.md.
Traceability Matrix
This matrix provides a single cross-reference for contract names, owning crate docs, and proof/test artifacts.
| Canonical Name | Crate Architecture Spec | Proof/Test Artifact |
|---|---|---|
InvariantSentMessagesHaveFacts | crates/aura-guards/ARCHITECTURE.md | verification/quint/transport.qnt |
InvariantFlowBudgetNonNegative | crates/aura-guards/ARCHITECTURE.md | verification/quint/transport.qnt |
InvariantContextIsolation | crates/aura-core/ARCHITECTURE.md, crates/aura-transport/ARCHITECTURE.md | verification/quint/transport.qnt |
InvariantSequenceMonotonic | crates/aura-transport/ARCHITECTURE.md | verification/quint/transport.qnt |
InvariantReceiptValidityWindow | crates/aura-rendezvous/ARCHITECTURE.md | verification/quint/epochs.qnt |
InvariantCrossEpochReplayPrevention | crates/aura-rendezvous/ARCHITECTURE.md | verification/quint/epochs.qnt |
InvariantNonceUnique | crates/aura-journal/ARCHITECTURE.md | verification/quint/journal/core.qnt |
InvariantVectorClockConsistent | crates/aura-anti-entropy/ARCHITECTURE.md | verification/quint/journal/anti_entropy.qnt |
InvariantUniqueCommitPerInstance | crates/aura-consensus/ARCHITECTURE.md | verification/quint/consensus/core.qnt, verification/lean/Aura/Proofs/Consensus/Agreement.lean |
InvariantCommitRequiresThreshold | crates/aura-consensus/ARCHITECTURE.md | verification/quint/consensus/core.qnt, verification/lean/Aura/Proofs/Consensus/Validity.lean |
InvariantEquivocatorsExcluded | crates/aura-consensus/ARCHITECTURE.md | verification/quint/consensus/core.qnt, verification/lean/Aura/Proofs/Consensus/Adversary.lean |
InvariantHonestMajorityCanCommit | crates/aura-consensus/ARCHITECTURE.md | verification/quint/consensus/adversary.qnt, verification/lean/Aura/Proofs/Consensus/Adversary.lean |
InvariantCompromisedNoncesExcluded | crates/aura-consensus/ARCHITECTURE.md | verification/quint/consensus/adversary.qnt |
Use just check-invariants to validate system invariants across the workspace.