fuel_tx/transaction/
metadata.rsuse alloc::vec::Vec;
use fuel_types::{
canonical::Serialize,
Bytes32,
ChainId,
};
use crate::{
field,
UniqueIdentifier,
ValidityError,
};
pub trait Cacheable {
fn is_computed(&self) -> bool;
fn precompute(&mut self, chain_id: &ChainId) -> Result<(), ValidityError>;
}
impl Cacheable for super::Transaction {
fn is_computed(&self) -> bool {
match self {
Self::Script(tx) => tx.is_computed(),
Self::Create(tx) => tx.is_computed(),
Self::Mint(tx) => tx.is_computed(),
Self::Upgrade(tx) => tx.is_computed(),
Self::Upload(tx) => tx.is_computed(),
Self::Blob(tx) => tx.is_computed(),
}
}
fn precompute(&mut self, chain_id: &ChainId) -> Result<(), ValidityError> {
match self {
Self::Script(tx) => tx.precompute(chain_id),
Self::Create(tx) => tx.precompute(chain_id),
Self::Mint(tx) => tx.precompute(chain_id),
Self::Upgrade(tx) => tx.precompute(chain_id),
Self::Upload(tx) => tx.precompute(chain_id),
Self::Blob(tx) => tx.precompute(chain_id),
}
}
}
#[derive(Debug, Clone, PartialEq, Eq, Hash)]
pub struct CommonMetadata {
pub id: Bytes32,
pub inputs_offset: usize,
pub inputs_offset_at: Vec<usize>,
pub inputs_predicate_offset_at: Vec<Option<(usize, usize)>>,
pub outputs_offset: usize,
pub outputs_offset_at: Vec<usize>,
pub witnesses_offset: usize,
pub witnesses_offset_at: Vec<usize>,
}
impl CommonMetadata {
pub fn compute<Tx>(tx: &Tx, chain_id: &ChainId) -> Result<Self, ValidityError>
where
Tx: UniqueIdentifier,
Tx: field::Inputs,
Tx: field::Outputs,
Tx: field::Witnesses,
{
use itertools::Itertools;
let id = tx.id(chain_id);
let inputs_predicate_offset_at = tx
.inputs()
.iter()
.enumerate()
.map(|(i, _)| tx.inputs_predicate_offset_at(i))
.collect_vec();
let mut offset = tx.inputs_offset();
let mut inputs_offset_at = Vec::new();
for (index, input) in tx.inputs().iter().enumerate() {
let i = offset;
offset = offset
.checked_add(input.size())
.ok_or(ValidityError::SerializedInputTooLarge { index })?;
inputs_offset_at.push(i);
}
let mut offset = tx.outputs_offset();
let mut outputs_offset_at = Vec::new();
for (index, output) in tx.outputs().iter().enumerate() {
let i = offset;
offset = offset
.checked_add(output.size())
.ok_or(ValidityError::SerializedOutputTooLarge { index })?;
outputs_offset_at.push(i);
}
let mut offset = tx.witnesses_offset();
let mut witnesses_offset_at = Vec::new();
for (index, witnesses) in tx.witnesses().iter().enumerate() {
let i = offset;
offset = offset
.checked_add(witnesses.size())
.ok_or(ValidityError::SerializedWitnessTooLarge { index })?;
witnesses_offset_at.push(i);
}
Ok(Self {
id,
inputs_offset: tx.inputs_offset(),
inputs_offset_at,
inputs_predicate_offset_at,
outputs_offset: tx.outputs_offset(),
outputs_offset_at,
witnesses_offset: tx.witnesses_offset(),
witnesses_offset_at,
})
}
}