llama_cpp_2/llama_batch.rs
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216
//! Safe wrapper around `llama_batch`.
use crate::token::LlamaToken;
use llama_cpp_sys_2::{llama_batch, llama_batch_free, llama_batch_init, llama_pos, llama_seq_id};
/// A safe wrapper around `llama_batch`.
#[derive(Debug)]
pub struct LlamaBatch {
/// The number of tokens the batch was allocated with. they are safe to write to - but not necessarily read from as they are not necessarily initialized
allocated: usize,
/// The logits that are initialized. Used by [`LlamaContext`] to ensure that only initialized logits are accessed.
pub(crate) initialized_logits: Vec<i32>,
#[allow(clippy::doc_markdown)]
/// The llama_cpp batch. always initialize by `llama_cpp_sys_2::llama_batch_init(allocated, <unknown>, <unknown>)`
pub(crate) llama_batch: llama_batch,
}
/// Errors that can occur when adding a token to a batch.
#[derive(thiserror::Error, Debug, PartialEq, Eq)]
pub enum BatchAddError {
/// There was not enough space in the batch to add the token.
#[error("Insufficient Space of {0}")]
InsufficientSpace(usize),
/// Empty buffer is provided for [`LlamaBatch::get_one`]
#[error("Empty buffer")]
EmptyBuffer,
}
impl LlamaBatch {
/// Clear the batch. This does not free the memory associated with the batch, but it does reset
/// the number of tokens to 0.
pub fn clear(&mut self) {
self.llama_batch.n_tokens = 0;
self.initialized_logits.clear();
}
/// add a token to the batch for sequences `seq_ids` at position `pos`. If `logits` is true, the
/// token will be initialized and can be read from after the next decode.
///
/// # Panics
///
/// - [`self.llama_batch.n_tokens`] does not fit into a usize
/// - [`seq_ids.len()`] does not fit into a [`llama_seq_id`]
///
/// # Errors
///
/// returns a error if there is insufficient space in the buffer
pub fn add(
&mut self,
LlamaToken(id): LlamaToken,
pos: llama_pos,
seq_ids: &[i32],
logits: bool,
) -> Result<(), BatchAddError> {
if self.allocated
< usize::try_from(self.n_tokens() + 1).expect("cannot fit n_tokens into a usize")
{
return Err(BatchAddError::InsufficientSpace(self.allocated));
}
let offset = self.llama_batch.n_tokens;
let offset_usize = usize::try_from(offset).expect("cannot fit n_tokens into a usize");
unsafe {
// batch.token [batch.n_tokens] = id;
self.llama_batch.token.add(offset_usize).write(id);
// batch.pos [batch.n_tokens] = pos,
self.llama_batch.pos.add(offset_usize).write(pos);
// batch.n_seq_id[batch.n_tokens] = seq_ids.size();
self.llama_batch.n_seq_id.add(offset_usize).write(
llama_seq_id::try_from(seq_ids.len())
.expect("cannot fit seq_ids.len() into a llama_seq_id"),
);
// for (size_t i = 0; i < seq_ids.size(); ++i) {
// batch.seq_id[batch.n_tokens][i] = seq_ids[i];
// }
for (i, seq_id) in seq_ids.iter().enumerate() {
let tmp = *self.llama_batch.seq_id.add(offset_usize);
tmp.add(i).write(*seq_id);
}
// batch.logits [batch.n_tokens] = logits;
self.llama_batch
.logits
.add(offset_usize)
.write(i8::from(logits));
}
if logits {
self.initialized_logits.push(offset);
} else {
self.initialized_logits.retain(|l| l != &offset);
}
// batch.n_tokens++;
self.llama_batch.n_tokens += 1;
Ok(())
}
/// Add a sequence of tokens to the batch for the given sequence id. If `logits_all` is true, the
/// tokens will be initialized and can be read from after the next decode.
///
/// Either way the last token in the sequence will have its logits set to `true`.
///
/// # Errors
///
/// Returns an error if there is insufficient space in the buffer
///
/// # Panics
///
/// - [`self.llama_batch.n_tokens`] does not fit into a [`usize`]
/// - [`n_tokens - 1`] does not fit into a [`llama_pos`]
pub fn add_sequence(
&mut self,
tokens: &[LlamaToken],
seq_id: i32,
logits_all: bool,
) -> Result<(), BatchAddError> {
let n_tokens_0 =
usize::try_from(self.llama_batch.n_tokens).expect("cannot fit n_tokens into a usize");
let n_tokens = tokens.len();
if self.allocated < n_tokens_0 + n_tokens {
return Err(BatchAddError::InsufficientSpace(self.allocated));
}
let last_index = llama_pos::try_from(n_tokens.saturating_sub(1))
.expect("cannot fit n_tokens into a llama_pos");
for (i, token) in (0..).zip(tokens.iter()) {
self.add(*token, i, &[seq_id], logits_all || i == last_index)?;
}
Ok(())
}
/// Create a new `LlamaBatch` that can contain up to `n_tokens` tokens.
///
/// # Arguments
///
/// - `n_tokens`: the maximum number of tokens that can be added to the batch
/// - `n_seq_max`: the maximum number of sequences that can be added to the batch (generally 1 unless you know what you are doing)
///
/// # Panics
///
/// Panics if `n_tokens` is greater than `i32::MAX`.
#[must_use]
pub fn new(n_tokens: usize, n_seq_max: i32) -> Self {
let n_tokens_i32 = i32::try_from(n_tokens).expect("cannot fit n_tokens into a i32");
let batch = unsafe { llama_batch_init(n_tokens_i32, 0, n_seq_max) };
LlamaBatch {
allocated: n_tokens,
initialized_logits: vec![],
llama_batch: batch,
}
}
/// ``llama_batch_get_one``
/// Return batch for single sequence of tokens
///
/// NOTE: this is a helper function to facilitate transition to the new batch API
///
/// # Errors
/// If the provided token buffer is empty.
///
/// # Panics
/// If the number of tokens in ``tokens`` exceeds [`i32::MAX`].
pub fn get_one(tokens: &[LlamaToken]) -> Result<Self, BatchAddError> {
if tokens.is_empty() {
return Err(BatchAddError::EmptyBuffer);
}
let batch = unsafe {
let ptr = tokens.as_ptr() as *mut i32;
llama_cpp_sys_2::llama_batch_get_one(
ptr,
tokens
.len()
.try_into()
.expect("number of tokens exceeds i32::MAX"),
)
};
let batch = Self {
allocated: 0,
initialized_logits: vec![(tokens.len() - 1)
.try_into()
.expect("number of tokens exceeds i32::MAX + 1")],
llama_batch: batch,
};
Ok(batch)
}
/// Returns the number of tokens in the batch.
#[must_use]
pub fn n_tokens(&self) -> i32 {
self.llama_batch.n_tokens
}
}
impl Drop for LlamaBatch {
/// Drops the `LlamaBatch`.
///
/// ```
/// # use llama_cpp_2::llama_batch::LlamaBatch;
/// # use std::error::Error;
/// # fn main() -> Result<(), Box<dyn Error>> {
/// let batch = LlamaBatch::new(512, 1);
/// // frees the memory associated with the batch. (allocated by llama.cpp)
/// drop(batch);
/// # Ok(())
/// # }
fn drop(&mut self) {
unsafe {
if self.allocated > 0 {
llama_batch_free(self.llama_batch);
}
}
}
}