Changes for version 0.04 - 2026-01-18
- Added KV Cache support for efficient incremental decoding - Lugh::KVCache
- Lugh::Inference - New create_kvcache() and forward_with_cache() methods
- New test file t/08-kvcache.t
Modules
Pure C LLM Inference Engine for Perl (built on ggml)
Memory Context for Tensor Allocation
Computation Graph for Tensor Operations
Transformer Forward Pass and Token Generation
KV Cache for efficient incremental decoding
GGUF Model Loading and Tensor Access
Tensor Operations for Neural Network Computation
N-Dimensional Tensor with ggml Backend
BPE Tokenizer for Text Encoding and Decoding