Changes for version 0.11 - 2026-01-21

  • Test coverage improvements
  • memory leak tests and fixes

Modules

Pure C LLM Inference Engine for Perl (built on ggml)
Memory Context for Tensor Allocation
Computation Graph for Tensor Operations
Transformer Forward Pass and Token Generation
KV Cache for efficient incremental decoding
Low-Rank Adaptation (LoRA) adapter support for Lugh
GGUF Model Loading and Tensor Access
Tensor Operations for Neural Network Computation
Chat Template Formatting for LLM Conversations
Quantization utilities for Lugh tensors
RoPE (Rotary Position Embedding) Scaling Configuration
Speculative decoding for faster LLM inference
N-Dimensional Tensor with ggml Backend
BPE Tokenizer for Text Encoding and Decoding