Overview
Thecactus convert command transforms models from HuggingFace format to Cactus format with quantization. Supports merging LoRA adapters into base models.
Syntax
Arguments
<model>- Model name or HuggingFace repository[output_dir]- Optional output directory (default:./weights/<model-name>)
Flags
—precision
Set the quantization precision level:INT4
Options:
INT4- 4-bit quantization (smallest size)INT8- 8-bit quantization (balanced)FP16- 16-bit floating point (highest quality)
—lora
Merge a LoRA adapter into the base model:- Local LoRA adapter directories
- HuggingFace LoRA repositories
- Multiple LoRA adapters (specify flag multiple times)
—token
Provide a HuggingFace API token for downloading source models:Examples
Conversion Process
The conversion pipeline includes:- Download - Fetch source model from HuggingFace (if needed)
- LoRA Merge - Apply LoRA adapters to base weights (if specified)
- Quantization - Convert to target precision level
- Optimization - Apply Cactus-specific optimizations
- Export - Write converted model to output directory
LoRA Adapter Format
Supported LoRA formats:Local Directory
HuggingFace Repository
Output Format
Converted models include:Use Cases
Fine-tuned Models
Convert your custom fine-tuned models:LoRA Experimentation
Test different LoRA combinations:Precision Optimization
Create multiple precision variants:See Also
Download Command
Download models without custom conversion
Run Command
Run converted models interactively