Trend models for DGX Spark β¨
Collection
Build in DGX Spark, for trended models β¨ β’ 9 items β’ Updated
You can easily run this model using the DGX-Spark-llama.cpp-Bench inference engine. It's pre-configured for high-performance inference on NVIDIA hardware (especially Blackwell/DGX Spark).
docker pull ghcr.io/sowilow/dgx-spark-llama.cpp-bench:latest
For detailed configuration and usage, visit the GitHub Repository.
This repository contains GGUF-quantized weights for LFM2.5-1.2B-Instruct, specifically optimized for NVIDIA Blackwell (DGX Spark) hardware.
This model is a quantized version of the original LiquidAI/LFM2.5-1.2B-Instruct and is subject to its original license.
lfm2.5-1.2b-instruct-q4_k_m.gguf: 4-bit quantized model.lfm2.5-1.2b-instruct-q8_0.gguf: 8-bit quantized model.Created using DGX-Spark-llama.cpp-Bench
4-bit
8-bit
Base model
LiquidAI/LFM2.5-1.2B-Base