How to use from the
Use from the
llama-cpp-python library
# !pip install llama-cpp-python

from llama_cpp import Llama

llm = Llama.from_pretrained(
	repo_id="darkmaniac7/tokforge-assets",
	filename="bge-small/bge-small-en-v1.5-q4_k_m.gguf",
)
output = llm(
	"Once upon a time,",
	max_tokens=512,
	echo=True
)
print(output)

tokforge-assets

Support assets used by TokForge.

This repository is not a standalone language model release. It stores auxiliary runtime assets that are useful to the app and related tooling.

Current contents

  • kokoro-en-v0_19/model.int8.onnx

Intended use

  • TokForge app support files
  • voice / auxiliary feature experiments
  • reproducible hosting of shared runtime assets

Limitations

  • Not a full model card for a standalone chat model
  • No Hugging Face Inference widget is expected here
  • Contents may evolve as TokForge asset packaging changes

TokForge

Downloads last month
171
GGUF
Model size
33.2M params
Architecture
bert
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support