Uzumaki
Narutoouz
AI & ML interests
None yet
Recent Activity
liked a model about 18 hours ago
z-lab/Qwen3.5-27B-DFlash new activity about 19 hours ago
google/gemma-4-31B-it:Guys please add the MTP to this model new activity about 19 hours ago
nvidia/Gemma-4-31B-IT-NVFP4:Why is this 4bit version has a 32.7 GB size?Organizations
Guys please add the MTP to this model
🔥 2
1
#50 opened about 19 hours ago
by
Narutoouz
Why is this 4bit version has a 32.7 GB size?
12
#3 opened 6 days ago
by
alexcardo
where is minimax 2.7
🔥 2
8
#54 opened 15 days ago
by
devops724
can we get minimax-m2.7
🤗 13
5
#49 opened 22 days ago
by
CHNtentes
Ideal Sampling parameters to reproduce benchmarks
1
#3 opened 15 days ago
by
Narutoouz
Feature Request: TFLite Q4/Q6/Q8 Quantizations for Nanbeige4.1-3B
1
#42 opened 24 days ago
by
Narutoouz
Need support for mlx inference
1
#1 opened 27 days ago
by
Narutoouz
please upload benchmarks
1
#2 opened 28 days ago
by
Narutoouz
mlx lm support
👍 1
#7 opened about 1 month ago
by
Narutoouz
Any Plans for an Instruct Model?
🤗🔥 6
6
#15 opened about 2 months ago
by
Ashacorporation
Model "thinks" for too long
👍 3
11
#12 opened about 2 months ago
by
Moisha1985
mlx version please
#1 opened about 1 month ago
by
Narutoouz
Insufficient context length
4
#2 opened about 1 month ago
by
X-SZM
please make mlx lm and gguf version
🚀 1
1
#1 opened about 1 month ago
by
Narutoouz
Can you make dwq 3bit and 4bit quant
#2 opened about 2 months ago
by
Narutoouz
can you make nvfp4 quant
#1 opened about 2 months ago
by
Narutoouz
please make 3 bit & 4 bit dwq quant of cerebras/MiniMax-M2.5-REAP-172B-A10B
#5 opened about 2 months ago
by
Narutoouz
can anybody make nvfp4 mlx quant ?
#2 opened about 2 months ago
by
Narutoouz
mlx lm and llama.cpp support
#9 opened about 2 months ago
by
Narutoouz
Support for mlx lm and llama.cpp
👍 1
#8 opened about 2 months ago
by
Narutoouz