Hugging Face
Models
Datasets
Spaces
Buckets
new
Docs
Enterprise
Pricing
Website
Tasks
HuggingChat
Collections
Languages
Organizations
Community
Blog
Posts
Daily Papers
Learn
Discord
Forum
GitHub
Solutions
Team & Enterprise
Hugging Face PRO
Enterprise Support
Inference Providers
Inference Endpoints
Storage Buckets
Log In
Sign Up
In a Training Loop 🔄
5
Eren Ekşi
PRO
ereniko
Follow
wop's profile picture
Alptekinege's profile picture
Fishtiks's profile picture
3 followers
·
5 following
AI & ML interests
None yet
Recent Activity
posted
an
update
2 days ago
Something concerning is happening with Nvidia's GPU pricing and benchmark transparency. Look at how other hardware generations work: a new iPhone costs roughly what the previous one did. Same with Samsung Galaxy phones, same with AMD Ryzen CPUs. Price increases are modest and the specs are straightforward to compare. Nvidia has broken from this pattern. Each new generation costs meaningfully more than the last, while the benchmarks get harder to interpret, not easier. Look at what's happening with the numbers. BF16 and FP16 performance figures, the ones most relevant for real workloads, are increasingly buried or absent. Sparsity is baked into headline numbers without always being clearly flagged. The B200 is being benchmarked as an NVLink pair rather than a single card in some comparisons. Each of these is defensible in isolation, but together they make it genuinely hard to do apples-to-apples comparisons across generations. The inference vs training tension is worth flagging too. Lower precision formats are good for inference but introduce real tradeoffs for training. As Nvidia pushes further into inference optimization, NVFP4 being the latest example, it raises a fair question about where training workloads go. NVFP4 training exists, but documentation and real-world results are thin enough that most teams aren't seriously evaluating it yet. The "ROCm is immature" narrative is also getting a bit stale. I've personally run workloads on the MI300X and honestly it felt smoother than I expected, even compared to CUDA in some cases. And the price to performance ratio isn't even close. AMD isn't a consolation prize anymore. At some point, labs and enterprises doing the actual buying will have enough alternatives like AMD MI300X, custom silicon, and inference-optimized ASICs to push back. Nvidia is standing on thin ice, and a few more steps like this might be enough to crack it.
posted
an
update
16 days ago
I don't know why, but lately there's been a growing problem on HuggingFace: the platform is filled to the brim with datasets of reasoning traces from large AI models. I feel like someone should address this, but the dataset tab is just full of reasoning and output traces from models like Claude Opus and the model tab is full of fine-tunes trained on these. What scares me most are the legal consequences of this, and the possibility that all models will start converging on the same tone because everyone is just fine-tuning on everyone else.
updated
a model
29 days ago
CaaLM/CaaLM-v1-GGUF
View all activity
Organizations
ereniko
's models
3
Sort: Recently updated
ereniko/Spamo-v1
Text Classification
•
67M
•
Updated
Mar 15
•
1
ereniko/SmolLLM2-135M-Code-ONNX
Updated
Mar 12
•
1
ereniko/SmolLLM2-135M-Code
Text Generation
•
Updated
Feb 27
•
4