Hugging Face
Models
Datasets
Spaces
Buckets
new
Docs
Enterprise
Pricing
Website
Tasks
HuggingChat
Collections
Languages
Organizations
Community
Blog
Posts
Daily Papers
Learn
Discord
Forum
GitHub
Solutions
Team & Enterprise
Hugging Face PRO
Enterprise Support
Inference Providers
Inference Endpoints
Storage Buckets
Log In
Sign Up
3
9
G. Cheuv
Cheuv
Follow
PhysiQuanty's profile picture
1 follower
·
6 following
AI & ML interests
None yet
Recent Activity
new
activity
3 days ago
Cheuv/Darwin-4B-Genesis-Q8_0-GGUF:
merge
liked
a model
19 days ago
mradermacher/Darwin-4B-Genesis-GGUF
reacted
to
SeaWolf-AI
's
post
with 🔥
23 days ago
🧬 Darwin V6: Diagnostic-Guided Evolutionary Model Merging We are releasing Darwin-31B-Opus — a reasoning-enhanced model merging Google's Gemma-4-31B-it and TeichAI's Claude Opus Distill using the Darwin V6 engine. Model: https://huggingface.co/FINAL-Bench/Darwin-31B-Opus Demo: https://huggingface.co/spaces/FINAL-Bench/Darwin-31B-Opus 🔬 What Darwin V6 Does Conventional merging tools (mergekit, etc.) apply a single ratio to all tensors. Set ratio=0.5 and all 1,188 tensors blend identically, with no distinction between which tensors matter for reasoning versus coding. Darwin V6 diagnoses both parents at the tensor level before merging. It measures Shannon entropy, standard deviation, and L2 norm for every tensor, then passes 5 diagnostic probes (REASONING, CODE, MATH, KNOWLEDGE, LANGUAGE) through the model to determine layer-wise functional importance. Each of the 1,188 tensors receives an independent optimal ratio. combined = static(entropy/std/norm) x 0.4 + probe(cosine_distance) x 0.6 final_ratio = mri_ratio x mri_trust + genome_ratio x (1 - mri_trust) When one parent is overwhelmingly superior for a tensor (ratio < 0.15 or > 0.85), Darwin transplants it directly without interpolation. The mri_trust parameter itself is optimized by CMA-ES evolutionary search, so optimal transplant intensity is determined automatically. After merging, a Health Check compares the child against both parents layer-by-layer to detect interference or function loss. 🧬 Parent Models Father: google/gemma-4-31B-it Mother: TeichAI/gemma-4-31B-it-Claude-Opus-Distill 🧬 Results Compared under identical conditions (same 50 questions, same seed, greedy, thinking mode): Father: 60.0% (30/50) Darwin-31B-Opus: 66.0% (33/50) — +10% relative improvement ARC-Challenge: 82.89% (loglikelihood, zero-shot, 200 questions) Optimal genome found by evolution: ffn_ratio=0.93 — FFN layers strongly favor Mother (Claude Opus Distill) block_5 (L50-L59)=0.86 and more...
View all activity
Organizations
None yet
Cheuv
's activity
All
Models
Datasets
Spaces
Buckets
Papers
Collections
Community
Posts
Upvotes
Likes
Articles
liked
a model
19 days ago
mradermacher/Darwin-4B-Genesis-GGUF
8B
•
Updated
1 day ago
•
3.11k
•
13
liked
5 models
24 days ago
FINAL-Bench/Darwin-4B-David
Text Generation
•
8B
•
Updated
1 day ago
•
240
•
40
FINAL-Bench/Darwin-9B-Opus
Text Generation
•
10B
•
Updated
1 day ago
•
405
•
28
FINAL-Bench/Darwin-4B-Genesis
Text Generation
•
8B
•
Updated
about 11 hours ago
•
622
•
31
FINAL-Bench/Darwin-27B-Opus
Text Generation
•
28B
•
Updated
1 day ago
•
119
•
32
FINAL-Bench/Darwin-31B-Opus
Text Generation
•
33B
•
Updated
1 day ago
•
453
•
43
liked
a model
25 days ago
FINAL-Bench/Darwin-2B-Opus
Text Generation
•
2B
•
Updated
1 day ago
•
1.49k
•
16
liked
2 models
about 1 month ago
mradermacher/Darwin-9B-Opus-GGUF
9B
•
Updated
Apr 5
•
2.09k
•
15
mradermacher/Darwin-4B-David-GGUF
8B
•
Updated
Apr 14
•
1.11k
•
14