File size: 7,924 Bytes
557c268
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9ef83c6
557c268
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7aaf3bc
2e52163
 
 
 
 
 
557c268
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
---
license: cc-by-4.0
task_categories:
- image-classification
- zero-shot-classification
tags:
- biology
- ecology
- wildlife
- camera-traps
- vision-transformers
- clustering
- zero-shot-learning
- biodiversity
- reproducibility
- benchmarking
- embeddings
- dinov3
- dinov2
- bioclip
- clip
- siglip
language:
- en
pretty_name: HUGO-Bench Paper Reproducibility Data
size_categories:
- 100K<n<1M
source_datasets:
- AI-EcoNet/HUGO-Bench
configs:
- config_name: primary_benchmarking
  data_files: primary_benchmarking/train-*.parquet
  default: true
- config_name: model_comparison
  data_files: model_comparison/train-*.parquet
- config_name: dimensionality_reduction
  data_files: dimensionality_reduction/train-*.parquet
- config_name: clustering_supervised
  data_files: clustering_supervised/train-*.parquet
- config_name: clustering_unsupervised
  data_files: clustering_unsupervised/train-*.parquet
- config_name: cluster_count_prediction
  data_files: cluster_count_prediction/train-*.parquet
- config_name: intra_species_variation
  data_files: intra_species_variation/train-*.parquet
- config_name: scaling_tests
  data_files: scaling_tests/train-*.parquet
- config_name: uneven_distribution
  data_files: uneven_distribution/train-*.parquet
- config_name: subsample_definitions
  data_files: subsample_definitions/train-*.parquet
- config_name: embeddings_dinov3_vith16plus
  data_files: embeddings_dinov3_vith16plus/train-*.parquet
- config_name: embeddings_dinov2_vitg14
  data_files: embeddings_dinov2_vitg14/train-*.parquet
- config_name: embeddings_bioclip2_vitl14
  data_files: embeddings_bioclip2_vitl14/train-*.parquet
- config_name: embeddings_clip_vitl14
  data_files: embeddings_clip_vitl14/train-*.parquet
- config_name: embeddings_siglip_vitb16
  data_files: embeddings_siglip_vitb16/train-*.parquet
---

# HUGO-Bench Paper Reproducibility

**Supplementary data and reproducibility materials for the paper:**

> **Vision Transformers for Zero-Shot Clustering of Animal Images: A Comparative Benchmarking Study** - https://arxiv.org/abs/2602.03894
>
> Hugo Markoff, Stefan Hein Bengtson, Michael Ørsted
>
> Aalborg University, Denmark

## Dataset Description

This repository contains complete experimental results, pre-computed embeddings, and execution logs from our comprehensive benchmarking study evaluating Vision Transformer models for zero-shot clustering of wildlife camera trap images.

### Related Resources

- **Source Images**: [AI-EcoNet/HUGO-Bench](https://huggingface.co/datasets/AI-EcoNet/HUGO-Bench) - 139,111 wildlife images
- **Code Repository**: [GitHub](https://github.com/AI-EcoNet/zeroshot_clustering_animal_images)   

## Repository Structure

```
├── primary_benchmarking/          # Main benchmark results (27,600 configurations)
├── model_comparison/              # Cross-model comparisons
├── dimensionality_reduction/      # UMAP/t-SNE/PCA analysis
├── clustering_supervised/         # Supervised clustering metrics
├── clustering_unsupervised/       # Unsupervised clustering results
├── cluster_count_prediction/      # Optimal cluster count analysis
├── intra_species_variation/       # Within-species cluster analysis
│   ├── train-*.parquet           # Analysis results
│   └── cluster_image_mappings.json  # Image-to-cluster assignments
├── scaling_tests/                 # Sample size scaling experiments
├── uneven_distribution/           # Class imbalance experiments
├── subsample_definitions/         # Reproducible subsample definitions
├── embeddings_*/                  # Pre-computed embeddings (5 models)
│   ├── embeddings_dinov3_vith16plus/  # 120K embeddings, 1280-dim
│   ├── embeddings_dinov2_vitg14/      # 120K embeddings, 1536-dim
│   ├── embeddings_bioclip2_vitl14/    # 120K embeddings, 768-dim
│   ├── embeddings_clip_vitl14/        # 120K embeddings, 768-dim
│   └── embeddings_siglip_vitb16/      # 120K embeddings, 768-dim
├── extreme_uneven_embeddings/     # Full dataset embeddings (PKL)
│   ├── aves_full_dinov3_embeddings.pkl      # 74,396 embeddings
│   └── mammalia_full_dinov3_embeddings.pkl  # 65,484 embeddings
└── execution_logs/                # Experiment execution logs
```

## Quick Start

### Load Primary Benchmark Results

```python
from datasets import load_dataset

# Load main benchmark results (27,600 configurations)
ds = load_dataset("AI-EcoNet/HUGO-Bench-Paper-Reproducibility", "primary_benchmarking")
print(f"Configurations: {len(ds['train'])}")
```

### Load Pre-computed Embeddings

```python
# Load DINOv3 embeddings (120,000 images)
embeddings = load_dataset(
    "AI-EcoNet/HUGO-Bench-Paper-Reproducibility", 
    "embeddings_dinov3_vith16plus"
)
print(f"Embeddings shape: {len(embeddings['train'])} x {len(embeddings['train'][0]['embedding'])}")
```

### Load Specific Analysis Results

```python
# Model comparison results
model_comp = load_dataset("AI-EcoNet/HUGO-Bench-Paper-Reproducibility", "model_comparison")

# Scaling test results
scaling = load_dataset("AI-EcoNet/HUGO-Bench-Paper-Reproducibility", "scaling_tests")

# Intra-species variation analysis
intra = load_dataset("AI-EcoNet/HUGO-Bench-Paper-Reproducibility", "intra_species_variation")
```

### Load Cluster Image Mappings

The intra-species analysis includes a mapping file showing which images belong to which clusters:

```python
from huggingface_hub import hf_hub_download
import json

# Download mapping file
mapping_file = hf_hub_download(
    "AI-EcoNet/HUGO-Bench-Paper-Reproducibility",
    "intra_species_variation/cluster_image_mappings.json",
    repo_type="dataset"
)

with open(mapping_file) as f:
    mappings = json.load(f)

# Structure: {species: {run: {cluster: [image_names]}}}
print(f"Species analyzed: {list(mappings.keys())}")
```

### Load Full Dataset Embeddings

For the extreme uneven distribution experiments, we provide full dataset embeddings:

```python
from huggingface_hub import hf_hub_download
import pickle

# Download Aves embeddings (74,396 images)
pkl_file = hf_hub_download(
    "AI-EcoNet/HUGO-Bench-Paper-Reproducibility",
    "extreme_uneven_embeddings/aves_full_dinov3_embeddings.pkl",
    repo_type="dataset"
)

with open(pkl_file, 'rb') as f:
    data = pickle.load(f)

print(f"Embeddings: {data['embeddings'].shape}")  # (74396, 1280)
print(f"Labels: {len(data['labels'])}")
print(f"Paths: {len(data['paths'])}")
```

## Experimental Setup

### Models Evaluated

| Model | Architecture | Embedding Dim | Pre-training |
|-------|-------------|---------------|--------------|
| DINOv3 | ViT-H/16+ | 1280 | Self-supervised |
| DINOv2 | ViT-G/14 | 1536 | Self-supervised |
| BioCLIP 2 | ViT-L/14 | 768 | Biology domain |
| CLIP | ViT-L/14 | 768 | Contrastive |
| SigLIP | ViT-B/16 | 768 | Sigmoid loss |

### Clustering Methods

- K-Means, DBSCAN, HDBSCAN, Agglomerative, Spectral
- GMM (Gaussian Mixture Models)
- With and without dimensionality reduction (UMAP, t-SNE, PCA)

### Evaluation Metrics

- **Supervised**: Adjusted Rand Index (ARI), Normalized Mutual Information (NMI), Accuracy, F1
- **Unsupervised**: Silhouette Score, Calinski-Harabasz Index, Davies-Bouldin Index

## Citation

If you use this dataset, please cite:

```bibtex
@misc{HUGO-Bench-Paper-Reproducibility_2026,
    author       = {Markoff, Hugo and Bengtson, Stefan Hein and Ørsted, Michael},
	title        = {HUGO-Bench-Paper-Reproducibility (Revision 31b0d17)},
	year         = 2026,
	url          = {https://huggingface.co/datasets/AI-EcoNet/HUGO-Bench-Paper-Reproducibility},
	doi          = {10.57967/hf/8191},
	publisher    = {Hugging Face}
```

## License

This dataset is released under the [CC-BY-4.0 License](https://creativecommons.org/licenses/by/4.0/).

## Contact

For questions or issues, please open an issue in this repository or contact the authors.