🏆 FungiTastic Benchmarks
FungiTastic is designed to push the limits of machine learning in real-world biological classification through a comprehensive suite of benchmarks. These challenges reflect the unique, dynamic, and fine-grained nature of fungal data in the wild.
Why Multiple Benchmarks?
- Realistic evaluation: Biological data are messy—seasonal, imbalanced, shifting, and full of new discoveries.
- Diverse ML challenges: Tackle closed-set, open-set, few-shot, domain adaptation, segmentation, and more.
- Multi-modal & cost-sensitive: Go beyond images, leverage metadata and context, and reason about real-world consequences (e.g., edible vs. poisonous).
🧩 Benchmark Tasks
Below is an overview of all the supported benchmarks.
Click the task name to jump to the dedicated section and learn more about the data splits, evaluation metrics, and baseline results.
Task | Description | Link |
---|---|---|
Closed-set Classification | Standard fine-grained classification—species from a fixed set. | Closed-set |
Open-set Classification | Identify when an observation belongs to a new, unseen species. | Open-set |
Few-shot Learning | Recognize rare species with only a handful of training samples. | Few-shot |
Chronological (Domain Shift) | Handle distribution shifts—train/test splits follow real observation dates. | Chronological |
Cost-sensitive Classification | Penalize mistakes by their real-world impact (e.g., toxic/edible errors). | Cost-sensitive |
Segmentation | Detect and segment key body parts of fungi for fine-grained recognition. | Segmentation |
🔍 Benchmark Descriptions
🔒 Closed-set Classification
- Goal: Assign the correct species label to each image from a fixed set of classes.
- Highlights: Long-tailed, visually similar classes; strong baseline models provided.
- 👉 Read more…
🌍 Open-set Classification
- Goal: Detect when a test observation is from a species not seen during training.
- Use-case: New/rare fungi are continually discovered in nature—your model should know when it’s unsure.
- 👉 Read more…
🎯 Few-shot Learning
- Goal: Correctly classify species with fewer than five training samples.
- Highlights: Rewards ability to generalize from very little data—crucial for rare/under-observed species.
- 👉 Read more…
⏳ Chronological / Domain Shift
- Goal: Evaluate robustness to distribution shift over time (e.g., due to seasonality, climate, or new locations).
- Highlights: Train on older data, validate/test on newer years; reflects real-world deployment scenarios.
- 👉 Read more…
⚖️ Cost-sensitive Classification
- Goal: Minimize "real-world" error costs. E.g., confusing a poisonous with an edible mushroom is much worse than the reverse.
- Highlights: Supports research into non-standard losses and safety-aware AI.
- 👉 Read more…
✂️ Segmentation
- Goal: Identify and segment morphological parts (caps, stems, gills, etc.) for selected species/groups.
- Highlights: Supports interpretable and part-aware models, with hand-checked masks in the Mini subset.
- 👉 Read more…
📈 Baselines & Results
For each benchmark, you’ll find: - Data splits & preparation instructions - Recommended metrics - Baseline architectures and results - Download links for splits & scripts
See the Baselines & Models section for implementation details and ready-to-use checkpoints.
📎 Quick Start
- See How to Use for step-by-step tutorials.
- Review Evaluation Protocols for metric details and leaderboard submission formats.
💡 Want to add a new benchmark or task? Open an issue or submit a pull request!