AgrI Challenge icon
AgrI Challenge
AgrI Challenge logo
Data-Centric AI Competition

AgrI Challenge

A Data-Centric Competition Framework for Agricultural Machine Learning

Twelve interdisciplinary teams independently collected 50,673 field images of 6 tree species over a 2-day campaign. We introduce the Cross-Team Validation (CTV) framework, a new paradigm for evaluating real-world model generalization.

0
Total Images
0
Teams
0
Tree Species
Scroll to explore
AgrI Challenge logo

What is AgrI Challenge?

A participant-led AI competition where each team independently collects their own field data — generating authentic, real-world distributional diversity.

Interdisciplinary Teams

Each of the 12 teams combined students from computing backgrounds (computer science, AI, data science) with students from ecological or agronomic backgrounds (agriculture, forestry, plant sciences). This mixed composition guaranteed high-quality species labeling grounded in domain expertise while supporting effective ML system design.

Competition Phases

Phase 01

Data Collection Phase

2 days📍 ENSA, El Harrach, Algiers

Teams independently collected field images of 6 tree species at the experimental facilities of École Nationale Supérieure Agronomique. Each team had full freedom over their collection devices, sampling strategies, environmental coverage, and imaging protocols, generating authentic domain diversity.

Phase 02

Model Development Phase

2 days📍 ENSIA, Algiers

Teams preprocessed, annotated, and trained machine learning models using only their own collected data at École Nationale Supérieure d'Intelligence Artificielle. AI specialists mentored teams on data preparation, model design, training, and evaluation.

Why AgrI Challenge?

Standard AI benchmarks fail in the field. AgrI Challenge was designed to close the gap between laboratory accuracy and real-world performance.

The Generalization Problem

Models trained on controlled laboratory datasets achieve over 99% accuracy on benchmarks, yet drop to 54% in real farm environments. This is not a model problem — it is a data problem.

The AgrI Challenge Approach

Rather than providing a fixed dataset, each team independently collects their own field data — generating authentic distributional diversity across devices, environments, and sampling strategies that mirrors real deployment conditions.

What the Results Show

Under single-team training (TOTO), models showed a mean validation-test gap of up to 16.2 percentage points — a direct consequence of narrow data diversity. When trained collaboratively across all 12 teams (LOTO), that gap collapsed by 82–84%, confirming that data diversity is the primary driver of robust AI generalization.

Where & Who

The AgrI Challenge was organized through a collaboration between two leading Algerian academic institutions, held across two campuses in Algiers.

ENSA logo
ENSA
El Harrach, Algiers, Algeria

École Nationale Supérieure Agronomique

Field data collection host, Data Collection Phase

A long-established agronomic institution in Algeria, founded in 1905. The experimental and teaching facilities at ENSA, El Harrach, provided access to representative agro-ecosystems and well-maintained plant collections.

ensa.dz
ENSIA logo
ENSIA
Algiers, Algeria

École Nationale Supérieure d'Intelligence Artificielle

Model development host, Model Development Phase

A national center of excellence dedicated to education and research in artificial intelligence and data science. ENSIA specialists in AI and machine learning mentored teams throughout the modeling phase.

ensia.edu.dz

Contact

For questions about the dataset, access requests, or research collaboration:

mohamed.brahimi@ensia.edu.dz

Corresponding author: Dr. Mohammed Brahimi, ENSIA, Algiers, Algeria.

Cross-Team Validation

A novel evaluation paradigm that treats each team's independently collected dataset as a distinct domain, revealing how models actually generalize.

What is CTV?

Cross-Team Validation (CTV) treats datasets collected by different teams as distinct domains for training and testing. Unlike traditional cross-validation that splits data randomly, CTV preserves natural domain boundaries created by different collection methodologies, environmental contexts, and device characteristics.

Each team's dataset represents a unique combination of factors, capturing authentic inter-domain variation that simulates real-world deployment scenarios.

Protocol A

TOTO

Train-on-One-Team-Only

A model is trained on a single team's collected data (70% train / 30% val split) and then evaluated on every other team's data independently. Measures single-source generalization.

For each team i:
Train→ Team i data (70%)
Val→ Team i data (30%)
Test→ All other teams j ≠ i
Protocol B

LOTO

Leave-One-Team-Out

A model is trained on the combined data of all teams except one, then evaluated on the held-out team's data. Measures collaborative multi-source generalization.

For each team i:
Train→ All teams except i (70%)
Val→ 30% of training set
Test→ Held-out team i

Why Standard Validation Is Not Enough

Standard cross-validation with random data splits does not expose a model's true generalization ability across genuinely different domains. CTV specifically reveals the validation–test gap that standard metrics miss.

“Validation accuracy was remarkably stable (≈98%) across all LOTO folds, while test accuracy varied by over 11 percentage points, confirming that validation accuracy alone is not a reliable proxy for real-world generalization.”

Dataset

50,673 field images spanning 6 tree species, collected by 12 independent teams and curated into a 47,367-image clean benchmark.

Tree Species

Carob
Ceratonia siliqua
Oak
Quercus spp.
Peruvian Pepper
Schinus molle
Ash
Fraxinus spp.
Pistachio
Pistacia vera
Tipuana
Tipuana tipu

Image Distribution by Team & Species

Full image count breakdown across all 12 teams and 6 species classes (Table 2 from the paper).

TeamCarobOakPeruvian PepperAshPistachioTipuanaTotal
AI-4o9601,6251,2219341,1111,4937,344
AiGro6866045955725416353,633
CACTUS4005976948037555523,801
CHAJARA6105494283082623832,540
GreenAI6757056095597225153,785
PLT1,0891,1311,0001,2979731,0666,556
RUSTICUS5656255055005556963,446
SMART AGRICULTURES6389231,6941,2141,0807726,321
Scorpions4224504074566204402,795
Condimenteum1,0481,0006551,0746061,1215,504
The Neural Ninjas2342622562362933611,642
Organization team5526577273923686103,306
Total7,8799,1288,7918,3457,8868,64450,673

Baseline Results

Dual architecture baselines evaluated under TOTO and LOTO protocols, providing benchmarks for future researchers using the dataset.

Baseline Architectures

DenseNet121

CNN
8M parameters

Uses dense connections between layers for efficient feature reuse. Pretrained on ImageNet-1K.

Efficient backbone

Swin Transformer

Vision Transformer
28M parameters

Computes self-attention within local windows with shifted windowing. Pretrained on ImageNet-1K.

Attention-based baseline
Training: AdamW (lr=1e-4, wd=1e-4) · Cosine annealing · Batch 32 · 20 epochs · 224×224px input · NVIDIA H100 NVL GPU

TOTO Protocol — Validation vs. Test Gap

DenseNet121

97.40%
Val Accuracy
81.19%
Test Accuracy
16.20%
Val–Test Gap

The gap between validation and test accuracy emerged by epoch 5 and remained stable throughout training, indicating distributional shift rather than overfitting.

Swin Transformer

98.59%
Val Accuracy
87.21%
Test Accuracy
11.37%
Val–Test Gap

The gap between validation and test accuracy emerged by epoch 5 and remained stable throughout training, indicating distributional shift rather than overfitting.

LOTO Protocol — Collaborative Training Impact

DenseNet121

TOTO
81.19%
LOTO
95.31%
Gain
+14.12pp
16.20%
Gap (TOTO)
2.82%
Gap (LOTO)
−82%
Gap reduction
Performance variance reduced by −40%

Swin Transformer

TOTO
87.21%
LOTO
97.04%
Gain
+9.83pp
11.37%
Gap (TOTO)
1.78%
Gap (LOTO)
−84%
Gap reduction
Performance variance reduced by −54%

Request Dataset Access

The AgrI Challenge dataset is available for non-commercial research purposes. Submit a request and we will review it and contact you.

The dataset contains 50,673 field images of 6 tree species collected by 12 teams under the AgrI Challenge framework. Access is granted for academic and research use only. Upon submission, your request will be reviewed and you will receive a response at the provided email address.

Data License & Access Conditions

By requesting access to the AgrI Challenge dataset you agree to all of the following conditions:

  • Non-Commercial Use Only. The dataset may only be used for academic, educational, and non-commercial research purposes. Any commercial use, including but not limited to product development, commercial services, or for-profit applications, is strictly prohibited.
  • No Redistribution. You may not share, publish, redistribute, sublicense, or make the dataset (or any portion of it) publicly available in any form. If a collaborator requires access, they must submit their own individual request to the authors.
  • Mandatory Citation. Any publication, report, or presentation that uses or references this dataset must cite the AgrI Challenge paper. See the Citation section for the correct BibTeX and APA formats.
  • No Derivative Datasets. You may not create and distribute datasets derived from the AgrI Challenge data without explicit written permission from the corresponding author (mohamed.brahimi@ensia.edu.dz).
  • Responsible Use. The dataset must be used in a manner consistent with applicable privacy laws and ethical research standards. You are responsible for ensuring your use complies with your institution's policies.

Cite This Work

If you use the AgrI Challenge dataset or the Cross-Team Validation framework in your research, please cite our paper.

AgrI Challenge: Cross-Team Insights from a Data-Centric AI Competition in Agricultural Vision

Brahimi, Laabassi, Hadj Ameur, Boutorh, Siab-Farsi, Khouani, Zouak, Bouziane, Lakhdari & Benghanem · 2026 · arXiv preprint

arXiv:XXXX.XXXXX(link available upon publication)
BibTeX
@article{brahimi2026agrichallenge, title = {AgrI Challenge: Cross-Team Insights from a Data-Centric AI Competition in Agricultural Vision}, author = {Brahimi, Mohammed and Laabassi, Karim and {Hadj Ameur}, Mohamed Seghir and Boutorh, Aicha and Siab-Farsi, Badia and Khouani, Amin and Zouak, Omar Farouk and Bouziane, Seif Eddine and Lakhdari, Kheira and Benghanem, Abdelkader Nabil}, journal = {arXiv preprint}, year = {2026}, url = {https://arxiv.org/abs/XXXX.XXXXX} }
APA Format
Brahimi, M., Laabassi, K., Hadj Ameur, M. S., Boutorh, A., Siab-Farsi, B., Khouani, A., Zouak, O. F., Bouziane, S. E., Lakhdari, K., & Benghanem, A. N. (2026). AgrI Challenge: Cross-Team Insights from a Data-Centric AI Competition in Agricultural Vision. arXiv preprint. https://arxiv.org/abs/XXXX.XXXXX

The arXiv ID will be updated upon paper publication.