AgriChain: Visually-Grounded Expert-Verified Reasoning for Interpretable Agricultural Vision–Language Models
Proceedings of the Fifteenth Language Resources and Evaluation Conference (LREC 2026)
Abstract
Accurate and interpretable plant disease diagnosis remains a key challenge for vision–language models in real agricultural settings. We present AgriChain, a new dataset of around 11,000 expert-curated leaf images covering a wide range of crops and diseases. Each image is paired with a disease label, a calibrated confidence score, and an expert-verified chain-of-thought explanation. Draft rationales were first generated by GPT-4o and then refined by a professional agricultural engineer using standard descriptors such as lesion color, margin, and distribution. Using these data, we fine-tune the open vision–language model Qwen-2.5-VL-3B to jointly identify diseases and explain its reasoning in a way that mirrors expert thinking. On a 1,000-image test set, our model reaches 73.1% accuracy and produces explanations that align closely with human expertise. These results show that expert-verified reasoning supervision enhances both performance and interpretability, bringing us closer to transparent and trustworthy AI tools for sustainable agriculture.To support reproducibility and further research, the dataset and code are publicly available at https://github.com/hazzanabeel12-netizen/agrichain.