Standardizing and Centralizing Datasets to Enable Efficient Training of Agricultural Deep Learning Models

08/04/2022
by   Amogh Joshi, et al.
8

In recent years, deep learning models have become the standard for agricultural computer vision. Such models are typically fine-tuned to agricultural tasks using model weights that were originally fit to more general, non-agricultural datasets. This lack of agriculture-specific fine-tuning potentially increases training time and resource use, and decreases model performance, leading an overall decrease in data efficiency. To overcome this limitation, we collect a wide range of existing public datasets for three distinct tasks, standardize them, and construct standard training and evaluation pipelines, providing us with a set of benchmarks and pretrained models. We then conduct a number of experiments using methods which are commonly used in deep learning tasks, but unexplored in their domain-specific applications for agriculture. Our experiments guide us in developing a number of approaches to improve data efficiency when training agricultural deep learning models, without large-scale modifications to existing pipelines. Our results demonstrate that even slight training modifications, such as using agricultural pretrained model weights, or adopting specific spatial augmentations into data processing pipelines, can significantly boost model performance and result in shorter convergence time, saving training resources. Furthermore, we find that even models trained on low-quality annotations can produce comparable levels of performance to their high-quality equivalents, suggesting that datasets with poor annotations can still be used for training, expanding the pool of currently available datasets. Our methods are broadly applicable throughout agricultural deep learning, and present high potential for significant data efficiency improvements.

READ FULL TEXT

page 2

page 5

page 6

page 9

page 14

research
03/28/2023

Large-scale pretraining on pathological images for fine-tuning of small pathological benchmarks

Pretraining a deep learning model on large image datasets is a standard ...
research
06/09/2021

Pretrained Encoders are All You Need

Data-efficiency and generalization are key challenges in deep learning a...
research
04/13/2023

Lossless Adaptation of Pretrained Vision Models For Robotic Manipulation

Recent works have shown that large models pretrained on common visual le...
research
07/20/2022

Large Scale Radio Frequency Signal Classification

Existing datasets used to train deep learning models for narrowband radi...
research
11/18/2016

ModelHub: Towards Unified Data and Lifecycle Management for Deep Learning

Deep learning has improved state-of-the-art results in many important fi...
research
05/17/2023

CWD30: A Comprehensive and Holistic Dataset for Crop Weed Recognition in Precision Agriculture

The growing demand for precision agriculture necessitates efficient and ...
research
06/28/2023

Effective Transfer of Pretrained Large Visual Model for Fabric Defect Segmentation via Specifc Knowledge Injection

Fabric defect segmentation is integral to textile quality control. Despi...

Please sign up or login with your details

Forgot password? Click here to reset