News Release

Artificial intelligence can now estimate rice yields, according to new study

Researchers train neural network models that can estimate rice yield from analyzing pre-harvest photographs

Peer-Reviewed Publication

Okayama University

Estimating rice yield with a convolutional neural network (CNN) model, using ground-based digital images

image: Researchers created a database of over 20,000 rice field images, and developed a CNN model that could instantaneously predict the yield of rice, using images from this database. The new method offers several advantages such as high precision, versatility and ease to use, as opposed to other time consuimg, labour intensive approached. view more 

Credit: Plant Phenomics

With the rise in global demand for staple crop products projected to substantially increase by 2050 due to population growth, rising per capita income, and the growing use of biofuels, it is necessary to adopt sustainable agricultural intensification practices in existing croplands to meet this demand. However, estimation processes currently employed in the global South remain inadequate. Traditional methods like self-reporting and crop cutting have their limitations, and remote sensing technologies are not fully utilized in this context.


However, recent advancements in artificial intelligence and machine learning, particularly deep learning with convolutional neural networks (CNNs), offer promising solutions here. To explore the scope of this new technology, researchers from Japan conducted a study focusing on rice. They used ground-based digital images taken at harvesting stage of the crop, combined with CNNs, to estimate rice yield. Their study appeared online on 29 June 2023 and was published on 28 July 2023 in Volume 5 of Plant Phenomics.


We started by conducting an extensive field campaign. We gathered rice canopy images and rough grain yield data from 20 locations in seven countries in order to create a comprehensive multinational database,” says Dr. Yu Tanaka, Associate Professor at the Graduate School of Environmental, Life, Natural Science and Technology, Okayama University, who led the study.

The images were captured using digital cameras which could gather the required data from a distance of 0.8–0.9 meters, vertically downwards from the rice canopy. With Dr. Kazuki Saito of the International Rice Research Institute (formerly Africa Rice Center) and other collaborators, the team successfully created a database of 4,820 yield data of harvesting plots and 22,067 images, encompassing various rice cultivars, production systems, and crop management practices.

Next, a CNN model was developed to estimate the grain yield for each of the collected images. The team used a visual-occlusion method to visualize the additive effect of different regions in the rice canopy images. It involved masking specific parts of the images and observing how the model's yield estimation changed in response to the masked regions. The insights gained from this method allowed the researchers to understand how the CNN model interpreted various features in the rice canopy images, influencing its accuracy and its ability to distinguish between yield-contributing components and non-contributing elements in the canopy.


The model performed well, explaining around 68%–69% of yield variation in the validation and test datasets. Study results highlighted the importance of panicles—loose-branching clusters of flowers—in yield estimation through occlusion-based visualization. The model could predict yield accurately during the ripening stage, recognizing mature panicles, and also detect cultivar and water management differences in yield in the prediction dataset. Its accuracy, however, decreased as image resolution decreased.


Nevertheless, the model proved robust, showing good accuracy at different shooting angles and times of day. “Overall, the developed CNN model demonstrated promising capabilities in estimating rough grain yield from rice canopy images across diverse environments and cultivars. Another appealing aspect is that it is highly cost effective and does not require labor-intensive crop cuts or complex remote-sensing technologies,” says Dr. Tanaka enthusiastically.


The study emphasizes the potential of CNN-based models for monitoring rice productivity at regional scales. However, the model's accuracy may vary under different conditions, and further research should focus on adapting the model to low-yielding and rainy environments. The AI-based method has also been made available to farmers and researchers through a simple smartphone application, thus greatly improving accessibility of the technology and its real-life applications. The name of this application is ‘HOJO’, and it is already available on iOS and Android. The researchers hope that their work will lead to better management of rice fields and assist accelerated breeding programs, contributing positively to global food production and sustainability initiatives.


About Okayama University, Japan

As one of the leading universities in Japan, Okayama University aims to create and establish a new paradigm for the sustainable development of the world. Okayama University offers a wide range of academic fields, which become the basis of the integrated graduate schools. This not only allows us to conduct the most advanced and up-to-date research, but also provides an enriching educational experience.



About Associate Professor Yu Tanaka from Okayama University, Japan

Dr. Yu Tanaka is an Associate Professor at the Graduate School of Environmental, Life, Natural Science and Technology at Okayama University. His research interests are environmental, agricultural, crop production science, and the use of artificial intelligence- based research methods in these disciplines. He has extensively worked on production, yield estimation, genetic variation studies involving the rice crop.

Disclaimer: AAAS and EurekAlert! are not responsible for the accuracy of news releases posted to EurekAlert! by contributing institutions or for the use of any information through the EurekAlert system.