Crop price and yield prediction using data science technique

https://doi.org/10.53730/ijhs.v6nS3.6953

Authors

  • Ritwik Saxena Dept of Computer Science and Engineering, SRM Insitute of Science and Technology, Kattankulathur
  • J. Selvin Paul Peter Associate Professor, SRM Insitute of Science and Technology, Kattankulathur
  • Shwetha Saxena Assistant Professor, Amity Business School, MP
  • Anubhav Sapra Dept of Computer Science and Engineering, SRM Insitute of Science and Technology, Kattankulathur

Abstract

Agriculture is primarily responsible for increasing the state's economic contribution around the world. The most significant agricultural fields, however, remain underdeveloped because of the absence of ecosystem control technology adoption. Crop output is not improving as a result of these issues, which has an impact on the farm economy. As a result, the plant yield prediction helps to support the growth of agricultural productivity. To address this issue, agricultural industries must use machine learning algorithms to forecast crop yield from a given dataset. In order to capture various pieces of information, the supervised machine learning technique must be employed to analyse the dataset. Some examples include variable identification, uni-variate analysis, bi-variate and multi-variate analysis, missing value treatments, and so on. A comparison of machine learning algorithms was performed to see which one was the most accurate at forecasting the most basic crop. The results reveal that the effectiveness of the proposed machine learning algorithm technique is commonly compared to the best accuracy using entropy calculation, precision, recall, F1 Score, sensitivity, and specificity.

Downloads

Download data is not yet available.

References

A. Mark Hall, “Feature selection for discrete and numeric class machine learning,” Comput. Sci., Univ. Waikato, pp. 359–366, Dec. 1999.

Guyon and A. Elisseeff, “An introduction to variable and feature selection,” J. Mach. Learn. Res., vol. 3, pp. 1157–1182, Jan. 2003.

R. Gilad-Bachrach, A. Navot, and N. Tishby, “Margin based feature selection–theory and algorithms,” in Proc. 21st Int. Conf. Mach. Learn. (ICML), 2004, p. 43.

P. S. Maya Gopal and R. Bhargavi, “Feature selection for yield prediction in boruta algorithm,” Int. J. Pure Appl. Math., vol. 118, no. 22, pp. 139–144, 2018.

S. Ji, S. Pan, X. Li, E. Cambria, G. Long, and Z. Huang, “Suicidal ideation detection: A review of machine learning methods and applications,” IEEE Trans. Comput. Social Syst., vol. 8, no. 1, pp. 214–226, Feb. 2021.

K. Ranjini, A. Suruliandi, and S. P. Raja, “An ensemble of heterogeneous incremental classifiers for assisted reproductive technology outcome prediction,” IEEE Trans. Comput. Social Syst.early access, Nov. 3, 2020, doi: 10.1109/TCSS.2020.3032640. [7] H. Liu and R. Setiono, “A probabilistic approach to feature selection-a filter solution,” in Proc. 13th Int. Conf. Int. Conf. Mach. Learn., vol. 96, 1996, pp. 319–327.

R. Kohavi and G. H. John, “Wrappers for feature subset selection,” Artif. Intell., vol. 97, nos. 1–2, pp. 273–324, Dec. 1997.

H. Wang, M. Taghi Khoshgoftaar, and K. Gao, “Ensemble feature selection technique for software quality classification,” in Proc. 22nd Int. Conf. Softw. Eng. Knowl. Eng., 2010, pp. 215–220.

Gregorutti, B. Michel, and P. Saint-Pierre, “Correlation and variable importance in random forests,” Statist. Comput., vol. 27, no. 3, pp. 659–678, May 2017.

M. A. Hall and G. Holmes, “Benchmarking attribute selection techniques for discrete class data mining,” IEEE Trans. Knowl. Data Eng., vol. 15, no. 6, pp. 1437–1447, Nov. 2003.

H. Liu and L. Yu, “Toward integrating feature selection algorithms for classification and clustering,” IEEE Trans. Knowl. Data Eng., vol. 17, no. 4, pp. 491–502, Apr. 2005. [13] P. M. Granitto, C. Furlanello, F. Biasioli, and F. Gasperi, “Recursive feature elimination with random forest for PTR-MS analysis of agroindustrial products,” Chemometric Intell. Lab. Syst., vol. 83, no. 2, pp. 83–90, Sep. 2006.

Araåzo-Azofra and J. M. Benítez, “Empirical study of feature selection methods in classification,” in Proc. 8th Int. Conf. Hybrid Intell. Syst., Barcelona, Spain, Sep. 2008, pp. 584–589.

Altmann, L. Toloåi, O. Sander, and T. Lengauer, “Permutation importance: A corrected feature importance measure,” Bioinformatics, vol. 26, no. 10, pp. 1340–1347, May 2010.

M. B. Kursa and W. R. Rudnicki, “Feature selection with the Boruta package,” J. Stat. Softw., vol. 36, no. 11, pp. 1–13, 2010.

G. Ruß and R. Kruse, “Feature selection for wheat yield prediction,” in Research and Development in Intelligent Systems. London, U.K.: Springer, 2010, pp. 465–478.

J. Camargo and A. Young, “Feature selection and non-linear classifiers: Effects on simultaneous motion recognition in upper limb,” IEEE Trans. Neural Syst. Rehabil. Eng., vol. 27, no. 4, pp. 743–750, Apr. 2019.

M. B. Kursa, A. Jankowski, and W. R. Rudnicki, “Boruta—A system for feature selection,” Fundam. Inf., vol. 101, no. 4, pp. 271–285, 2010.

R. Rajasheker Pullanagari, G. Kereszturi, and I. Yule, “Integrating airborne hyperspectral, topographic, and soil data for estimating pasture quality using recursive feature elimination with random forest regression,” Remote Sens., vol. 10, no. 7, pp. 1117–1130, 2018.

Choudhary, S. Kolhe, and H. Rajkamal, “Performance Evaluation of feature selection methods for Mobile devices,” Int. J. Eng. Res. Appl., vol. 3, no. 6, pp. 587–594, 2013.

F. Balducci, D. Impedovo, and G. Pirlo, “Machine learning applications on agricultural datasets for smart farm enhancement,” Machine, vol. 6, no. 3, pp. 38–59, 2018.

J. Weston, S. Mukherjee, O. Chapelle, M. Pontil, T. Poggio, and V. Vapnik, “Feature selection for SVMs,” in Proc. Adv. Neural Inf. Process. Syst., vol. 13, 2001, pp. 668–674.

Bahl et al., “Recursive feature elimination in random forest classification supports nanomaterial grouping,” NanoImpact, vol. 15, Mar. 2019, Art. no. 100179.

Published

05-05-2022

How to Cite

Saxena, R., Peter, J. S. P., Saxena, S., & Sapra, A. (2022). Crop price and yield prediction using data science technique. International Journal of Health Sciences, 6(S3), 4777–4793. https://doi.org/10.53730/ijhs.v6nS3.6953

Issue

Section

Peer Review Articles

Most read articles by the same author(s)