Developing an Efficient Predictive Model Based on ML and DL Approaches to Detect Diabetes
DOI:
https://doi.org/10.31449/inf.v45i3.3041Abstract
During the last decade, some important progress on machine learning ML area have been made, especially with the apparition of a new subfield called deep learning DL and CNN networks (Convolutional Neural Networks). This new tendency is used to perform much more sophisticated algorithms allowing high performance in many disciplines such as: pattern recognition, image classification, computer vision, as well as other supervised and unsupervised classification tasks. In this work, we have developed an automatic classifier that permits to classify a number of diabetic patients based on some blood characteristics by using ML and DL approaches. Initially, we have proceeded to the classification task using many ML algorithms. Then we proposed a simple CNN model composed of many layers. Finally, we established a comparison between ML and DL algorithms. For programming task, we have used Python, Tensorflow and Keras which are the most used in the field.References
Lee H., Grosse R., Ranganath R. , and Ng A.Y(2009). Convolutional deep belief networks for scalable unsupervised learning of hierarchical representations. In Proceedings of the 26th Annual International Conference on Machine Learning, pages 609–616. ACM.
Pinto N., Doukhan D., DiCarlo J.J. , and Cox D.D. (2009) A high-throughput screening approach to discovering good forms of biologically inspired visual representation. PLoS computational biology, 5(11):e1000579.
Turaga S.C., Murray J.F., Jain V. Roth F., Helmstaedter M., Briggman K., Denk W., and Seung H.S. (2010). Convolutional networks can learn to generate affinity graphs for image segmentation. Neural Computation, 22(2):511–538.
Abadi B., Agarwal M., Barham A, P, Brevdo E, Chen Z, Citro C, Corrado GS, Davis A, Dean J, Devin M, Ghemawat S, Goodfellow I, Harp A, Irving G, Isard M, Jia Y, Jozefowicz R, Kaiser L, Kudlur M, Levenberg J, Mané D, Monga R, Moore S, Murray D, Olah C, Schuster M, Shlens J, Steiner B, Sutskever I, Talwar K, Tucker P, Vanhoucke V, Vasudevan V, Viégas F, Vinyals O, Warden P, Wattenberg M, Wicke M, Yu Y, Zheng X. (2018) TensorFlow: large-scale machine learning on heterogeneous systems 2015. http://tensorflow.org/. Accessed 1 Nov 2018.
Theano Development Team. (2016). Theano: a Python framework for fast computation of mathematical expressions. arXiv e-prints arXiv:1605.02688.
Chollet F, et al (2018). Keras. 2015. https://keras.io. Accessed 1 Nov 2018.
Paszke A, Gross S, Chintala S, Chanan G, Yang E, DeVito Z, Lin Z, Desmaison A, Antiga L, Lerer A (2017). Automatic differentiation in pytorch. In: NIPS-W.
Chetlur S, Woolley C, Vandermersch P, Cohen J, Tran J, Catanzaro B, Shelhamer, E. cudnn (2014). Efficient primitives for deep learning.
Krizhevsky A, Sutskever I, Hinton GE (2012). Imagenet classification with deep convolutional neural networks. In: Neural information processing systems. p. 25.
Fukushima K. (1980). Neocognitron: a self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position. Biol Cybern. 1980;36(4):193–202.
LeCun Y, Bottou L, Bengio Y, Haffner P (1998). Gradient-based learning applied to document recognition. Proc IEEE. 86(11):2278–324.
Witten IH, Frank E, Hall MA, Pal CJ (2016). Data mining, Fourth Edition: Practical machine learning tools and techniques. 4th ed. San Francisco: Morgan Kaufmann Publishers Inc.
Goodfellow I, Bengio Y, Courville A (2016). Deep learning. Cambridge: The MIT Press; 2016.
Minar MR, Naher J. (2018). Recent advances in deep learning: an overview. arXiv:1807.08169.
LeCun Y, Bengio Y, Hinton G (2015). Deep learning. Nature;521:436.
Schmidhuber J. (2015). Deep learning in neural networks: an overview. Neural Net;61:85–117.
Rumelhart DE, Hinton GE, Williams RJ. (1986). Learning representations by back-propagating errors. Nature;323:533.
Y. Le Cun,, Y., Boser B., Denker J.S., Henderson D., Howard R.E., Hubbard W., Jackel L.D. (1990) :Handwritten digit recognition with a back-propagation network. In Advances in neural information processing systems.
LeCun Y, Boser B, Denker JS, Henderson D, Howard RE, Hubbard W, Jackel LD. (1989). Backpropagation applied to handwritten zip code recognition. Neural Comput;1(4):541–51.
Hinton GE, Osindero S, Teh Y-W. (2006). A fast learning algorithm for deep belief nets. Neural Comput;18(7):1527–54. https://doi.org/10.1162/neco.2006.18.7.1527.
Bengio Y, Lamblin P, Popovici D, Larochelle H. (2006) Greedy layer-wise training of deep networks. In: Proceedings of the 19th international conference on neural information processing systems. NIPS’06. MIT Press, Cambridge, MA, USA. p. 153–60.
Russakovsky O, Deng J, Su H, Krause J, Satheesh S, Ma S, Huang Z, Karpathy A, Khosla A, Bernstein M, Berg AC, Fei-Fei L. (2015). ImageNet large scale visual recognition challenge. Int J Comput Vision (IJCV);115(3):211–52. https://doi.org/10.1007/s11263-015-0816-y.
Kumar M. (2016). An incorporation of artificial intelligence capabilities in cloud computing. Int J Eng Comput Sci. https://doi.org/10.18535/ijecs/v5i11.63.
Saiyeda A, Mir MA. (2017). Cloud computing for deep learning analytics: a survey of current trends and challenges. Int J Adv Res Comput Sci;8(2):68–72.
Dumbill E. (2012). What is big data?: an introduction to the big data landscape. http://radar.oreilly.com/2012/01/what-is-big-data.html
Najafabadi MM, Villanustre F, Khoshgoftaar TM, Seliya N, Wald R, Muharemagic E. (2015). Deep learning applications and challenges in big data analytics. J Big Data; 2(1):1. https://doi.org/10.1186/s40537-014-0007-7.
Hinton G, Salakhutdinov R. (2011). Discovering binary codes for documents by learning deep generative models. Top Cogn Sci.;3(1):74–91.
Salakhutdinov R, Hinton G. (2009). Semantic hashing. Int J Approx Reason;50(7):969–78.
Downloads
Published
How to Cite
Issue
Section
License
I assign to Informatica, An International Journal of Computing and Informatics ("Journal") the copyright in the manuscript identified above and any additional material (figures, tables, illustrations, software or other information intended for publication) submitted as part of or as a supplement to the manuscript ("Paper") in all forms and media throughout the world, in all languages, for the full term of copyright, effective when and if the article is accepted for publication. This transfer includes the right to reproduce and/or to distribute the Paper to other journals or digital libraries in electronic and online forms and systems.
I understand that I retain the rights to use the pre-prints, off-prints, accepted manuscript and published journal Paper for personal use, scholarly purposes and internal institutional use.
In certain cases, I can ask for retaining the publishing rights of the Paper. The Journal can permit or deny the request for publishing rights, to which I fully agree.
I declare that the submitted Paper is original, has been written by the stated authors and has not been published elsewhere nor is currently being considered for publication by any other journal and will not be submitted for such review while under review by this Journal. The Paper contains no material that violates proprietary rights of any other person or entity. I have obtained written permission from copyright owners for any excerpts from copyrighted works that are included and have credited the sources in my article. I have informed the co-author(s) of the terms of this publishing agreement.
Copyright © Slovenian Society Informatika