An Efficient Grocery Detection System Using HYOLO-NAS Deep Learning Model for Visually Impaired People

  • Chhabra Payal
    Dept. of Computer Science and Engineering, Maharishi Markandeshwar (Deemed to be University) M.M Engineering College
  • Goyal Sonali
    Dept. of Computer Science and Engineering, Maharishi Markandeshwar (Deemed to be University) M.M Engineering College

Search this article

Description

Real-time application of object detections are common, and the area of computer vision dramatically benefits from them. Recognizing grocery items poses a more significant challenge for blind individuals compared to those with normal vision. For that purpose, an effective model HYOLO-NAS, is used to detect groceries to aid the visually impaired by seamlessly converting text to audio messages. In the proposed work, Neural Architecture Search technology is used to dynamically update the weights that design child neural networks with the highest accuracy. The hyperparameter tuning on the child network involves adjusting the learning rate, number of epochs, and L2 regularization of weight decay with an Adaptive Moment Estimation optimizer. Google’s Text-to-Speech (gTTS) transforms text into speech signals. After doing many inference experiments, the Hypertuned YOLO-NAS grocery detection model is introduced. The experimental results show that optimized HYOLO-NAS outperforms various detection algorithms with mAP0.5 reaching 96.80% on Grozi-120 and 97.61% on the Retail Product dataset.

Journal

  • Evergreen

    Evergreen 11 (3), 1990-2003, 2024-09

    Transdiscilinary Research and Education Center for Green Technologies, Kyushu University

Details 詳細情報について

  • CRID
    1390301763614376320
  • DOI
    10.5109/7236846
  • ISSN
    24325953
    21890420
  • HANDLE
    2324/7236846
  • Text Lang
    en
  • Article Type
    journal article
  • Data Source
    • JaLC
    • IRDB
    • Crossref
  • Abstract License Flag
    Allowed

Report a problem

Back to top