An Efficient Grocery Detection System Using HYOLO-NAS Deep Learning Model for Visually Impaired People

  • Chhabra Payal
    Dept. of Computer Science and Engineering, Maharishi Markandeshwar (Deemed to be University) M.M Engineering College
  • Goyal Sonali
    Dept. of Computer Science and Engineering, Maharishi Markandeshwar (Deemed to be University) M.M Engineering College

この論文をさがす

説明

Real-time application of object detections are common, and the area of computer vision dramatically benefits from them. Recognizing grocery items poses a more significant challenge for blind individuals compared to those with normal vision. For that purpose, an effective model HYOLO-NAS, is used to detect groceries to aid the visually impaired by seamlessly converting text to audio messages. In the proposed work, Neural Architecture Search technology is used to dynamically update the weights that design child neural networks with the highest accuracy. The hyperparameter tuning on the child network involves adjusting the learning rate, number of epochs, and L2 regularization of weight decay with an Adaptive Moment Estimation optimizer. Google’s Text-to-Speech (gTTS) transforms text into speech signals. After doing many inference experiments, the Hypertuned YOLO-NAS grocery detection model is introduced. The experimental results show that optimized HYOLO-NAS outperforms various detection algorithms with mAP0.5 reaching 96.80% on Grozi-120 and 97.61% on the Retail Product dataset.

収録刊行物

  • Evergreen

    Evergreen 11 (3), 1990-2003, 2024-09

    九州大学グリーンテクノロジー研究教育センター

詳細情報 詳細情報について

  • CRID
    1390301763614376320
  • DOI
    10.5109/7236846
  • ISSN
    24325953
    21890420
  • HANDLE
    2324/7236846
  • 本文言語コード
    en
  • 資料種別
    journal article
  • データソース種別
    • JaLC
    • IRDB
    • Crossref
  • 抄録ライセンスフラグ
    使用可

問題の指摘

ページトップへ