Driver Behavior State Recognition based on Silence Removal Speech

被引:0
|
作者
Kamaruddin, Norhaslinda [1 ]
Rahman, Abdul Wahab Abdul [2 ]
Halim, Khairul Ikhwan Mohamad [1 ]
Noh, Muhammad Hafiq Iqmal Mohd [1 ]
机构
[1] Univ Teknol MARA Melaka, Kampus Jasin, Merlimau 77300, Melaka, Malaysia
[2] Int Islamic Univ Malaysia, Kulliyyah Informat & Commun Technol, Kuala Lumpur, Malaysia
来源
2016 INTERNATIONAL CONFERENCE ON INFORMATICS AND COMPUTING (ICIC) | 2016年
关键词
driver behavior state; silence removal; Zero Crossing Rate; Short Term Energy; Mel Frequency Cepstral Coefficient; Multi Layer Perceptron;
D O I
暂无
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Numerous researches have linked driver behavior to the cause of accident and some studies are concentrated into different input providing practical preventive measures. Nonetheless speech has been found to be a suitable input source in understanding and analyzing driver's behavior state due to the underlying emotional information when the driver speaks and such changes can be measured. However, the massive amount of driving speech data may hinder optimal performance of processing and analyzing the data due to the computational complexity and time constraint. This paper presents a silence removal approach using Short Term Energy (STE) and Zero Crossing Rate (ZCR) prior to extracting the relevant features in order to reduce the computational time in a vehicular environment. Mel Frequency Cepstral Coefficient (MFCC) feature extraction method coupled with Multi Layer Perceptron (MLP) classifier are employed to get the driver behavior state recognition performance. Experimental results demonstrated that the proposed approach is able to obtain comparable performance with accuracy ranging between 58.7% and 76.6% to differentiate four driver behavior states, namely; talking through cell telephone phone, out-burst laughing, sleepy and normal driving. It is envisages that such engine can be extended for a more comprehensive driver behavior identification system that may acts as an embedded warning system for sleepy driver.
引用
收藏
页码:186 / 191
页数:6
相关论文
共 50 条
  • [41] THE STATE-OF-THE-ART IN SPEECH RECOGNITION
    BISIANI, R
    TRENDS IN NEUROSCIENCES, 1985, 8 (01) : 9 - 11
  • [42] Study and Application of Silence Model Adaptation for Use in Telephone Speech Recognition System
    Novotny, Jan
    Sovka, Pavel
    Uhlir, Jan
    RADIOENGINEERING, 2004, 13 (03) : 1 - 6
  • [43] Noisy speech recognition based on speech enhancement
    Wang, Xia
    Tang, Hongmei
    Zhao, Xiaoqun
    SNPD 2007: EIGHTH ACIS INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING, ARTIFICIAL INTELLIGENCE, NETWORKING, AND PARALLEL/DISTRIBUTED COMPUTING, VOL 3, PROCEEDINGS, 2007, : 713 - +
  • [44] Dynamical Energy-Based Speech/Silence Detector for Speech Enhancement Applications
    Sakhnov, Kirill
    Verteletskaya, Ekaterina
    Simak, Boris
    WORLD CONGRESS ON ENGINEERING 2009, VOLS I AND II, 2009, : 801 - 806
  • [45] A Survey of EEG-Based Driver State and Behavior Detection for Intelligent Vehicles
    Ju, Jiawei
    Li, Hongqi
    IEEE TRANSACTIONS ON BIOMETRICS, BEHAVIOR, AND IDENTITY SCIENCE, 2024, 6 (03): : 420 - 434
  • [46] Speech is silver, silence is golden? Examining state activity in international negotiations
    Diana Panke
    The Review of International Organizations, 2017, 12 : 121 - 146
  • [47] Speech is silver, silence is golden? Examining state activity in international negotiations
    Panke, Diana
    REVIEW OF INTERNATIONAL ORGANIZATIONS, 2017, 12 (01): : 121 - 146
  • [48] A lightweight and explainable model for driver abnormal behavior recognition
    Hao, Jingbin
    Sun, Xiaokai
    Liu, Xinhua
    Hua, Dezheng
    Hu, Jianhua
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2025, 139
  • [49] End-to-End Speech Recognition and Disfluency Removal
    Lou, Paria Jamshid
    Johnson, Mark
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 2051 - 2061
  • [50] Integrated bias removal techniques for robust speech recognition
    Lawrence, C
    Rahim, M
    COMPUTER SPEECH AND LANGUAGE, 1999, 13 (03): : 283 - 298