Displaying publications 41 - 60 of 167 in total

Abstract:
Sort:
  1. Sinin Hamdan, Iran Amri Musoddiq, Ahmad Fauzi Musib, Marini Sawawi
    MyJurnal
    The tone of peking 1, 2, 3, 5, 6, 1’ was investigated using time-frequency analysis (TFA). The frequencies were measured using PicoScope oscilloscope, Melda analyzer in Cubase version 9 and Adobe version 3. Three different approaches for time-frequency analysis were used: Fourier spectra (using PicoScope), spectromorphology (using Melda analyzer) and spectrograms (using Adobe). Fourier spectra only identify intensity-frequency within entire signals, while spectromorphology identify the changes of intensity-frequency spectrum at fixed time and Adobe spectrograms identify the frequency with time. PicoScope reading produces the spectra of the fundamental and overtone frequencies in the entire sound. These overtones are non-harmonic since they are non-integral multiples of the fundamental. The fundamental frequencies of peking 1, 2, 3, 5, 6 were 1066Hz (C6), 1178Hz (D6), 1342Hz (E6), 1599Hz (G6) and 1793Hz (A6) respectively while peking 1’was 2123Hz (C7) i.e. one octave higher than peking 1. Melda analyzer reading proved that all peking sustained the initial fundamental frequency and overtone at t=0 until 2s. TFA from Adobe reading provides a description of the sound in the time-frequency plane. From TFA, peking 1, 2 and 6 exhibited a much gentler attack and more rapid decay than peking 3, 5 and 1’.
    Matched MeSH terms: Speech Acoustics
  2. Da'u A, Salim N
    PeerJ Comput Sci, 2019;5:e191.
    PMID: 33816844 DOI: 10.7717/peerj-cs.191
    Aspect extraction is a subtask of sentiment analysis that deals with identifying opinion targets in an opinionated text. Existing approaches to aspect extraction typically rely on using handcrafted features, linear and integrated network architectures. Although these methods can achieve good performances, they are time-consuming and often very complicated. In real-life systems, a simple model with competitive results is generally more effective and preferable over complicated models. In this paper, we present a multichannel convolutional neural network for aspect extraction. The model consists of a deep convolutional neural network with two input channels: a word embedding channel which aims to encode semantic information of the words and a part of speech (POS) tag embedding channel to facilitate the sequential tagging process. To get the vector representation of words, we initialized the word embedding channel and the POS channel using pretrained word2vec and one-hot-vector of POS tags, respectively. Both the word embedding and the POS embedding vectors were fed into the convolutional layer and concatenated to a one-dimensional vector, which is finally pooled and processed using a Softmax function for sequence labeling. We finally conducted a series of experiments using four different datasets. The results indicated better performance compared to the baseline models.
    Matched MeSH terms: Speech
  3. Weng, Brandon Chai An
    Borneo Akademika, 2020;4(4):1-8.
    MyJurnal
    Words have a habit of appearing in recurrent patterns. These recurring patterns may take the
    form of phrasal verbs, collocations, and other multi-word expressions (MWEs). Since these
    patterns constantly re-occur in both speech and writing, it would seem prudent to teach
    vocabulary to ESL learners in the manner in which words actually present themselves: in
    typical chunks. A common example of such a chunk would be “to deal with the problem”, in
    which “deal with” and “problem” are collocates. This paper contains three sections. The first
    examines what phrasal verbs and collocates are in the first place, and why they are particularly
    important for learners. The second part is a review of recent studies that support the teaching
    of MWEs and chunking pedagogy. This paper concludes by discussing the potential of
    incorporating chunking pedagogy in one’s own ESL vocabulary teaching.
    Matched MeSH terms: Speech
  4. Ahmad, R., Ishlah, W., Norie A.
    MyJurnal
    The ultimate goal of treatment of glottic malignancy is eradication of the cancer and preservation of maximum function. Ideally this would mean return of normal speech, respiration and deglutition. The foundation of vertical partial laryngectomy (VPL) is based on the knowledge of laryngeal anatomy and an understanding of how cancers invade the larynx. The VPL surgery of the glottic cancer is based on oncological sound principles, which will be further disscussed in this article. For the purpose of description we briefly presented 3 of our cases that underwent vertical partial laryngectomy for glottic cancer.
    Matched MeSH terms: Speech
  5. Al-Haddad, S.A.R., Samad, S.A., Hussain, A., Ishak, K.A., Noor, A.O.A.
    ASM Science Journal, 2008;2(1):75-81.
    MyJurnal
    Robustness is a key issue in speech recognition. A speech recognition algorithm for Malay digits from zero to nine and an algorithm for noise cancellation by using recursive least squares (RLS) is proposed in this article. This system consisted of speech processing inclusive of digit margin and recognition using zero crossing and energy calculations. Mel-frequency cepstral coefficient vectors were used to provide an estimate of the vocal tract filter. Meanwhile dynamic time warping was used to detect the nearest recorded voice with appropriate global constraint. The global constraint was used to set a valid search region because the variation of the speech rate of the speaker was considered to be limited in a reasonable range which meant that it could prune the unreasonable search space. The algorithm was tested on speech samples that were recorded as part of a Malay corpus. The results showed that the algorithm managed to recognize almost 80.5% of the Malay digits for all recorded words. The addition of a RLS noise canceller in the preprocessing stage increased the accuracy to 94.1%.
    Matched MeSH terms: Speech
  6. Mohd. Izhan Mohd. Yusoff, Mohd. Rizam Abu Bakar, Abu Hassan Shaari Mohd. Nor
    MyJurnal
    Expectation Maximization (EM) algorithm has experienced a significant increase in terms of usage in many fields of study. In this paper, the performance of the said algorithm in finding the Maximum Likelihood for the Gaussian Mixed Models (GMM), a probabilistic model normally used in fraud detection and recognizing a person’s voice in speech recognition field, is shown and discussed. At the end of the paper, some suggestions for future research works will also be given.
    Matched MeSH terms: Speech
  7. Noor Alaudin Abdul Wahab, Wan Fazlina Wan Hashim
    MyJurnal
    Garis panduan saringan telinga tengah oleh American Speech-Language-Hearing Association (ASHA) telah menyarankan agar pendekatan kuantitatif digunakan untuk menginterpretasi keputusan timpanogram. Berbanding pendekatan kualitatif yang digunapakai untuk menginterpretasi timpanogram sebelum ini, pendekatan kuantitatif adalah lebih sesuai kerana ianya bersifat objektif serta dapat diaplikasi dengan peralatan timpanometer komersial yang terdapat di pasaran sekarang. Kajian ini bertujuan untuk mendapatkan nilai mutlak ciri timpanometri iaitu puncak statik admitan akustik dikompensasi (Puncak Ytm), kelebaran timpanogram (TW) dan isipadu salur telinga luar (Vea) di kalangan kanak-kanak pra-sekolah normal dan Sindrom Down. Lima belas kanak-kanak normal dan 12 kanak-kanak Sindrom Down yang memenuhi kriteria pemilihan subjek terlibat di dalam kajian ini. Nilai purata Puncak Ytm dan Vea bagi kanak-kanak normal masing-masing adalah 0.36 mmho dan 0.57 cm3 manakala kanak-kanak Sindrom Down mencatat nilai purata 0.14 mmho dan 0.38 cm3. Kanak-kanak normal dan Sindrom Down masingmasing mencatatkan nilai purata TW 99.73 dan 148.65 daPa. Analisis statistik menunjukkan terdapat perbezaan yang signifikan di antara kedua-dua kumpulan subjek bagi ketiga-tiga nilai parameter timpanometri yang diperolehi.
    Matched MeSH terms: American Speech-Language-Hearing Association
  8. Ling, Booi C.
    Ann Dent, 1995;2(1):-.
    MyJurnal
    Temporary obturator is necessary for maxillectomy procedures to minimise functional disabilities of speech and the swallowing of food and fluid following surgery.l.2 It also provides a base to hold surgical packing and/ or tissue conditioners in place. This packing helps to restore function and counter act contraction during healing of the ,tissues at the operated site. Dental clinicians are often faced with the problem of surgical staffs requesting an obturator for their patients at short notice.This article describes a fast technique of constructing an immediate surgical obturator.
    Matched MeSH terms: Speech
  9. Jalaei B, Azmi MHAM, Zakaria MN
    Braz J Otorhinolaryngol, 2018 05 17;85(4):486-493.
    PMID: 29858160 DOI: 10.1016/j.bjorl.2018.04.005
    INTRODUCTION: Binaurally evoked auditory evoked potentials have good diagnostic values when testing subjects with central auditory deficits. The literature on speech-evoked auditory brainstem response evoked by binaural stimulation is in fact limited. Gender disparities in speech-evoked auditory brainstem response results have been consistently noted but the magnitude of gender difference has not been reported.

    OBJECTIVE: The present study aimed to compare the magnitude of gender difference in speech-evoked auditory brainstem response results between monaural and binaural stimulations.

    METHODS: A total of 34 healthy Asian adults aged 19-30 years participated in this comparative study. Eighteen of them were females (mean age=23.6±2.3 years) and the remaining sixteen were males (mean age=22.0±2.3 years). For each subject, speech-evoked auditory brainstem response was recorded with the synthesized syllable /da/ presented monaurally and binaurally.

    RESULTS: While latencies were not affected (p>0.05), the binaural stimulation produced statistically higher speech-evoked auditory brainstem response amplitudes than the monaural stimulation (p<0.05). As revealed by large effect sizes (d>0.80), substantive gender differences were noted in most of speech-evoked auditory brainstem response peaks for both stimulation modes.

    CONCLUSION: The magnitude of gender difference between the two stimulation modes revealed some distinct patterns. Based on these clinically significant results, gender-specific normative data are highly recommended when using speech-evoked auditory brainstem response for clinical and future applications. The preliminary normative data provided in the present study can serve as the reference for future studies on this test among Asian adults.

    Matched MeSH terms: Speech/physiology*; Speech Perception/physiology*
  10. Jalaei B, Zakaria MN, Mohd Azmi MH, Nik Othman NA, Sidek D
    Ann Otol Rhinol Laryngol, 2017 Apr;126(4):290-295.
    PMID: 28177264 DOI: 10.1177/0003489417690169
    OBJECTIVES: Gender disparities in speech-evoked auditory brainstem response (speech-ABR) outcomes have been reported, but the literature is limited. The present study was performed to further verify this issue and determine the influence of head size on speech-ABR results between genders.

    METHODS: Twenty-nine healthy Malaysian subjects (14 males and 15 females) aged 19 to 30 years participated in this study. After measuring the head circumference, speech-ABR was recorded by using synthesized syllable /da/ from the right ear of each participant. Speech-ABR peaks amplitudes, peaks latencies, and composite onset measures were computed and analyzed.

    RESULTS: Significant gender disparities were noted in the transient component but not in the sustained component of speech-ABR. Statistically higher V/A amplitudes and less steeper V/A slopes were found in females. These gender differences were partially affected after controlling for the head size.

    CONCLUSIONS: Head size is not the main contributing factor for gender disparities in speech-ABR outcomes. Gender-specific normative data can be useful when recording speech-ABR for clinical purposes.

    Matched MeSH terms: Speech*; Speech Perception/physiology*
  11. Vong E, Wilson L, Lincoln M
    J Fluency Disord, 2016 09;49:29-39.
    PMID: 27638190 DOI: 10.1016/j.jfludis.2016.07.003
    PURPOSE: This study investigated the outcomes of implementing the Lidcombe Program, an evidence-based early intervention for stuttering, with four preschool children in Malaysia. Early stuttering intervention is currently underdeveloped in Malaysia, where stuttering treatment is often more assertion-based than evidence-based. Therefore, introducing an evidence-based early stuttering intervention is an important milestone for Malaysian preschoolers who stutter.

    METHOD: The participants ranged from 3 years 3 months to 4 years 9 months at the start of the study. Beyond-clinic speech samples were obtained at 1 month and 1 week pretreatment and immediately post-Stage 1, and at 1 month, 3 months, 6 months and 12 months post-Stage 1.

    RESULTS: Two participants, who were bilingual, achieved near-zero levels of stuttering at 12 months posttreatment. Near zero levels of stuttering were also present in their untreated languages. One participant withdrew due to reasons not connected with the research or treatment. The remaining participant, who presented with severe stuttering, completed Stage 1 but had some relapse in Stage 2 and demonstrated mild stuttering 12 months post-Stage 1.

    CONCLUSIONS: The outcomes were achieved without the need to significantly adapt Lidcombe Program procedures to Malaysian culture. Further research to continue evaluation of the Lidcombe Program with Malaysian families and to estimate proportion of those who will respond is warranted.

    Matched MeSH terms: Speech Production Measurement/methods; Speech Therapy/methods*
  12. Nashrah Maamor, Nurul Huda Bani, Wan Syafira Ishak, Siti Zamratol Mai-Sarah Mukari, Kalaivani Chellapan
    Jurnal Sains Kesihatan Malaysia, 2018;16(101):215-215.
    MyJurnal
    This paper describes the systematic process followed in the development of culturally appropriate equalized speech-in-noise sentences suitable for use in an adaptive Speech-In-Noise training protocol for adults in Malaysia. The process involved three iterative phases of development. They were (1) analysis, (2) design and (3) development phases. In the analysis phase, important variables that needed to be considered for speech-in-noise materials were identified through literature review and discussion with the experts in the field. Next, in the design phase, the compilation and formation of sentences, evaluation of naturalness and recording of the speech materials were done. The last phase was the development phase which involved the evaluation of performance intensity function and equalization of intelligibility. The final outcome of these phases were 171 sentences with equal intelligibility that can be used interchangeably in a speech-in-noise training protocol for adults in Malaysia.
    Matched MeSH terms: Speech Intelligibility
  13. Yusof Y, Mukari SZS, Dzulkifli MA, Chellapan K, Ahmad K, Ishak I, et al.
    Geriatr Gerontol Int, 2019 Aug;19(8):768-773.
    PMID: 31237107 DOI: 10.1111/ggi.13710
    AIM: To evaluate the efficacy of a newly developed auditory-cognitive training system on speech recognition, central auditory processing and cognition among older adults with normal cognition (NC) and with neurocognitive impairment (NCI).

    METHODS: A double-blind quasi-experiment was carried out on NC (n = 43) and NCI (n = 33) groups. Participants in each group were randomly assigned into treatment and control programs groups. The treatment group underwent auditory-cognitive training, whereas the control group was assigned to watch documentary videos, three times per week, for 8 consecutive weeks. Study outcomes that included Montreal Cognitive Assessment, Malay Hearing in Noise Test, Dichotic Digit Test, Gaps in Noise Test and Pitch Pattern Sequence Test were measured at 4-week intervals at baseline, and weeks 4, 8 and 12.

    RESULTS: Mixed design anova showed significant training effects in total Montreal Cognitive Assessment and Dichotic Digit Test in both groups, NC (P 

    Matched MeSH terms: Audiometry, Speech/methods; Speech Perception*
  14. Zakaria MN, Jalaei B
    Int J Pediatr Otorhinolaryngol, 2017 Nov;102:28-31.
    PMID: 29106871 DOI: 10.1016/j.ijporl.2017.08.033
    OBJECTIVE: Auditory brainstem responses evoked by complex stimuli such as speech syllables have been studied in normal subjects and subjects with compromised auditory functions. The stability of speech-evoked auditory brainstem response (speech-ABR) when tested over time has been reported but the literature is limited. The present study was carried out to determine the test-retest reliability of speech-ABR in healthy children at a low sensation level.

    METHODS: Seventeen healthy children (6 boys, 11 girls) aged from 5 to 9 years (mean = 6.8 ± 3.3 years) were tested in two sessions separated by a 3-month period. The stimulus used was a 40-ms syllable /da/ presented at 30 dB sensation level.

    RESULTS: As revealed by pair t-test and intra-class correlation (ICC) analyses, peak latencies, peak amplitudes and composite onset measures of speech-ABR were found to be highly replicable. Compared to other parameters, higher ICC values were noted for peak latencies of speech-ABR.

    CONCLUSION: The present study was the first to report the test-retest reliability of speech-ABR recorded at low stimulation levels in healthy children. Due to its good stability, it can be used as an objective indicator for assessing the effectiveness of auditory rehabilitation in hearing-impaired children in future studies.

    Matched MeSH terms: Speech/physiology; Speech Perception/physiology*
  15. Lee OW, Mao D, Savkovic B, Wunderlich J, Nicholls N, Jeffreys E, et al.
    Ear Hear, 2023 01 14;44(4):776-786.
    PMID: 36706073 DOI: 10.1097/AUD.0000000000001325
    OBJECTIVES: Cardiac responses (e.g., heart rate changes) due to an autonomous response to sensory stimuli have been reported in several studies. This study investigated whether heart rate information extracted from functional near-infrared spectroscopy (fNIRS) data can be used to assess the discrimination of speech sounds in sleeping infants. This study also investigated the adaptation of the heart rate response over multiple, sequential stimulus presentations.

    DESIGN: fNIRS data were recorded from 23 infants with no known hearing loss, aged 2 to 10 months. Speech syllables were presented using a habituation/dishabituation test paradigm: the infant's heart rate response was first habituated by repeating blocks of one speech sound; then, the heart rate response was dishabituated with the contrasting (novel) speech sound. This stimulus presentation sequence was repeated for as long as the infants were asleep.

    RESULTS: The group-level average heart rate response to the novel stimulus was greater than that to the habituated first sound, indicating that sleeping infants were able to discriminate the speech sound contrast. A significant adaptation of the heart rate responses was seen over the session duration.

    CONCLUSION: The dishabituation response could be a valuable marker for speech discrimination, especially when used in conjunction with the fNIRS hemodynamic response.

    Matched MeSH terms: Speech
  16. Fraundorf SH, Watson DG, Benjamin AS
    Psychol Aging, 2012 Mar;27(1):88-98.
    PMID: 21639646 DOI: 10.1037/a0024138
    In two experiments, we investigated age-related changes in how prosodic pitch accents affect memory. Participants listened to recorded discourses that contained two contrasts between pairs of items (e.g., one story contrasted British scientists with French scientists and Malaysia with Indonesia). The end of each discourse referred to one item from each pair; these references received a pitch accent that either denoted contrast (L + H* in the ToBI system) or did not (H*). A contrastive accent on a particular pair improved later recognition memory equally for young and older adults. However, older adults showed decreased memory if the other pair received a contrastive accent (Experiment 1). Young adults with low working memory performance also showed this penalty (Experiment 2). These results suggest that pitch accents guide processing resources to important information for both older and younger adults but diminish memory for less important information in groups with reduced resources, including older adults.
    Matched MeSH terms: Speech Acoustics*; Speech Perception/physiology*
  17. Mustafa MB, Ainon RN
    J Acoust Soc Am, 2013 Oct;134(4):3057-66.
    PMID: 24116440 DOI: 10.1121/1.4818741
    The ability of speech synthesis system to synthesize emotional speech enhances the user's experience when using this kind of system and its related applications. However, the development of an emotional speech synthesis system is a daunting task in view of the complexity of human emotional speech. The more recent state-of-the-art speech synthesis systems, such as the one based on hidden Markov models, can synthesize emotional speech with acceptable naturalness with the use of a good emotional speech acoustic model. However, building an emotional speech acoustic model requires adequate resources including segment-phonetic labels of emotional speech, which is a problem for many under-resourced languages, including Malay. This research shows how it is possible to build an emotional speech acoustic model for Malay with minimal resources. To achieve this objective, two forms of initialization methods were considered: iterative training using the deterministic annealing expectation maximization algorithm and the isolated unit training. The seed model for the automatic segmentation is a neutral speech acoustic model, which was transformed to target emotion using two transformation techniques: model adaptation and context-dependent boundary refinement. Two forms of evaluation have been performed: an objective evaluation measuring the prosody error and a listening evaluation to measure the naturalness of the synthesized emotional speech.
    Matched MeSH terms: Speech Acoustics*; Speech Perception*; Speech Production Measurement
  18. Ooi CC, Wong AM
    Int J Speech Lang Pathol, 2012 Dec;14(6):499-508.
    PMID: 23039126 DOI: 10.3109/17549507.2012.712159
    One reason why specific language impairment (SLI) is grossly under-identified in Malaysia is the absence of locally- developed norm-referenced language assessment tools for its multilingual and multicultural population. Spontaneous language samples provide quantitative information for language assessment, and useful descriptive information on child language development in complex language and cultural environments. This research consisted of two studies and investigated the use of measures obtained from English conversational samples among bilingual Chinese-English Malaysian preschoolers. The research found that the language sample measures were sensitive to developmental changes in this population and could identify SLI. The first study examined the relationship between age and mean length of utterance (MLU(w)), lexical diversity (D), and the index of productive syntax (IPSyn) among 52 typically-developing (TD) children aged between 3;4-6;9. Analyses showed a significant linear relationship between age and D (r = .450), the IPsyn (r = .441), and MLU(w) (r = .318). The second study compared the same measures obtained from 10 children with SLI, aged between 3;8-5;11, and their age-matched controls. The children with SLI had significantly shorter MLU(w) and lower IPSyn scores than the TD children. These findings suggest that utterance length and syntax production can be potential clinical markers of SLI in Chinese-English Malaysian children.
    Matched MeSH terms: Speech Acoustics*; Speech-Language Pathology/methods*; Speech Production Measurement*
  19. Zourmand A, Mirhassani SM, Ting HN, Bux SI, Ng KH, Bilgen M, et al.
    Biomed Eng Online, 2014;13:103.
    PMID: 25060583 DOI: 10.1186/1475-925X-13-103
    The phonetic properties of six Malay vowels are investigated using magnetic resonance imaging (MRI) to visualize the vocal tract in order to obtain dynamic articulatory parameters during speech production. To resolve image blurring due to the tongue movement during the scanning process, a method based on active contour extraction is used to track tongue contours. The proposed method efficiently tracks tongue contours despite the partial blurring of MRI images. Consequently, the articulatory parameters that are effectively measured as tongue movement is observed, and the specific shape of the tongue and its position for all six uttered Malay vowels are determined.Speech rehabilitation procedure demands some kind of visual perceivable prototype of speech articulation. To investigate the validity of the measured articulatory parameters based on acoustic theory of speech production, an acoustic analysis based on the uttered vowels by subjects has been performed. As the acoustic speech and articulatory parameters of uttered speech were examined, a correlation between formant frequencies and articulatory parameters was observed. The experiments reported a positive correlation between the constriction location of the tongue body and the first formant frequency, as well as a negative correlation between the constriction location of the tongue tip and the second formant frequency. The results demonstrate that the proposed method is an effective tool for the dynamic study of speech production.
    Matched MeSH terms: Speech/physiology*
  20. Ahmad K, Yan Y, Bless DM
    J Voice, 2012 Mar;26(2):239-53.
    PMID: 21621975 DOI: 10.1016/j.jvoice.2011.02.001
    The purpose of the study was to investigate relationships between vocal fold vibrations and voice quality. Laryngeal images obtained from high-speed digital imaging (HSDI) were examined for their open-closed timing characteristics and perturbation values. A customized software delineated the glottal edges and used the Hilbert transform-based method of analysis to provide objective quantification of glottal perturbation. Overlay tracings of the transformed glottal cycles provided visual patterns on the overall vibratory dynamics. In this paper, we described the use of this method in looking at vibratory characteristics of a group of young female speakers (N=23). We found that, females with no voice complaints and who had been perceived to have normal voices were not a homogeneous group in terms of their glottal vibratory patterns during phonation. Their vibratory patterns showed characteristics similar to exemplar voices targeted to be clear (50%), pressed (27%), breathy (15%), or a mixed quality (8%). Perturbation range in terms of cycle-to-cycle frequency and amplitude was small and did not discriminate patterns. All these patterns yielded perceptually normal voices suggesting that in normal young speakers, the level of perturbation may be more important to the judgment than the actual pattern of closure.
    Matched MeSH terms: Speech/physiology
Filters
Contact Us

Please provide feedback to Administrator ([email protected])

External Links