SlideShare una empresa de Scribd logo
1 de 14
Descargar para leer sin conexión
INTERNATIONAL JOURNAL OF ELECTRONICS AND
 International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN
 0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME
  COMMUNICATION ENGINEERING & TECHNOLOGY (IJECET)

ISSN 0976 – 6464(Print)
ISSN 0976 – 6472(Online)
Volume 3, Issue 2, July- September (2012), pp. 116-129
                                                                       IJECET
© IAEME: www.iaeme.com/ijecet.html                                   ©IAEME
Journal Impact Factor (2012): 3.5930 (Calculated by GISI)
www.jifactor.com




   COCHLEAR IMPLANT ACOUSTIC SIMULATION MODEL BASED
               ON CRITICAL BAND FILTERS

                                P Mahalakshmi1, M R Reddy2
            Bio-medical Engineering Group, Department of Applied Mechanics,
        Indian Institute of Technology Madras, Chennai-600 036, Tamil Nadu, India.
                1
                  E-mail: maha_50@yahoo.com, 2E-mail: rsreddy@iitm.ac.in


 ABSTRACT

 Cochlear Implants (CI) are surgically implanted biomedical devices that can provide
 hearing to profoundly deaf people by direct electrical stimulation of the auditory nerve.
 Choosing the optimal number of electrodes for the improved perception of speech is an
 especially difficult problem for users of CI, as the prostheses system depends on various
 parameters and processing strategies. It has widely been recognized that using the auditory
 models in cochlear prostheses might significantly improve the speech perception. In this
 paper, we propose a new CI acoustic simulation model based on a 16-channel Critical Band
 Finite-Impulse Response (CB-FIR) filter bank and its performance is discussed. The place
 of stimulation in the cochlear electrode array is identified through the computation and
 selection of significant spectral band energies. We also investigate the temporal
 information through the simulation model based algorithm. The amplitude levels of the
 pulses are identified by envelope detection method. The simulation model is tested with
 TIMIT database for 15 phonemes. The experimental results indicated that, a reduced
 number of channels are only required for stimulation of cochlear electrodes, one at a time,
 thereby minimizing channel interaction. The proposed model should be helpful for the
 development of novel speech processing strategies to improve the speech perception of
 cochlear implantees.

 Keywords

 Auditory model, cochlear implant, speech processing, speech perception, spectral analysis,
 acoustic simulation.


     1. INTRODUCTION
 A cochlear implant (CI) is an auditory neural prosthesis for restoring hearing function in
 patients with sensori-neural hearing loss. Hearing restoration is achieved by electrically


                                              116
International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN
0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME
stimulating the auditory nerve [1]. The electrical stimulation pulse parameters are derived
from incoming speech by speech processors contained within the CI devices [2]. Several CI
devices developed over the years have the following features in common. It is comprised of
external/ and internal units. The external unit includes microphone, speech processor and
transmitter. The internal unit consists of receiver-stimulator and an electrode array
embedded in cochlea [3]. The sound is picked up by the microphone and processed by the
signal/speech processor that analyzes sound and encodes the results into instructions. By
RF (radio frequency) signal, the transmission system transmits the instructions to the
implanted receiver. The internal circuit decodes the data, then generates electrical stimulus
and drives the electrode array to stimulate auditory nerve fibers at different places [4]. The
speech processor is responsible for decomposing the input audio signals into different
frequency bands and formulating the most appropriate stimulation pattern to the electrodes
[5]. The design of speech processors must be such that it extracts the speech parameters
that are essential for intelligibility and then encode them for electrical stimulation [6].
Speech understanding requires the analysis of various speech components such as vowels
and consonants. The estimate of perceptual importance of these component features will
lead to intelligible speech stimuli [7]. This speech stimuli depends on various parameters
such as speech material, filter cutoff frequencies, order of band pass and low pass filters,
number of channels, number of electrodes, type and parameters of the compression
function, pulse width and pulse duration of modulating pulses, sequence of channel
stimulation [8].

   Various speech processing strategies have been developed and reported in literature over
time for cochlear prostheses which include Continuous Interleaved Sampling (CIS),
Spectral Peak (SPEAK), Advanced Combination Encoder (ACE), Spectral Maxima Sound
Processor (SMSP), Simultaneous Analog Strategy (SAS) and various formant based
strategies [9,10]. Most implant manufacturers provide research speech processors for use
on human subjects that allow researchers to develop and test new signal processing
algorithms [11]. Much research has therefore been directed at improving the performance
and major successes have been achieved since they were first implemented in the 1970s.
The number of implanted electrodes has grown from 1 to as many as 24 in current implants
[12]. However, current cochlear implants are not very successful in delivering phase
information that might be responsible for the reduced ability of CI users to perceive music.

   Spectral enhancement is another major concern that has benefits of improving
performance in noise. It is possible that this is one of the reasons for why the human
auditory system with a biological cochlea front end is known to have good performance in
noise[13]. Spectral enhancement is a natural consequence of our trying to mimic the
operation of the biological cochlea. Also, the effect of electrode interactions and current
spreading limits the spectral resolution to a great degree. A common speech processing
strategy used in implants and in speech-recognition systems, employs a mel cepstrum filter
bank. In implants, the number of functioning electrodes used for stimulation is between 8-
24 and having more electrodes is often not useful due to spatial interactions amongst the
electrodes [14].

   Our work proposes an acoustic simulation model for cochlear implants that uses a 16-
channel CB-FIR filter bank. Spectral enhancement and audibility could be improved by
using a set of independent filters with independent and easily programmable parameters.
Input speech signals are filtered using this filter bank, then rectified, low-pass filtered and
the amplitude levels are estimated. Spectral energies are computed for various speech
components such as phonemes from each channel. The most significant channels, which
are few in number, constituting maximum signal energy are selected thus aiding to
minimize channel interactions. These channel energies are used in estimating the
stimulation pulse level to be presented to the selected cochlear electrodes.

                                             117
International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN
0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME

   The outline of the paper is as follows. In Section 2, we describe the essential ideas behind
the choice of filter bank strategy, the improved design aspects and how various speech
components are assessed for its speech intelligibility. In Section 3, we present Matlab
simulation data for the designed filter bank system with speech input that illustrate the
working and benefits of the system. In Section 4, we illustrate and discuss the results of our
proposed cochlear implant acoustic simulation model through the spectral and temporal
analysis performed for various speech components and paving a way for the electrical
stimulation procedure for cochlear implants. In Section 5, we conclude by summarizing our
contributions.

   2. MATERIALS AND METHODS
2.1 Digital speech processing
  The greatest common block of all recognition systems is the signal-processing front end,
which converts the speech waveform to some type of parametric representation for further
analysis and processing. Parametrical representation of speech signal uses short-time
energy and zero crossing rates [15]. Speech signals are more consistently analyzed
spectrally than in the time domain and human hearing pays more attention to spectral
aspects of speech [16]. Filter bank representation is one such spectral analysis method that
uses a set of band pass filters, each analyzing a different range of frequencies of the input
speech. There are two main motivations for the filter bank representation. First, the position
of maximum displacement along the basilar membrane for stimuli such as pure tones is
proportional to the logarithm of the frequency of the tone. This hypothesis is part of a
theory of hearing called the ‘place’ theory. Second, the experiments in human perception
have shown that frequencies of a complex sound within a certain bandwidth of some
nominal frequency cannot be individually identified. When one of the components of this
sound falls outside this bandwidth, it can be individually distinguished. This bandwidth is
referred to as the critical bandwidth [17].

  The digital speech processing proposed for ‘cochlear implant - acoustic simulation
model’ is built around a critical band filter bank followed by rectifiers and low pass filters.
The cascade of a rectifier and a low pass filter yields an envelope detector. Input sound
signals are filtered using a digital band-pass FIR (Finite Impulse Response) filter bank,
with one filter for each electrode of the implant. The number of filter bands depends on the
number of stimulation channels to be considered in the implant of cochlear prostheses
system [18]. The filter bank was designed to have linear frequency spacing at low center
frequencies and logarithmic spacing at high center frequencies in correspondence to the
properties of cochlear filters [19]. In accordance with Rekha et al., [20], the human speech
frequency, approximately 100 to 3500 Hz, is allocated to filter bands using critical band
scale. This frequency range is covered in 16 bands, which is shown in Table 1. In the
signal analysis application, filter bank model with non-uniform spacing has the potential to
offer better performance [21].

  The general specifications of the filter bank include the stop band attenuation (e.g., 40-
50dB), number of channels (e.g., 16) and the pass band spacing. For the advanced cochlear
devices, we specify the need for a linear phase finite-impulse response (FIR) filter bank as
opposed to the nonlinear phase infinite-impulse response (IIR) filter bank for the following
reasons [22].
     • FIR filters are intrinsically stable
     • They can easily realize various features that are not possible or are difficult to
       achieve with IIR filters
     • To eliminate channel interactions.




                                             118
International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN
0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME
To achieve more stop-band attenuation, a higher filter order is required with consequent
cost in CPU usage. We impose higher stop band attenuation ( –60dB) compared to –50dB
achieved by the current devices.

Table 1. Critical bandwidth as a function of center frequency: Non-uniform, non-overlapping band-
pass filter cutoff frequencies that correspond to cochlea filter properties are listed here.

                             Channel        Center         Frequency
                             Number        Frequency          Band
                                              (Hz)            (Hz)
                                 1             150          125-175
                                 2             250          225-275
                                 3             350          325-375
                                 4             450          420-480
                                 5             570          530-605
                                 6             700          655-745
                                 7             840          790-890
                                 8           1000           940-1060
                                 9           1170          1105-1235
                                 10          1370          1285-1455
                                 11          1600          1505-1695
                                 12          1850          1745-1955
                                 13          2150          2005-2295
                                 14          2500          2345-2655
                                 15          2900          2705-3095
                                 16          3400          3145-3655
2.2 FIR filter design using kaiser window
  The problem of designing an FIR filter consists of determining the impulse response
sequence, h[n], or its system function H(z), so that given requirements on the filter
response are satisfied. The filter requirements are usually specified in the frequency
domain, and only this case is considered here [23].              The frequency response
       jω
 H d (e ) corresponding to the impulse response h[n] is expressed in a fourier series as
                 ∞
  H d (e jω ) = ∑ hd (n) e− jωn
               n=−∞
                                                     (1)
and known as fourier coefficients having infinite length. FIR filter is obtained by truncating
the infinite fourier series at n = ± 
                                       N −1 
                                           ,   where N is the length of the desired sequence.
                                       2 
Abrupt truncation of the fourier series results in oscillations in pass-band and stop-band
regions of the filter. An optimal way to reduce these oscillations is to use an appropriate
finite-length window w(n) , which controls the overall filter to yield a smooth frequency
response. Therefore, a window should possess some of the following spectral
characteristics [24,25].
(i)The main lobe width in the frequency response of the window should be narrow [17]
(ii) The ripple ratio should be small
(iii)The side-lobes should decrease in energy rapidly as ω tends to π.
(iv)To adequately block the stop-band frequencies, it is necessary to have good stop-


                                                  119
International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN
0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME
    band attenuation.

  The symmetrical frequency spectrum of an odd length window w(nT ) can be written as
[25],
                                 ( N −1) 2
       W (e jωT ) = w(0) + 2             ∑      w(nT ) Cos( wnT )
                                         n =1                               (2)
Where T = 2π ω s is the sampling period.

   Standard windows such as Rectangular, Hamming and Hanning have only one
independent parameter namely, the ‘window length’ N , which controls the ‘transition
width’ of the filter. An increase in window length decreases the transition width of the
filter. But the minimum stop band attenuation of the filter is independent of N . In order to
overcome this, a Kaiser window is used which has an adjustable parameter β that controls
the stop-band attenuation and an independent parameter N , that controls the transition
width [26]. Kaiser window is widely used for the spectral analysis and FIR filter design
applications.

The design of band pass filter using Kaiser Window has been carried out based on the
following steps [24,26,27].
Step 1:
  The length ‘N’ of the band pass filter is computed using Eq (3).
     −20log10 ( δ pδ s ) −13
N ≅
        14.6(ωs − ω p ) 2π
                                                       (3)
where, δ p and δs are the pass band and stop band ripples respectively and ω p and ωs are
the pass band and stop band edge frequencies respectively.
The design is based on the narrower of the two transition bandwidths ( Bt ) , that is
 Bt = min[(ω p − ωs ), (ωs − ω p )], where (ωs ,ωp ) and (ωp ,ωs ) are the initial and later
                1       1            2          2                   1   1         2     2

transition edge frequencies of the band pass filter. The cut-off frequencies are given by
ωc1 = ω p1 − Bt 2 and ωc2 = ω p2 + Bt 2.

Step 2:
The Kaiser window sequence is given by
           
w [n] = I0 
            (
            β 1 − [2 n / ( N − 1) ]2
                                           )
                                      , − ( N − 1) / 2 ≤ n ≤ ( N − 1) / 2
                   I0 ( β )
                                                            (4)
where β is an adjustable parameter, and I0(x) is the modified zeroth order Bessel function,
which can be expanded into the following series.
                ∞  x/2 r        2
                      ( )    
I 0 ( x) = 1 +  ∑   
                     r!
                             
                             
               r = 1                                                                (5)
which is seen to be positive for all real values of x.




                                                                120
International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN
0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME
Step 3:
The desired minimum stop band attenuation is calculated by α s = - 20 log10 δ s . β         is
computed from Eq.(6).

     0.1102 α − 8.7 ,
               (            )                for α s > 50,
             s
                    0.4
               (           )                    (         )
β =  0.5842 α s − 21 + 0.007886 α s − 21 , for 21 ≤ α s ≤ 50,
    
     0,                                   for α s < 21.
                                                                             (6)

Step 4:
 The coefficients of linear phase filter with delay γ =  N −1 are computed using
                                                        
                                                              
                                                                       2 
                        ωc − ωc 
                        2      1 
                                     ,                    for n = 0
         
                           π     
hd (n) =                                                                   (7)
          [ Sin ωc    (n − γ ) − Sin ωc (n − γ ) ]
                  2                       1
                                                      ,   for n ≠ 0
         
                          π (n − γ )
Step 5:
  The impulse response h(n) is obtained by multiplying the desired impulse response hd(n)
of the ideal filter and the window coefficients generated in Step 2 to yield the coefficients
of FIR filter.
h( n ) = hd ( n) .w( n)                                        (8)
Step 6:
 Filter specifications are usually given in terms of the magnitude response |H(ejω)|. The
magnitude function of FIR filter when the impulse response is symmetric and when N is
odd is generated by
                            N −1 
                                 
                            2 
            N −1                        N −1     
H (ω ) = h 
            2 
                 +         ∑         2h 
                                           2
                                                − n  Cos (ω n)
                                                    
                                                                              (9)
                             n =1
Filters designed with this window had the level of stop-band approximately equal to         –
60dB.
2.3 CI-Acoustic Simulation Model
  The schematic of the proposed CI acoustic simulation model is shown in Figure 1. The
input speech signal is passed through a bank of 16 band pass filters, whose coverage spans
the human speech frequency range. The spectral output is obtained by the summation of
independent band pass filters responses. Spectral energies are estimated for each of 16
channels. Temporal analysis is done by full-wave rectification and low pass filtering the
decomposed signal in each band. The stimulus amplitude levels are estimated through
envelope detection procedure.




                                                              121
International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN
0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME




  Fig.1.Schematic of the proposed CI-acoustic simulation model: Spectral energies are estimated by
a bank of band pass filters, followed by temporal estimation by rectification and envelope detection,
determining the place and stimulus level information.

2.3.1 Rectification
  Speech signals are complex patterns that do not permit direct analysis of their wave
shape. The envelope of the overall signal carries important information. The envelope of a
signal can be derived by obtaining the absolute value of the signal at each time instant, i.e.,
performing full-wave rectification [28]. A FIR low pass filter of order 100 and a cut-off
frequency of 400 Hz are used to obtain smooth envelopes of speech signals [29].
2.3.2 Envelope detection
  Envelope detection is done using a mathematical method called ‘Linear Interpolation’.
This technique is used to detect the envelope of the low pass filter output. In order to obtain
smooth envelope of the low pass filter output, the estimation of values for intermediate
time instances are carried out. A point in the low pass filter output is interpolated using Eq
(9).
           ( y2 − y1 )( x − x1 )
y = y1 +
                ( x2 − x1 )
                                                               (10)
where x1, x2 are the two time instances and y1, y2 are their corresponding amplitude values
estimated from the output of low pass filter. If these quantities are known, then for a
desired intermediate time instant x, a new value y can be obtained. Matlab code is
developed for interpolation and stimulus amplitude estimation.
2.4 Speech sounds and features
  The perceptually important information of a speech signal needs to be preserved in order
to assess its intelligibility. Speech sounds called phonemes are produced depending on the
state of input excitation (vocal cords). Voiced speech (e.g., vowels) is produced if the vocal
cords are tensed and vibrate periodically. Unvoiced speech (e.g., consonants) is produced if
the vocal cords are not vibrating and the resulting waveform is aperiodic or random [15].
Based on the position of the tongue, vowel articulatory configurations are produced and
classified as front vowel /i/, mid vowel /a/ and back vowel /u/. A ‘diphthong’ is a complex
speech sound that begins with one vowel and gradually changes to another vowel within
the same syllable (egs., /ay/, /oy/, /aw/, /ey/). Voiced (/b/, /d/, /g/) and unvoiced stop
consonants (/p/, /t/, /k/) are produced by the complete stoppage and sudden release of the
breath. Nasal consonants (/m/, /n/) are produced with glottal excitation and the vocal tract
totally constricted. In the design of speech analysis system, how the spectral and temporal

                                                122
International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN
0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME
properties of a speech signal affect its perception need to be studied [30]. In this paper, we
have estimated the spectral energy to separate the voiced and unvoiced parts of speech and
also to determine the stimulus pulse level to be given to the cochlear electrodes, using the
developed ‘CI acoustic simulation model’ discussed in the earlier section. Table II lists the
classification of phonemes into sound classes taken for the study.

Table 2. Classification of the standard phonemes into sound classes: Three vowels: front, mid, back
corresponding to the position of the tongue hump; four diphthongs, two nasal consonants; three
voiced and unvoiced stop consonants are listed here.
                  S.No    Type of sound               Phonetic symbol
                  1       Vowels                      /a/, /i/, /u/
                  2       Diphthongs                  /aw /, /oy/, /ay/, /ey/
                  3       Nasal consonants            /m/, /n/
                  4       Voiced consonants           /b/, /d/, /g/
                  5       Unvoiced consonants         /k/, /t/, /p/

2.5 Speech data
  The speech data used for the current study is obtained from TIMIT database which
consists of phonemes, sampled at 16 KHz. In a total of 48 standard phonemes, 15
phonemes as listed in Table 2, are used for the analysis.

     3. EXPERIMENT
  The function of signal processor in CI devices primarily consists of filter bank model that
divides an input speech signal into a number of frequency bands in order to extract signal
strength in each band for exciting the implanted electrodes accordingly. As per the
frequency bands shown in Table 1, the input speech signals are digitally filtered using
critical band FIR filter based on the specifications discussed in section 2.2. MATLAB code
is written to generate filter coefficients by using Kaiser Window. We adopt the linear-
phase FIR filter and use its symmetrical property to reduce the number of multiplications
by half [24]. Figure 2 depicts the magnitude response of 16 channels of the filter bank;
each channel is an 877th-order FIR filter sampled at 16KHz. The stop-band attenuation
obtained is approximately –60dB for a β value of 6.

  A CI acoustic simulation model is constructed and simulation experiment is conducted to
investigate its contribution for speech perception. All the test materials are analyzed for its
dominant energy levels between 125Hz and 3655 Hz. Each utterance of speech signal for
the vowels /a/, /i/, /u/, diphthongs /ay/, /oy/, /aw/, /ey/, voiced /b/, /d/, /g/, unvoiced stop
consonants /p/, /t/, /k/, and nasal consonants /m/, /n/, are passed through the 16
independent channels. Spectral energies are estimated from each band and the amplitude
levels are estimated through envelope detection.




                                               123
International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN
0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME




Fig. 2. Magnitude response of the 16 channel band pass filter: Each channel exhibiting –
60dB stop band attenuation.

4. RESULTS AND DISCUSSION
   Using the acoustic model discussed in section 2.3, it is proposed that a reduced number
of channels is selected for stimulation based on the speech energy obtained in different
channels; in order to avoid inter-channel interference among the cochlear electrodes. The
experiment is performed for all the phonemes listed in section 2.5. To select the appropriate
stimulation electrodes, the spectral energies of the model output are estimated for each of
16 channels, where each phoneme is analyzed for 10 utterances. The following sections
discuss the results obtained through spectral and temporal analysis based on the CI acoustic
simulation model.
4.1 Vowels
  Here an illustration is discussed for selecting the stimulation electrodes based on the
channel energies. Typical energy levels of vowel ‘a’ for ten utterances are represented in
Table 3. The sixteen rows represent the sixteen channels’ energy and columns U1 through
U10 represent utterance of each subject. The energy values are normalized across sixteen
channels for each utterance. It is inferred from Table 3 that, nine utterances show
maximum energy (energy > 0.9) in five and six channels. This shows that maximum speech
information is present in the frequency range of channels 5 and 6 for vowel /a/. If we form
a criterion that only channels having normalized energy level > 0.9 are used for
stimulation, then for vowel ‘a’, only electrodes 5 and 6 are to be stimulated. This
corresponds to a frequency range of 530Hz-605Hz and 655Hz-745Hz. All other channel
outputs that possess less than 0.9 of the normalized energy value may not contain much of
speech information and these corresponding electrodes need not be stimulated.
  The output speech spectra of 16 channel BPF for vowel /a/ for utterance 1 are illustrated
in Figure 3a. Two dominant peaks are observed at center frequencies 570Hz and 700Hz,
which correspond to channels 5 and 6. The experiment is performed for the remaining
vowels /i/ and /u/. The front vowel /i/ showed > 0.9 of the normalized energy in channels
13 and 14, which correspond to the center frequencies 2150Hz and 2500Hz. For back
vowel /u/, channels 4 and 11 which correspond to 450Hz and 1600Hz were identified.
These significant channels only would be used for the excitation of electrodes.




                                             124
International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN
0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME


Table 3. Maximum energy levels corresponding to the utterances of vowel /a/: Distribution of speech
energy in 16 channels for ten utterances.

 Channel   U1       U2       U3       U4       U5       U6       U7       U8       U9       U10
 1         1.5      1.3      0.2      0.2      1.7      0.6      1.2      0.6      0.1      4.4
 2         1.1      2.5      0.2      1.1      4.9      2.9      1.4      4.4      2.1      7
 3         1.9      0.4      1.2      1.5      1.9      2.4      0.2      7.2      2.9      4.2
 4         1.5      1        6.9      6.2      1.4      3.9      5.5      10.3     3.8      8.5
 5         10.2     3.1      7.6      29       6.9      9.9      24.8     20.6     16.5     26
 6         12.2     5.9      7.5      17.3     9.4      14.2     5.6      42.5     25.1     30
 7         5.1      3        2        5.1      5.5      15.5     2.5      14.4     14       6.8
 8         2.9      3.1      2.4      4.1      8.8      13.9     2.6      9.1      26.1     5.5
 9         3.5      3.4      3.2      5.2      13.1     4.5      1.4      9        20.9     6
 10        6.4      1.4      8.8      18.3     12.9     1.8      2.9      32.8     4.6      19
 11        4.9      0.2      8.3      5        7.1      1.2      12.9     25.6     2.6      9.2
 12        2.1      0.1      2.5      3.2      4.9      1        6.6      10.3     3.3      4.1
 13        2.1      0.4      2.7      6.8      7.2      3        8.6      19.2     6.4      7.3
 14        4.3      1.1      6        1.7      5.5      1.8      5.1      13.6     1.3      10.2
 15        2.2      0.6      2.6      1.1      5.2      1.4      2.5      5.8      1.8      3.9
 16        0.9      1.4      4.7      3.4      5.4      3.1      9.2      10.9     1.5      8.1


4.2 Diphthongs
 The experiment is performed for four diphthongs and the observations are made as follows.
For /aw/ and /oy/, two channels 5 and 7 are found to have normalized energy level > 0.9.
These channels only need to be stimulated for diphthong /aw/ and /oy/. The output speech
spectra for /aw/ are illustrated in Figure 3b which shows only two peaks at center frequencies
570Hz and 840Hz that correspond to channels 5 and 7 respectively. Similarly, for /ay/, the
normalized energy level > 0.9 is observed in channel 6 corresponding to a center frequency of
700Hz. For /ey/, it is channel 5, which correspond to a center frequency of 570Hz.
4.3 Nasal consonants
 For nasal consonants /m/ and /n/, channel 2 is found to have normalized energy level >
0.9. Therefore channel 2 needs to be stimulated for nasal consonants /m/ and /n/. The BPF
output speech spectra of nasal consonant /m/ are shown in Figure 3c. A peak is observed in
channel 2 which correspond to a frequency range of 225Hz-275Hz.
4.4 Voiced and unvoiced stop consonants
  The energy levels vary greatly with the frequency content in both voiced and unvoiced stop
consonants. The experimental procedure is repeated for voiced (/b/, /d/, /g/) and unvoiced (/k/,
/t/, /p/) stop consonants. The results of the significant energy levels are indicated across
channels 5, 11 and 16 for /b/, channel 16 for /d/ and channels 10, 11 for /g/. For /k/, it is
observed as channels 10 and 11; for /t/, it is channels 15 and 16 and for /p/, it is channels 10
and 14. The BPF output speech spectra of unvoiced stop consonant /t/ are shown in Figure 3d,
which shows peaks at channels 15 and 16 that correspond to center frequencies 2900Hz and
3400Hz.




                                              125
International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN
0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME




Fig. 3. Spectral analysis outputs showing the energy peaks for various phonemes (a) Two dominant
peaks are observed in channels 5 and 6 for vowel /a/, for utterance 1 (b) Diphthong /aw/ exhibiting
channels 5 and 7 (c) A peak in channel 2 for nasal consonant /m/ (d) A peak in channel 15 for
unvoiced stop consonant /t/.

  From the CI acoustic simulation model, it is evident that, low frequencies are significant for
vowels and high frequencies are significant for consonants. The events of the consonants are
consistent across different subjects, despite that the parameters such as frequency and strength
may change to a certain degree. The investigations further demonstrated that, a maximum of
three channels are only required for the stimulation of cochlear electrodes for each of the
phonemes listed in Table 2. Hence it is inferred that the spectral energies in all 16 channels
are not required to be utilized for stimulation of cochlear electrodes. By selecting a few
channels for stimulation, our proposed model would help in reducing inter channel
interactions among the electrodes. The computation of spectral energies gives a measure of
the number of electrodes required for stimulation. The identified channels through spectral
analysis are subjected to temporal analysis, in order to get the amplitude information.
4.5 Electrical stimulation procedure
  The electrodes in the cochlear array are stimulated by means of biphasic pulses [29]. A
biphasic pulse of 15µs width with a repetition rate of 250 µs which produces a stimulation
rate of 4000 pulses per second are generated. The ‘place’ of stimulation in the cochlear
electrode array is obtained by band energy levels in various channels estimated from the
spectral analysis using the CI acoustic simulation model as discussed in section 2.2. The
selected channels whose normalized energy >0.9 are subjected to temporal analysis. The
amplitudes of the pulses are derived by extracting the envelopes of band passed filter outputs
which are obtained by full-wave rectification and low-pass filtering as discussed in section
2.3. The amplitude estimates are useful in fixing the amplitude of the pulses. i.e, the pulse
amplitudes are made proportional to the amplitude of the envelope output. Trains of biphasic
pulses thus generated are to be delivered to the cochlear electrodes. In our study, the pulses
are successfully generated for all the phonemes listed in Table 2, with amplitudes proportional
to the outputs of selected channels.
  An illustration for vowel /a/ depicting the various stages of temporal outputs in channel 6
(having normalized energy >0.9) is shown in Figure 4(a-d), which correspond to the input

                                               126
International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN
0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME


speech signal, band pass filtered output, rectified and low-pass filtered output, and the
extracted envelope output. From this extracted envelope output, the stimulus pulse amplitude
estimates are derived. Figure 4e shows the stimulation pulse energy levels for vowel /a/ in
channel 6.




Fig. 4. Various stages of temporal outputs in channel 6 for vowel /a/.
(a) Input speech signal of vowel /a/ (b) Band pass filtered output in channel 6 (c) Illustration of the
rectified and low-pass filtered outputs (d) the envelope tracing of the low pass filter (e) Amplitude
level of biphasic stimulation pulses (not expanded).

5.   CONCLUSION
  In this study, the performance of cochlear implant acoustic simulation model is discussed.
We have designed a 16-channel FIR filter bank based on critical band spacing, using Kaiser
Window. Higher stop band attenuation and exhibiting linear phase property are the two
advantages of the critical band FIR filter design, which is utilized in the proposed CI-acoustic
simulation model. This model estimates the spectral output using the designed CB-FIR filter
bank and temporal output by envelope detection method through which the ‘place’ and
‘stimulus level’ information are obtained respectively. The spectral band energies of 15
speech components from the TIMIT data base are analyzed and the most significant channels
are identified and selected, by computing the maximum energy values for the phonemes taken
for the study. The selected channels are subjected to temporal analysis and the amplitudes are
extracted by the envelope detection procedure. The simulation results of the selected channels


                                                 127
International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN
0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME


for different phonemes, demonstrated a good discrimination between vowels, diphthongs,
voiced and unvoiced consonants. The other advantage of the proposed model is, it yields a
way for minimizing the channel interactions as a reduced number of electrodes only are
selected for stimulation. The overall results from the acoustic simulation experiment showed
that the CB-FIR filter bank model based speech processor is useful in selecting the channels
and estimating the required stimulation pulse level to be generated and presented to the
cochlear electrodes.

 REFERENCES
 [1] Philipos C. Loizou (1999), “Introduction to Cochlear Implants”, IEEE Engineering in
      Medicine and Biology, pp. 32-42.
 [2] Kyung Hwan Kim, Sung Jen Choi, Jin Ho Kim, Doo Hee Kim (2009), “An Improved
      Speech Processing Strategy for Cochlear Implants based on an Active Non linear
      Filter Bank Model of the Biological Cochlea”, IEEE Transactions on biomedical
      engineering, 56(3): 828-836.
 [3] Lan Tian, Zhengxin Hou, Zhimim Pang (2007), “The Application and Computer
      Simulation of Multi-Channel Cochlear Implant based on all Phase DFT Filter”,
      ICASSP, Proceedings of the IEEE (IV), pp.701-704.
 [4] Francis A. Spelman (1999), “The Past, Present and Future of Cochlear Prostheses:
      Accomplishments and Challenges in treating Sensori-neural Deafness through
      Electrical Stimulation”, IEEE Engineering in Medicine and Biology, pp. 27-33.
 [5] Waldo Nogueira, Andras Katai, Tamas Harczos, Frank Klefenz, Andreas Buechner,
      Bernd Edler (2007), “An Auditory Model Based Strategy for Cochlear Implants”,
      Proceedings of the IEEE EMBS Conference, pp. 4127-4130.
 [6] Chris J. James, Peter J.Blamey, Lois Martin, Brett Swanson, Yvette Just, David
      Macfarlane (2002), “Adaptive Dynamic Range Optimization for Cochlear Implants: A
      Preliminary Study”, Ear and Hearing, Vol.23, pp. 49S-58S.
 [7] Fan-Gang Zeng, Stephen Rebscher, William Harrison, Xiaoan Sun, Haihong Feng
      (2008), “Cochlear Implants: System Design, Integration and Evaluation, IEEE
      Reviews in Biomedical Engineering”, Vol.1, pp. 115-142.
 [8] Eric D.Marsman, Gordon A.Carichner, Sundus Kubba, Richard B. Brown (2006), “A
      DSP Architecture of Cochlear Implants”, ISCAS, Proceedings of the IEEE, pp. 657-
      666.
 [9] Talha J. Ahmad, Hussnain Ali, Muhammed Asim Ajaz and Shoab A.Khan (2009),
      “Efficient Algorithm Development of CIS Speech Processing Strategy for Cochlear
      Implants”, Proceedings of the IEEE EMBS Conference, pp. 1270-1273.
[10] Talha Jamal Ahmad, Hussnain Ali, Muhammed Asim Ajaz, Shoab A.Khan (2009), “A
      DSK based Simplified Speech Processing Module for Cochlear Implant Research”,
      ICASSP, Proceedings of the IEEE, pp. 369-372.
[11] Douglas Kim, Vanishree Gopalakrishna, Song Guo, Hoi Lee, Murat Torlak, Nasser
      Kehtarnavaz, Arthur lobo, Philipos C. Loizou (2011), “On the Design of a Flexible
      Stimulator for Animal Studies in Auditory Prostheses”, Biomedical Signal Processing
      and Control, Vol.6, pp. 346-355.
[12] Ji-Jon Sit, Andrea M.Simonson, Andrew J.Oxenham, Michael A.Faltys, Rahul
      Sarpeshkar (2007), “A Low-Power Asynchronous Interleaved Sampling Algorithm for
      Cochlear Implants that Encodes Envelope and Phase Information”, IEEE Transactions
      on Biomedical Engineering, 54(1): 138-149.
[13] L.Turicchia and R.Sarpeshkar (2005), “A Bio-Inspired Companding Strategy for
      Spectral Enhancement”, IEEE Transactions on Speech and Audioprocessing, 13(2):
      243-253.
[14] Rahul Sarpeshkar, Christopher Salthouse, Ji-Jon Sit, Michael W.Baker, Serhii M.Zark,
      Timothy K. T.Lorenzo Turicchia, Stephanie Balster (2005), “An Ultra-Low-Power

                                             128
International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN
0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME


       Programmable Analog Bionic Ear Processor”, IEEE transactions on biomedical
       engineering, 52(4): 711-727.
[15]   Lawrence Rabiner (2009),          Biing – Hwang Juang,          B.  Yegnanarayana,
       “Fundamentals of Speech Recognition”, Pearson Education, Dorling Kindersley India.
[16]   Douglas O’Shaughnessy (2001), “Speech Communications: Human and Machine”,
       Second Edition, Universities Press, Hyderabad.
[17]   Joseph W. Picone (1993), “Signal Modeling Techniques in Speech Recognition”,
       Proceedings of the IEEE, Vol.81, No.9, pp. 1215-1247.
[18]   Ahmed Ben Hamida (1999), “An Adjustable Filter-Bank Based Algorithm for Hearing
       Aid Systems”, Proceedings of the IEEE, pp. 1187-1192.
[19]   Han Xian Hua, Nie Kaibao (2001), “Implementation of Spectral Maxima Sound
       Processing for Cochlear Implants by Using Bark Scale Frequency Band Partition”,
       Proceedings of the EMBS, pp. 2097-2101.
[20]   Rekha V. Dundur, M.V. Latte, S.Y. Kulkarni, M.K. Venkatesha (2008), “Digital Filter
       for Cochlear Implant Implemented on a Field-Programmable Gate Array”,
       Proceedings of World Academy of Science, Engineering and Technology, Vol.33, pp.
       468-472.
[21]   S.S.Yin, S.C.Chan, X.M.Xie (2006), “On the Theory and Design of a Class of
       Recombination Non-uniform Filter Banks with Low-Delay FIR and IIR Filters”,
       ISCAS, Proceedings of the IEEE, pp. 895-898.
[22]   Yutaka Yamamoto, Brian D.O.Anderson, Masaaki Nagahara, Yoko Koyanagi (2003),
       “Optimizing FIR Approximation for Discrete-time IIR filters”, IEEE Signal
       Processing Letters, 10(9): 273-276.
[23]   A.Enis Cetin, Omer N.Gerek, Yasemin Yardimci (1997), “Equiripple FIR Filter
       Design by the FFT Algorithm”, IEEE Signal Processing Magazine, pp. 60-64.
[24]   Sanjit K.Mitra (2008), “Digital Signal Processing:A Computer-based approach”, Third
       Edition, Tata McGraw-Hill Publishing Company Limited, Newyork.
[25]   Kemal Avci, Arif Nacaroglu (2009), “Cosh Window Family and its Application to
       FIR Filter Design”, Int. J.Electron. Commun (AEU), Vol.63,pp. 907-916.
[26]   Tapio Saramaki (1989), “A Class of Window Functions with Nearly Minimum
       Sidelobe Energy for Designing FIR Filters”, Proceedings of the ISCAS, pp. 359-362.
[27]   P. Mahalakshmi, M R Reddy (2010), “Signal Analysis by Using FIR Filter Banks in
       Cochlear Implant Prostheses”, Proceedings of the IEEE International Conference on
       Systems in Medicine and Biology, No.89, pp. 253-258.
[28]   Rangaraj M. Rangayyan (2002), “Biomedical Signal Analysis: A Case Study
       Approach”, John Wiley & Sons Inc., Newyork.
[29]   Philipos C. Loizou (1999), “Signal Processing Techniques for Cochlear Implants: A
       Review of Progress in Deriving Electrical Stimuli From the Speech Signals”, IEEE
       Engineering in Medicine and Biology, pp. 34-45.
[30]   Thomas F. Quatieri (2004), “Discrete-time Speech Signal Processing: Principles and
       Practice”, Pearson Education, India.




                                             129

Más contenido relacionado

La actualidad más candente

Malayalam Isolated Digit Recognition using HMM and PLP cepstral coefficient
Malayalam Isolated Digit Recognition using HMM and PLP cepstral coefficientMalayalam Isolated Digit Recognition using HMM and PLP cepstral coefficient
Malayalam Isolated Digit Recognition using HMM and PLP cepstral coefficientijait
 
Speech coding strategies in CI
Speech coding strategies in CISpeech coding strategies in CI
Speech coding strategies in CIDr.Ebtessam Nada
 
Utterance Based Speaker Identification Using ANN
Utterance Based Speaker Identification Using ANNUtterance Based Speaker Identification Using ANN
Utterance Based Speaker Identification Using ANNIJCSEA Journal
 
Utterance Based Speaker Identification Using ANN
Utterance Based Speaker Identification Using ANNUtterance Based Speaker Identification Using ANN
Utterance Based Speaker Identification Using ANNIJCSEA Journal
 
CI mapping troubleshooting and problem solving
CI mapping troubleshooting and problem solvingCI mapping troubleshooting and problem solving
CI mapping troubleshooting and problem solvingDr.Ebtessam Nada
 
01 8445 speech enhancement
01 8445 speech enhancement01 8445 speech enhancement
01 8445 speech enhancementIAESIJEECS
 
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...IJERD Editor
 
IRJET- Survey on Efficient Signal Processing Techniques for Speech Enhancement
IRJET- Survey on Efficient Signal Processing Techniques for Speech EnhancementIRJET- Survey on Efficient Signal Processing Techniques for Speech Enhancement
IRJET- Survey on Efficient Signal Processing Techniques for Speech EnhancementIRJET Journal
 
Dynamic Audio-Visual Client Recognition modelling
Dynamic Audio-Visual Client Recognition modellingDynamic Audio-Visual Client Recognition modelling
Dynamic Audio-Visual Client Recognition modellingCSCJournals
 
IRJET- Performance Comparison of Cognitive Radio Network by Spectrum Sensing ...
IRJET- Performance Comparison of Cognitive Radio Network by Spectrum Sensing ...IRJET- Performance Comparison of Cognitive Radio Network by Spectrum Sensing ...
IRJET- Performance Comparison of Cognitive Radio Network by Spectrum Sensing ...IRJET Journal
 
Estimating the quality of digitally transmitted speech over satellite communi...
Estimating the quality of digitally transmitted speech over satellite communi...Estimating the quality of digitally transmitted speech over satellite communi...
Estimating the quality of digitally transmitted speech over satellite communi...Alexander Decker
 
LPC Models and Different Speech Enhancement Techniques- A Review
LPC Models and Different Speech Enhancement Techniques- A ReviewLPC Models and Different Speech Enhancement Techniques- A Review
LPC Models and Different Speech Enhancement Techniques- A Reviewijiert bestjournal
 
Optimized BER for channel equalizer using cuckoo search and neural network
Optimized BER for channel equalizer using cuckoo search and neural network Optimized BER for channel equalizer using cuckoo search and neural network
Optimized BER for channel equalizer using cuckoo search and neural network IJECEIAES
 
HRTF Project Report
HRTF Project ReportHRTF Project Report
HRTF Project ReportManu Mitra
 

La actualidad más candente (16)

Malayalam Isolated Digit Recognition using HMM and PLP cepstral coefficient
Malayalam Isolated Digit Recognition using HMM and PLP cepstral coefficientMalayalam Isolated Digit Recognition using HMM and PLP cepstral coefficient
Malayalam Isolated Digit Recognition using HMM and PLP cepstral coefficient
 
Speech coding strategies in CI
Speech coding strategies in CISpeech coding strategies in CI
Speech coding strategies in CI
 
Utterance Based Speaker Identification Using ANN
Utterance Based Speaker Identification Using ANNUtterance Based Speaker Identification Using ANN
Utterance Based Speaker Identification Using ANN
 
Utterance Based Speaker Identification Using ANN
Utterance Based Speaker Identification Using ANNUtterance Based Speaker Identification Using ANN
Utterance Based Speaker Identification Using ANN
 
CI mapping troubleshooting and problem solving
CI mapping troubleshooting and problem solvingCI mapping troubleshooting and problem solving
CI mapping troubleshooting and problem solving
 
19 ijcse-01227
19 ijcse-0122719 ijcse-01227
19 ijcse-01227
 
01 8445 speech enhancement
01 8445 speech enhancement01 8445 speech enhancement
01 8445 speech enhancement
 
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
 
IRJET- Survey on Efficient Signal Processing Techniques for Speech Enhancement
IRJET- Survey on Efficient Signal Processing Techniques for Speech EnhancementIRJET- Survey on Efficient Signal Processing Techniques for Speech Enhancement
IRJET- Survey on Efficient Signal Processing Techniques for Speech Enhancement
 
Dy36749754
Dy36749754Dy36749754
Dy36749754
 
Dynamic Audio-Visual Client Recognition modelling
Dynamic Audio-Visual Client Recognition modellingDynamic Audio-Visual Client Recognition modelling
Dynamic Audio-Visual Client Recognition modelling
 
IRJET- Performance Comparison of Cognitive Radio Network by Spectrum Sensing ...
IRJET- Performance Comparison of Cognitive Radio Network by Spectrum Sensing ...IRJET- Performance Comparison of Cognitive Radio Network by Spectrum Sensing ...
IRJET- Performance Comparison of Cognitive Radio Network by Spectrum Sensing ...
 
Estimating the quality of digitally transmitted speech over satellite communi...
Estimating the quality of digitally transmitted speech over satellite communi...Estimating the quality of digitally transmitted speech over satellite communi...
Estimating the quality of digitally transmitted speech over satellite communi...
 
LPC Models and Different Speech Enhancement Techniques- A Review
LPC Models and Different Speech Enhancement Techniques- A ReviewLPC Models and Different Speech Enhancement Techniques- A Review
LPC Models and Different Speech Enhancement Techniques- A Review
 
Optimized BER for channel equalizer using cuckoo search and neural network
Optimized BER for channel equalizer using cuckoo search and neural network Optimized BER for channel equalizer using cuckoo search and neural network
Optimized BER for channel equalizer using cuckoo search and neural network
 
HRTF Project Report
HRTF Project ReportHRTF Project Report
HRTF Project Report
 

Destacado

C:\users\nachitaypatty\documents\nacha\aspectos fundamentales a considerar al...
C:\users\nachitaypatty\documents\nacha\aspectos fundamentales a considerar al...C:\users\nachitaypatty\documents\nacha\aspectos fundamentales a considerar al...
C:\users\nachitaypatty\documents\nacha\aspectos fundamentales a considerar al...María Ignacia Zamorano Zuloaga
 
El color morado para el vestuario
El color morado para el vestuarioEl color morado para el vestuario
El color morado para el vestuarioliesel9208
 
Investigation of outdoor path loss models for wireless communication in bhuj
Investigation of outdoor path loss models for wireless communication in bhujInvestigation of outdoor path loss models for wireless communication in bhuj
Investigation of outdoor path loss models for wireless communication in bhujIAEME Publication
 
A study on users of moodle through sarasin model
A study on users of moodle through sarasin modelA study on users of moodle through sarasin model
A study on users of moodle through sarasin modelIAEME Publication
 
Relevance vector machine based prediction of mrrand sr for electro chemical m...
Relevance vector machine based prediction of mrrand sr for electro chemical m...Relevance vector machine based prediction of mrrand sr for electro chemical m...
Relevance vector machine based prediction of mrrand sr for electro chemical m...IAEME Publication
 
Production of nylon 6 fr lever using an injection moulding tool and identific...
Production of nylon 6 fr lever using an injection moulding tool and identific...Production of nylon 6 fr lever using an injection moulding tool and identific...
Production of nylon 6 fr lever using an injection moulding tool and identific...IAEME Publication
 
Protection for images transmission
Protection for images transmissionProtection for images transmission
Protection for images transmissionIAEME Publication
 
A data mining approach for prediction of heart disease using neural networks
A data mining approach for prediction of heart disease using neural networksA data mining approach for prediction of heart disease using neural networks
A data mining approach for prediction of heart disease using neural networksIAEME Publication
 

Destacado (9)

C:\users\nachitaypatty\documents\nacha\aspectos fundamentales a considerar al...
C:\users\nachitaypatty\documents\nacha\aspectos fundamentales a considerar al...C:\users\nachitaypatty\documents\nacha\aspectos fundamentales a considerar al...
C:\users\nachitaypatty\documents\nacha\aspectos fundamentales a considerar al...
 
El color morado para el vestuario
El color morado para el vestuarioEl color morado para el vestuario
El color morado para el vestuario
 
Investigation of outdoor path loss models for wireless communication in bhuj
Investigation of outdoor path loss models for wireless communication in bhujInvestigation of outdoor path loss models for wireless communication in bhuj
Investigation of outdoor path loss models for wireless communication in bhuj
 
A study on users of moodle through sarasin model
A study on users of moodle through sarasin modelA study on users of moodle through sarasin model
A study on users of moodle through sarasin model
 
Relevance vector machine based prediction of mrrand sr for electro chemical m...
Relevance vector machine based prediction of mrrand sr for electro chemical m...Relevance vector machine based prediction of mrrand sr for electro chemical m...
Relevance vector machine based prediction of mrrand sr for electro chemical m...
 
Production of nylon 6 fr lever using an injection moulding tool and identific...
Production of nylon 6 fr lever using an injection moulding tool and identific...Production of nylon 6 fr lever using an injection moulding tool and identific...
Production of nylon 6 fr lever using an injection moulding tool and identific...
 
Protection for images transmission
Protection for images transmissionProtection for images transmission
Protection for images transmission
 
Presente
PresentePresente
Presente
 
A data mining approach for prediction of heart disease using neural networks
A data mining approach for prediction of heart disease using neural networksA data mining approach for prediction of heart disease using neural networks
A data mining approach for prediction of heart disease using neural networks
 

Similar a Cochlear implant acoustic simulation model based on critical band filters

Speech processing strategies for cochlear prostheses the past, present and fu...
Speech processing strategies for cochlear prostheses the past, present and fu...Speech processing strategies for cochlear prostheses the past, present and fu...
Speech processing strategies for cochlear prostheses the past, present and fu...iaemedu
 
Speech processing strategies for cochlear prostheses
Speech processing strategies for cochlear prosthesesSpeech processing strategies for cochlear prostheses
Speech processing strategies for cochlear prosthesesiaemedu
 
Speech processing strategies for cochlear prostheses the past, present and fu...
Speech processing strategies for cochlear prostheses the past, present and fu...Speech processing strategies for cochlear prostheses the past, present and fu...
Speech processing strategies for cochlear prostheses the past, present and fu...iaemedu
 
Speech processing strategies for cochlear prostheses
Speech processing strategies for cochlear prosthesesSpeech processing strategies for cochlear prostheses
Speech processing strategies for cochlear prosthesesIAEME Publication
 
Speech processing strategies for cochlear prostheses
Speech processing strategies for cochlear prosthesesSpeech processing strategies for cochlear prostheses
Speech processing strategies for cochlear prosthesesiaemedu
 
Intelligent Arabic letters speech recognition system based on mel frequency c...
Intelligent Arabic letters speech recognition system based on mel frequency c...Intelligent Arabic letters speech recognition system based on mel frequency c...
Intelligent Arabic letters speech recognition system based on mel frequency c...IJECEIAES
 
Unspoken Words Recognition: A Review
Unspoken Words Recognition: A ReviewUnspoken Words Recognition: A Review
Unspoken Words Recognition: A Reviewidescitation
 
SPEECH RECOGNITION USING SONOGRAM AND AANN
SPEECH RECOGNITION USING SONOGRAM AND AANNSPEECH RECOGNITION USING SONOGRAM AND AANN
SPEECH RECOGNITION USING SONOGRAM AND AANNAM Publications
 
Wbi real time synthesiser
Wbi real time synthesiserWbi real time synthesiser
Wbi real time synthesiserPriyanka Yadav
 
Single Channel Speech Enhancement using Wiener Filter and Compressive Sensing
Single Channel Speech Enhancement using Wiener Filter and Compressive Sensing Single Channel Speech Enhancement using Wiener Filter and Compressive Sensing
Single Channel Speech Enhancement using Wiener Filter and Compressive Sensing IJECEIAES
 
Novel Approach of Implementing Psychoacoustic model for MPEG-1 Audio
Novel Approach of Implementing Psychoacoustic model for MPEG-1 AudioNovel Approach of Implementing Psychoacoustic model for MPEG-1 Audio
Novel Approach of Implementing Psychoacoustic model for MPEG-1 Audioinventy
 
Design of single channel portable eeg
Design of single channel portable eegDesign of single channel portable eeg
Design of single channel portable eegijbesjournal
 
Application of Digital Signal Processing In Echo Cancellation: A Survey
Application of Digital Signal Processing In Echo Cancellation: A SurveyApplication of Digital Signal Processing In Echo Cancellation: A Survey
Application of Digital Signal Processing In Echo Cancellation: A SurveyEditor IJCATR
 
VOWEL PHONEME RECOGNITION BASED ON AVERAGE ENERGY INFORMATION IN THE ZEROCROS...
VOWEL PHONEME RECOGNITION BASED ON AVERAGE ENERGY INFORMATION IN THE ZEROCROS...VOWEL PHONEME RECOGNITION BASED ON AVERAGE ENERGY INFORMATION IN THE ZEROCROS...
VOWEL PHONEME RECOGNITION BASED ON AVERAGE ENERGY INFORMATION IN THE ZEROCROS...ijistjournal
 
Design and Analysis of Optimum Performance Pacemaker Telemetry Antenna
Design and Analysis of Optimum Performance Pacemaker Telemetry AntennaDesign and Analysis of Optimum Performance Pacemaker Telemetry Antenna
Design and Analysis of Optimum Performance Pacemaker Telemetry AntennaTELKOMNIKA JOURNAL
 
Research on Wireless Earphone System using Multiple Auditory Filters
Research on Wireless Earphone System using Multiple Auditory FiltersResearch on Wireless Earphone System using Multiple Auditory Filters
Research on Wireless Earphone System using Multiple Auditory Filtersijtsrd
 

Similar a Cochlear implant acoustic simulation model based on critical band filters (20)

Speech processing strategies for cochlear prostheses the past, present and fu...
Speech processing strategies for cochlear prostheses the past, present and fu...Speech processing strategies for cochlear prostheses the past, present and fu...
Speech processing strategies for cochlear prostheses the past, present and fu...
 
Speech processing strategies for cochlear prostheses
Speech processing strategies for cochlear prosthesesSpeech processing strategies for cochlear prostheses
Speech processing strategies for cochlear prostheses
 
Speech processing strategies for cochlear prostheses the past, present and fu...
Speech processing strategies for cochlear prostheses the past, present and fu...Speech processing strategies for cochlear prostheses the past, present and fu...
Speech processing strategies for cochlear prostheses the past, present and fu...
 
Speech processing strategies for cochlear prostheses
Speech processing strategies for cochlear prosthesesSpeech processing strategies for cochlear prostheses
Speech processing strategies for cochlear prostheses
 
Speech processing strategies for cochlear prostheses
Speech processing strategies for cochlear prosthesesSpeech processing strategies for cochlear prostheses
Speech processing strategies for cochlear prostheses
 
Intelligent Arabic letters speech recognition system based on mel frequency c...
Intelligent Arabic letters speech recognition system based on mel frequency c...Intelligent Arabic letters speech recognition system based on mel frequency c...
Intelligent Arabic letters speech recognition system based on mel frequency c...
 
1801 1805
1801 18051801 1805
1801 1805
 
1801 1805
1801 18051801 1805
1801 1805
 
Unspoken Words Recognition: A Review
Unspoken Words Recognition: A ReviewUnspoken Words Recognition: A Review
Unspoken Words Recognition: A Review
 
SPEECH RECOGNITION USING SONOGRAM AND AANN
SPEECH RECOGNITION USING SONOGRAM AND AANNSPEECH RECOGNITION USING SONOGRAM AND AANN
SPEECH RECOGNITION USING SONOGRAM AND AANN
 
Sensors 20-00904-v2
Sensors 20-00904-v2Sensors 20-00904-v2
Sensors 20-00904-v2
 
Wbi real time synthesiser
Wbi real time synthesiserWbi real time synthesiser
Wbi real time synthesiser
 
Single Channel Speech Enhancement using Wiener Filter and Compressive Sensing
Single Channel Speech Enhancement using Wiener Filter and Compressive Sensing Single Channel Speech Enhancement using Wiener Filter and Compressive Sensing
Single Channel Speech Enhancement using Wiener Filter and Compressive Sensing
 
40120140503008
4012014050300840120140503008
40120140503008
 
Novel Approach of Implementing Psychoacoustic model for MPEG-1 Audio
Novel Approach of Implementing Psychoacoustic model for MPEG-1 AudioNovel Approach of Implementing Psychoacoustic model for MPEG-1 Audio
Novel Approach of Implementing Psychoacoustic model for MPEG-1 Audio
 
Design of single channel portable eeg
Design of single channel portable eegDesign of single channel portable eeg
Design of single channel portable eeg
 
Application of Digital Signal Processing In Echo Cancellation: A Survey
Application of Digital Signal Processing In Echo Cancellation: A SurveyApplication of Digital Signal Processing In Echo Cancellation: A Survey
Application of Digital Signal Processing In Echo Cancellation: A Survey
 
VOWEL PHONEME RECOGNITION BASED ON AVERAGE ENERGY INFORMATION IN THE ZEROCROS...
VOWEL PHONEME RECOGNITION BASED ON AVERAGE ENERGY INFORMATION IN THE ZEROCROS...VOWEL PHONEME RECOGNITION BASED ON AVERAGE ENERGY INFORMATION IN THE ZEROCROS...
VOWEL PHONEME RECOGNITION BASED ON AVERAGE ENERGY INFORMATION IN THE ZEROCROS...
 
Design and Analysis of Optimum Performance Pacemaker Telemetry Antenna
Design and Analysis of Optimum Performance Pacemaker Telemetry AntennaDesign and Analysis of Optimum Performance Pacemaker Telemetry Antenna
Design and Analysis of Optimum Performance Pacemaker Telemetry Antenna
 
Research on Wireless Earphone System using Multiple Auditory Filters
Research on Wireless Earphone System using Multiple Auditory FiltersResearch on Wireless Earphone System using Multiple Auditory Filters
Research on Wireless Earphone System using Multiple Auditory Filters
 

Más de IAEME Publication

IAEME_Publication_Call_for_Paper_September_2022.pdf
IAEME_Publication_Call_for_Paper_September_2022.pdfIAEME_Publication_Call_for_Paper_September_2022.pdf
IAEME_Publication_Call_for_Paper_September_2022.pdfIAEME Publication
 
MODELING AND ANALYSIS OF SURFACE ROUGHNESS AND WHITE LATER THICKNESS IN WIRE-...
MODELING AND ANALYSIS OF SURFACE ROUGHNESS AND WHITE LATER THICKNESS IN WIRE-...MODELING AND ANALYSIS OF SURFACE ROUGHNESS AND WHITE LATER THICKNESS IN WIRE-...
MODELING AND ANALYSIS OF SURFACE ROUGHNESS AND WHITE LATER THICKNESS IN WIRE-...IAEME Publication
 
A STUDY ON THE REASONS FOR TRANSGENDER TO BECOME ENTREPRENEURS
A STUDY ON THE REASONS FOR TRANSGENDER TO BECOME ENTREPRENEURSA STUDY ON THE REASONS FOR TRANSGENDER TO BECOME ENTREPRENEURS
A STUDY ON THE REASONS FOR TRANSGENDER TO BECOME ENTREPRENEURSIAEME Publication
 
BROAD UNEXPOSED SKILLS OF TRANSGENDER ENTREPRENEURS
BROAD UNEXPOSED SKILLS OF TRANSGENDER ENTREPRENEURSBROAD UNEXPOSED SKILLS OF TRANSGENDER ENTREPRENEURS
BROAD UNEXPOSED SKILLS OF TRANSGENDER ENTREPRENEURSIAEME Publication
 
DETERMINANTS AFFECTING THE USER'S INTENTION TO USE MOBILE BANKING APPLICATIONS
DETERMINANTS AFFECTING THE USER'S INTENTION TO USE MOBILE BANKING APPLICATIONSDETERMINANTS AFFECTING THE USER'S INTENTION TO USE MOBILE BANKING APPLICATIONS
DETERMINANTS AFFECTING THE USER'S INTENTION TO USE MOBILE BANKING APPLICATIONSIAEME Publication
 
ANALYSE THE USER PREDILECTION ON GPAY AND PHONEPE FOR DIGITAL TRANSACTIONS
ANALYSE THE USER PREDILECTION ON GPAY AND PHONEPE FOR DIGITAL TRANSACTIONSANALYSE THE USER PREDILECTION ON GPAY AND PHONEPE FOR DIGITAL TRANSACTIONS
ANALYSE THE USER PREDILECTION ON GPAY AND PHONEPE FOR DIGITAL TRANSACTIONSIAEME Publication
 
VOICE BASED ATM FOR VISUALLY IMPAIRED USING ARDUINO
VOICE BASED ATM FOR VISUALLY IMPAIRED USING ARDUINOVOICE BASED ATM FOR VISUALLY IMPAIRED USING ARDUINO
VOICE BASED ATM FOR VISUALLY IMPAIRED USING ARDUINOIAEME Publication
 
IMPACT OF EMOTIONAL INTELLIGENCE ON HUMAN RESOURCE MANAGEMENT PRACTICES AMONG...
IMPACT OF EMOTIONAL INTELLIGENCE ON HUMAN RESOURCE MANAGEMENT PRACTICES AMONG...IMPACT OF EMOTIONAL INTELLIGENCE ON HUMAN RESOURCE MANAGEMENT PRACTICES AMONG...
IMPACT OF EMOTIONAL INTELLIGENCE ON HUMAN RESOURCE MANAGEMENT PRACTICES AMONG...IAEME Publication
 
VISUALISING AGING PARENTS & THEIR CLOSE CARERS LIFE JOURNEY IN AGING ECONOMY
VISUALISING AGING PARENTS & THEIR CLOSE CARERS LIFE JOURNEY IN AGING ECONOMYVISUALISING AGING PARENTS & THEIR CLOSE CARERS LIFE JOURNEY IN AGING ECONOMY
VISUALISING AGING PARENTS & THEIR CLOSE CARERS LIFE JOURNEY IN AGING ECONOMYIAEME Publication
 
A STUDY ON THE IMPACT OF ORGANIZATIONAL CULTURE ON THE EFFECTIVENESS OF PERFO...
A STUDY ON THE IMPACT OF ORGANIZATIONAL CULTURE ON THE EFFECTIVENESS OF PERFO...A STUDY ON THE IMPACT OF ORGANIZATIONAL CULTURE ON THE EFFECTIVENESS OF PERFO...
A STUDY ON THE IMPACT OF ORGANIZATIONAL CULTURE ON THE EFFECTIVENESS OF PERFO...IAEME Publication
 
GANDHI ON NON-VIOLENT POLICE
GANDHI ON NON-VIOLENT POLICEGANDHI ON NON-VIOLENT POLICE
GANDHI ON NON-VIOLENT POLICEIAEME Publication
 
A STUDY ON TALENT MANAGEMENT AND ITS IMPACT ON EMPLOYEE RETENTION IN SELECTED...
A STUDY ON TALENT MANAGEMENT AND ITS IMPACT ON EMPLOYEE RETENTION IN SELECTED...A STUDY ON TALENT MANAGEMENT AND ITS IMPACT ON EMPLOYEE RETENTION IN SELECTED...
A STUDY ON TALENT MANAGEMENT AND ITS IMPACT ON EMPLOYEE RETENTION IN SELECTED...IAEME Publication
 
ATTRITION IN THE IT INDUSTRY DURING COVID-19 PANDEMIC: LINKING EMOTIONAL INTE...
ATTRITION IN THE IT INDUSTRY DURING COVID-19 PANDEMIC: LINKING EMOTIONAL INTE...ATTRITION IN THE IT INDUSTRY DURING COVID-19 PANDEMIC: LINKING EMOTIONAL INTE...
ATTRITION IN THE IT INDUSTRY DURING COVID-19 PANDEMIC: LINKING EMOTIONAL INTE...IAEME Publication
 
INFLUENCE OF TALENT MANAGEMENT PRACTICES ON ORGANIZATIONAL PERFORMANCE A STUD...
INFLUENCE OF TALENT MANAGEMENT PRACTICES ON ORGANIZATIONAL PERFORMANCE A STUD...INFLUENCE OF TALENT MANAGEMENT PRACTICES ON ORGANIZATIONAL PERFORMANCE A STUD...
INFLUENCE OF TALENT MANAGEMENT PRACTICES ON ORGANIZATIONAL PERFORMANCE A STUD...IAEME Publication
 
A STUDY OF VARIOUS TYPES OF LOANS OF SELECTED PUBLIC AND PRIVATE SECTOR BANKS...
A STUDY OF VARIOUS TYPES OF LOANS OF SELECTED PUBLIC AND PRIVATE SECTOR BANKS...A STUDY OF VARIOUS TYPES OF LOANS OF SELECTED PUBLIC AND PRIVATE SECTOR BANKS...
A STUDY OF VARIOUS TYPES OF LOANS OF SELECTED PUBLIC AND PRIVATE SECTOR BANKS...IAEME Publication
 
EXPERIMENTAL STUDY OF MECHANICAL AND TRIBOLOGICAL RELATION OF NYLON/BaSO4 POL...
EXPERIMENTAL STUDY OF MECHANICAL AND TRIBOLOGICAL RELATION OF NYLON/BaSO4 POL...EXPERIMENTAL STUDY OF MECHANICAL AND TRIBOLOGICAL RELATION OF NYLON/BaSO4 POL...
EXPERIMENTAL STUDY OF MECHANICAL AND TRIBOLOGICAL RELATION OF NYLON/BaSO4 POL...IAEME Publication
 
ROLE OF SOCIAL ENTREPRENEURSHIP IN RURAL DEVELOPMENT OF INDIA - PROBLEMS AND ...
ROLE OF SOCIAL ENTREPRENEURSHIP IN RURAL DEVELOPMENT OF INDIA - PROBLEMS AND ...ROLE OF SOCIAL ENTREPRENEURSHIP IN RURAL DEVELOPMENT OF INDIA - PROBLEMS AND ...
ROLE OF SOCIAL ENTREPRENEURSHIP IN RURAL DEVELOPMENT OF INDIA - PROBLEMS AND ...IAEME Publication
 
OPTIMAL RECONFIGURATION OF POWER DISTRIBUTION RADIAL NETWORK USING HYBRID MET...
OPTIMAL RECONFIGURATION OF POWER DISTRIBUTION RADIAL NETWORK USING HYBRID MET...OPTIMAL RECONFIGURATION OF POWER DISTRIBUTION RADIAL NETWORK USING HYBRID MET...
OPTIMAL RECONFIGURATION OF POWER DISTRIBUTION RADIAL NETWORK USING HYBRID MET...IAEME Publication
 
APPLICATION OF FRUGAL APPROACH FOR PRODUCTIVITY IMPROVEMENT - A CASE STUDY OF...
APPLICATION OF FRUGAL APPROACH FOR PRODUCTIVITY IMPROVEMENT - A CASE STUDY OF...APPLICATION OF FRUGAL APPROACH FOR PRODUCTIVITY IMPROVEMENT - A CASE STUDY OF...
APPLICATION OF FRUGAL APPROACH FOR PRODUCTIVITY IMPROVEMENT - A CASE STUDY OF...IAEME Publication
 
A MULTIPLE – CHANNEL QUEUING MODELS ON FUZZY ENVIRONMENT
A MULTIPLE – CHANNEL QUEUING MODELS ON FUZZY ENVIRONMENTA MULTIPLE – CHANNEL QUEUING MODELS ON FUZZY ENVIRONMENT
A MULTIPLE – CHANNEL QUEUING MODELS ON FUZZY ENVIRONMENTIAEME Publication
 

Más de IAEME Publication (20)

IAEME_Publication_Call_for_Paper_September_2022.pdf
IAEME_Publication_Call_for_Paper_September_2022.pdfIAEME_Publication_Call_for_Paper_September_2022.pdf
IAEME_Publication_Call_for_Paper_September_2022.pdf
 
MODELING AND ANALYSIS OF SURFACE ROUGHNESS AND WHITE LATER THICKNESS IN WIRE-...
MODELING AND ANALYSIS OF SURFACE ROUGHNESS AND WHITE LATER THICKNESS IN WIRE-...MODELING AND ANALYSIS OF SURFACE ROUGHNESS AND WHITE LATER THICKNESS IN WIRE-...
MODELING AND ANALYSIS OF SURFACE ROUGHNESS AND WHITE LATER THICKNESS IN WIRE-...
 
A STUDY ON THE REASONS FOR TRANSGENDER TO BECOME ENTREPRENEURS
A STUDY ON THE REASONS FOR TRANSGENDER TO BECOME ENTREPRENEURSA STUDY ON THE REASONS FOR TRANSGENDER TO BECOME ENTREPRENEURS
A STUDY ON THE REASONS FOR TRANSGENDER TO BECOME ENTREPRENEURS
 
BROAD UNEXPOSED SKILLS OF TRANSGENDER ENTREPRENEURS
BROAD UNEXPOSED SKILLS OF TRANSGENDER ENTREPRENEURSBROAD UNEXPOSED SKILLS OF TRANSGENDER ENTREPRENEURS
BROAD UNEXPOSED SKILLS OF TRANSGENDER ENTREPRENEURS
 
DETERMINANTS AFFECTING THE USER'S INTENTION TO USE MOBILE BANKING APPLICATIONS
DETERMINANTS AFFECTING THE USER'S INTENTION TO USE MOBILE BANKING APPLICATIONSDETERMINANTS AFFECTING THE USER'S INTENTION TO USE MOBILE BANKING APPLICATIONS
DETERMINANTS AFFECTING THE USER'S INTENTION TO USE MOBILE BANKING APPLICATIONS
 
ANALYSE THE USER PREDILECTION ON GPAY AND PHONEPE FOR DIGITAL TRANSACTIONS
ANALYSE THE USER PREDILECTION ON GPAY AND PHONEPE FOR DIGITAL TRANSACTIONSANALYSE THE USER PREDILECTION ON GPAY AND PHONEPE FOR DIGITAL TRANSACTIONS
ANALYSE THE USER PREDILECTION ON GPAY AND PHONEPE FOR DIGITAL TRANSACTIONS
 
VOICE BASED ATM FOR VISUALLY IMPAIRED USING ARDUINO
VOICE BASED ATM FOR VISUALLY IMPAIRED USING ARDUINOVOICE BASED ATM FOR VISUALLY IMPAIRED USING ARDUINO
VOICE BASED ATM FOR VISUALLY IMPAIRED USING ARDUINO
 
IMPACT OF EMOTIONAL INTELLIGENCE ON HUMAN RESOURCE MANAGEMENT PRACTICES AMONG...
IMPACT OF EMOTIONAL INTELLIGENCE ON HUMAN RESOURCE MANAGEMENT PRACTICES AMONG...IMPACT OF EMOTIONAL INTELLIGENCE ON HUMAN RESOURCE MANAGEMENT PRACTICES AMONG...
IMPACT OF EMOTIONAL INTELLIGENCE ON HUMAN RESOURCE MANAGEMENT PRACTICES AMONG...
 
VISUALISING AGING PARENTS & THEIR CLOSE CARERS LIFE JOURNEY IN AGING ECONOMY
VISUALISING AGING PARENTS & THEIR CLOSE CARERS LIFE JOURNEY IN AGING ECONOMYVISUALISING AGING PARENTS & THEIR CLOSE CARERS LIFE JOURNEY IN AGING ECONOMY
VISUALISING AGING PARENTS & THEIR CLOSE CARERS LIFE JOURNEY IN AGING ECONOMY
 
A STUDY ON THE IMPACT OF ORGANIZATIONAL CULTURE ON THE EFFECTIVENESS OF PERFO...
A STUDY ON THE IMPACT OF ORGANIZATIONAL CULTURE ON THE EFFECTIVENESS OF PERFO...A STUDY ON THE IMPACT OF ORGANIZATIONAL CULTURE ON THE EFFECTIVENESS OF PERFO...
A STUDY ON THE IMPACT OF ORGANIZATIONAL CULTURE ON THE EFFECTIVENESS OF PERFO...
 
GANDHI ON NON-VIOLENT POLICE
GANDHI ON NON-VIOLENT POLICEGANDHI ON NON-VIOLENT POLICE
GANDHI ON NON-VIOLENT POLICE
 
A STUDY ON TALENT MANAGEMENT AND ITS IMPACT ON EMPLOYEE RETENTION IN SELECTED...
A STUDY ON TALENT MANAGEMENT AND ITS IMPACT ON EMPLOYEE RETENTION IN SELECTED...A STUDY ON TALENT MANAGEMENT AND ITS IMPACT ON EMPLOYEE RETENTION IN SELECTED...
A STUDY ON TALENT MANAGEMENT AND ITS IMPACT ON EMPLOYEE RETENTION IN SELECTED...
 
ATTRITION IN THE IT INDUSTRY DURING COVID-19 PANDEMIC: LINKING EMOTIONAL INTE...
ATTRITION IN THE IT INDUSTRY DURING COVID-19 PANDEMIC: LINKING EMOTIONAL INTE...ATTRITION IN THE IT INDUSTRY DURING COVID-19 PANDEMIC: LINKING EMOTIONAL INTE...
ATTRITION IN THE IT INDUSTRY DURING COVID-19 PANDEMIC: LINKING EMOTIONAL INTE...
 
INFLUENCE OF TALENT MANAGEMENT PRACTICES ON ORGANIZATIONAL PERFORMANCE A STUD...
INFLUENCE OF TALENT MANAGEMENT PRACTICES ON ORGANIZATIONAL PERFORMANCE A STUD...INFLUENCE OF TALENT MANAGEMENT PRACTICES ON ORGANIZATIONAL PERFORMANCE A STUD...
INFLUENCE OF TALENT MANAGEMENT PRACTICES ON ORGANIZATIONAL PERFORMANCE A STUD...
 
A STUDY OF VARIOUS TYPES OF LOANS OF SELECTED PUBLIC AND PRIVATE SECTOR BANKS...
A STUDY OF VARIOUS TYPES OF LOANS OF SELECTED PUBLIC AND PRIVATE SECTOR BANKS...A STUDY OF VARIOUS TYPES OF LOANS OF SELECTED PUBLIC AND PRIVATE SECTOR BANKS...
A STUDY OF VARIOUS TYPES OF LOANS OF SELECTED PUBLIC AND PRIVATE SECTOR BANKS...
 
EXPERIMENTAL STUDY OF MECHANICAL AND TRIBOLOGICAL RELATION OF NYLON/BaSO4 POL...
EXPERIMENTAL STUDY OF MECHANICAL AND TRIBOLOGICAL RELATION OF NYLON/BaSO4 POL...EXPERIMENTAL STUDY OF MECHANICAL AND TRIBOLOGICAL RELATION OF NYLON/BaSO4 POL...
EXPERIMENTAL STUDY OF MECHANICAL AND TRIBOLOGICAL RELATION OF NYLON/BaSO4 POL...
 
ROLE OF SOCIAL ENTREPRENEURSHIP IN RURAL DEVELOPMENT OF INDIA - PROBLEMS AND ...
ROLE OF SOCIAL ENTREPRENEURSHIP IN RURAL DEVELOPMENT OF INDIA - PROBLEMS AND ...ROLE OF SOCIAL ENTREPRENEURSHIP IN RURAL DEVELOPMENT OF INDIA - PROBLEMS AND ...
ROLE OF SOCIAL ENTREPRENEURSHIP IN RURAL DEVELOPMENT OF INDIA - PROBLEMS AND ...
 
OPTIMAL RECONFIGURATION OF POWER DISTRIBUTION RADIAL NETWORK USING HYBRID MET...
OPTIMAL RECONFIGURATION OF POWER DISTRIBUTION RADIAL NETWORK USING HYBRID MET...OPTIMAL RECONFIGURATION OF POWER DISTRIBUTION RADIAL NETWORK USING HYBRID MET...
OPTIMAL RECONFIGURATION OF POWER DISTRIBUTION RADIAL NETWORK USING HYBRID MET...
 
APPLICATION OF FRUGAL APPROACH FOR PRODUCTIVITY IMPROVEMENT - A CASE STUDY OF...
APPLICATION OF FRUGAL APPROACH FOR PRODUCTIVITY IMPROVEMENT - A CASE STUDY OF...APPLICATION OF FRUGAL APPROACH FOR PRODUCTIVITY IMPROVEMENT - A CASE STUDY OF...
APPLICATION OF FRUGAL APPROACH FOR PRODUCTIVITY IMPROVEMENT - A CASE STUDY OF...
 
A MULTIPLE – CHANNEL QUEUING MODELS ON FUZZY ENVIRONMENT
A MULTIPLE – CHANNEL QUEUING MODELS ON FUZZY ENVIRONMENTA MULTIPLE – CHANNEL QUEUING MODELS ON FUZZY ENVIRONMENT
A MULTIPLE – CHANNEL QUEUING MODELS ON FUZZY ENVIRONMENT
 

Cochlear implant acoustic simulation model based on critical band filters

  • 1. INTERNATIONAL JOURNAL OF ELECTRONICS AND International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN 0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME COMMUNICATION ENGINEERING & TECHNOLOGY (IJECET) ISSN 0976 – 6464(Print) ISSN 0976 – 6472(Online) Volume 3, Issue 2, July- September (2012), pp. 116-129 IJECET © IAEME: www.iaeme.com/ijecet.html ©IAEME Journal Impact Factor (2012): 3.5930 (Calculated by GISI) www.jifactor.com COCHLEAR IMPLANT ACOUSTIC SIMULATION MODEL BASED ON CRITICAL BAND FILTERS P Mahalakshmi1, M R Reddy2 Bio-medical Engineering Group, Department of Applied Mechanics, Indian Institute of Technology Madras, Chennai-600 036, Tamil Nadu, India. 1 E-mail: maha_50@yahoo.com, 2E-mail: rsreddy@iitm.ac.in ABSTRACT Cochlear Implants (CI) are surgically implanted biomedical devices that can provide hearing to profoundly deaf people by direct electrical stimulation of the auditory nerve. Choosing the optimal number of electrodes for the improved perception of speech is an especially difficult problem for users of CI, as the prostheses system depends on various parameters and processing strategies. It has widely been recognized that using the auditory models in cochlear prostheses might significantly improve the speech perception. In this paper, we propose a new CI acoustic simulation model based on a 16-channel Critical Band Finite-Impulse Response (CB-FIR) filter bank and its performance is discussed. The place of stimulation in the cochlear electrode array is identified through the computation and selection of significant spectral band energies. We also investigate the temporal information through the simulation model based algorithm. The amplitude levels of the pulses are identified by envelope detection method. The simulation model is tested with TIMIT database for 15 phonemes. The experimental results indicated that, a reduced number of channels are only required for stimulation of cochlear electrodes, one at a time, thereby minimizing channel interaction. The proposed model should be helpful for the development of novel speech processing strategies to improve the speech perception of cochlear implantees. Keywords Auditory model, cochlear implant, speech processing, speech perception, spectral analysis, acoustic simulation. 1. INTRODUCTION A cochlear implant (CI) is an auditory neural prosthesis for restoring hearing function in patients with sensori-neural hearing loss. Hearing restoration is achieved by electrically 116
  • 2. International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN 0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME stimulating the auditory nerve [1]. The electrical stimulation pulse parameters are derived from incoming speech by speech processors contained within the CI devices [2]. Several CI devices developed over the years have the following features in common. It is comprised of external/ and internal units. The external unit includes microphone, speech processor and transmitter. The internal unit consists of receiver-stimulator and an electrode array embedded in cochlea [3]. The sound is picked up by the microphone and processed by the signal/speech processor that analyzes sound and encodes the results into instructions. By RF (radio frequency) signal, the transmission system transmits the instructions to the implanted receiver. The internal circuit decodes the data, then generates electrical stimulus and drives the electrode array to stimulate auditory nerve fibers at different places [4]. The speech processor is responsible for decomposing the input audio signals into different frequency bands and formulating the most appropriate stimulation pattern to the electrodes [5]. The design of speech processors must be such that it extracts the speech parameters that are essential for intelligibility and then encode them for electrical stimulation [6]. Speech understanding requires the analysis of various speech components such as vowels and consonants. The estimate of perceptual importance of these component features will lead to intelligible speech stimuli [7]. This speech stimuli depends on various parameters such as speech material, filter cutoff frequencies, order of band pass and low pass filters, number of channels, number of electrodes, type and parameters of the compression function, pulse width and pulse duration of modulating pulses, sequence of channel stimulation [8]. Various speech processing strategies have been developed and reported in literature over time for cochlear prostheses which include Continuous Interleaved Sampling (CIS), Spectral Peak (SPEAK), Advanced Combination Encoder (ACE), Spectral Maxima Sound Processor (SMSP), Simultaneous Analog Strategy (SAS) and various formant based strategies [9,10]. Most implant manufacturers provide research speech processors for use on human subjects that allow researchers to develop and test new signal processing algorithms [11]. Much research has therefore been directed at improving the performance and major successes have been achieved since they were first implemented in the 1970s. The number of implanted electrodes has grown from 1 to as many as 24 in current implants [12]. However, current cochlear implants are not very successful in delivering phase information that might be responsible for the reduced ability of CI users to perceive music. Spectral enhancement is another major concern that has benefits of improving performance in noise. It is possible that this is one of the reasons for why the human auditory system with a biological cochlea front end is known to have good performance in noise[13]. Spectral enhancement is a natural consequence of our trying to mimic the operation of the biological cochlea. Also, the effect of electrode interactions and current spreading limits the spectral resolution to a great degree. A common speech processing strategy used in implants and in speech-recognition systems, employs a mel cepstrum filter bank. In implants, the number of functioning electrodes used for stimulation is between 8- 24 and having more electrodes is often not useful due to spatial interactions amongst the electrodes [14]. Our work proposes an acoustic simulation model for cochlear implants that uses a 16- channel CB-FIR filter bank. Spectral enhancement and audibility could be improved by using a set of independent filters with independent and easily programmable parameters. Input speech signals are filtered using this filter bank, then rectified, low-pass filtered and the amplitude levels are estimated. Spectral energies are computed for various speech components such as phonemes from each channel. The most significant channels, which are few in number, constituting maximum signal energy are selected thus aiding to minimize channel interactions. These channel energies are used in estimating the stimulation pulse level to be presented to the selected cochlear electrodes. 117
  • 3. International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN 0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME The outline of the paper is as follows. In Section 2, we describe the essential ideas behind the choice of filter bank strategy, the improved design aspects and how various speech components are assessed for its speech intelligibility. In Section 3, we present Matlab simulation data for the designed filter bank system with speech input that illustrate the working and benefits of the system. In Section 4, we illustrate and discuss the results of our proposed cochlear implant acoustic simulation model through the spectral and temporal analysis performed for various speech components and paving a way for the electrical stimulation procedure for cochlear implants. In Section 5, we conclude by summarizing our contributions. 2. MATERIALS AND METHODS 2.1 Digital speech processing The greatest common block of all recognition systems is the signal-processing front end, which converts the speech waveform to some type of parametric representation for further analysis and processing. Parametrical representation of speech signal uses short-time energy and zero crossing rates [15]. Speech signals are more consistently analyzed spectrally than in the time domain and human hearing pays more attention to spectral aspects of speech [16]. Filter bank representation is one such spectral analysis method that uses a set of band pass filters, each analyzing a different range of frequencies of the input speech. There are two main motivations for the filter bank representation. First, the position of maximum displacement along the basilar membrane for stimuli such as pure tones is proportional to the logarithm of the frequency of the tone. This hypothesis is part of a theory of hearing called the ‘place’ theory. Second, the experiments in human perception have shown that frequencies of a complex sound within a certain bandwidth of some nominal frequency cannot be individually identified. When one of the components of this sound falls outside this bandwidth, it can be individually distinguished. This bandwidth is referred to as the critical bandwidth [17]. The digital speech processing proposed for ‘cochlear implant - acoustic simulation model’ is built around a critical band filter bank followed by rectifiers and low pass filters. The cascade of a rectifier and a low pass filter yields an envelope detector. Input sound signals are filtered using a digital band-pass FIR (Finite Impulse Response) filter bank, with one filter for each electrode of the implant. The number of filter bands depends on the number of stimulation channels to be considered in the implant of cochlear prostheses system [18]. The filter bank was designed to have linear frequency spacing at low center frequencies and logarithmic spacing at high center frequencies in correspondence to the properties of cochlear filters [19]. In accordance with Rekha et al., [20], the human speech frequency, approximately 100 to 3500 Hz, is allocated to filter bands using critical band scale. This frequency range is covered in 16 bands, which is shown in Table 1. In the signal analysis application, filter bank model with non-uniform spacing has the potential to offer better performance [21]. The general specifications of the filter bank include the stop band attenuation (e.g., 40- 50dB), number of channels (e.g., 16) and the pass band spacing. For the advanced cochlear devices, we specify the need for a linear phase finite-impulse response (FIR) filter bank as opposed to the nonlinear phase infinite-impulse response (IIR) filter bank for the following reasons [22]. • FIR filters are intrinsically stable • They can easily realize various features that are not possible or are difficult to achieve with IIR filters • To eliminate channel interactions. 118
  • 4. International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN 0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME To achieve more stop-band attenuation, a higher filter order is required with consequent cost in CPU usage. We impose higher stop band attenuation ( –60dB) compared to –50dB achieved by the current devices. Table 1. Critical bandwidth as a function of center frequency: Non-uniform, non-overlapping band- pass filter cutoff frequencies that correspond to cochlea filter properties are listed here. Channel Center Frequency Number Frequency Band (Hz) (Hz) 1 150 125-175 2 250 225-275 3 350 325-375 4 450 420-480 5 570 530-605 6 700 655-745 7 840 790-890 8 1000 940-1060 9 1170 1105-1235 10 1370 1285-1455 11 1600 1505-1695 12 1850 1745-1955 13 2150 2005-2295 14 2500 2345-2655 15 2900 2705-3095 16 3400 3145-3655 2.2 FIR filter design using kaiser window The problem of designing an FIR filter consists of determining the impulse response sequence, h[n], or its system function H(z), so that given requirements on the filter response are satisfied. The filter requirements are usually specified in the frequency domain, and only this case is considered here [23]. The frequency response jω H d (e ) corresponding to the impulse response h[n] is expressed in a fourier series as ∞ H d (e jω ) = ∑ hd (n) e− jωn n=−∞ (1) and known as fourier coefficients having infinite length. FIR filter is obtained by truncating the infinite fourier series at n = ±  N −1   , where N is the length of the desired sequence.  2  Abrupt truncation of the fourier series results in oscillations in pass-band and stop-band regions of the filter. An optimal way to reduce these oscillations is to use an appropriate finite-length window w(n) , which controls the overall filter to yield a smooth frequency response. Therefore, a window should possess some of the following spectral characteristics [24,25]. (i)The main lobe width in the frequency response of the window should be narrow [17] (ii) The ripple ratio should be small (iii)The side-lobes should decrease in energy rapidly as ω tends to π. (iv)To adequately block the stop-band frequencies, it is necessary to have good stop- 119
  • 5. International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN 0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME band attenuation. The symmetrical frequency spectrum of an odd length window w(nT ) can be written as [25], ( N −1) 2 W (e jωT ) = w(0) + 2 ∑ w(nT ) Cos( wnT ) n =1 (2) Where T = 2π ω s is the sampling period. Standard windows such as Rectangular, Hamming and Hanning have only one independent parameter namely, the ‘window length’ N , which controls the ‘transition width’ of the filter. An increase in window length decreases the transition width of the filter. But the minimum stop band attenuation of the filter is independent of N . In order to overcome this, a Kaiser window is used which has an adjustable parameter β that controls the stop-band attenuation and an independent parameter N , that controls the transition width [26]. Kaiser window is widely used for the spectral analysis and FIR filter design applications. The design of band pass filter using Kaiser Window has been carried out based on the following steps [24,26,27]. Step 1: The length ‘N’ of the band pass filter is computed using Eq (3). −20log10 ( δ pδ s ) −13 N ≅ 14.6(ωs − ω p ) 2π (3) where, δ p and δs are the pass band and stop band ripples respectively and ω p and ωs are the pass band and stop band edge frequencies respectively. The design is based on the narrower of the two transition bandwidths ( Bt ) , that is Bt = min[(ω p − ωs ), (ωs − ω p )], where (ωs ,ωp ) and (ωp ,ωs ) are the initial and later 1 1 2 2 1 1 2 2 transition edge frequencies of the band pass filter. The cut-off frequencies are given by ωc1 = ω p1 − Bt 2 and ωc2 = ω p2 + Bt 2. Step 2: The Kaiser window sequence is given by  w [n] = I0  (  β 1 − [2 n / ( N − 1) ]2 ) , − ( N − 1) / 2 ≤ n ≤ ( N − 1) / 2  I0 ( β )  (4) where β is an adjustable parameter, and I0(x) is the modified zeroth order Bessel function, which can be expanded into the following series. ∞  x/2 r 2 ( )  I 0 ( x) = 1 + ∑   r!   r = 1  (5) which is seen to be positive for all real values of x. 120
  • 6. International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN 0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME Step 3: The desired minimum stop band attenuation is calculated by α s = - 20 log10 δ s . β is computed from Eq.(6).  0.1102 α − 8.7 , ( ) for α s > 50,  s  0.4 ( ) ( ) β =  0.5842 α s − 21 + 0.007886 α s − 21 , for 21 ≤ α s ≤ 50,   0, for α s < 21.  (6) Step 4: The coefficients of linear phase filter with delay γ =  N −1 are computed using    2    ωc − ωc    2 1  , for n = 0    π  hd (n) =    (7)  [ Sin ωc (n − γ ) − Sin ωc (n − γ ) ]  2 1 , for n ≠ 0   π (n − γ ) Step 5: The impulse response h(n) is obtained by multiplying the desired impulse response hd(n) of the ideal filter and the window coefficients generated in Step 2 to yield the coefficients of FIR filter. h( n ) = hd ( n) .w( n) (8) Step 6: Filter specifications are usually given in terms of the magnitude response |H(ejω)|. The magnitude function of FIR filter when the impulse response is symmetric and when N is odd is generated by  N −1     2   N −1  N −1  H (ω ) = h   2  + ∑ 2h   2 − n  Cos (ω n)  (9) n =1 Filters designed with this window had the level of stop-band approximately equal to – 60dB. 2.3 CI-Acoustic Simulation Model The schematic of the proposed CI acoustic simulation model is shown in Figure 1. The input speech signal is passed through a bank of 16 band pass filters, whose coverage spans the human speech frequency range. The spectral output is obtained by the summation of independent band pass filters responses. Spectral energies are estimated for each of 16 channels. Temporal analysis is done by full-wave rectification and low pass filtering the decomposed signal in each band. The stimulus amplitude levels are estimated through envelope detection procedure. 121
  • 7. International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN 0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME Fig.1.Schematic of the proposed CI-acoustic simulation model: Spectral energies are estimated by a bank of band pass filters, followed by temporal estimation by rectification and envelope detection, determining the place and stimulus level information. 2.3.1 Rectification Speech signals are complex patterns that do not permit direct analysis of their wave shape. The envelope of the overall signal carries important information. The envelope of a signal can be derived by obtaining the absolute value of the signal at each time instant, i.e., performing full-wave rectification [28]. A FIR low pass filter of order 100 and a cut-off frequency of 400 Hz are used to obtain smooth envelopes of speech signals [29]. 2.3.2 Envelope detection Envelope detection is done using a mathematical method called ‘Linear Interpolation’. This technique is used to detect the envelope of the low pass filter output. In order to obtain smooth envelope of the low pass filter output, the estimation of values for intermediate time instances are carried out. A point in the low pass filter output is interpolated using Eq (9). ( y2 − y1 )( x − x1 ) y = y1 + ( x2 − x1 ) (10) where x1, x2 are the two time instances and y1, y2 are their corresponding amplitude values estimated from the output of low pass filter. If these quantities are known, then for a desired intermediate time instant x, a new value y can be obtained. Matlab code is developed for interpolation and stimulus amplitude estimation. 2.4 Speech sounds and features The perceptually important information of a speech signal needs to be preserved in order to assess its intelligibility. Speech sounds called phonemes are produced depending on the state of input excitation (vocal cords). Voiced speech (e.g., vowels) is produced if the vocal cords are tensed and vibrate periodically. Unvoiced speech (e.g., consonants) is produced if the vocal cords are not vibrating and the resulting waveform is aperiodic or random [15]. Based on the position of the tongue, vowel articulatory configurations are produced and classified as front vowel /i/, mid vowel /a/ and back vowel /u/. A ‘diphthong’ is a complex speech sound that begins with one vowel and gradually changes to another vowel within the same syllable (egs., /ay/, /oy/, /aw/, /ey/). Voiced (/b/, /d/, /g/) and unvoiced stop consonants (/p/, /t/, /k/) are produced by the complete stoppage and sudden release of the breath. Nasal consonants (/m/, /n/) are produced with glottal excitation and the vocal tract totally constricted. In the design of speech analysis system, how the spectral and temporal 122
  • 8. International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN 0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME properties of a speech signal affect its perception need to be studied [30]. In this paper, we have estimated the spectral energy to separate the voiced and unvoiced parts of speech and also to determine the stimulus pulse level to be given to the cochlear electrodes, using the developed ‘CI acoustic simulation model’ discussed in the earlier section. Table II lists the classification of phonemes into sound classes taken for the study. Table 2. Classification of the standard phonemes into sound classes: Three vowels: front, mid, back corresponding to the position of the tongue hump; four diphthongs, two nasal consonants; three voiced and unvoiced stop consonants are listed here. S.No Type of sound Phonetic symbol 1 Vowels /a/, /i/, /u/ 2 Diphthongs /aw /, /oy/, /ay/, /ey/ 3 Nasal consonants /m/, /n/ 4 Voiced consonants /b/, /d/, /g/ 5 Unvoiced consonants /k/, /t/, /p/ 2.5 Speech data The speech data used for the current study is obtained from TIMIT database which consists of phonemes, sampled at 16 KHz. In a total of 48 standard phonemes, 15 phonemes as listed in Table 2, are used for the analysis. 3. EXPERIMENT The function of signal processor in CI devices primarily consists of filter bank model that divides an input speech signal into a number of frequency bands in order to extract signal strength in each band for exciting the implanted electrodes accordingly. As per the frequency bands shown in Table 1, the input speech signals are digitally filtered using critical band FIR filter based on the specifications discussed in section 2.2. MATLAB code is written to generate filter coefficients by using Kaiser Window. We adopt the linear- phase FIR filter and use its symmetrical property to reduce the number of multiplications by half [24]. Figure 2 depicts the magnitude response of 16 channels of the filter bank; each channel is an 877th-order FIR filter sampled at 16KHz. The stop-band attenuation obtained is approximately –60dB for a β value of 6. A CI acoustic simulation model is constructed and simulation experiment is conducted to investigate its contribution for speech perception. All the test materials are analyzed for its dominant energy levels between 125Hz and 3655 Hz. Each utterance of speech signal for the vowels /a/, /i/, /u/, diphthongs /ay/, /oy/, /aw/, /ey/, voiced /b/, /d/, /g/, unvoiced stop consonants /p/, /t/, /k/, and nasal consonants /m/, /n/, are passed through the 16 independent channels. Spectral energies are estimated from each band and the amplitude levels are estimated through envelope detection. 123
  • 9. International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN 0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME Fig. 2. Magnitude response of the 16 channel band pass filter: Each channel exhibiting – 60dB stop band attenuation. 4. RESULTS AND DISCUSSION Using the acoustic model discussed in section 2.3, it is proposed that a reduced number of channels is selected for stimulation based on the speech energy obtained in different channels; in order to avoid inter-channel interference among the cochlear electrodes. The experiment is performed for all the phonemes listed in section 2.5. To select the appropriate stimulation electrodes, the spectral energies of the model output are estimated for each of 16 channels, where each phoneme is analyzed for 10 utterances. The following sections discuss the results obtained through spectral and temporal analysis based on the CI acoustic simulation model. 4.1 Vowels Here an illustration is discussed for selecting the stimulation electrodes based on the channel energies. Typical energy levels of vowel ‘a’ for ten utterances are represented in Table 3. The sixteen rows represent the sixteen channels’ energy and columns U1 through U10 represent utterance of each subject. The energy values are normalized across sixteen channels for each utterance. It is inferred from Table 3 that, nine utterances show maximum energy (energy > 0.9) in five and six channels. This shows that maximum speech information is present in the frequency range of channels 5 and 6 for vowel /a/. If we form a criterion that only channels having normalized energy level > 0.9 are used for stimulation, then for vowel ‘a’, only electrodes 5 and 6 are to be stimulated. This corresponds to a frequency range of 530Hz-605Hz and 655Hz-745Hz. All other channel outputs that possess less than 0.9 of the normalized energy value may not contain much of speech information and these corresponding electrodes need not be stimulated. The output speech spectra of 16 channel BPF for vowel /a/ for utterance 1 are illustrated in Figure 3a. Two dominant peaks are observed at center frequencies 570Hz and 700Hz, which correspond to channels 5 and 6. The experiment is performed for the remaining vowels /i/ and /u/. The front vowel /i/ showed > 0.9 of the normalized energy in channels 13 and 14, which correspond to the center frequencies 2150Hz and 2500Hz. For back vowel /u/, channels 4 and 11 which correspond to 450Hz and 1600Hz were identified. These significant channels only would be used for the excitation of electrodes. 124
  • 10. International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN 0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME Table 3. Maximum energy levels corresponding to the utterances of vowel /a/: Distribution of speech energy in 16 channels for ten utterances. Channel U1 U2 U3 U4 U5 U6 U7 U8 U9 U10 1 1.5 1.3 0.2 0.2 1.7 0.6 1.2 0.6 0.1 4.4 2 1.1 2.5 0.2 1.1 4.9 2.9 1.4 4.4 2.1 7 3 1.9 0.4 1.2 1.5 1.9 2.4 0.2 7.2 2.9 4.2 4 1.5 1 6.9 6.2 1.4 3.9 5.5 10.3 3.8 8.5 5 10.2 3.1 7.6 29 6.9 9.9 24.8 20.6 16.5 26 6 12.2 5.9 7.5 17.3 9.4 14.2 5.6 42.5 25.1 30 7 5.1 3 2 5.1 5.5 15.5 2.5 14.4 14 6.8 8 2.9 3.1 2.4 4.1 8.8 13.9 2.6 9.1 26.1 5.5 9 3.5 3.4 3.2 5.2 13.1 4.5 1.4 9 20.9 6 10 6.4 1.4 8.8 18.3 12.9 1.8 2.9 32.8 4.6 19 11 4.9 0.2 8.3 5 7.1 1.2 12.9 25.6 2.6 9.2 12 2.1 0.1 2.5 3.2 4.9 1 6.6 10.3 3.3 4.1 13 2.1 0.4 2.7 6.8 7.2 3 8.6 19.2 6.4 7.3 14 4.3 1.1 6 1.7 5.5 1.8 5.1 13.6 1.3 10.2 15 2.2 0.6 2.6 1.1 5.2 1.4 2.5 5.8 1.8 3.9 16 0.9 1.4 4.7 3.4 5.4 3.1 9.2 10.9 1.5 8.1 4.2 Diphthongs The experiment is performed for four diphthongs and the observations are made as follows. For /aw/ and /oy/, two channels 5 and 7 are found to have normalized energy level > 0.9. These channels only need to be stimulated for diphthong /aw/ and /oy/. The output speech spectra for /aw/ are illustrated in Figure 3b which shows only two peaks at center frequencies 570Hz and 840Hz that correspond to channels 5 and 7 respectively. Similarly, for /ay/, the normalized energy level > 0.9 is observed in channel 6 corresponding to a center frequency of 700Hz. For /ey/, it is channel 5, which correspond to a center frequency of 570Hz. 4.3 Nasal consonants For nasal consonants /m/ and /n/, channel 2 is found to have normalized energy level > 0.9. Therefore channel 2 needs to be stimulated for nasal consonants /m/ and /n/. The BPF output speech spectra of nasal consonant /m/ are shown in Figure 3c. A peak is observed in channel 2 which correspond to a frequency range of 225Hz-275Hz. 4.4 Voiced and unvoiced stop consonants The energy levels vary greatly with the frequency content in both voiced and unvoiced stop consonants. The experimental procedure is repeated for voiced (/b/, /d/, /g/) and unvoiced (/k/, /t/, /p/) stop consonants. The results of the significant energy levels are indicated across channels 5, 11 and 16 for /b/, channel 16 for /d/ and channels 10, 11 for /g/. For /k/, it is observed as channels 10 and 11; for /t/, it is channels 15 and 16 and for /p/, it is channels 10 and 14. The BPF output speech spectra of unvoiced stop consonant /t/ are shown in Figure 3d, which shows peaks at channels 15 and 16 that correspond to center frequencies 2900Hz and 3400Hz. 125
  • 11. International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN 0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME Fig. 3. Spectral analysis outputs showing the energy peaks for various phonemes (a) Two dominant peaks are observed in channels 5 and 6 for vowel /a/, for utterance 1 (b) Diphthong /aw/ exhibiting channels 5 and 7 (c) A peak in channel 2 for nasal consonant /m/ (d) A peak in channel 15 for unvoiced stop consonant /t/. From the CI acoustic simulation model, it is evident that, low frequencies are significant for vowels and high frequencies are significant for consonants. The events of the consonants are consistent across different subjects, despite that the parameters such as frequency and strength may change to a certain degree. The investigations further demonstrated that, a maximum of three channels are only required for the stimulation of cochlear electrodes for each of the phonemes listed in Table 2. Hence it is inferred that the spectral energies in all 16 channels are not required to be utilized for stimulation of cochlear electrodes. By selecting a few channels for stimulation, our proposed model would help in reducing inter channel interactions among the electrodes. The computation of spectral energies gives a measure of the number of electrodes required for stimulation. The identified channels through spectral analysis are subjected to temporal analysis, in order to get the amplitude information. 4.5 Electrical stimulation procedure The electrodes in the cochlear array are stimulated by means of biphasic pulses [29]. A biphasic pulse of 15µs width with a repetition rate of 250 µs which produces a stimulation rate of 4000 pulses per second are generated. The ‘place’ of stimulation in the cochlear electrode array is obtained by band energy levels in various channels estimated from the spectral analysis using the CI acoustic simulation model as discussed in section 2.2. The selected channels whose normalized energy >0.9 are subjected to temporal analysis. The amplitudes of the pulses are derived by extracting the envelopes of band passed filter outputs which are obtained by full-wave rectification and low-pass filtering as discussed in section 2.3. The amplitude estimates are useful in fixing the amplitude of the pulses. i.e, the pulse amplitudes are made proportional to the amplitude of the envelope output. Trains of biphasic pulses thus generated are to be delivered to the cochlear electrodes. In our study, the pulses are successfully generated for all the phonemes listed in Table 2, with amplitudes proportional to the outputs of selected channels. An illustration for vowel /a/ depicting the various stages of temporal outputs in channel 6 (having normalized energy >0.9) is shown in Figure 4(a-d), which correspond to the input 126
  • 12. International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN 0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME speech signal, band pass filtered output, rectified and low-pass filtered output, and the extracted envelope output. From this extracted envelope output, the stimulus pulse amplitude estimates are derived. Figure 4e shows the stimulation pulse energy levels for vowel /a/ in channel 6. Fig. 4. Various stages of temporal outputs in channel 6 for vowel /a/. (a) Input speech signal of vowel /a/ (b) Band pass filtered output in channel 6 (c) Illustration of the rectified and low-pass filtered outputs (d) the envelope tracing of the low pass filter (e) Amplitude level of biphasic stimulation pulses (not expanded). 5. CONCLUSION In this study, the performance of cochlear implant acoustic simulation model is discussed. We have designed a 16-channel FIR filter bank based on critical band spacing, using Kaiser Window. Higher stop band attenuation and exhibiting linear phase property are the two advantages of the critical band FIR filter design, which is utilized in the proposed CI-acoustic simulation model. This model estimates the spectral output using the designed CB-FIR filter bank and temporal output by envelope detection method through which the ‘place’ and ‘stimulus level’ information are obtained respectively. The spectral band energies of 15 speech components from the TIMIT data base are analyzed and the most significant channels are identified and selected, by computing the maximum energy values for the phonemes taken for the study. The selected channels are subjected to temporal analysis and the amplitudes are extracted by the envelope detection procedure. The simulation results of the selected channels 127
  • 13. International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN 0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME for different phonemes, demonstrated a good discrimination between vowels, diphthongs, voiced and unvoiced consonants. The other advantage of the proposed model is, it yields a way for minimizing the channel interactions as a reduced number of electrodes only are selected for stimulation. The overall results from the acoustic simulation experiment showed that the CB-FIR filter bank model based speech processor is useful in selecting the channels and estimating the required stimulation pulse level to be generated and presented to the cochlear electrodes. REFERENCES [1] Philipos C. Loizou (1999), “Introduction to Cochlear Implants”, IEEE Engineering in Medicine and Biology, pp. 32-42. [2] Kyung Hwan Kim, Sung Jen Choi, Jin Ho Kim, Doo Hee Kim (2009), “An Improved Speech Processing Strategy for Cochlear Implants based on an Active Non linear Filter Bank Model of the Biological Cochlea”, IEEE Transactions on biomedical engineering, 56(3): 828-836. [3] Lan Tian, Zhengxin Hou, Zhimim Pang (2007), “The Application and Computer Simulation of Multi-Channel Cochlear Implant based on all Phase DFT Filter”, ICASSP, Proceedings of the IEEE (IV), pp.701-704. [4] Francis A. Spelman (1999), “The Past, Present and Future of Cochlear Prostheses: Accomplishments and Challenges in treating Sensori-neural Deafness through Electrical Stimulation”, IEEE Engineering in Medicine and Biology, pp. 27-33. [5] Waldo Nogueira, Andras Katai, Tamas Harczos, Frank Klefenz, Andreas Buechner, Bernd Edler (2007), “An Auditory Model Based Strategy for Cochlear Implants”, Proceedings of the IEEE EMBS Conference, pp. 4127-4130. [6] Chris J. James, Peter J.Blamey, Lois Martin, Brett Swanson, Yvette Just, David Macfarlane (2002), “Adaptive Dynamic Range Optimization for Cochlear Implants: A Preliminary Study”, Ear and Hearing, Vol.23, pp. 49S-58S. [7] Fan-Gang Zeng, Stephen Rebscher, William Harrison, Xiaoan Sun, Haihong Feng (2008), “Cochlear Implants: System Design, Integration and Evaluation, IEEE Reviews in Biomedical Engineering”, Vol.1, pp. 115-142. [8] Eric D.Marsman, Gordon A.Carichner, Sundus Kubba, Richard B. Brown (2006), “A DSP Architecture of Cochlear Implants”, ISCAS, Proceedings of the IEEE, pp. 657- 666. [9] Talha J. Ahmad, Hussnain Ali, Muhammed Asim Ajaz and Shoab A.Khan (2009), “Efficient Algorithm Development of CIS Speech Processing Strategy for Cochlear Implants”, Proceedings of the IEEE EMBS Conference, pp. 1270-1273. [10] Talha Jamal Ahmad, Hussnain Ali, Muhammed Asim Ajaz, Shoab A.Khan (2009), “A DSK based Simplified Speech Processing Module for Cochlear Implant Research”, ICASSP, Proceedings of the IEEE, pp. 369-372. [11] Douglas Kim, Vanishree Gopalakrishna, Song Guo, Hoi Lee, Murat Torlak, Nasser Kehtarnavaz, Arthur lobo, Philipos C. Loizou (2011), “On the Design of a Flexible Stimulator for Animal Studies in Auditory Prostheses”, Biomedical Signal Processing and Control, Vol.6, pp. 346-355. [12] Ji-Jon Sit, Andrea M.Simonson, Andrew J.Oxenham, Michael A.Faltys, Rahul Sarpeshkar (2007), “A Low-Power Asynchronous Interleaved Sampling Algorithm for Cochlear Implants that Encodes Envelope and Phase Information”, IEEE Transactions on Biomedical Engineering, 54(1): 138-149. [13] L.Turicchia and R.Sarpeshkar (2005), “A Bio-Inspired Companding Strategy for Spectral Enhancement”, IEEE Transactions on Speech and Audioprocessing, 13(2): 243-253. [14] Rahul Sarpeshkar, Christopher Salthouse, Ji-Jon Sit, Michael W.Baker, Serhii M.Zark, Timothy K. T.Lorenzo Turicchia, Stephanie Balster (2005), “An Ultra-Low-Power 128
  • 14. International Journal of Electronics and Communication Engineering & Technology (IJECET), ISSN 0976 – 6464(Print), ISSN 0976 – 6472(Online) Volume 3, Issue 2, July-September (2012), © IAEME Programmable Analog Bionic Ear Processor”, IEEE transactions on biomedical engineering, 52(4): 711-727. [15] Lawrence Rabiner (2009), Biing – Hwang Juang, B. Yegnanarayana, “Fundamentals of Speech Recognition”, Pearson Education, Dorling Kindersley India. [16] Douglas O’Shaughnessy (2001), “Speech Communications: Human and Machine”, Second Edition, Universities Press, Hyderabad. [17] Joseph W. Picone (1993), “Signal Modeling Techniques in Speech Recognition”, Proceedings of the IEEE, Vol.81, No.9, pp. 1215-1247. [18] Ahmed Ben Hamida (1999), “An Adjustable Filter-Bank Based Algorithm for Hearing Aid Systems”, Proceedings of the IEEE, pp. 1187-1192. [19] Han Xian Hua, Nie Kaibao (2001), “Implementation of Spectral Maxima Sound Processing for Cochlear Implants by Using Bark Scale Frequency Band Partition”, Proceedings of the EMBS, pp. 2097-2101. [20] Rekha V. Dundur, M.V. Latte, S.Y. Kulkarni, M.K. Venkatesha (2008), “Digital Filter for Cochlear Implant Implemented on a Field-Programmable Gate Array”, Proceedings of World Academy of Science, Engineering and Technology, Vol.33, pp. 468-472. [21] S.S.Yin, S.C.Chan, X.M.Xie (2006), “On the Theory and Design of a Class of Recombination Non-uniform Filter Banks with Low-Delay FIR and IIR Filters”, ISCAS, Proceedings of the IEEE, pp. 895-898. [22] Yutaka Yamamoto, Brian D.O.Anderson, Masaaki Nagahara, Yoko Koyanagi (2003), “Optimizing FIR Approximation for Discrete-time IIR filters”, IEEE Signal Processing Letters, 10(9): 273-276. [23] A.Enis Cetin, Omer N.Gerek, Yasemin Yardimci (1997), “Equiripple FIR Filter Design by the FFT Algorithm”, IEEE Signal Processing Magazine, pp. 60-64. [24] Sanjit K.Mitra (2008), “Digital Signal Processing:A Computer-based approach”, Third Edition, Tata McGraw-Hill Publishing Company Limited, Newyork. [25] Kemal Avci, Arif Nacaroglu (2009), “Cosh Window Family and its Application to FIR Filter Design”, Int. J.Electron. Commun (AEU), Vol.63,pp. 907-916. [26] Tapio Saramaki (1989), “A Class of Window Functions with Nearly Minimum Sidelobe Energy for Designing FIR Filters”, Proceedings of the ISCAS, pp. 359-362. [27] P. Mahalakshmi, M R Reddy (2010), “Signal Analysis by Using FIR Filter Banks in Cochlear Implant Prostheses”, Proceedings of the IEEE International Conference on Systems in Medicine and Biology, No.89, pp. 253-258. [28] Rangaraj M. Rangayyan (2002), “Biomedical Signal Analysis: A Case Study Approach”, John Wiley & Sons Inc., Newyork. [29] Philipos C. Loizou (1999), “Signal Processing Techniques for Cochlear Implants: A Review of Progress in Deriving Electrical Stimuli From the Speech Signals”, IEEE Engineering in Medicine and Biology, pp. 34-45. [30] Thomas F. Quatieri (2004), “Discrete-time Speech Signal Processing: Principles and Practice”, Pearson Education, India. 129