SlideShare una empresa de Scribd logo
1 de 31
CHANNEL EQUALISATION
           BY-   AJIT KUMAR PANDA
                 POONAN SAHOO
                 SAYANTAN DAS
                 SURAJ CHOUDHURY
THREATS IN DIGITAL
COMMUNICATION

 •There are four main threats in the process of
 digital communication

 Inter Symbol Interference (ISI)

 Multipath Propagation

 Co-channel Interference

 Presence of noise in the channel
INTER SYMBOL INTERFERENCE:
 Inter Symbol Interference
    in Digital Transmission

 Inter-symbol interference
    (ISI) arises when the data
    transmitted through the
    channel is dispersive, in
    which each received pulse
    is affected somewhat by
    adjacent pulses and due to
    which interference occurs
    in the transmitted signals.

   It is difficult to recover the
    original data from one
    channel sample
CO-CHANNEL INTERFERENCE:
 Co-channel Interference (CCI) and
  Adjacent Channel Interference
  (ACI) occur in communication
  systems due to multiple access
  techniques using space, frequency or
  time.


 CCI occurs in cellular radio and dual-
  polarized microwave radio, for
  efficient utilization of the allocated
  channels frequencies by reusing the
  frequencies in different cells.
MULTI-PATH PROPAGATION:
   Within telecommunication channels multiple paths of propagation commonly occur.
    In practical terms this is equivalent to transmitting the same signal through a number
    of separate channels, each having a different attenuation and delay.
    Consider an open-air radio transmission channel that has three propagation
    paths, as illustrated in Fig. These could be
     -     Direct
    -      Earth Bound
    -      Sky Bound
   Fig1.2b describes how a receiver picks up the
     transmitted data. The direct signal is received
    first whilst the earth and sky bound are delayed.
    All three of the signals are attenuated with the sky path suffering the most.
   Multipath interference between consecutively transmitted signals will take place if
    one signal is received whilst the previous signal is still being detected.
    In Fig1.2 this would occur if the symbol transmission rate is greater than 1/τ where, τ
    represents transmission delay. Because bandwidth efficiency leads to high data
    rates, multi-path interference commonly occurs.
EQUALIZER
 Equalization is the process to remove ISI and noise effects from the
    channel
   It is located at the receiver end of the channel




 It is an inverse filter placed at the front end of the receiver
 The transfer function of the equalizer is just inverse of the transfer
    function of the channel
 Equalization is an iterative process of reducing the mean square
    error the difference between desired response and output of filter
    used in equalizer
TYPES OF EQUALIZERS:

 Equalizers are of two types

                                  LINEAR
                EQUALIZERS
                                NON LINEAR


 Linear equalizers aim at reducing ISI in linear channels
  using various algorithms like Least Mean
  Square(LMS), Recursive Least Square(RLS) and
  normalized LMS
 Non linear equalizers equalize non-linear channels. They
  mainly use Neural Networks(NN) and Multilayer
  Perception(MLP) algorithms for equalization
Linear Adaptive Filters:




 An adaptive filter is a computational device that
  attempts to model the relationship between two signals
  in real time in an iterative manner
 Here the output is compared to the desired signal and
  accordingly the parameters of adaptive filter are varied
  and so it is known as self designing filter.
Applications of Adaptive Filters:
Identification
 Used to provide a linear model of an unknown plant




 Parameters
       u=input of adaptive filter=input to plant
       y=output of adaptive filter
       d=desired response=output of plant
       ---> e=d-y=estimation error
          Applications:
          System identification
Applications of Adaptive Filters:
Inverse Modeling
     Used to provide an inverse model of an unknown plant




 Parameters
        u=input of adaptive filter=output to plant
        y=output of adaptive filter
        d=desired response=delayed system input
        e=d-y=estimation error
         Applications:
              Channel Equalization
The channel equalization
model:
Stochastic Gradient
 Approach:
    Most commonly used type of Adaptive Filters
    Define cost function as mean-squared error
          Difference between filter output and desired response
    Based on the method of steepest descent
        Move towards the minimum on the error surface to get to minimum
        Requires the gradient of the error surface to be known
    Most popular adaptation algorithm is LMS
        Derived from steepest descent
        Doesn’t require gradient to be know: it is estimated at every iteration
    Least-Mean-Square (LMS) Algorithm

update value             old value                   learning -         tap
                                                                                   error
of tap - weigth          of tap - weight             rate              input
                                                                                   signal
vector                   vector                      parameter vector
LMS algorithm
•   Introduced by Widrow & Hoff in 1959
•   Simple, no matrices calculation involved in the adaptation
•   In the family of stochastic gradient algorithms
•   Approximation of the steepest – descent method
•   Based on the MMSE criterion.(Minimum Mean square Error)
•   Adaptive process containing two important signals:
•       1.) Filtering process, producing output signal.
•       2.) Desired signal (Training sequence)
•   Adaptive process: recursive adjustment of filter tap
    weights
Least-Mean-Square (LMS)
Algorithm continued....
 The LMS Algorithm consists of two basic processes that is
  followed in the adaptive equalization processes:
    Training : It refers to adapting to the training sequence
    Tracking: keeps track of the changing characteristics of the
     channel.
LMS Algorithm Steps:
                       M 1
 Filter output   zn                *
                             u n k wk n
                       k 0


 Estimation error     en      dn     zn

                               wk n    1   wk n   un   k e* n
 Tap-weight adaptation
Derivation of the LMS MSE
expression:
 Error=E=(x(n)-x(n)’)
 Square error=E=(x(n)-x(n)’)2
 Using minimum mean square error criterion , we differentiate the
  expression
        dE/dw=d/dw((x(n)-x(n)’)2)
 Applying chain rule and substitution of x(n)’ ,we get
      dE/dw=2(x(n)-x(n)’)*d/dw(x(n)- Ʃw*s(n-i))
      dE/dw=2(e(n))(s(n-i))
 From this we can derive an update equation for every new
  sample n using steepest descent and gradient method as
          w(n+1)= -u*(dE/dw)
     so,w(n+1)=2*u*e(n)*s(n-i)
          for i=0,1,2,3...........
Stability of LMS:
 The LMS algorithm is convergent in the mean square if and only
   if the step-size parameter satisfy

                                     1
                          0
                                     m ax


  Here max is the largest eigen value of the correlation matrix of
  the input data.
                                                          1
 More practical test for stability is    0
                                                        input signal power

 The value of step size has to be a trade off between fast
  convergence rates and less steady state misadjustment.
    Larger values for step size
           Increases adaptation rate (faster adaptation)
           Increases residual mean-squared error
LMS-Pros & cons:

 LMS – Advantage:
    • Simplicity of implementation
    • Not neglecting the noise like Zero forcing equalizer
    • Stable and robust performance against different signal
      conditions

 LMS – Disadvantage:
     Slow Convergence

     Demands using of training sequence as reference ,thus
      decreasing the communication BW.
NLMS-Normalised LMS
algorithm
 Is mainly required to provide better performance than LMS
    as LMS performance is slow
   Uses normalization technique to provide a variable step size
     as step size ‘u’ is divided by instantaneous signal power thus
    providing more stability and faster convergence.
   Is equivalent to running the LMS recursion for a new sample of
    inputs every time recursion or the NLMS operation is carried out.
    W(n+1) = w(n) + (1/xT(n)x(n)) * e(n) x(n)
   The step size value for the input vector is calculated
    µ (n) = 1/xT(n)x(n)
   The filter tap weights are updated in preparation for the next
    iteration
    W(n+1) = w(n) + 2*µ (n) * e(n) * x(n)
Results for LMS algorithm:




 Convergence is faster with increased step size .
  Plot is for noise=30 dB
Results for NLMS algorithm:




 Convergence is faster in case of NLMS
  algorithm
 It provides a more stable output.
NON-LINEAR CHANNEL
EQUALISATION
Need For Non-Linear
Equalizer:
 Linear Equalizers do not perform well on channels
  having deep spectral nulls in the pass band.

 To compensate distortion linear equalizer places too
  much gain in the vicinity of spectral nulls thereby
  enhancing the noise present in these frequencies.

 BER is better in Non-linear channel equalizer

 Linear equalizer-inverse problem
   Non-linear equalizer-pattern classification
Non-Linear Channel
Equalizer:
  t k denotes a sequence of
  T spaced complex symbols
  of an BPSK
  constellation, where 1/T
  denotes the symbol rate and
  k denotes the discrete time
  index.
 A widely used model for a
  linear dispersive channel is
  an FIR filter whose output at
        th
  the kN instant is given by
         h-1
                                  Schematic Diagram of a Non-Linear
                                  Wireless Digital Communication system
        ∑
    ak= i=0 hi * t k-i            with channel equalizer
Continued…
   where
            hi- denotes the FIR filter weights
            Nh- denotes the FIR order.

   Considering the channel to be a nonlinear one the NL block introduces
    channel nonlinearity to the filter output.

   The transmitted signal t k after being passed through the nonlinear
    channel and added with the additive noise arrives at the receiver, which
    is denoted by r k .The received signal at the kth time instant is given by r k.

   The purpose of equalizer attached at the receiver front end is to recover
    the transmitted sequence t k or its delayed version t k-1 ,where t is the
    propagation delay associated with the physical channel.
Neural Network:
 Started in 1800s as an effort to
  describe how human mind
  performs.

 It is applied to computational
  models with Turing ‘s B-type
  machine and Perceptron

 A neural network is a massively
  parallel distributed processor made
  up of simple processing units, which
  has a natural propensity for storing
  experimental knowledge and making
  it available for use.
Continued:
   Today in general form a neural network is a
    machine that is designed by using electronic
    components or is simulated in software on
    a digital computer.

   To achieve good performance, neural
    networks employ a massive interconnection of
    simple computing cells referred to as
    “Neurons” or “processing units”

   The procedure is called a learning
    algorithm, the function of which is to modify the
    synaptic weights of the network in an orderly
    fashion to attain a desired design objective.
    McCulloch and Pitts have developed the neural
    networks for different computing machines.

Artificial Neural Network:
 Artificial Neural Network (ANN)
  have become a powerful tool for
  many complex applications including
  functional approximation, nonlinear
  system identification, motor
  control, pattern recognition, adaptive
  channel equalization and
  optimization.

 ANN is capable of performing
  nonlinear mapping between the
  input and output space due to its
  large parallel interconnection
  between different layers and the
  nonlinear processing characteristics.
Continued:
 An artificial neuron basically consists of a
  computing element that performs the
  weighted sum of the input signal and the
  connecting weight. The weighted sum is
  added with the bias called threshold and
  the resultant signal is passed through a
  nonlinear activation function. Common
  types of activation functions are sigmoid and
  hyperbolic tangent.
 Each neuron is associated with three
  parameters whose learning can be adjusted.
  These are the connecting weights, the bias
  and the slope of the nonlinear function.
 For the structural point of view a NN may
  be single layer or it may be multilayer
Multi-layer Perceptron:
 The perceptron is a single level
  connection of McCulloch-Pitts
  neurons is called as Single-layer
  feed forward networks.
 The network is capable of
  linearly separating the input
  vectors into pattern of classes by
  a hyper plane. Similarly many
  perceptrons can be connected in
  layers
 To provide a MLP network, the
  input signal propagates through
  the network in a forward
  direction, on a layer-by-layer
  basis. This network has been
  applied successfully to solve
  diverse problems.
Continued…
 Generally MLP is trained using popular error back-
  propagation algorithm.
 The scheme of MLP using four layers is shown. Si
  represent the inputs s1, s2, ….. , sn to the network, and yk
  represents the output of the final layer of the neural
  network.
 The connecting weights between the input to the first
  hidden layer, first to second hidden layer and the second
  hidden layer to the output layers are represented by W i
  ,W ji ,W kj respectively.
 The final output layer of the MLP may be expressed as

Más contenido relacionado

La actualidad más candente

Adaptive equalization
Adaptive equalizationAdaptive equalization
Adaptive equalizationKamal Bhatt
 
Pulse modulation
Pulse modulationPulse modulation
Pulse modulationstk_gpg
 
Matched filter detection
Matched filter detectionMatched filter detection
Matched filter detectionSURYA DEEPAK
 
7. log distance and log normal shadowing
7. log distance and log normal shadowing7. log distance and log normal shadowing
7. log distance and log normal shadowingJAIGANESH SEKAR
 
Lecture Notes on Adaptive Signal Processing-1.pdf
Lecture Notes on Adaptive Signal Processing-1.pdfLecture Notes on Adaptive Signal Processing-1.pdf
Lecture Notes on Adaptive Signal Processing-1.pdfVishalPusadkar1
 
Digital communications
Digital communicationsDigital communications
Digital communicationsAllanki Rao
 
M ary psk modulation
M ary psk modulationM ary psk modulation
M ary psk modulationAhmed Diaa
 
Companding & Pulse Code Modulation
Companding & Pulse Code ModulationCompanding & Pulse Code Modulation
Companding & Pulse Code ModulationYeshudas Muttu
 
4.5 equalizers and its types
4.5   equalizers and its types4.5   equalizers and its types
4.5 equalizers and its typesJAIGANESH SEKAR
 
Pulse amplitude modulation
Pulse amplitude modulationPulse amplitude modulation
Pulse amplitude modulationVishal kakade
 
Digital modulation techniques
Digital modulation techniquesDigital modulation techniques
Digital modulation techniquessrkrishna341
 
Pulse amplitude modulation (PAM)
Pulse amplitude modulation (PAM)Pulse amplitude modulation (PAM)
Pulse amplitude modulation (PAM)Mugadha Bane
 
Linear prediction
Linear predictionLinear prediction
Linear predictionUma Rajaram
 

La actualidad más candente (20)

Adaptive equalization
Adaptive equalizationAdaptive equalization
Adaptive equalization
 
Pulse modulation
Pulse modulationPulse modulation
Pulse modulation
 
quantization
quantizationquantization
quantization
 
Matched filter detection
Matched filter detectionMatched filter detection
Matched filter detection
 
7. log distance and log normal shadowing
7. log distance and log normal shadowing7. log distance and log normal shadowing
7. log distance and log normal shadowing
 
Pcm
PcmPcm
Pcm
 
Lecture Notes on Adaptive Signal Processing-1.pdf
Lecture Notes on Adaptive Signal Processing-1.pdfLecture Notes on Adaptive Signal Processing-1.pdf
Lecture Notes on Adaptive Signal Processing-1.pdf
 
Digital communications
Digital communicationsDigital communications
Digital communications
 
Design of Filters PPT
Design of Filters PPTDesign of Filters PPT
Design of Filters PPT
 
Windowing ofdm
Windowing ofdmWindowing ofdm
Windowing ofdm
 
M ary psk modulation
M ary psk modulationM ary psk modulation
M ary psk modulation
 
Companding & Pulse Code Modulation
Companding & Pulse Code ModulationCompanding & Pulse Code Modulation
Companding & Pulse Code Modulation
 
4.5 equalizers and its types
4.5   equalizers and its types4.5   equalizers and its types
4.5 equalizers and its types
 
Pulse amplitude modulation
Pulse amplitude modulationPulse amplitude modulation
Pulse amplitude modulation
 
Digital modulation techniques
Digital modulation techniquesDigital modulation techniques
Digital modulation techniques
 
Matched filter
Matched filterMatched filter
Matched filter
 
Pulse amplitude modulation (PAM)
Pulse amplitude modulation (PAM)Pulse amplitude modulation (PAM)
Pulse amplitude modulation (PAM)
 
Linear prediction
Linear predictionLinear prediction
Linear prediction
 
Frequency translation
Frequency translationFrequency translation
Frequency translation
 
Digital communication unit 1
Digital communication unit 1Digital communication unit 1
Digital communication unit 1
 

Similar a Channel Equalisation

Equalization with the help of Non OMA.pptx
Equalization with the help of Non OMA.pptxEqualization with the help of Non OMA.pptx
Equalization with the help of Non OMA.pptxUditJain156267
 
Eqalization and diversity
Eqalization and diversityEqalization and diversity
Eqalization and diversityerindrasen
 
Introduction to adaptive filtering and its applications.ppt
Introduction to adaptive filtering and its applications.pptIntroduction to adaptive filtering and its applications.ppt
Introduction to adaptive filtering and its applications.pptdebeshidutta2
 
Performance Comparison of Modified Variable Step Size Leaky LMS Algorithm for...
Performance Comparison of Modified Variable Step Size Leaky LMS Algorithm for...Performance Comparison of Modified Variable Step Size Leaky LMS Algorithm for...
Performance Comparison of Modified Variable Step Size Leaky LMS Algorithm for...ijcnac
 
Adaptive Channel Equalization for Nonlinear Channels using Signed Regressor F...
Adaptive Channel Equalization for Nonlinear Channels using Signed Regressor F...Adaptive Channel Equalization for Nonlinear Channels using Signed Regressor F...
Adaptive Channel Equalization for Nonlinear Channels using Signed Regressor F...IDES Editor
 
A Decisive Filtering Selection Approach For Improved Performance Active Noise...
A Decisive Filtering Selection Approach For Improved Performance Active Noise...A Decisive Filtering Selection Approach For Improved Performance Active Noise...
A Decisive Filtering Selection Approach For Improved Performance Active Noise...IOSR Journals
 
_Pulse-Modulation-Techniqnes.pdf
_Pulse-Modulation-Techniqnes.pdf_Pulse-Modulation-Techniqnes.pdf
_Pulse-Modulation-Techniqnes.pdfSoyallRobi
 
BER Analysis ofImpulse Noise inOFDM System Using LMS,NLMS&RLS
BER Analysis ofImpulse Noise inOFDM System Using LMS,NLMS&RLSBER Analysis ofImpulse Noise inOFDM System Using LMS,NLMS&RLS
BER Analysis ofImpulse Noise inOFDM System Using LMS,NLMS&RLSiosrjce
 
PONDICHERRY UNIVERSITY DEPARTMENT OF ELECTRONICS ENGINEERING.pdf
PONDICHERRY UNIVERSITY DEPARTMENT OF ELECTRONICS ENGINEERING.pdfPONDICHERRY UNIVERSITY DEPARTMENT OF ELECTRONICS ENGINEERING.pdf
PONDICHERRY UNIVERSITY DEPARTMENT OF ELECTRONICS ENGINEERING.pdfAWANISHKUMAR84
 
Comparison of different Sub-Band Adaptive Noise Canceller with LMS and RLS
Comparison of different Sub-Band Adaptive Noise Canceller with LMS and RLSComparison of different Sub-Band Adaptive Noise Canceller with LMS and RLS
Comparison of different Sub-Band Adaptive Noise Canceller with LMS and RLSijsrd.com
 
Effects of Weight Approximation Methods on Performance of Digital Beamforming...
Effects of Weight Approximation Methods on Performance of Digital Beamforming...Effects of Weight Approximation Methods on Performance of Digital Beamforming...
Effects of Weight Approximation Methods on Performance of Digital Beamforming...IOSR Journals
 
Ch4 1 Data communication and networking by neha g. kurale
Ch4 1 Data communication and networking by neha g. kuraleCh4 1 Data communication and networking by neha g. kurale
Ch4 1 Data communication and networking by neha g. kuraleNeha Kurale
 
Low power vlsi implementation adaptive noise cancellor based on least means s...
Low power vlsi implementation adaptive noise cancellor based on least means s...Low power vlsi implementation adaptive noise cancellor based on least means s...
Low power vlsi implementation adaptive noise cancellor based on least means s...shaik chand basha
 
Noice canclellation using adaptive filters with adpative algorithms(LMS,NLMS,...
Noice canclellation using adaptive filters with adpative algorithms(LMS,NLMS,...Noice canclellation using adaptive filters with adpative algorithms(LMS,NLMS,...
Noice canclellation using adaptive filters with adpative algorithms(LMS,NLMS,...Brati Sundar Nanda
 

Similar a Channel Equalisation (20)

Unit iv wcn main
Unit iv wcn mainUnit iv wcn main
Unit iv wcn main
 
Adaptive Equalization
Adaptive EqualizationAdaptive Equalization
Adaptive Equalization
 
Equalization with the help of Non OMA.pptx
Equalization with the help of Non OMA.pptxEqualization with the help of Non OMA.pptx
Equalization with the help of Non OMA.pptx
 
Eqalization and diversity
Eqalization and diversityEqalization and diversity
Eqalization and diversity
 
Introduction to adaptive filtering and its applications.ppt
Introduction to adaptive filtering and its applications.pptIntroduction to adaptive filtering and its applications.ppt
Introduction to adaptive filtering and its applications.ppt
 
Performance Comparison of Modified Variable Step Size Leaky LMS Algorithm for...
Performance Comparison of Modified Variable Step Size Leaky LMS Algorithm for...Performance Comparison of Modified Variable Step Size Leaky LMS Algorithm for...
Performance Comparison of Modified Variable Step Size Leaky LMS Algorithm for...
 
Adaptive Filtering.ppt
Adaptive Filtering.pptAdaptive Filtering.ppt
Adaptive Filtering.ppt
 
Adaptive Channel Equalization for Nonlinear Channels using Signed Regressor F...
Adaptive Channel Equalization for Nonlinear Channels using Signed Regressor F...Adaptive Channel Equalization for Nonlinear Channels using Signed Regressor F...
Adaptive Channel Equalization for Nonlinear Channels using Signed Regressor F...
 
A Decisive Filtering Selection Approach For Improved Performance Active Noise...
A Decisive Filtering Selection Approach For Improved Performance Active Noise...A Decisive Filtering Selection Approach For Improved Performance Active Noise...
A Decisive Filtering Selection Approach For Improved Performance Active Noise...
 
_Pulse-Modulation-Techniqnes.pdf
_Pulse-Modulation-Techniqnes.pdf_Pulse-Modulation-Techniqnes.pdf
_Pulse-Modulation-Techniqnes.pdf
 
I017325055
I017325055I017325055
I017325055
 
BER Analysis ofImpulse Noise inOFDM System Using LMS,NLMS&RLS
BER Analysis ofImpulse Noise inOFDM System Using LMS,NLMS&RLSBER Analysis ofImpulse Noise inOFDM System Using LMS,NLMS&RLS
BER Analysis ofImpulse Noise inOFDM System Using LMS,NLMS&RLS
 
LMS .pdf
LMS .pdfLMS .pdf
LMS .pdf
 
PONDICHERRY UNIVERSITY DEPARTMENT OF ELECTRONICS ENGINEERING.pdf
PONDICHERRY UNIVERSITY DEPARTMENT OF ELECTRONICS ENGINEERING.pdfPONDICHERRY UNIVERSITY DEPARTMENT OF ELECTRONICS ENGINEERING.pdf
PONDICHERRY UNIVERSITY DEPARTMENT OF ELECTRONICS ENGINEERING.pdf
 
Comparison of different Sub-Band Adaptive Noise Canceller with LMS and RLS
Comparison of different Sub-Band Adaptive Noise Canceller with LMS and RLSComparison of different Sub-Band Adaptive Noise Canceller with LMS and RLS
Comparison of different Sub-Band Adaptive Noise Canceller with LMS and RLS
 
Effects of Weight Approximation Methods on Performance of Digital Beamforming...
Effects of Weight Approximation Methods on Performance of Digital Beamforming...Effects of Weight Approximation Methods on Performance of Digital Beamforming...
Effects of Weight Approximation Methods on Performance of Digital Beamforming...
 
Ch4 1 Data communication and networking by neha g. kurale
Ch4 1 Data communication and networking by neha g. kuraleCh4 1 Data communication and networking by neha g. kurale
Ch4 1 Data communication and networking by neha g. kurale
 
Low power vlsi implementation adaptive noise cancellor based on least means s...
Low power vlsi implementation adaptive noise cancellor based on least means s...Low power vlsi implementation adaptive noise cancellor based on least means s...
Low power vlsi implementation adaptive noise cancellor based on least means s...
 
Noice canclellation using adaptive filters with adpative algorithms(LMS,NLMS,...
Noice canclellation using adaptive filters with adpative algorithms(LMS,NLMS,...Noice canclellation using adaptive filters with adpative algorithms(LMS,NLMS,...
Noice canclellation using adaptive filters with adpative algorithms(LMS,NLMS,...
 
Pulse Code Modulation
Pulse Code ModulationPulse Code Modulation
Pulse Code Modulation
 

Último

TrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law DevelopmentsTrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law DevelopmentsTrustArc
 
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, AdobeApidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobeapidays
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024The Digital Insurer
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerThousandEyes
 
Manulife - Insurer Innovation Award 2024
Manulife - Insurer Innovation Award 2024Manulife - Insurer Innovation Award 2024
Manulife - Insurer Innovation Award 2024The Digital Insurer
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024The Digital Insurer
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024The Digital Insurer
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationSafe Software
 
Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024The Digital Insurer
 
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot TakeoffStrategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoffsammart93
 
Boost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfBoost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfsudhanshuwaghmare1
 
AWS Community Day CPH - Three problems of Terraform
AWS Community Day CPH - Three problems of TerraformAWS Community Day CPH - Three problems of Terraform
AWS Community Day CPH - Three problems of TerraformAndrey Devyatkin
 
Top 5 Benefits OF Using Muvi Live Paywall For Live Streams
Top 5 Benefits OF Using Muvi Live Paywall For Live StreamsTop 5 Benefits OF Using Muvi Live Paywall For Live Streams
Top 5 Benefits OF Using Muvi Live Paywall For Live StreamsRoshan Dwivedi
 
Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024The Digital Insurer
 
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...apidays
 
MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024MIND CTI
 
Strategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherStrategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherRemote DBA Services
 
Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsJoaquim Jorge
 
Automating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps ScriptAutomating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps Scriptwesley chun
 

Último (20)

TrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law DevelopmentsTrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
 
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, AdobeApidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
Manulife - Insurer Innovation Award 2024
Manulife - Insurer Innovation Award 2024Manulife - Insurer Innovation Award 2024
Manulife - Insurer Innovation Award 2024
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
 
Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024
 
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot TakeoffStrategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
 
Boost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfBoost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdf
 
AWS Community Day CPH - Three problems of Terraform
AWS Community Day CPH - Three problems of TerraformAWS Community Day CPH - Three problems of Terraform
AWS Community Day CPH - Three problems of Terraform
 
Top 5 Benefits OF Using Muvi Live Paywall For Live Streams
Top 5 Benefits OF Using Muvi Live Paywall For Live StreamsTop 5 Benefits OF Using Muvi Live Paywall For Live Streams
Top 5 Benefits OF Using Muvi Live Paywall For Live Streams
 
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
 
Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024
 
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
 
MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024
 
Strategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherStrategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a Fresher
 
Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and Myths
 
Automating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps ScriptAutomating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps Script
 

Channel Equalisation

  • 1. CHANNEL EQUALISATION BY- AJIT KUMAR PANDA POONAN SAHOO SAYANTAN DAS SURAJ CHOUDHURY
  • 2. THREATS IN DIGITAL COMMUNICATION •There are four main threats in the process of digital communication Inter Symbol Interference (ISI) Multipath Propagation Co-channel Interference Presence of noise in the channel
  • 3. INTER SYMBOL INTERFERENCE:  Inter Symbol Interference in Digital Transmission  Inter-symbol interference (ISI) arises when the data transmitted through the channel is dispersive, in which each received pulse is affected somewhat by adjacent pulses and due to which interference occurs in the transmitted signals.  It is difficult to recover the original data from one channel sample
  • 4. CO-CHANNEL INTERFERENCE:  Co-channel Interference (CCI) and Adjacent Channel Interference (ACI) occur in communication systems due to multiple access techniques using space, frequency or time.  CCI occurs in cellular radio and dual- polarized microwave radio, for efficient utilization of the allocated channels frequencies by reusing the frequencies in different cells.
  • 5. MULTI-PATH PROPAGATION:  Within telecommunication channels multiple paths of propagation commonly occur. In practical terms this is equivalent to transmitting the same signal through a number of separate channels, each having a different attenuation and delay.  Consider an open-air radio transmission channel that has three propagation paths, as illustrated in Fig. These could be - Direct - Earth Bound - Sky Bound  Fig1.2b describes how a receiver picks up the transmitted data. The direct signal is received first whilst the earth and sky bound are delayed.  All three of the signals are attenuated with the sky path suffering the most.  Multipath interference between consecutively transmitted signals will take place if one signal is received whilst the previous signal is still being detected.  In Fig1.2 this would occur if the symbol transmission rate is greater than 1/τ where, τ represents transmission delay. Because bandwidth efficiency leads to high data rates, multi-path interference commonly occurs.
  • 6. EQUALIZER  Equalization is the process to remove ISI and noise effects from the channel  It is located at the receiver end of the channel  It is an inverse filter placed at the front end of the receiver  The transfer function of the equalizer is just inverse of the transfer function of the channel  Equalization is an iterative process of reducing the mean square error the difference between desired response and output of filter used in equalizer
  • 7. TYPES OF EQUALIZERS:  Equalizers are of two types LINEAR EQUALIZERS NON LINEAR  Linear equalizers aim at reducing ISI in linear channels using various algorithms like Least Mean Square(LMS), Recursive Least Square(RLS) and normalized LMS  Non linear equalizers equalize non-linear channels. They mainly use Neural Networks(NN) and Multilayer Perception(MLP) algorithms for equalization
  • 8. Linear Adaptive Filters:  An adaptive filter is a computational device that attempts to model the relationship between two signals in real time in an iterative manner  Here the output is compared to the desired signal and accordingly the parameters of adaptive filter are varied and so it is known as self designing filter.
  • 9. Applications of Adaptive Filters: Identification  Used to provide a linear model of an unknown plant  Parameters  u=input of adaptive filter=input to plant  y=output of adaptive filter  d=desired response=output of plant  ---> e=d-y=estimation error  Applications:  System identification
  • 10. Applications of Adaptive Filters: Inverse Modeling  Used to provide an inverse model of an unknown plant  Parameters  u=input of adaptive filter=output to plant  y=output of adaptive filter  d=desired response=delayed system input  e=d-y=estimation error  Applications:  Channel Equalization
  • 12. Stochastic Gradient Approach:  Most commonly used type of Adaptive Filters  Define cost function as mean-squared error  Difference between filter output and desired response  Based on the method of steepest descent  Move towards the minimum on the error surface to get to minimum  Requires the gradient of the error surface to be known  Most popular adaptation algorithm is LMS  Derived from steepest descent  Doesn’t require gradient to be know: it is estimated at every iteration  Least-Mean-Square (LMS) Algorithm update value old value learning - tap error of tap - weigth of tap - weight rate input signal vector vector parameter vector
  • 13. LMS algorithm • Introduced by Widrow & Hoff in 1959 • Simple, no matrices calculation involved in the adaptation • In the family of stochastic gradient algorithms • Approximation of the steepest – descent method • Based on the MMSE criterion.(Minimum Mean square Error) • Adaptive process containing two important signals: • 1.) Filtering process, producing output signal. • 2.) Desired signal (Training sequence) • Adaptive process: recursive adjustment of filter tap weights
  • 14. Least-Mean-Square (LMS) Algorithm continued....  The LMS Algorithm consists of two basic processes that is followed in the adaptive equalization processes:  Training : It refers to adapting to the training sequence  Tracking: keeps track of the changing characteristics of the channel.
  • 15. LMS Algorithm Steps: M 1  Filter output zn * u n k wk n k 0  Estimation error en dn zn wk n 1 wk n un k e* n  Tap-weight adaptation
  • 16. Derivation of the LMS MSE expression:  Error=E=(x(n)-x(n)’)  Square error=E=(x(n)-x(n)’)2  Using minimum mean square error criterion , we differentiate the expression dE/dw=d/dw((x(n)-x(n)’)2)  Applying chain rule and substitution of x(n)’ ,we get dE/dw=2(x(n)-x(n)’)*d/dw(x(n)- Ʃw*s(n-i)) dE/dw=2(e(n))(s(n-i))  From this we can derive an update equation for every new sample n using steepest descent and gradient method as w(n+1)= -u*(dE/dw) so,w(n+1)=2*u*e(n)*s(n-i) for i=0,1,2,3...........
  • 17. Stability of LMS:  The LMS algorithm is convergent in the mean square if and only if the step-size parameter satisfy 1 0 m ax Here max is the largest eigen value of the correlation matrix of the input data. 1  More practical test for stability is 0 input signal power  The value of step size has to be a trade off between fast convergence rates and less steady state misadjustment.  Larger values for step size  Increases adaptation rate (faster adaptation)  Increases residual mean-squared error
  • 18. LMS-Pros & cons:  LMS – Advantage: • Simplicity of implementation • Not neglecting the noise like Zero forcing equalizer • Stable and robust performance against different signal conditions  LMS – Disadvantage:  Slow Convergence  Demands using of training sequence as reference ,thus decreasing the communication BW.
  • 19. NLMS-Normalised LMS algorithm  Is mainly required to provide better performance than LMS as LMS performance is slow  Uses normalization technique to provide a variable step size as step size ‘u’ is divided by instantaneous signal power thus providing more stability and faster convergence.  Is equivalent to running the LMS recursion for a new sample of inputs every time recursion or the NLMS operation is carried out. W(n+1) = w(n) + (1/xT(n)x(n)) * e(n) x(n)  The step size value for the input vector is calculated µ (n) = 1/xT(n)x(n)  The filter tap weights are updated in preparation for the next iteration W(n+1) = w(n) + 2*µ (n) * e(n) * x(n)
  • 20. Results for LMS algorithm:  Convergence is faster with increased step size . Plot is for noise=30 dB
  • 21. Results for NLMS algorithm:  Convergence is faster in case of NLMS algorithm  It provides a more stable output.
  • 23. Need For Non-Linear Equalizer:  Linear Equalizers do not perform well on channels having deep spectral nulls in the pass band.  To compensate distortion linear equalizer places too much gain in the vicinity of spectral nulls thereby enhancing the noise present in these frequencies.  BER is better in Non-linear channel equalizer  Linear equalizer-inverse problem Non-linear equalizer-pattern classification
  • 24. Non-Linear Channel Equalizer:  t k denotes a sequence of T spaced complex symbols of an BPSK constellation, where 1/T denotes the symbol rate and k denotes the discrete time index.  A widely used model for a linear dispersive channel is an FIR filter whose output at th the kN instant is given by h-1 Schematic Diagram of a Non-Linear Wireless Digital Communication system ∑ ak= i=0 hi * t k-i with channel equalizer
  • 25. Continued…  where hi- denotes the FIR filter weights Nh- denotes the FIR order.  Considering the channel to be a nonlinear one the NL block introduces channel nonlinearity to the filter output.  The transmitted signal t k after being passed through the nonlinear channel and added with the additive noise arrives at the receiver, which is denoted by r k .The received signal at the kth time instant is given by r k.  The purpose of equalizer attached at the receiver front end is to recover the transmitted sequence t k or its delayed version t k-1 ,where t is the propagation delay associated with the physical channel.
  • 26. Neural Network:  Started in 1800s as an effort to describe how human mind performs.  It is applied to computational models with Turing ‘s B-type machine and Perceptron  A neural network is a massively parallel distributed processor made up of simple processing units, which has a natural propensity for storing experimental knowledge and making it available for use.
  • 27. Continued:  Today in general form a neural network is a machine that is designed by using electronic components or is simulated in software on a digital computer.  To achieve good performance, neural networks employ a massive interconnection of simple computing cells referred to as “Neurons” or “processing units”  The procedure is called a learning algorithm, the function of which is to modify the synaptic weights of the network in an orderly fashion to attain a desired design objective. McCulloch and Pitts have developed the neural networks for different computing machines. 
  • 28. Artificial Neural Network:  Artificial Neural Network (ANN) have become a powerful tool for many complex applications including functional approximation, nonlinear system identification, motor control, pattern recognition, adaptive channel equalization and optimization.  ANN is capable of performing nonlinear mapping between the input and output space due to its large parallel interconnection between different layers and the nonlinear processing characteristics.
  • 29. Continued:  An artificial neuron basically consists of a computing element that performs the weighted sum of the input signal and the connecting weight. The weighted sum is added with the bias called threshold and the resultant signal is passed through a nonlinear activation function. Common types of activation functions are sigmoid and hyperbolic tangent.  Each neuron is associated with three parameters whose learning can be adjusted. These are the connecting weights, the bias and the slope of the nonlinear function.  For the structural point of view a NN may be single layer or it may be multilayer
  • 30. Multi-layer Perceptron:  The perceptron is a single level connection of McCulloch-Pitts neurons is called as Single-layer feed forward networks.  The network is capable of linearly separating the input vectors into pattern of classes by a hyper plane. Similarly many perceptrons can be connected in layers  To provide a MLP network, the input signal propagates through the network in a forward direction, on a layer-by-layer basis. This network has been applied successfully to solve diverse problems.
  • 31. Continued…  Generally MLP is trained using popular error back- propagation algorithm.  The scheme of MLP using four layers is shown. Si represent the inputs s1, s2, ….. , sn to the network, and yk represents the output of the final layer of the neural network.  The connecting weights between the input to the first hidden layer, first to second hidden layer and the second hidden layer to the output layers are represented by W i ,W ji ,W kj respectively.  The final output layer of the MLP may be expressed as