fir filter difference in 8-bit and 12-bit modes

Status
Not open for further replies.

mohsen68sh

Junior Member level 1
Joined
Jun 10, 2016
Messages
16
Helped
0
Reputation
0
Reaction score
0
Trophy points
1
Activity points
129
hello
i designed a low-pass filter in fdatool like this:

when i use it in 12-bit configuration output is like below:

this output looks like my input and seems to be correct.
but when i use that filter in 8-bit configuration the output is like this

my output is some steps!!!
is this normal?
why it happens?
tanx in advance
 

This will be from clipping and saturation. It's pretty normal.
 

Rounding and precision issues can be very significant in digital signal processing.

For example I inherited an IIR filter in a control application and eventually realized how poorly configured it was for the task. Small signals (and as an error amplifier that's all it should see) multiplied by small coefficients rounded to zero, essentially negating that coefficient and meaning that the transfer function/step response varied significantly with signal amplitude.

The solution was to add additional fractional bits. But with fixed point math and/or large dynamic range in coefficient values these problems can be very significant. It's very important to evaluate real life performance which simulate the constraints of the real system.
 

Status
Not open for further replies.
Cookies are required to use this site. You must accept them to continue using the site. Learn more…