Dear All,

I have an OFDM TX block connected to an OFDM RX block through a 'Frequency
Selective Fading Model' block, implemented in GRC. The parameters used in
the model are:

Num sinusoids (for sum of sinusoids): 8
Max Doppler: 0
LOS Model: NLOS
PDP Delays: (0, 1.6, 2.8)
PDP Magnitudes: (0.4, 0.4, 0.2)
Num Taps: 100

When a packet is received by the channel payload equalizer block, I simply
print out the "Initial Taps", which are the initial channel estimated, and
plot the channel as a function of subcarrier index. I observe that at
subcarrier 0 (the midpoint, since subcarrier indices are centered around
0), there is a large phase shift. In my experiment, the fft_len is 64, used
as follows:

Subcarrier (-32 through -27): Unused
Subcarrier (-26 through -1) : Carry data/pilots
Subcarrier 0                : Unused
Subcarrier (+1 through +26) : Carry data/pilots
Subcarrier (+27 through +31): Unused

What is the cause of this large discontinuity between subcarriers (-1 and
+1)? Obviously, ignore subcarrier 0 since it is unused. I've pasted the
'Initial Taps' below. The columns are: subcarrier index, real part, imag
part, magnitude of channel, respectively.

  -32 0 0 0  -31 0 0 0  -30 0 0 0  -29 0 0 0  -28 0 0 0  -27 0 0 0  -26
-0.241 -0.245 0.343  -25 -0.164 -0.203 0.261  -24 -0.08 -0.183 0.199  -23
0.006 -0.185 0.185  -22 0.087 -0.21 0.228  -21 0.157 -0.255 0.3  -20 0.212
-0.318 0.382  -19 0.247 -0.391 0.463  -18 0.262 -0.47 0.538  -17 0.254
-0.549 0.605  -16 0.225 -0.622 0.662  -15 0.178 -0.683 0.706  -14 0.117
-0.728 0.737  -13 0.046 -0.754 0.755  -12 -0.028 -0.758 0.758  -11 -0.099
-0.741 0.748  -10 -0.163 -0.706 0.724  -9 -0.211 -0.653 0.686  -8 -0.244
-0.589 0.638  -7 -0.255 -0.519 0.578  -6 -0.246 -0.448 0.511  -5 -0.213
-0.381 0.437  -4 -0.165 -0.327 0.366  -3 -0.098 -0.289 0.306  -2 -0.016
-0.265 0.266  -1 0.079 -0.264 0.275  0 0 0 0  1 0.232 -0.506 0.557  2 0.191
-0.587 0.617  3 0.143 -0.654 0.67  4 0.079 -0.7 0.704  5 0.01 -0.727 0.727
6 -0.065 -0.729 0.732  7 -0.135 -0.712 0.724  8 -0.199 -0.675 0.703  9 -0.25
-0.62 0.669  10 -0.282 -0.552 0.62  11 -0.295 -0.478 0.561  12 -0.285 -0.401
0.492  13 -0.254 -0.329 0.416  14 -0.204 -0.268 0.337  15 -0.138 -0.221 0.26
16 -0.06 -0.195 0.204  17 0.022 -0.189 0.191  18 0.105 -0.207 0.232  19
0.181 -0.246 0.306  20 0.245 -0.306 0.392  21 0.29 -0.38 0.478  22 0.316
-0.465 0.562  23 0.318 -0.554 0.639  24 0.296 -0.641 0.706  25 0.253 -0.72
0.763  26 0.189 -0.786 0.808  27 0 0 0  28 0 0 0  29 0 0 0  30 0 0 0  31 0 0
0
Why is there such a large phase shift taking place around 0? Any pointers
would be greatly appreciated.

This effect does not show up with a standard frequency xlating fir filter
used instead of this frequency selective fading model block. Nor does the
problem show up when the experiment is run over USRP hardware instead of
the channel model block.

Thanks in advance!

Best regards,
Aditya
_______________________________________________
Discuss-gnuradio mailing list
Discuss-gnuradio@gnu.org
https://lists.gnu.org/mailman/listinfo/discuss-gnuradio

Reply via email to