• Tidak ada hasil yang ditemukan

in Partial Fulfillment of the Requirements for the Degree of

N/A
N/A
Protected

Academic year: 2023

Membagikan "in Partial Fulfillment of the Requirements for the Degree of"

Copied!
173
0
0

Teks penuh

From an algorithmic point of view, memory reduction techniques for a parallel turbo decoder are also presented in this work. Therefore, our research goal is to design an efficient turbo decoder that can support such higher throughputs for future wireless communication systems.

Background

Similarly, an excellent theoretical justification of the near-optimal turbo code error rate is provided by Benedetto et al. 1.2 (b) shows the basic block diagram of a turbo decoder, which is an integration of component MAP decoders with a pseudo-random interleaver and de-interleaver.

Figure 1.2: Basic block diagrams of (a) turbo encoder (b) turbo decoder.
Figure 1.2: Basic block diagrams of (a) turbo encoder (b) turbo decoder.

Design Perspective

Such MAP decoders are fundamentally based on BCJR (Bahl Cocke Jelinek Raviv) algorithm which works on the principle of lattice graph [18], and it processes a-priori probabilities of systematic and parity bits to produce a-posteriori probability values ​​of the transmitted information bits . Such extrinsic information values ​​are passed between these MAP decoders and are iteratively processed together with a-priori probability values ​​to produce error-free a-posteriori probabilities of the transmitted bits.

Figure 1.3: Variation of silicon areas with the technology scaling from 1000 nm to 22 nm CMOS processes.
Figure 1.3: Variation of silicon areas with the technology scaling from 1000 nm to 22 nm CMOS processes.

Contributions

Comparative failure rate curves are plotted with unlisted values ​​from software simulations and the values ​​obtained from the hardware implementations.

Organization of the Thesis

Performance and throughput analysis of turbo decoder for the physical layer of DVB-SH standard. A comprehensive analysis of the coding performance of turbo code for AWGN (additive white Gaussian noise) and frequency-selective fading channels with different modulation schemes conforming to the DVB-SH standard is presented.

Overview of DVB-SH Physical Layer

Transmitter

The 'symbol interleaver' unit is fed with QPSK or 16-QAM modulated symbols and is used for mapping these modulated symbols with pilot symbols for different IFFT sizes. Symbol interleaver unit incorporates lead symbols with the modulated symbols to produce Nf parallel symbols, where Nf is the size of IFFT.

Receiver

The soft demodulation process produces the LLR (Logarithmic Likelihood Ratio) of the a priori probabilities for the transmitted bits. The de-pointed LLR values ​​of the a priori probabilities of the transmitted bits are fed to a turbo decoder which undergoes an iterative decoding process to generate the final LLR values ​​of the a posteriori probabilities.

Fig. 2.1. These faded plus noisy analog signals are converted into discrete values using ADC (analog to digital converter) and fed to the receiver base-band system
Fig. 2.1. These faded plus noisy analog signals are converted into discrete values using ADC (analog to digital converter) and fed to the receiver base-band system

Performance and Throughput Analysis

  • Performance Analysis of Turbo Decoder in AWGN and Frequency Selective
  • Performance Analysis of Turbo Decoder for Different Decoding Iterations 18
  • Performance Analysis of Turbo Decoder for Different MAP Algorithms
  • Performance Analysis of Turbo Decoder for Different Code Rates

The coding performance of the turbo decoder with QPSK modulation for different iterations in the frequency-selective channel model with ITUR fading for a code rate of 1/2 is shown in Fig. The coding efficiency of the turbo decoder for different sliding window sizes with the QPSK modulation scheme for code rate 1/2 in the frequency-selective fading channel is shown in Fig.

Figure 2.2: Organization of an OFDM symbol at the transmitter-side using 1K-IFFT, where QPSK/16-QAM modulated symbols are concatenated with pilot-symbols and cyclic-prefix.
Figure 2.2: Organization of an OFDM symbol at the transmitter-side using 1K-IFFT, where QPSK/16-QAM modulated symbols are concatenated with pilot-symbols and cyclic-prefix.

Summary

Architecture for SISO unit of turbo decoder based on the selected simplified MAP algorithm is presented. Turbo decoder architecture and its integral parts, such as SISO unit based on a simplified MAP algorithm and QPP interleaver are presented in section-3.3.

Figure 3.1: A conventional parallel-architecture of turbo decoder which iteratively processes input-soft-values to produce decoded-bits.
Figure 3.1: A conventional parallel-architecture of turbo decoder which iteratively processes input-soft-values to produce decoded-bits.

Comparative Study

Overview of Simplified MAP Algorithms

Section 3.2 presents a brief discussion of the reported simplified MAP algorithms and their architecture and BER performance comparison. Similarly, the simplified MAP algorithm based on MSE [38] could be another candidate in terms of BER performance and implementation complexity.

Comparative Analysis of Architectures

The approximation of the simplified MAP algorithm based on max ford PWLA for r=3 and r=4 as shown in Table 3.1 is further reduced to more simplified approximations. Finally, the value C3 = max{0,(ln) is added with C1 to realize the maxmac value for the simplified MSE-based MAP algorithm as shown in Fig.

as max(Ψ d 1 , Ψ 2 ) ≈ max mac = max (Ψ 1 , Ψ 2 )+max{0, (ln 2−0.5×|∆|)}, as shown in Table 3.1
as max(Ψ d 1 , Ψ 2 ) ≈ max mac = max (Ψ 1 , Ψ 2 )+max{0, (ln 2−0.5×|∆|)}, as shown in Table 3.1

Performance Analysis

From this we can conclude that both architectures are suitable for high-speed turbo decoder implementations. Therefore, a simplified MAP algorithm based on the maxred3 approximation is chosen to implement the turbo decoder, which provides near-optimal BER performance and has an architecture suitable for high-speed applications.

Turbo Decoder Architecture

  • SISO Architecture
  • SISO Scheduling
  • Analysis of Memory Requirement
  • Interleaver Design
  • Decoder Architecture

Simultaneously, the previously stored parent branch metrics of SW1 are obtained through the p1 gates of the DP-SRAMs and fed into the BMR unit before the FSMC unit via REG2, as shown in Fig. Simultaneously, SW2's parent branch metrics are received through the p2 gates of the DP-SRAMs and fed to the FSMC unit to calculate forward state metrics for SW2.

Figure 3.5: High-level architecture of SISO unit which is an integration of various sub-blocks like BMC, BMR, FSMC, BSMC, DBSMC, LCU, DP-SRAMs and SRAMs.
Figure 3.5: High-level architecture of SISO unit which is an integration of various sub-blocks like BMC, BMR, FSMC, BSMC, DBSMC, LCU, DP-SRAMs and SRAMs.

VLSI Implementation, Application and Comparison

VLSI Implementation

Similarly, two input blocks are used for clock and enable signals, and one output block is assigned to deliver decoded bits from turbo decoder. Design criteria such as core area, power consumption and maximum operating clock frequency of an implemented turbo decoder design at 130 nm technology node are presented in Table 3.4.

Figure 3.9: Chip-layout of turbo decoder implemented at 130 nm CMOS technology node.
Figure 3.9: Chip-layout of turbo decoder implemented at 130 nm CMOS technology node.

Possible Applications

On the other hand, using radix-4 configuration (rad-4) with parallel SISO units for turbo decoder architecture can achieve throughput of approx. 110 Mbps, 220 Mbps, 320 Mbps and 425 Mbps at a clock frequency of 303 MHz for P = 2, 4, 6 and 8, as shown in fig. It is proportional to throughput achieved, current consumed and iterations performed by the turbo decoder.

Figure 3.10: Plots of achievable throughputs with respect to operating clock frequencies for various configurations of turbo decoder.
Figure 3.10: Plots of achievable throughputs with respect to operating clock frequencies for various configurations of turbo decoder.

Comparison of Results

However, the parallel architectures are better than the proposed non-parallel-radix-2 architecture in terms of throughput.

Memory-Reduced MAP Decoding for Parallel Turbo Decoders

Theoretical Background

Applying Bayes' rule and assuming that the channel is memoryless and discrete, an expression forP(s0, s, y) from (3.12) can be rewritten as.

RSWMAP Algorithm

This relationship is expressed by the a-posteriori transition probability from (3.13), where the backward state metric is denoted as. At k=M and from (3.18), the initial value of the backward state metric that starts the backward recursion can be expressed as.

Mathematical Reformulation of Branch Metric Equations

Thus, if nγ represents the quantization of the branch metric, then the SISO unit architecture must accommodate memory to store 2×M×2n×nγ bits for the parent branch metrics. Unlike the conventional method, the SISO unit, which is based on the branch metric reformulation of this work, needs to store 2×M×nγbits for the branch metric.

Fig. 3.11 shows the percentages of SBMSs achieved by proposed and reported works. Arch- Arch-1 presented in [53] has achieved a saving due to reduced memory required for forward state metrics, and its SBMS is 50%
Fig. 3.11 shows the percentages of SBMSs achieved by proposed and reported works. Arch- Arch-1 presented in [53] has achieved a saving due to reduced memory required for forward state metrics, and its SBMS is 50%

Architecture and Scheduling of SISO Unit

Architecture

3.12, the estimated backward state metrics from the BRFE submodule are fed to the dummy state metric computation (DSMC) submodule, which is used in the dummy backward recursion process for MAP decoding. Outputs from the DSMC submodule are fed consecutively to the BSMC submodule, which is also an SMC unit.

Figure 3.12: High-level architecture of SISO unit based on RSWMAP algorithm and refor- refor-mulation of branch metric equation.
Figure 3.12: High-level architecture of SISO unit based on RSWMAP algorithm and refor- refor-mulation of branch metric equation.

Scheduling

Backward state metric evaluation and virtual backward recursion are then performed for the second sliding window. In addition, the BMC submodule calculates the branch metric for the fifth sliding window and stores it in MEM2.

Figure 3.14: Timing-chart that illustrates scheduling of MAP decoding based on the sug- sug-gested memory-reduced techniques.
Figure 3.14: Timing-chart that illustrates scheduling of MAP decoding based on the sug- sug-gested memory-reduced techniques.

Comparative Analysis of Memory Requirement

In order to estimate the memory savings in a parallel turbo decoder using SISO units based on branch metric reformulation, Fig. The percentage improvements achieved by the parallel turbo decoder for different values ​​of P are shown in Fig.

Figure 3.15: Memory required by parallel turbo decoder architectures using branch-metric reformulation, SWBCJR and BCJR algorithms based SISO units
Figure 3.15: Memory required by parallel turbo decoder architectures using branch-metric reformulation, SWBCJR and BCJR algorithms based SISO units

Performance Analysis, Implementation Trade-offs and Comparison

BER Performance

SWBCJR algorithm based turbo decoder with M=32 has similar BER performance to that of RSWMAP algorithm based turbo decoder with M=24.

Implementation Trade-offs

From the previous BER analysis, it is observed that the parallel turbo decoder based on the RSWMAP algorithm can give the optimal BER performance forM=24 instead of M=32 required by the decoder based on the SWBCJR algorithm. 3.18, while a maximum of 44.14% of hardware resources are saved, due to the memory reduction in the parallel turbo decoder, forP=64.

Summary

Recently, a maximum data rate has been achieved on the 3GPP-LTE standard with parallel turbo decoder implemented in [29]. The primary motive of this work is to devise parallel turbo-decoder architecture for high-throughput applications.

Theoretical Background

It generates a-prioriLLRvaluesλsk, λp1k and λp2k for the transmitted systematic and parity bits, respectively, and are fed into the turbo decoder via the serial-to-parallel converter. Finally, the a-posterior LLR values ​​that are generated by the turbo decoder are processed by the rest of the baseband sub-modules and the sequence of decoded bits Vk is obtained, as shown in Fig.

Figure 4.2: (a) Trellis graph with N stages and N s trellis states. (b) Scheduling of sliding window technique for LBCJR algorithm, where x-axis and y-axis represent time and
Figure 4.2: (a) Trellis graph with N stages and N s trellis states. (b) Scheduling of sliding window technique for LBCJR algorithm, where x-axis and y-axis represent time and

Proposed Techniques

A Modified Sliding Window Approach

At the same time, the second unjoined backward recursion is started from the initialized stage grids=5 by calculating the new set{Bk=4}u=2. At the same time, the fourth unjoined backward recursion begins with the calculation of the new set{Bk=6}u=4.

Figure 4.3: Illustration of un-grouped backward recursions in four-state trellis graph, with M =4, for trellis stages k =1 and k =2.
Figure 4.3: Illustration of un-grouped backward recursions in four-state trellis graph, with M =4, for trellis stages k =1 and k =2.

A State Metric Normalization Technique

The critical paths of ACSUs based on the proposed approach, modulo and subtractive normalization techniques are highlighted in the figure. Subsequently, the SMCU used in this work consumes 17.87% less silicon area than the SMCU based on the subtractive normalization technique.

Figure 4.5: (a) An ACSU for modulo normalization technique [28] (b) An ACSU for suggested normalization technique (c) An ACSU for subtractive normalization technique [24] (d) Part of a trellis graph with N s =8 showing (k -1) th and k th trellis stages and
Figure 4.5: (a) An ACSU for modulo normalization technique [28] (b) An ACSU for suggested normalization technique (c) An ACSU for subtractive normalization technique [24] (d) Part of a trellis graph with N s =8 showing (k -1) th and k th trellis stages and

Decoder Architectures and Scheduling

MAP Decoder Architecture and Scheduling

In the fifth clock cycle, branch statistics of the Γk=4 set are launched from RE2 and are used by SMCU1 together with initial values ​​of backward state statistics of LUT to calculate backward state statistics of {Bk=3}u=1 for the first non- grouped backward recursion and then stored in RE8, as shown in figure. It can be noted that the similar pattern of calculations for branching and status metrics is performed for successive trellis stages, as shown in Figure 2.

Figure 4.7: Launched values of state and branch metric sets as well as a-posteriori LLRs by different registers of MAP decoder in successive clock cycles.
Figure 4.7: Launched values of state and branch metric sets as well as a-posteriori LLRs by different registers of MAP decoder in successive clock cycles.

Retimed and Deep-pipelined Decoder Architecture

These individually pipelined units are incorporated into the MAP decoder design to become a deeply pipelined architecture as shown in the figure. The MAP decoder architecture presented in [90] is based on a two-dimensional radix-4×4 retiming ACSU.

Figure 4.8: (a) Data-flow-graph of retimed SMCU for computing N s =4 forward state metrics.
Figure 4.8: (a) Data-flow-graph of retimed SMCU for computing N s =4 forward state metrics.

Parallel Turbo Decoder Architecture

Data output from the memory bank is fed to slave network and routed to 8×MAP decoders. During the first half iteration, the input a-prioriLLRvaluesλsk andλp1k are sequentially retrieved from memory banks and fed to 8 × MAP decoders.

Figure 4.13: Parallel turbo decoder architecture with 8 × MAP decoders.
Figure 4.13: Parallel turbo decoder architecture with 8 × MAP decoders.

Performance Analysis, VLSI Implementation and Comparison of Parallel Turbo

The parallel turbo decoder architecture, with P=8, using the suggested MAP decoder design, is implemented in 90 nm CMOS process. On the other hand, the parallel turbo decoder with P=64 of this work has 38.4% better throughput compared to the work [49] which is simulated after rendering in 90 nm CMOS process.

Figure 4.15: BER performance in AWGN channel using BPSK modulation for a low effective code-rate of 1/3, N=6144 (f 1 =263, f 2 =480), M =32, P =8 and ω=1
Figure 4.15: BER performance in AWGN channel using BPSK modulation for a low effective code-rate of 1/3, N=6144 (f 1 =263, f 2 =480), M =32, P =8 and ω=1

Summary

Software Model

Communication System

At the transmitter side, the randomly generated bit sequence (Uk) is encoded using a convolutional encoder with transfer function {1, (1+D+D3)/(1+D2+D3)}. Puncturer can produce a sequence of bits (Upun) for arbitrary code rates depending on the puncturing pattern used [98, 99].

Figure 5.2: Software model of communication system for testing the MAP/turbo decoder in MATLAB environment.
Figure 5.2: Software model of communication system for testing the MAP/turbo decoder in MATLAB environment.

BER Performance Evaluation

Thus, quantization and saturation processes are required for fixed-point representation of real-valued a-prioriLLRs (λsk, λp1k, and λp2k). In this paper, we have chosen the values ​​of (nB,nP) as (5, 2) bits and (7, 3) bits to represent the fixed-point test vectors of a-priori input LLR values ​​for MAP and turbo decoders respectively .

FPGA Implementation and Verification of MAP Decoder

Implementation

Testing

The netlists of these ILA and ICON cores can be easily integrated with the target IMD core. The configuration file containing the built-in netlist of the IMD core with the ILA and ICON cores is dumped into the FPGA board.

Figure 5.5: FPGA on-board integration of suggested MAP decoder-design with memories containing the fixed point soft values x and xp1.
Figure 5.5: FPGA on-board integration of suggested MAP decoder-design with memories containing the fixed point soft values x and xp1.

Performance Evaluation

Such performance degradation is due to the implementation of a fixed point MAP decoder compared to simulated values ​​in which the accuracy used to represent a number is very high. The BER performance of the implemented MAP decoder can be improved by increasing the number of bits for the fixed point representation.

Implementation, Testing and Performance

On-board fractional PLL (phase-locked loop) is used for generating the clock for RAMs and hardware prototype of parallel turbo-decoder. BER performance from simulation BER performance from hardware prototype BER performance from hardware prototype BER performance from simulation.

Figure 5.9: Schematic of test-plan for the hardware prototype of parallel turbo decoder using FPGA and logic analyzer.
Figure 5.9: Schematic of test-plan for the hardware prototype of parallel turbo decoder using FPGA and logic analyzer.

Summary

Next, the BER plots of the parallel turbo decoder hardware prototype were presented and compared with the simulated BER curve of the turbo decoder. Finally, a high power and energy efficient turbo parallel decoder was conceived for future wireless communication systems.

Figure 6.1: Comparative plots of energy efficiency, area consumed and throughput achieved by the proposed and the state-of-the-art works.
Figure 6.1: Comparative plots of energy efficiency, area consumed and throughput achieved by the proposed and the state-of-the-art works.

Future Directions

One of the crucial steps is to include the .synopsys dc.setup file in the working directory, as this creates an environment for the Synopsys-DC tool to run. Figure A.2 shows snapshots of some parts of the reports generated by the Synopsys-DC tool.

Figure A.1: GUI invoked by Synopsys-VCS tool for logical and functional verification of the digital design.
Figure A.1: GUI invoked by Synopsys-VCS tool for logical and functional verification of the digital design.

Backend Design Flow

Glavieux, “Near-optimal error-correction coding and decoding: turbo codes,” IEEE Transactions on Communications , vol. Masera, "State metric compression techniques for turbo decoder architectures," IEEE Transactions on Circuits and Systems I: Regular Papers, vol.

Figure A.5: GUI of SOC-Encounter after importing standard-cells, hard-macros and pads.
Figure A.5: GUI of SOC-Encounter after importing standard-cells, hard-macros and pads.

Gambar

Figure 1.1: Ever increasing peak data rates of various wireless communication standards which include turbo code as their error-correcting codes.
Figure 2.1: System level architecture for the physical layer of DVB-SH-A wireless communi- communi-cation standard.
Figure 2.5: Coding performances of turbo code for different iterations in AWGN channel for a code rate of 1/2.
Figure 3.1: A conventional parallel-architecture of turbo decoder which iteratively processes input-soft-values to produce decoded-bits.
+7

Referensi

Dokumen terkait

Octal-to-binary 8-line-to-3-line Encoder Fig 9-13 Logic circuit for an octal-to-binary 8-line-to-3-line encoder For proper operation only one input should be encoder.. For proper