Single User or Multiple User?

Similar documents
MULTILEVEL CODING (MLC) with multistage decoding

Digital Television Lecture 5

Chapter 3 Convolutional Codes and Trellis Coded Modulation

Multilevel RS/Convolutional Concatenated Coded QAM for Hybrid IBOC-AM Broadcasting

n Based on the decision rule Po- Ning Chapter Po- Ning Chapter

Capacity-Approaching Bandwidth-Efficient Coded Modulation Schemes Based on Low-Density Parity-Check Codes

Notes 15: Concatenated Codes, Turbo Codes and Iterative Processing

Recent Progress in Mobile Transmission

Physical Layer: Modulation, FEC. Wireless Networks: Guevara Noubir. S2001, COM3525 Wireless Networks Lecture 3, 1

THE idea behind constellation shaping is that signals with

A Capacity Achieving and Low Complexity Multilevel Coding Scheme for ISI Channels

Incremental Redundancy Via Check Splitting

Department of Electronic Engineering FINAL YEAR PROJECT REPORT

Volume 2, Issue 9, September 2014 International Journal of Advance Research in Computer Science and Management Studies

Outline. Communications Engineering 1

EFFECTIVE CHANNEL CODING OF SERIALLY CONCATENATED ENCODERS AND CPM OVER AWGN AND RICIAN CHANNELS

Study of Turbo Coded OFDM over Fading Channel

Using TCM Techniques to Decrease BER Without Bandwidth Compromise. Using TCM Techniques to Decrease BER Without Bandwidth Compromise. nutaq.

Physical-Layer Network Coding Using GF(q) Forward Error Correction Codes

Performance comparison of convolutional and block turbo codes

Constellation Shaping for LDPC-Coded APSK

Advanced channel coding : a good basis. Alexandre Giulietti, on behalf of the team

Frequency-Hopped Spread-Spectrum

CT-516 Advanced Digital Communications

A Survey of Advanced FEC Systems

ECE 8771, Information Theory & Coding for Digital Communications Summer 2010 Syllabus & Outline (Draft 1 - May 12, 2010)

Robust Reed Solomon Coded MPSK Modulation

Optimized Codes for the Binary Coded Side-Information Problem

A REVIEW OF CONSTELLATION SHAPING AND BICM-ID OF LDPC CODES FOR DVB-S2 SYSTEMS

AN INTRODUCTION TO ERROR CORRECTING CODES Part 2

MULTILEVEL RS/CONVOLUTIONAL CONCATENATED CODED QAM FOR HYBRID IBOC-AM BROADCASTING

Course Developer: Ranjan Bose, IIT Delhi

IDMA Technology and Comparison survey of Interleavers

An Iterative Noncoherent Relay Receiver for the Two-way Relay Channel

Digital Communications I: Modulation and Coding Course. Term Catharina Logothetis Lecture 12

LDPC codes for OFDM over an Inter-symbol Interference Channel

Decoding of Block Turbo Codes

Coding and Modulation

Goa, India, October Question: 4/15 SOURCE 1 : IBM. G.gen: Low-density parity-check codes for DSL transmission.

An Improved Design of Gallager Mapping for LDPC-coded BICM-ID System

Lecture 17 Components Principles of Error Control Borivoje Nikolic March 16, 2004.

Improved concatenated (RS-CC) for OFDM systems

ECE 6640 Digital Communications

Convolutional Coding Using Booth Algorithm For Application in Wireless Communication

Joint Relaying and Network Coding in Wireless Networks

Error Correcting Codes for Cooperative Broadcasting

Error Control Coding. Aaron Gulliver Dept. of Electrical and Computer Engineering University of Victoria

Contents Chapter 1: Introduction... 2

IN 1993, powerful so-called turbo codes were introduced [1]

Polar Codes for Magnetic Recording Channels

Chapter 1 Coding for Reliable Digital Transmission and Storage

High-Rate Non-Binary Product Codes

On Performance Improvements with Odd-Power (Cross) QAM Mappings in Wireless Networks

Polar Codes for Probabilistic Amplitude Shaping

TSTE17 System Design, CDIO. General project hints. Behavioral Model. General project hints, cont. Lecture 5. Required documents Modulation, cont.

Serial Concatenation of LDPC Codes and Differentially Encoded Modulations. M. Franceschini, G. Ferrari, R. Raheli and A. Curtoni

Closing the Gap to the Capacity of APSK: Constellation Shaping and Degree Distributions

Performance of Nonuniform M-ary QAM Constellation on Nonlinear Channels

Code Design for Incremental Redundancy Hybrid ARQ

EELE 6333: Wireless Commuications

REVIEW OF COOPERATIVE SCHEMES BASED ON DISTRIBUTED CODING STRATEGY

ERROR CONTROL CODING From Theory to Practice

PERFORMANCE OF TWO LEVEL TURBO CODED 4-ARY CPFSK SYSTEMS OVER AWGN AND FADING CHANNELS

An Improved Rate Matching Method for DVB Systems Through Pilot Bit Insertion

Lab/Project Error Control Coding using LDPC Codes and HARQ

Ultra high speed optical transmission using subcarrier-multiplexed four-dimensional LDPCcoded

ATSC 3.0 Physical Layer Overview

Lecture 4: Wireless Physical Layer: Channel Coding. Mythili Vutukuru CS 653 Spring 2014 Jan 16, Thursday

Statistical Communication Theory

Punctured vs Rateless Codes for Hybrid ARQ

International Journal of Digital Application & Contemporary research Website: (Volume 1, Issue 7, February 2013)

Lecture 12: Summary Advanced Digital Communications (EQ2410) 1

Performance and Complexity Tradeoffs of Space-Time Modulation and Coding Schemes

Bit-permuted coded modulation for polar codes

LDPC Decoding: VLSI Architectures and Implementations

Performance of Channel Coded Noncoherent Systems: Modulation Choice, Information Rate, and Markov Chain Monte Carlo Detection

Dual-Mode Decoding of Product Codes with Application to Tape Storage

Improvement Of Block Product Turbo Coding By Using A New Concept Of Soft Hamming Decoder

Bit-Interleaved Coded Modulation: Low Complexity Decoding

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 50, NO. 1, JANUARY

Coded noncoherent communication with amplitude/phase modulation: from Shannon theory to practical turbo architectures

Performance Optimization of Hybrid Combination of LDPC and RS Codes Using Image Transmission System Over Fading Channels

Orthogonal vs Non-Orthogonal Multiple Access with Finite Input Alphabet and Finite Bandwidth

Power Efficiency of LDPC Codes under Hard and Soft Decision QAM Modulated OFDM

Trellis-Coded Modulation [TCM]

Low Complexity Decoding of Bit-Interleaved Coded Modulation for M-ary QAM

ECE 6640 Digital Communications

Bit-Interleaved Coded Modulation for Delay-Constrained Mobile Communication Channels

On the Capacity Regions of Two-Way Diamond. Channels

Modulation and Coding Tradeoffs

Spreading Codes and Characteristics. Error Correction Codes

Turbo Codes for Pulse Position Modulation: Applying BCJR algorithm on PPM signals

IEEE C /02R1. IEEE Mobile Broadband Wireless Access <

Channel Coding for IEEE e Mobile WiMAX

The BICM Capacity of Coherent Continuous-Phase Frequency Shift Keying

designing the inner codes Turbo decoding performance of the spectrally efficient RSCC codes is further evaluated in both the additive white Gaussian n

Lecture #2. EE 471C / EE 381K-17 Wireless Communication Lab. Professor Robert W. Heath Jr.

Joint Viterbi Detector/Decoder for Satellite Comms.

Feedback via Message Passing in Interference Channels

Master s Thesis Defense

Transcription:

Single User or Multiple User? Speaker: Xiao Ma maxiao@mail.sysu.edu.cn Dept. Electronics and Comm. Eng. Sun Yat-sen University March 19, 2013 Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 1 / 80

Outline 1 Single- and Multi-User Communication 2 Superposition Coded Modulation 3 Kite Codes 4 Block Markov Superposition Transmission 5 Conclusions Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 2 / 80

Outline 1 Single- and Multi-User Communication 2 Superposition Coded Modulation 3 Kite Codes 4 Block Markov Superposition Transmission 5 Conclusions Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 3 / 80

Single-User Communication System C. E. Shannon, A mathematical theory of communication, Bell Sys. Tech. Journal, 27, 379-423, 623-656, 1948. Digital Communication Framework noise source encoder channel decoder sink Figure: Block diagram of communication system. A source is nothing more than and nothing less than an arbitrary random process; The task of the encoder is to transform the output from the source into signals that matched to the channel, which can be split into two parts: Source encoder: Everything is binary! Channel encoder: Key techniques in the physical layer. A channel transforms an input to an output in a random manner dominated by a probability transition law. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 4 / 80

Single-User Communication System Shannon showed that a channel can be characterized by a parameter, C, called the channel capacity, which is a measure of how much information the channel can convey. The Channel Coding Theorem codes exist that provide reliable communication provided that the code rate satisfies R< C; conversely, if R> C, there exists no code that provides reliable communication. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 5 / 80

Single-User Communication System Capacity of the Ideal AWGN Channel Consider the discrete-time channel model, where E[X 2 t ]<P and Z t (0,σ 2 ). Y t = X t +Z t, The capacity of the channel is given by C= 1 2 log 2(1+ P ) bits/channel symbol. σ2 Frequently, one is interested in a channel capacity in units of bits per second rather than bits per channel symbol, C= W log 2 (1+ P σ 2 ) bits/second. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 6 / 80

Single-User Communication System 9 8 7 1.53 db 128QAM 256QAM Capacity(bits/symbol) 6 5 4 3 2 1 C = log(1 + SNR) 64QAM 32QAM 16QAM 8PSK QPSK 0 10 5 0 5 10 15 20 25 30 35 SNR [db] Figure: Capacity versus SNR curves for selected modulation schemes. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 7 / 80

Power Efficiency Traditional Codes Hamming codes, Golay codes, Reed-Muller codes Bose-Chaudhuri-Hocquenghem codes, Reed-Solomon codes Convolutional Codes Capacity-Approaching Codes Turbo codes Low-density parity-check (LDPC) codes Repeat-accumulate codes Accumulate-repeat-accumulate codes Concatenated zigzag codes, concatenated tree codes Precoded concatenated zigzag codes Convolutional LDPC codes Polar codes Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 8 / 80

Bandwidth Efficiency Existing Coded Modulation Schemes Trellis-coded modulation (TCM): proposed by Ungerboeck in 1982; Bit-interleaved coded modulation (BICM): proposed by Zehavi in 1992 for coding for fading channels; Multilevel codes (MLC): first proposed by H. Imai in 1977. Capacity-Approaching Coded Modulation Schemes Turbo-TCM schemes: two (or multiple) TCM codes are concatenated in the same fashion as binary turbo codes BICM with iterative decoding The output stream of a binary (turbo or LDPC) encoder is bit-interleaved and then mapped to an M-ary constellation The de-mapper is viewed as a APP decoder MLC with iterative multistage decoding Superimposed binary codes Coded modulation using non-binary LDPC codes Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 9 / 80

Typical Multi-User Channels Multiple-Access Channel x1 x2 xm p( y x1, x2,..., xm ) y two (or more) senders send information to a common receiver; senders must contend not only with the receiver noise but with interference from each other as well. Broadcast Channel y 1 x p( y1, y2,.., ym x) y 2 y m one sender send information to two or more receivers; the basic problem is to find the set of simultaneously achievable rates for communication. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 10 / 80

Typical Multi-User Channels Relay Channel one sender and one receiver with a number of relays; relays help the communication from the sender to the receiver. Interference Channel x1 p( y1, y2 x1, x2) y 1 Sender 1 wishes to send information to receiver 1 without caring what receiver 2 receives or understands; two senders and two receivers; x2 y 2 Similarly with sender 2 and receiver 2. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 11 / 80

Typical Multi-User Channels Two-Way Channel x1 x2 p( y1, y2 x1, x2) y 1 y 2 The two-way channel is very similar to the interference channel; Sender 1 is attached to receiver 2 and sender 2 is attached to receiver 1; This channel introduces another fundamental aspect of network information theory: namely, feedback; Feedback enables the senders to use the partial information that each has about the other s message to cooperate with each other. Question Can we apply the strategies in the multi-user communication system to the single-user communication system? Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 12 / 80

Outline 1 Single- and Multi-User Communication 2 Superposition Coded Modulation 3 Kite Codes 4 Block Markov Superposition Transmission 5 Conclusions Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 13 / 80

Superposition Coded Modulation Gaussian Multiple-Access Channel Two senders, X 1 and X 2, communicate to the single receiver, Y. The received signal at time t is Y t = X 1t +X 2t +Z t, where Z t (0,N); We assume that there is a power constraint P j on sender j. Interpretation of the Corner Points Point A: the maximum rate achievable C ( P 1 N ) from sender 1 to the receiver when sender 2 is not sending any information. Point B: decoding as a two-stage process The receiver decodes the second sender, considering the first sender as part of P the noise. This decoding will have low probability of error if R 2 < C ( 2 ); P 1 +N After the second sender has been decoded successfully, it can be subtracted out and the first sender can be decoded correctly if R 1 < C ( P 1 ). N Points C and D correspond to B and A, respectively, with the roles of the senders reversed. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 14 / 80

Superposition Coded Modulation Main Ideas Key technology: successive cancellation decoding (consider the other as noise; subtract its effect); Higher spectral efficiency is achieved for larger number of users. We proposed a coded modulation using superimposed binary codes. The Multilevel Coding/Sigma-Mapping Scheme Partition the information sequence into many subsequences; Each subsequence is encoded by a component code; Each coded sequence is then randomly-interleaved; All the random-interleaved versions are then mapped to a signal sequence by a sigma-mapper [MA04]X. Ma and Li Ping, Coded modulation using superimposed binary codes, IEEE Trans. Inform. Theory, 2004 Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 15 / 80

Multilevel Coding/Sigma-Mapping Scheme Multilevel Coding/Sigma-Mapping Scheme Component codes: turbo-like codes or LDPC codes; Power-allocation strategy: A simulation-based recursive search algorithm; Gaussian approximation power allocation. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 16 / 80

Multilevel Coding/Sigma-Mapping Scheme Normal Graph Figure: A normal graph for the multilevel coding/sigma-mapping scheme. Decoding three kinds of nodes: C, and ; messages are processed and exchanged over the normal graph. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 17 / 80

Simulation Result Using ten Brink s doped code of length 500000 as component codes. Figure: Performance of the four-level coding/sigma-mapping system with coding rate of 2 bits/dim. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 18 / 80

Summary 1 The multilevel coding/sigma-mapping scheme is an instance of MLC scheme. For the conventional MLC with lattice-based constellations and set-partitioning-based bit mappings, different levels are usually protected by codes with different rates. 2 In contrast, by choosing appropriate amplitudesα i in the multilevel coding/sigma-mapping systems, the component codes at different levels can be the same. 3 The multilevel coding/sigma-mapping system can be treated as a multiuser system by viewing one level as one user. So it is not surprising that most important methods and results for the multiuser system are applicable here. 4 Since the cooperation among different users is perfect, we are able to play more at both the transmitter and the receiver. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 19 / 80

Outline 1 Single- and Multi-User Communication 2 Superposition Coded Modulation 3 Kite Codes 4 Block Markov Superposition Transmission 5 Conclusions Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 20 / 80

Time Varying Channels Gaussian Broadcast Channels A sender of power P and two distant receivers; Y 1 = X+Z 1 and Y 2 = X+Z 2, where Z 1 and Z 2 are arbitrarily correlated Gaussian random variables with variances N 1 and N 2, respectively. Assume that N 1 < N 2. That is, the receiver Y 1 sees a better channel. The message consists of common information for both users and private information for Y 1. The results of the broadcast channel can be applied to the case of a single-user channel with an unknown distribution. The objective is to get at least the minimum information through when the channel is bad and to get some extra information through when the channel is good. Time Varying Channels different channel states at different time; adaptive coding scheme: rateless coding. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 21 / 80

Rateless Coding A coding method that can generate potentially infinite parity bits for any given fixed-length sequence. Existing Rateless Codes LT-codes; Raptor codes.... Motivations Raptor codes are optimized by degree distribution for erasure channels; No universal degree distributions exist for AWGN channels. How to construct good codes for AWGN channels with arbitrarily designated coding rate? We will propose a class of rateless codes for AWGN channels. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 22 / 80

Kite Codes An ensemble of Kite codes with dimension k, denoted byk[,k;p], is specified by a real sequence p, called p-sequence. Encoder of Kite Codes c= (v 0,,v k 1,w 0,w 1,,w t, ) $ -) (,' + & *) % /. 0!!"# $ Initially, load information sequence of length k into a buffer. At time t 0, randomly choose, with success probability p t, several bits from the buffer. Calculate the XOR of these chosen bits and use it to drive the accumulator to generate a parity bit w t. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 23 / 80

Decoder of Kite Codes For any given n k, the prefix code with length n of a Kite codek[,k;p] is denoted byk[n,k] and also called Kite code. From the encoding process of Kite codes, we can see that the parity-check matrix of Kite codes has the following form. 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 By choosing p t 0.5, we can construct Kite codes as LDPC codes. If so, the receiver can perform the iterative sum-product decoding algorithm. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 24 / 80

Relations Between Kite Codes and Existing Codes A specific Kite code (a realization of Kite code) is a kind of LDPC code, which is closely related to generalized IRA code. A specific Kite code can also be considered as a partially serially concatenated code with a systematic LDGM code as outer code and an accumulator as an inner code. However, as an ensemble, Kite codes are new. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 25 / 80

Relations Between Codes Ensembles A binary linear code ensemble is a probability space (C,Q) a sample spacecand a probability assignment Q(C) to each C C. Each sample C C is a binary linear code, and the probability Q(C) is usually implicitly determined by a random construction method. 1 Code ensemble C g : random generator matrix G of size k n. 2 Code ensemble C h : random parity-check matrix H of size (n k) n. 3 Code ensemble C s : random parity-check matrix H s = [P,I n k ]. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 26 / 80

Kite code is new as an ensemble An ensemble of Kite codes (of length n) with p t < 1/2 has the same sample space as that of C s but different probability assignment to each code. An ensemble of general LDPC code is specified by the a pair of degree distributions λ(x )= λ i x i and ρ(x )= ρ i x i, whereλ i andρ i are fractions. Givenλ i > 0, there must exist nodes of degree i. These fractions are fixed. An ensemble of Kite code is specified by the p-sequence. Its degree distributions are λ(x )= λ i x i and ρ(x )= ρ i x i, whereλ i andρ i are probabilities. Even ifλ i > 0, it is possible for a specific Kite code to have no nodes of degree i. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 27 / 80

Design of Kite Code Original Problem Evidently, the performance of Kite codes is determined by the p-sequence. The whole p-sequence should be optimized jointly such that all the prefix codes of Kite codes are good enough, which is too complex to implement. 4 3 2 2 2 2 2 6 5 2 2 2 2 2 2 2 2 2 2 8 7 2 2 2 2 2 2 2 2 2 2 Too complex due to too many (may be infinite) variables involved in. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 28 / 80

Design of Kite Code: A Simple Idea (layer by layer) Partitioning the coding rate into 9 subintervals. 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 < = > ; : A B F G C H @? E D Firstly, we choose q 9 such that the prefix codek[ k/0.9,k] is as good as possible. Secondly, we choose q 8 with fixed q 9 such that the prefix code K[ k/0.8,k] is as good as possible. Thirdly, we choose q 7 with fixed (q 9,q 8 ) such that the prefix code K[ k/0.7,k] is as good as possible. we choose q 1 with fixed (q 9,q 8,,q 2 ) such that the prefix codek[ k/0.1,k] is as good as possible. At each step, it is a one-dimensional optimization problem and can be implemented with density evolution or simulations. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 29 / 80

Numerical Results With data length k= 1890 and rates from 0.1 to 0.9, we have the following curves. 10 0 10 1 10 2 BER (Bit Error Rate) 10 3 10 4 10 5 10 6 10 7 6 4 2 0 2 4 6 8 SNR (Signal to Noise Ratio) = 10log 10 (1/σ 2 ) [db] Issue: Error floors. [MA11]X. Ma et al, Serial Concatenation of RS Codes with Kite Codes: Performance Analysis, Iterative Decoding and Design, http://arxiv.org/abs/1104.4927, 2011 [BAI11]B. Bai, B. Bai, X. Ma, Semi-random Kite Codes over Fading Channels, AINA 2011 Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 30 / 80

Numerical Results (continued) With k= 50000, we utilize RS code as outer codes to lower down the error floor. 1 0.9 capacity simulation results 0.8 0.7 rate [bits/bpsk] 0.6 0.5 0.4 0.3 0.2 0.1 6 4 2 0 2 4 6 8 10 12 SNR [db] Issue: Relative large gap between the performance and the Shannon limits. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 31 / 80

Improved Design of Kite Codes 1 Issue I: In the high-rate region, there exist error floors, which is caused by the existence of all-zero (or extremely-low-weight) columns in the randomly generated matrix H v. 2 Issue II: In the low-rate region, there exists a relatively large gap between the performances of the Kite codes and the Shannon limits. 3 Issue III: The optimized p-sequence depends on the data length k. The objective of this work is to solve these issues in simple ways. We partition the coding rates into 20 intervals. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 32 / 80

Row-weight concentration algorithm: to lower down the error floor. Given the parity-check matrix constructed layer by layer, we swap the 1 s and 0 s within each layer as follows. The i-th layer: Method: swap the 1 in the position (highest weight row,highest weight column) with the 0 in the position (lowest weight row, lowest weight column). Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 33 / 80

Accumulator randomization algorithm: to mitigate the performance loss. To introduce more randomness in the dual-diagonal matrix. This is done layer by layer. The current parity-check bit depends randomly on previous parity-check bits. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 34 / 80

Improved design: Numerical result Improved Kite codes with data length k= 1890 are constructed and the performances are shown as below. 10 0 10 1 Kite code Improved Kite code 10 2 BER (Bit Error Rate) 10 3 10 4 10 5 10 6 10 7 10 8 10 5 0 5 10 SNR (Signal to Noise Ratio) = 10log 10 (1/σ 2 ) [db] Remark: lower error-floors, better performances. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 35 / 80

Improved design: Numerical result Improved Kite codes with data length k= 3780 are constructed and the performances are shown as below. 10 0 10 1 k = 3780 k = 1890 10 2 BER (Bit Error Rate) 10 3 10 4 10 5 10 6 R = 0.5 10 7 R = 0.2 R = 0.8 10 8 10 5 0 5 10 SNR (Signal to Noise Ratio) = 10log 10 (1/σ 2 ) [db] Remark: lower error-floors, better performances. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 36 / 80

Improved design: Numerical result 10 0 10 1 Kite codes Kite codes modified by Row concentration Kite codes modified by Row concentration and Accumulator randomization 10 2 BER (Bit Error R ate) 10 3 10 4 10 5 10 6 10 7 R = 0.2 R = 0.5 R = 0.8 10 8 5 0 5 10 SNR (Signal to Noise Ratio) = 10log 10 (1/σ 2 ) [db] 1 In high-rate region: the row weight concentration algorithm lowers down the error-floors. 2 In low-rate region: the row weight concentration algorithm and the accumulator randomization algorithm gain about 0.9 db. 3 In moderate-rate region: no much gain. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 37 / 80

Empirical Formula of p-sequence To accelerate the design of improved Kite codes, we present the following empirical formula for the p-sequence. q l = 1 ( ) 1.65 k (1.5 0.05l) 6+2.0 for 1 l 19. 10 0 10 1 Greedy optimizing algorithm Empirical formula p sequence value 10 2 10 3 k = 1890 k = 3780 10 4 0 0.2 0.4 0.6 0.8 1 Coding Rate Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 38 / 80

Improved design: Constructing procedure The procedure to construct an improved Kite code: 1 Calculate the p-sequence according to the empirical formula; 2 Randomly generate the parity-check matrix according to the p-sequence; 3 Conduct the Row weight concentration algorithm; 4 Conduct the Accumulator randomization algorithm. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 39 / 80

Improved design: Numerical result A Kite code with data length k= 9450. The average decoding rates (at zero error probability) of this improved Kite code over AWGN channels is shown below. 1 0.9 0.8 Capacity Improved Kite code k=9450 RS Kite code k=50000 0.7 Rate [bits/bpsk] 0.6 0.5 0.4 0.3 0.2 0.1 0 10 5 0 5 10 SNR (signal to Noise Ratio) = 10log 10 (1/σ 2 ) [db] Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 40 / 80

Rate-compatible codes and adaptive coded modulation The system model for adaptive coded modulation is shown below. Source Modified Kite Encoder c[n] Modulator (Gray Mapper) x AWGN channel Sink Modified Kite Decoder ^ c[n] Demodulator (Bit Metric Calculator) y Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 41 / 80

Rate-compatible codes and adaptive coded modulation The average decoding spectral efficiency (at zero error probability) of the improved Kite code with data length k= 9450 over AWGN channels. 9 Capacity 8 Simulation results 1.53dB 256QAM 7 Capacity (bits/symbol) 6 5 4 3 128QAM C = log(1+snr) 64QAM 32QAM 16QAM 8PSK 2 QPSK 1 0 10 5 0 5 10 15 20 25 30 35 SNR [db] Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 42 / 80

Summary Given any code length k, any code rate r, we construct a well-performed binary LDPC code. Application 1 Broadcasting common information; 2 Adaptive coded modulation; 3 Easily extended to group code[ma11b]; 4 Joint source-channel code[yang12]; 5 Useful for research. [MA11a]X. Ma et al, Kite codes over groups, ITW 2011 [YANG12]Z. Yang, S. Zhao, X. Ma and B. Bai A new joint source-channel coding scheme based on nested lattice codes, IEEE Communication Letters 2012 Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 43 / 80

Pair of rates: (0.48, 0.88). Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 44 / 80

A rateless transmission scheme for two-user Gaussian broadcast channels. X IJKLMNO LP Q RQSSTKNUVTSN WLMN q r st u vw bcdefg hcicjk `TaJQZ [LTJS \NZNKSLO kcidel kmnomdpm X Y ZQSSTKN [LTJS \N]^NJKN _ Y KLMNM ZQSSTKN [LTJS \N]^NJKN Figure: Encoding structure of the two-way lattice-kite code. Figure: Gaussian broadcast channel. two receivers, R 1 and R 2 with sinal-to-noise ratios (SNR) SNR 1 and SNR 2, respectively; we assume that SNR=SNR 1 SNR 2 > 0. That is, receiver R 1 sees a better channel. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 45 / 80

y} Ÿ Ÿ y{ Ÿ ž œ yx š š y ª«ª«ª«ª«y} x xyz z zyz { {yz yz } }yz ~ ƒ ˆ Š Œ Ž Figure: Bandwidth efficiency of the proposed rateless transmission scheme for TU-GBC. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 46 / 80

Outline 1 Single- and Multi-User Communication 2 Superposition Coded Modulation 3 Kite Codes 4 Block Markov Superposition Transmission 5 Conclusions Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 47 / 80

Block Markov Superposition Transmission Gaussian Relay Channel A sender X and an ultimate intended receiver Y; The Gaussian relay channel is given by Y 1 = X+Z 1 Y = X+Z 1 +X 1 +Z 2, where Z 1 and Z 2 are independent zero-mean Gaussian random variables with variance N 1 and N 2, respectively; The encoding allowed by the relay is the causal sequence X 1i = f i (Y 11,Y 12,...,Y 1i 1 ); Sender X has power P and sender X 1 has power P 1. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 48 / 80

Block Markov Superposition Transmission Capacity of the Gaussian Relay Channel The capacity is whereα=1 α. C=max min{ C ( P+P 1+2 αpp1 ),C ( αp ) }, 0 α 1 N 1 +N 2 N 1 Basic Techniques for the Proof of Achievability Random coding; List codes; Slepian-Wolf partitioning; Coding for the cooperative multiple-access channel; Superposition coding; Block Markov encoding at the relay and transmitter. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 49 / 80

Block Markov Superposition Transmission Superposition Block Markov Encoding (SBME) in the Relay Channel The data are equally grouped into B blocks; Initially, the source broadcasts a codeword that corresponds to the first data block; Then the source and the relay cooperatively transmit more information about the first data block; In the meanwhile, the source superimposes a codeword that corresponds to the second data block; Finally, the destination recovers the first data block from the two successive received blocks; After removing the effect of the first data block, the system returns to the initial state; This process iterates B+1 times until all B blocks of data are sent successfully. We apply a similar strategy (SBME) to the single-user communication system, resulting in the block Markov superposition transmission (BMST) scheme. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 50 / 80

Block Markov Superposition Transmission BMST scheme The data are equally grouped into B blocks; Initially, the transmitter sends a codeword that corresponds to the first data block; Since the short code is weak, the receiver is unable to recover reliably the data from the current received block. Hence the transmitter transmits the codeword (in its interleaved version) one more time. In the meanwhile, a fresh codeword that corresponds to the second data block is superimposed on the second block transmission. Finally, the receiver recovers the first data block from the two successive received blocks. After removing the effect of the first data block, the system returns to the initial state; This process iterates B+1 times until all B blocks of data are sent successfully. [MA13]X. Ma et al, Obtaining extra coding gain for short codes by block Markov superposition transmission, submitted to ISIT, 2013 Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 51 / 80

Block Markov Superposition Transmission Encoding + + + + w (1) w (2) w (m-1) w (m) u (t) C D D D v (t) v (t-1) v (t-2) v (t-m+1) v (t-m) Figure: Encoding structure of BMST with memory m. Recursive Encoding of BMST 1 Initialization: For t< 0, set v (t) = 0 F n 2. 2 Recursion: For t= 0, 1,, L 1, Encode u (t) into v (t) F n 2 by the encoding algorithm of the basic code C; For 1 i m, interleave v (t i) by the i-th interleaverπ i into w (i) ; Compute c (t) = v (t) + 1 i m w (i), which is taken as the t-th block of transmission. 3 Termination: For t= L,L+1,,L+m 1, set u (t) = 0 F k 2 and compute c (t) recursively. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 52 / 80 m-1 m c (t)

Block Markov Superposition Transmission Normal Graph C (0) C (1) C (2) C (3) C (4) C (5) + + + + 1 2 1 2 1 2 1 2 V (0) = C V (1) = C V (2) = C V (3) = C a decoding layer U (0) U (1) U (2) U (3) Decoding Figure: The normal graph of a code with L=4 and m= 2. an iterative sliding-window decoding algorithm is used; four types of nodes: C,=,+, and ; messages are processed and passed through different decoding layers forward and backward over the normal graph; Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 53 / 80

Coding Gain Analysis of the BMST Genie-Aided Lower Bound on BER The performance of the BMST under MAP decoding is determined by Pr{u (t) j y}= Pr{u y}pr{u (t) j u,y}, u where the summation is over u ={u (i),t m i t+m,i t}; the BER performance can be lower-bounded by f n (γ b ) f o (γ b +10log 10 (m+1) 10log 10 (1+m/L)); noticing that Pr{u y} 1 for the transmitted data block u in the low error rate region, we can expect that asγ b increases. f n (γ b ) f o (γ b +10log 10 (m+1) 10log 10 (1+m/L)) the maximum coding gain can be 10log 10 (m+1) db for large L in the low error rate region. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 54 / 80

Simulation Result 10 0 CC, k = 50, n = 104 10 1 10 2 BER 10 3 BCJR 10 4 10 5 10 6 0 1 2 3 4 5 6 7 E b /N 0 (db) Figure: Coding gain analysis of the BMST system. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 55 / 80

Simulation Result 10 0 10 1 CC, k = 50, n = 104 lower bound for m = 1 10 2 BER 10 3 BCJR 10 4 10log 10 (2) 10 5 10 6 0 1 2 3 4 5 6 7 E b /N 0 (db) Figure: Coding gain analysis of the BMST system. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 56 / 80

Simulation Result 10 0 10 1 CC, k = 50, n = 104 lower bound for m = 1 10 2 BER 10 3 BCJR 10 4 10log 10 (2) 10 5 10log 10 (2) 10log 10 (20/19) 10 6 0 1 2 3 4 5 6 7 E b /N 0 (db) Figure: Coding gain analysis of the BMST system. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 57 / 80

Simulation Result 10 0 10 1 CC, k = 50, n = 104 lower bound for m = 1 upper bound for m = 1 10 2 BER 10 3 BCJR 10 4 10log 10 (2) 10 5 10log 10 (2) 10log 10 (20/19) 10 6 0 1 2 3 4 5 6 7 E b /N 0 (db) Figure: Coding gain analysis of the BMST system. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 58 / 80

Simulation Result 10 0 10 1 10 2 CC, k = 50, n = 104 lower bound for m = 1 upper bound for m = 1 simulation, m = 1 BER 10 3 BCJR 10 4 10log 10 (2) 10 5 10log 10 (2) 10log 10 (20/19) 10 6 0 1 2 3 4 5 6 7 E b /N 0 (db) Figure: Coding gain analysis of the BMST system. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 59 / 80

Simulation Result 10 0 10 1 Shannon limit of rate 1/2 CRC + CC, k = 10000, n = 20068 10 2 BER 10 3 10 4 10 5 CRC + list Viterbi BCJR only 10 6 10 7 0 1 2 3 4 5 6 7 E b /N 0 (db) Figure: The basic code C is a concatenated code of dimension k= 10000 and length n= 20068, where the outer code is a 32-bit CRC code and the inner code is a terminated 4-state (2, 1, 2) convolutional code. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 60 / 80

Simulation Result 10 0 10 1 Shannon limit of rate 1/2 CRC + CC, k = 10000, n = 20068 10 2 BER 10 3 10 4 10 5 10log 10 (2) CRC + list Viterbi BCJR only 10 6 10 7 0 1 2 3 4 5 6 7 E b /N 0 (db) Figure: The basic code C is a concatenated code of dimension k= 10000 and length n= 20068, where the outer code is a 32-bit CRC code and the inner code is a terminated 4-state (2, 1, 2) convolutional code. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 61 / 80

Simulation Result 10 0 10 1 Shannon limit of rate 1/2 CRC + CC, k = 10000, n = 20068 m = 1, d = 7 10 2 BER 10 3 10 4 10 5 10log 10 (2) CRC + list Viterbi BCJR only 10 6 10 7 0 1 2 3 4 5 6 7 E b /N 0 (db) Figure: The basic code C is a concatenated code of dimension k= 10000 and length n= 20068, where the outer code is a 32-bit CRC code and the inner code is a terminated 4-state (2, 1, 2) convolutional code. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 62 / 80

Simulation Result 10 0 10 1 Shannon limit of rate 1/2 CRC + CC, k = 10000, n = 20068 m = 1, d = 7 10 2 BER 10 3 10 4 10 5 10log CRC + list Viterbi 10 (2) BCJR only 10log 10 (3) 10 6 10 7 0 1 2 3 4 5 6 7 E b /N 0 (db) Figure: The basic code C is a concatenated code of dimension k= 10000 and length n= 20068, where the outer code is a 32-bit CRC code and the inner code is a terminated 4-state (2, 1, 2) convolutional code. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 63 / 80

Simulation Result BER 10 0 10 1 10 2 10 3 10 4 10 5 Shannon limit of rate 1/2 CRC + CC, k = 10000, n = 20068 m = 1, d = 7 m = 2, d = 2 m = 2, d = 7 m = 2, d = 7 with list Viterbi 10log 10 (2) 10log 10 (3) CRC + list Viterbi BCJR only 10 6 10 7 0 1 2 3 4 5 6 7 E b /N 0 (db) Figure: The basic code C is a concatenated code of dimension k= 10000 and length n= 20068, where the outer code is a 32-bit CRC code and the inner code is a terminated 4-state (2, 1, 2) convolutional code. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 64 / 80

Simulation Result BER 10 0 10 1 10 2 10 3 10 4 10 5 Shannon limit of rate 1/2 CRC + CC, k = 10000, n = 20068 m = 1, d = 7 m = 2, d = 2 m = 2, d = 7 m = 2, d = 7 with list Viterbi 10log 10 (2) 10log 10 (3) CRC + list Viterbi BCJR only 10 6 10log 10 (4) 10 7 0 1 2 3 4 5 6 7 E b /N 0 (db) Figure: The basic code C is a concatenated code of dimension k= 10000 and length n= 20068, where the outer code is a 32-bit CRC code and the inner code is a terminated 4-state (2, 1, 2) convolutional code. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 65 / 80

Simulation Result BER 10 0 10 1 10 2 10 3 10 4 10 5 Shannon limit of rate 1/2 CRC + CC, k = 10000, n = 20068 m = 1, d = 7 m = 2, d = 2 m = 2, d = 7 m = 2, d = 7 with list Viterbi m = 3, d = 7 10log 10 (2) 10log 10 (3) CRC + list Viterbi BCJR only 10 6 10log 10 (4) 10 7 0 1 2 3 4 5 6 7 E b /N 0 (db) Figure: The basic code C is a concatenated code of dimension k= 10000 and length n= 20068, where the outer code is a 32-bit CRC code and the inner code is a terminated 4-state (2, 1, 2) convolutional code. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 66 / 80

Simulation Result 10 0 10 1 Shannon limit of rate 4/7 CRC + Hamming, k = 11968, n = 21000 10 2 BER 10 3 10 4 CRC + list Viterbi BCJR only 10 5 10 6 0 1 2 3 4 5 6 7 8 E b /N 0 (db) Figure: The basic code C is a concatenated code of dimension k= 11968 and length n= 21000, where the outer code is a 32-bit CRC code and the inner code is the Cartesian product of Hamming code [7,4] 3000. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 67 / 80

Simulation Result 10 0 10 1 Shannon limit of rate 4/7 CRC + Hamming, k = 11968, n = 21000 10 2 BER 10 3 10 4 10log 10 (2) CRC + list Viterbi BCJR only 10 5 10 6 0 1 2 3 4 5 6 7 8 E b /N 0 (db) Figure: The basic code C is a concatenated code of dimension k= 11968 and length n= 21000, where the outer code is a 32-bit CRC code and the inner code is the Cartesian product of Hamming code [7,4] 3000. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 68 / 80

Simulation Result 10 0 10 1 Shannon limit of rate 4/7 CRC + Hamming, k = 11968, n = 21000 m = 1, d = 7 10 2 BER 10 3 10 4 10log 10 (2) CRC + list Viterbi BCJR only 10 5 10 6 0 1 2 3 4 5 6 7 8 E b /N 0 (db) Figure: The basic code C is a concatenated code of dimension k= 11968 and length n= 21000, where the outer code is a 32-bit CRC code and the inner code is the Cartesian product of Hamming code [7,4] 3000. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 69 / 80

Simulation Result 10 0 10 1 Shannon limit of rate 4/7 CRC + Hamming, k = 11968, n = 21000 m = 1, d = 7 10 2 BER 10 3 10 4 10 5 CRC + list Viterbi BCJR only 10log 10 (2) 10log 10 (3) 10 6 0 1 2 3 4 5 6 7 8 E b /N 0 (db) Figure: The basic code C is a concatenated code of dimension k= 11968 and length n= 21000, where the outer code is a 32-bit CRC code and the inner code is the Cartesian product of Hamming code [7,4] 3000. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 70 / 80

Simulation Result BER 10 0 10 1 10 2 10 3 10 4 10 5 Shannon limit of rate 4/7 CRC + Hamming, k = 11968, n = 21000 m = 1, d = 7 m = 2, d = 7 m = 2, d = 7 with list Viterbi 10log 10 (2) CRC + list Viterbi 10log 10 (3) BCJR only 10 6 0 1 2 3 4 5 6 7 8 E b /N 0 (db) Figure: The basic code C is a concatenated code of dimension k= 11968 and length n= 21000, where the outer code is a 32-bit CRC code and the inner code is the Cartesian product of Hamming code [7,4] 3000. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 71 / 80

Simulation Result BER 10 0 10 1 10 2 10 3 10 4 10 5 Shannon limit of rate 4/7 CRC + Hamming, k = 11968, n = 21000 m = 1, d = 7 m = 2, d = 7 m = 2, d = 7 with list Viterbi 10log 10 (2) CRC + list Viterbi 10log 10 (3) BCJR only 10log 10 (4) 10 6 0 1 2 3 4 5 6 7 8 E b /N 0 (db) Figure: The basic code C is a concatenated code of dimension k= 11968 and length n= 21000, where the outer code is a 32-bit CRC code and the inner code is the Cartesian product of Hamming code [7,4] 3000. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 72 / 80

Simulation Result BER 10 0 10 1 10 2 10 3 10 4 10 5 Shannon limit of rate 4/7 CRC + Hamming, k = 11968, n = 21000 m = 1, d = 7 m = 2, d = 7 m = 2, d = 7 with list Viterbi m = 3, d = 7 10log 10 (2) CRC + list Viterbi 10log 10 (3) BCJR only 10log 10 (4) 10 6 0 1 2 3 4 5 6 7 8 E b /N 0 (db) Figure: The basic code C is a concatenated code of dimension k= 11968 and length n= 21000, where the outer code is a 32-bit CRC code and the inner code is the Cartesian product of Hamming code [7,4] 3000. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 73 / 80

Simulation Result BER 10 0 10 1 10 2 10 3 Shannon limit of rate 4/7 CRC + Hamming, k = 11968, n = 21000 m = 1, d = 7 m = 2, d = 7 m = 2, d = 7 with list Viterbi m = 3, d = 7 CRC + list Viterbi BCJR only 10 4 10 5 10 6 10log 10 (2) 10log 10 (3) 10log 10 (4) 10log 10 (5) 0 1 2 3 4 5 6 7 8 E b /N 0 (db) Figure: The basic code C is a concatenated code of dimension k= 11968 and length n= 21000, where the outer code is a 32-bit CRC code and the inner code is the Cartesian product of Hamming code [7,4] 3000. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 74 / 80

Simulation Result BER 10 0 10 1 10 2 10 3 Shannon limit of rate 4/7 CRC + Hamming, k = 11968, n = 21000 m = 1, d = 7 m = 2, d = 7 m = 2, d = 7 with list Viterbi m = 3, d = 7 m = 4, d = 7 CRC + list Viterbi BCJR only 10 4 10 5 10 6 10log 10 (2) 10log 10 (3) 10log 10 (4) 10log 10 (5) 0 1 2 3 4 5 6 7 8 E b /N 0 (db) Figure: The basic code C is a concatenated code of dimension k= 11968 and length n= 21000, where the outer code is a 32-bit CRC code and the inner code is the Cartesian product of Hamming code [7,4] 3000. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 75 / 80

Simulation Result Performance of the BMST System 10 0 RS + CC, k = 1784, n = 4092 10 1 BER 10 2 10 3 BM + BCJR 10 4 10 5 10 6 1 1.5 2 2.5 3 E b /N 0 (db) Figure: The basic code C is the Consultative Committee on Space Data System (CCSDS) standard code of dimension k= 1784 and length n= 4092, where the outer code is a [255, 223] Reed-Solomon (RS) code overf 256 and the inner code is a terminated 64-state (2, 1, 6) convolutional code. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 76 / 80

Simulation Result Performance of the BMST System 10 0 10 1 RS + CC, k = 1784, n = 4092 m = 1, d = 4 BER 10 2 10 3 BM + BCJR 10 4 10 5 10 6 1 1.5 2 2.5 3 E b /N 0 (db) Figure: The basic code C is the Consultative Committee on Space Data System (CCSDS) standard code of dimension k= 1784 and length n= 4092, where the outer code is a [255, 223] Reed-Solomon (RS) code overf 256 and the inner code is a terminated 64-state (2, 1, 6) convolutional code. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 77 / 80

Outline 1 Single- and Multi-User Communication 2 Superposition Coded Modulation 3 Kite Codes 4 Block Markov Superposition Transmission 5 Conclusions Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 78 / 80

Conclusions Superposition Coded Modulation We proposed a coded modulation system using superimposed binary codes; Using the unequal power-allocations and the Gaussian-approximation-based suboptimal demapping algorithm, coded modulation with high bandwidth efficiency can be implemented. Kite Codes We proposed a kind of rateless codes for AWGN channels; A greedy optimization was presented to optimize Kite codes; Three methods were presented either to improve the performance of Kite codes, or to accelerate the design of Kite codes; Possible applications of Kite codes were investigated. Block Markov Superposition Transmission We presented a new method for constructing long codes from short codes; The encoding process can be as fast as the short code, while the decoding has a fixed delay. Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 79 / 80

Acknowledgements Thank You for Your Attention! Xiao Ma (SYSU) Coding Group Guangzhou, February 2013 80 / 80