A Bit of network information theory

Similar documents
Wireless Network Information Flow

Information flow over wireless networks: a deterministic approach

Opportunistic network communications

Multicasting over Multiple-Access Networks

Joint Relaying and Network Coding in Wireless Networks

Interference: An Information Theoretic View

Relay Scheduling and Interference Cancellation for Quantize-Map-and-Forward Cooperative Relaying

Coding for Noisy Networks

Symmetric Decentralized Interference Channels with Noisy Feedback

Approximately Optimal Wireless Broadcasting

Information Flow in Wireless Networks

Network Information Theory

Computing and Communications 2. Information Theory -Channel Capacity

Degrees of Freedom of Bursty Multiple Access Channels with a Relay

Multi-user Two-way Deterministic Modulo 2 Adder Channels When Adaptation Is Useless

Variations on the Index Coding Problem: Pliable Index Coding and Caching

On Optimum Communication Cost for Joint Compression and Dispersive Information Routing

How (Information Theoretically) Optimal Are Distributed Decisions?

ECE 4400:693 - Information Theory

Wireless Network Coding with Local Network Views: Coded Layer Scheduling

Scheduling in omnidirectional relay wireless networks

Feedback via Message Passing in Interference Channels

On the Capacity Regions of Two-Way Diamond. Channels

3432 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 53, NO. 10, OCTOBER 2007

Interference Management in Wireless Networks

State of the Cognitive Interference Channel

Exploiting Interference through Cooperation and Cognition

Analog network coding in the high-snr regime

Cooperation in Wireless Networks

On the Achievable Diversity-vs-Multiplexing Tradeoff in Cooperative Channels

Role of a Relay in Bursty Multiple Access Channels

The Multi-way Relay Channel

Chapter 10. User Cooperative Communications

5984 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 56, NO. 12, DECEMBER 2010

Optimum Power Allocation in Cooperative Networks

SHANNON S source channel separation theorem states

EE 8510: Multi-user Information Theory

On the Capacity of Multi-Hop Wireless Networks with Partial Network Knowledge

On Information Theoretic Interference Games With More Than Two Users

Broadcast Networks with Layered Decoding and Layered Secrecy: Theory and Applications

arxiv: v1 [cs.it] 12 Jan 2011

Diversity and Freedom: A Fundamental Tradeoff in Multiple Antenna Channels

DELAY CONSTRAINED MULTIMEDIA COMMUNICATIONS: COMPARING SOURCE-CHANNEL APPROACHES FOR QUASI-STATIC FADING CHANNELS. A Thesis

The Reachback Channel in Wireless Sensor Networks

Index Terms Deterministic channel model, Gaussian interference channel, successive decoding, sum-rate maximization.

Information Theoretic Analysis of Cognitive Radio Systems

Two Models for Noisy Feedback in MIMO Channels

SHANNON showed that feedback does not increase the capacity

Bounds on Achievable Rates for Cooperative Channel Coding

On Fading Broadcast Channels with Partial Channel State Information at the Transmitter

Uplink Multicell Processing with Limited Backhaul via Successive Interference Cancellation

The Degrees of Freedom of Full-Duplex. Bi-directional Interference Networks with and without a MIMO Relay

On Achieving Local View Capacity Via Maximal Independent Graph Scheduling

Overlay Systems. Results around Improved Scheme Transmission for Achievable Rates. Outer Bound. Transmission Strategy Pieces

On Coding for Cooperative Data Exchange

Computing functions over wireless networks

Interference Mitigation Through Limited Transmitter Cooperation I-Hsiang Wang, Student Member, IEEE, and David N. C.

TWO-WAY communication between two nodes was first

Degrees of Freedom in Multiuser MIMO

On Event Signal Reconstruction in Wireless Sensor Networks

From Wireless Network Coding to Matroids. Rico Zenklusen

Optimal Coded Information Network Design and Management via Improved Characterizations of the Binary Entropy Function

CONSIDER a sensor network of nodes taking

Multiuser Information Theory and Wireless Communications. Professor in Charge: Toby Berger Principal Lecturer: Jun Chen

DoF Analysis in a Two-Layered Heterogeneous Wireless Interference Network

The Z Channel. Nihar Jindal Department of Electrical Engineering Stanford University, Stanford, CA

WIRELESS or wired link failures are of a nonergodic nature

THE multi-way relay channel [4] is a fundamental building

Message Passing in Distributed Wireless Networks

Efficient Codes using Channel Polarization!

Degrees of Freedom of the MIMO X Channel

Information Theory: the Day after Yesterday

Performance Enhancement of Interference Alignment Techniques for MIMO Multi Cell Networks

Green Codes : Energy-efficient short-range communication

Causal state amplification

An Introduction to Distributed Channel Coding

Rab Nawaz. Prof. Zhang Wenyi

On the Unicast Capacity of Stationary Multi-channel Multi-radio Wireless Networks: Separability and Multi-channel Routing

Routing versus Network Coding in Erasure Networks with Broadcast and Interference Constraints

Degrees of Freedom of Multi-hop MIMO Broadcast Networks with Delayed CSIT

Capacity of Two-Way Linear Deterministic Diamond Channel

Block Markov Encoding & Decoding

State-Dependent Relay Channel: Achievable Rate and Capacity of a Semideterministic Class

Course Developer: Ranjan Bose, IIT Delhi

Communication Theory II

Detection and Estimation in Wireless Sensor Networks

MIMO III: Channel Capacity, Interference Alignment

OPTIMAL POWER ALLOCATION FOR MULTIPLE ACCESS CHANNEL

Chapter Number. Parameter Estimation Over Noisy Communication Channels in Distributed Sensor Networks

Natasha Devroye, Mai Vu, and Vahid Tarokh ] Cognitive Radio Networks. [Highlights of information theoretic limits, models, and design]

Quantize-Map-and-Forward Relaying: Coding and System Design

Yale University Department of Computer Science

Fractional Cooperation and the Max-Min Rate in a Multi-Source Cooperative Network

Cooperative Diversity in Wireless Networks: Efficient Protocols and Outage Behavior

A unified graphical approach to

Cloud-Based Cell Associations

Distributed Source Coding: A New Paradigm for Wireless Video?

On Achieving Local View Capacity Via Maximal Independent Graph Scheduling

On Secure Signaling for the Gaussian Multiple Access Wire-Tap Channel

Physical Layer Security for Wireless Networks

Transcription:

Š#/,% 0/,94%#(.)15% A Bit of network information theory Suhas Diggavi 1 Email: suhas.diggavi@epfl.ch URL: http://licos.epfl.ch Parts of talk are joint work with S. Avestimehr 2, S. Mohajer 1, C. Tian 3, D. Tse 2 1 Ecole Polytechnique Fédérale de Lausanne (EPFL), Lausanne, Switzerland 2 University of California, Berkeley, California, USA 3 AT&T Shannon Laboratories, Florham Park, New Jersey, USA December 29th 2008

The holy grail Information flow over shared networks Unicast, multicast, multiple unicast (multicommodity flow). Significant progress for graphs (routing, network coding etc). Less understood for flows over wireless networks. Network data compression: sensor networks. Some successes in side-information coding: Slepian-Wolf, Wyner Ziv etc. Many unresolved questions: Distributed source coding, multiple description coding. Question: How can we make progress to fundamentally characterize flow of information over networks?

Approximate characterizations Š#/,% 0/,94%#(.)15% Philosophy: Gain insight into central difficulties of problem by identifying underlying deterministic/lossless structures. Goal: Use the insight of underlying problem to get (provable) approximate characterization for noisy/lossy problem. Underlying problem should be characterized exactly to give insight into solution structure for general case. Universal approximation: Approximation should depend only on the problem structure and not on parameters (like channel gains, distortions etc.). Question: Can we identify the appropriate underlying problems and use them to get provable (universal) approximations.

Overall agenda Central theme: Obtain (universal) approximate characterizations for network flow problems. Talk outline: Wireless relay networks Focus on signal interactions study deterministic networks. Characterize deterministic networks approximation for noisy case. coding Identify related underlying lossless problem study multi-level source coding. Use multi-level rate region approximation for lossy case.

Key distinctions between wired and wireless channels Broadcast: Transmit signal potentially received by multiple receivers.

Key distinctions between wired and wireless channels Broadcast: Transmit signal potentially received by multiple receivers. Multiple access: Transmitted signals mix at the receivers.

Key distinctions between wired and wireless channels Broadcast: Transmit signal potentially received by multiple receivers. Multiple access: Transmitted signals mix at the receivers. High dynamic range: Large range in relative signal strengths.

Key distinctions between wired and wireless channels Broadcast: Transmit signal potentially received by multiple receivers. Multiple access: Transmitted signals mix at the receivers. High dynamic range: Large range in relative signal strengths. Implications: Complex signal interactions at different signal levels.

Key distinctions between wired and wireless channels Broadcast: Transmit signal potentially received by multiple receivers. Multiple access: Transmitted signals mix at the receivers. High dynamic range: Large range in relative signal strengths. Implications: Complex signal interactions at different signal levels. Interacting signals from nodes contain information (not to be treated as noise).

Key distinctions between wired and wireless channels Broadcast: Transmit signal potentially received by multiple receivers. Multiple access: Transmitted signals mix at the receivers. High dynamic range: Large range in relative signal strengths. Implications: Complex signal interactions at different signal levels. Interacting signals from nodes contain information (not to be treated as noise). Question: Can we develop cooperative mechanisms to utilize signal interaction?

Signal interaction: Gaussian wireless networks Source 1 Receiver 1 Source 2 Receiver 2 y j [t] = i H ij x i [t] + z j [t] Broadcast because transmission x i is heard by all receivers. Multiple access because transmitted signals from all nodes mix linearly at the receiver j. Dynamic range depends on relative strengths of H ij.

Signal interaction: Gaussian wireless networks Source 1 Receiver 1 Source 2 Receiver 2 y j [t] = i H ij x i [t] + z j [t] Broadcast because transmission x i is heard by all receivers. Multiple access because transmitted signals from all nodes mix linearly at the receiver j. Dynamic range depends on relative strengths of H ij. Question: Can we characterize capacity of such networks?

Gaussian network capacity: state of knowledge Resolved: Point-to-point channel, multiple access channel, broadcast channel (private messages).

Gaussian network capacity: state of knowledge Resolved: Point-to-point channel, multiple access channel, broadcast channel (private messages). Unresolved: Relay Source Sink RELAY CHANNEL: Cover, El Gamal (1979)

Gaussian network capacity: state of knowledge Resolved: Point-to-point channel, multiple access channel, broadcast channel (private messages). Unresolved: Relay Message 1 Decode message 1 Tx 1 Rx 1 Source Sink Message 2 Tx 2 Decode message 2 Rx 2 RELAY CHANNEL: Cover, El Gamal (1979) INTERFERENCE CHANNEL: Han Kobayashi (1981)

Gaussian network capacity: state of knowledge Resolved: Point-to-point channel, multiple access channel, broadcast channel (private messages). Unresolved: Relay Message 1 Decode message 1 Tx 1 Rx 1 Source Sink Message 2 Tx 2 Decode message 2 Rx 2 RELAY CHANNEL: Cover, El Gamal (1979) INTERFERENCE CHANNEL: Han Kobayashi (1981) Question: Thirty years have gone by... How can we make progress from here?

Simplify the model Focus on signal interaction not noise Observation: Success of network coding was through examination of flow on wireline networks, a special deterministic channel.

Simplify the model Focus on signal interaction not noise Observation: Success of network coding was through examination of flow on wireline networks, a special deterministic channel. Idea: Many wireless systems are interference rather than noise limited. Use deterministic channel model to focus on signal interaction and not noise.

Simplify the model Focus on signal interaction not noise Observation: Success of network coding was through examination of flow on wireline networks, a special deterministic channel. Idea: Hope: Many wireless systems are interference rather than noise limited. Use deterministic channel model to focus on signal interaction and not noise. Deterministic models more tractable. Use insight to obtain approximate characterizations for noisy (Gaussian) networks.

Simplify the model Focus on signal interaction not noise Observation: Success of network coding was through examination of flow on wireline networks, a special deterministic channel. Idea: Hope: Many wireless systems are interference rather than noise limited. Use deterministic channel model to focus on signal interaction and not noise. Deterministic models more tractable. Use insight to obtain approximate characterizations for noisy (Gaussian) networks. Question: Can we develop relevant models and analyze networks with deterministic signal interactions to get the insights?

Agenda: Introduce deterministic channel model. Motivate the utility of deterministic model with examples. Develop achievable rates for general deterministic relay networks Characterizations for linear finite field deterministic models.

Agenda: Introduce deterministic channel model. Motivate the utility of deterministic model with examples. Develop achievable rates for general deterministic relay networks Characterizations for linear finite field deterministic models. Connection to wireless networks: Use insights on achievability of deterministic networks to obtain approximate characterization of noisy relay networks.

Example 1: Point-to-point link Gaussian Deterministic y = 2 α/2 x + z A B Capacity is log(1 + 2 α ) α log 2 assuming unit variance noise. Receiver observes α most significant bits of transmitted signal. Number of levels received shows scale of channel strength. Scale important when signals interact in broadcast and multiple access.

Example 2: Multiple access channel DETERMINISTIC MODEL Rx GAUSSIAN MODEL B 1 D R 2 y = 2 α1/2 x 1 + 2 α2/2 x 2 + z. n 2 log(1 + SNR 2 ) B 2 Mod 2 addition log(1 + SNR 1 ) n 1 R 1

Example 3: Scalar broadcast channel Tx GAUSSIAN MODEL DETERMINISTIC MODEL A 1 S y 1 = 2 α1/2 x + z 1, y 2 = 2 α2/2 x + z 2 R 2 A 2 n 2 log(1 + SNR 2 ) Approximation of 1 bit log(1 + SNR 1 ) n 1 R 1

Relay channel: deterministic approximation GAUSSIAN RELAY CHANNEL R S 00 00 00 0 11 11 11 1 00 00 00 0000 11 11 11 1111 0000000000000 1111111111111 D

Relay channel: deterministic approximation GAUSSIAN RELAY CHANNEL R R S 00 00 00 11 11 11 00 00 00 0000 11 11 11 1111 0 1 0000000000000 11111111111110 1 D S D DETERMINISTIC MODEL

Relay channel: deterministic approximation GAUSSIAN RELAY CHANNEL R S 00 00 00 0 11 11 11 1 00 00 00 0000 11 11 11 1111 0000000000000 1111111111111 D S R D DETERMINISTIC MODEL Source cut = 3 bits

Relay channel: deterministic approximation GAUSSIAN RELAY CHANNEL R R S 00 00 00 11 11 11 00 00 00 0000 11 11 11 1111 0 1 0000000000000 11111111111110 1 D S D DETERMINISTIC MODEL Destination cut = 3 bits

Relay channel: deterministic approximation GAUSSIAN RELAY CHANNEL R R S 00 00 00 11 11 11 00 00 00 0000 11 11 11 1111 0 1 0000000000000 11111111111110 1 D S D DETERMINISTIC MODEL

Relay channel: deterministic approximation GAUSSIAN RELAY CHANNEL R R S 00 00 00 11 11 11 00 00 00 0000 11 11 11 1111 0 1 0000000000000 11111111111110 1 D S D DETERMINISTIC MODEL

Relay channel: deterministic approximation GAUSSIAN RELAY CHANNEL R R S 00 00 00 11 11 11 00 00 00 0000 11 11 11 1111 0 1 0000000000000 11111111111110 1 D S D DETERMINISTIC MODEL

Relay channel: deterministic approximation GAUSSIAN RELAY CHANNEL R R S 00 00 00 11 11 11 00 00 00 0000 11 11 11 1111 0 1 0000000000000 11111111111110 1 D S D DETERMINISTIC MODEL Cut-set bound achievable. Decode and forward is optimal.

Relay channel: deterministic approximation GAUSSIAN RELAY CHANNEL R S 00 00 00 0 11 11 11 1 00 00 00 0000 11 11 11 1111 0000000000000 1111111111111 D R S D gap 30 20 10 1 0 0.5 h RD 2 / h SD 2 0 10 30 20 10 20 0 10 h SR 2 / h SD 2 20 30 30 Result: Gap from cut-set less than 1 bit, on average much less. DETERMINISTIC MODEL Cut-set bound achievable. Decode and forward is optimal.

Diamond network Gaussian A h SA1 h A1 D S D h SA2 h A2 D B

Diamond network Deterministic Cut value = 3 Gaussian A A S h SA1 h A1 D D S h SA2 h A2 D D B B

Diamond network Deterministic Gaussian A A Cut value = 3 h SA1 h A1 D S D S h SA2 h A2 D D B B

Diamond network Gaussian A Deterministic Cut value = 6 A h SA1 h A1 D S D S h SA2 h A2 D D B B

Diamond network Gaussian A Deterministic A Cut value = 3 h SA1 h A1 D S D S h SA2 h A2 D D B B

Diamond network Deterministic Gaussian A A h SA1 h A1 D S D S h SA2 h A2 D D B Result: Gap from cut-set less 1 bit. Cut-set bound achievable. Partial decode-forward is optimal. B

Two-layer network Gaussian A 1 h A1 B 1 B 1 S h S,A1 h S,A2 h A1 B 2 h A2 B 1 h B1 D D hb2 D A 2 h A2 B B 2 2

Two-layer network Gaussian Deterministic A 1 h A1 B 1 B 1 h S,A1 h B1 D A1 B1 h A1 B 2 S D h S,A2 h A2 B 1 hb2 D S D A 2 h A2 B B 2 2 A2 B2 Cut-set bound achievable. Linear map and forward is optimal.

Two-layer network Gaussian Deterministic A 1 h A1 B 1 B 1 h S,A1 h B1 D A1 B1 h A1 B 2 S D h S,A2 h A2 B 1 hb2 D S D A 2 h A2 B B 2 2 Result: Gap from cut-set less than constant number of bits. A2 Cut-set bound achievable. Linear map and forward is optimal. B2

Questions Is the cut-set bound achievable for the deterministic model in arbitrary networks?

Questions Is the cut-set bound achievable for the deterministic model in arbitrary networks? What is the structure of the optimal strategy?

Questions Is the cut-set bound achievable for the deterministic model in arbitrary networks? What is the structure of the optimal strategy? Can we use insight from deterministic analysis to get approximately optimal strategy for Gaussian networks?

Algebraic representation S A1 b 1 y 1 2 b 2 y 2 b y 3 3 b y 4 4 b 5 y 5 S = 6 4 A2 B1 B2 D 0 0... 0 0 1 0 0... 0 0 1 0... 0........... 3 7 5 c 1 c 2 c 3 c 4 c 5 S is shift matrix of size q = max i,j n i,j. y B1 = y 1 y 2 y 3 y 4 y 5 = 0 0 b 1 b 2 b 3 0 0 0 c 1 c 2 = S5 3 x A1 S 5 2 x A2 = S 5 3 b S 5 2 c

Generalizations Linear finite field model Channel from i to j is described by channel matrix G ij operating over F 2. Received signal at node j: y j [t] = N G ij x i [t] i=1 Special case: our model given in examples G ij = S q α ij General deterministic network: y[t] = G(x 1 [t],..., x N [t]) Observation: Wireline networks are a special case.

Information-theoretic cut-set A1 B1 S D A2 B2 Ω Λ D Ω c Cut: Separates S from D Cut transfer matrix G Ω,Ω c: Transfer function from nodes in Ω to Ω c.

Cutset upper bound General relay network: C relay C = max min I(X Ω ; Y Ω c X Ω c) p(x 1,...,x N ) Ω

Cutset upper bound General relay network: C relay C = General deterministic relay network: C relay C = max min I(X Ω ; Y Ω c X Ω c) p(x 1,...,x N ) Ω max min H(Y Ω c X Ω c) p(x 1,...,x N ) Ω

Cutset upper bound General relay network: C relay C = General deterministic relay network: C relay C = max min I(X Ω ; Y Ω c X Ω c) p(x 1,...,x N ) Ω max min H(Y Ω c X Ω c) p(x 1,...,x N ) Ω Linear finite field network: Optimal input distribution x 1,..., x N independent and uniform C relay C = min Ω rank(g Ω,Ω c) where G Ω,Ω c is the transfer matrix X Ω Y Ω c.

Main results: Deterministic relay networks Theorem (Avestimehr, Diggavi and Tse, 2007) Given a general deterministic relay network (with broadcast and multiple access), we can achieve all rates R upto Multicast information flow: Qmax min H(Y Ω c X Ω c) i p(x i ) Ω Theorem (Avestimehr, Diggavi and Tse, 2007) Given a general deterministic relay network (with broadcast and multiple access), we can achieve all rates R from S multicasting to all destinations D D up to, Qmax min min H(Y Ω c X Ω c) i V p(x i) D D Ω Λ D

Application Linear deterministic models Corollary (Avestimehr, Diggavi and Tse, 2007) Given a linear finite-field relay network (with broadcast and multiple access), the capacity C of such a relay network is given by, Multicast information flow: C = min Ω Λ D rank(g Ω,Ω c). Corollary (Avestimehr, Diggavi and Tse, 2007) Given a linear finite-field relay network (with broadcast and multiple access), the multicast capacity C of such a relay network is given by, C = min D D min Ω Λ D rank(g Ω,Ω c).

Consequences: Deterministic Relay Networks General deterministic networks: Cutset upper bound was C relay max p(x 1,...,x N ) min Ω H(Y Ω c X Ω c) = achievable if optimum was product distribution. Linear finite field model: Cutset upper bound was C relay min Ω rank(g Ω,Ω c) = cutset bound achievable For wireline graph model rank(g Ω,Ω c) is number of links crossing the cut. Observation: We have a generalization of Ford-Fulkerson max-flow min-cut theorem to linear finite field relay networks with broadcast and multiple access.

Main results: Gaussian relay networks Theorem (Avestimehr, Diggavi and Tse, 2007) Given a Gaussian relay network, G, we can achieve all rates R up to C κ. Therefore the capacity of this network satisfies C κ C C, where C is the cut-set upper bound on the capacity of G, and κ is a constant independent of channel gains. Theorem (Multicast information flow) Given a Gaussian relay network, G, we can achieve all multicast rates R up to C mcast κ, i.e., for C mcast = min D D C D, C mcast κ C C mcast

Ingredients and insights Main steps: Gaussian strategy Relay operation: Quantize received signal at noise-level. Relay function: Random mapping from received quantized signal to transmitted signal. Handle unequal (multiple) paths between nodes like inter-symbol interference. Consequences: With probabilistic method we demonstrate min-cut achievability for linear deterministic networks. Gaussian networks constant gap independent of SNR operating point. Engineering insight of (almost) optimal coding strategies.

Compound relay networks Compound model: Channel realizations from a set h i,j H i,j, unknown to sender. Observations: Theorem Relay strategy does not depend on the channel realization. Overall network from source to destination behaves like a compound channel. Utilize point-to-point compound channel ideas get approximate characterization for compound network. Given a compound Gaussian relay network the capacity C cn satisfies C cn κ C cn C cn, where C cn = max p({xj } j V ) inf h H min Ω ΛD I(Y Ω c; X Ω X Ω c).

: Open questions and extensions Extensions: Outage set behavior for full duplex networks. Analysis of half-duplex systems with fixed transmit fractions. Ergodic channel variations. Open questions: D-M trade-off for channel dependent half-duplex systems. Tightening gap to cut-set bound. Use deterministic model directly to get Gaussian result.

Extensions of deterministic approach Interference channel: Successfully used to generate approximate characterization (Bresler and Tse, 2007), K -user interference channel: Used to demonstrate new phenomenon of interference alignment (Bresler-Tse, 2007, Jafar 2007). Relay-interference networks: Extension of multiple unicast to wireless networks (Mohajer, Diggavi, Fragouli and Tse, 2008). Wireless network secrecy: Used to demonstrate secrecy over networks (Diggavi, Perron and Telatar, 2008). Network data compression: Identify correct multi-terminal lossless structures to get approximations next topic.

Problem statement Encoder architecture Main results Extensions and discussion coding: route diversity Route 1 SERVER Route 2 CLIENT Route 3 Route diversity: Multiple routes from source to destination. Goal: Graceful degradation in performance with route failures = multiple description source coding. Generate binary streams with rate constraints with distortion guarantees when only subset of routes succeed.

Symmetric multiple description Problem statement Encoder architecture Main results Extensions and discussion X X, X,..., 1, 2 3 X n D 100

Symmetric multiple description Problem statement Encoder architecture Main results Extensions and discussion X X, X,..., 1, 2 3 X n D 0

Symmetric multiple description Problem statement Encoder architecture Main results Extensions and discussion X X, X,..., 1, 2 3 X n!"#$%#&' (!"#$%#&' )!"#$%#&' * ++, D 0

Symmetric multiple description Problem statement Encoder architecture Main results Extensions and discussion X X, X,..., 1, 2 3 X n -./234256 7 -./234256 8 -./234256 9 ::; D 110

Symmetric multiple description Problem statement Encoder architecture Main results Extensions and discussion X X, X,..., 1, 2 3 X n <=>?@ABCADE F <=>?@ABCADE G <=>?@ABCADE H IJI D 1

Symmetric multiple description Problem statement Encoder architecture Main results Extensions and discussion X X, X,..., 1, 2 3 X n KLMNOPQRPST U KLMNOPQRPST V KLMNOPQRPST W XYY D 1

Symmetric multiple description Problem statement Encoder architecture Main results Extensions and discussion X X, X,..., 1, 2 3 X n Z[\]^_`a_bc d Z[\]^_`a_bc e Z[\]^_`a_bc f ggg D 111

Symmetric multiple description Problem statement Encoder architecture Main results Extensions and discussion X X, X,..., 1, 2 3 X n hijklmnompq r hijklmnompq s hijklmnompq t uvv v uv vvu uuv uv u vuu uuu = = = = D100 D0 D0 D110 D1 D1 111 D Goal: Characterize tuple (R 1, R 2, R 3, D 1, D 2, D 3 ), for Gaussian quadratic source, where R i are rates on descriptions and D i are distortions for i successful descriptions.

Problem statement Encoder architecture Main results Extensions and discussion Simple architecture: lossless multilevel source codes a 1 D 1 a 2 U 1 SOURCE Layered Encoder a p b 1 D 2 U 2 b q c 1 c 2 D 3 U 3 Symmetric multi-level coding

Overall strategy Problem statement Encoder architecture Main results Extensions and discussion Approach: Identify underlying lossless coding problem and solve rate region. Use polytopic lossless rate region as template. Derive outer bound using intuition from the template. Technical ideas: New lower bounding technique: Expand auxiliary random variable space to K 1 for K -descriptions. Intuition: Each auxiliary variable captures distortion level. Use structure for auxiliary variables to obtain lower bound to match inner bound region.

Problem statement Encoder architecture Main results Extensions and discussion An approximate characterization for SR-MLD R 3 w{y ~~x 2R + R + R 1 w{y }z 2 3 R + 1 + R2 R3 R 1 + R 3 wxyzxz{ R 2 R 1 Bottom line: Simple architecture almost optimal!

Problem statement Encoder architecture Main results Extensions and discussion Improved approximation using binning scheme R 3 R + 1 + R2 R3 2R + R + R 1 2 3 R 1 + R 3 ƒ R 2 R 1 have been generalized to K > 3 using bounding hyperplanes specification.

Extensions Problem statement Encoder architecture Main results Extensions and discussion Symmetric MD Lower bound extension to arbitrary K description problem (DCC 2008). Approximation for K > 3 rate-region using the symmetric MLD (K > 3) insights (ISIT 2008). Extension to non-gaussian sources. Asymmetric problem Solved underlying 3-description lossless asymmetric multi-level coding rate region (DCC 2008) Used insight to approximate asymmetric Gaussian MD rate region (ISIT 2008).

Program: Hope: Focus on underlying deterministic or lossless coding problem this identification is a central challenge. Obtain exact characterization of underlying problem. Use insight to obtain approximation of rate region of noisy/lossy problem. The program will yield insight to network flow problems. Exposes the central difficulties, solution insights and new schemes? Approximations may be sufficient for engineering practice. Papers/preprints: http://licos.epfl.ch