I have posted a corrected version of my ISIT paper on randomized coding for channels whose state can depend non-causally on the transmitted codeword. It turns out the result as presented there was incorrect. Perhaps if I get a little more time I’ll write something expository on my little nook of information theory.
Author Archives: Anand Sarwate
office safety
When running a fast optimization tool such as CVX, it is important to wear a helmet to protect you in case of crashes or attacks by monkeys:
SF Fringe : The Sewers
On Tuesday R and I saw The Sewers, a production from the New York theater company Banana Bag and Bodice. The play was part of the SF Fringe Festival, and is running through this weekend. I really enjoyed this play — the visual elements, disjointed narrative, and intense performances were exactly the kind of theater I’ve been itching to see:
This show is a conjuring act; an entire, albeit, tiny village by the name of The Sewers mysteriously appears one night in the theatre. All the children are dead. An acid plant in a barn. A triangular shaped love tryst. This show is a tour de force by manipulation.
If you are a fan of Richard Foreman, you will like it for sure, I think — it was done at the Ontological Theater in NY. The best thing about it is that it affects you in a way that is hard to articulate, and the process of trying to express to yourself what you think about it is fodder for hours of thought. Go see it!
ITW 2007 (Tahoe) : Day Two
- Plenary Talk: Source and Channel Coding: Theory, Practice and System Aspects (Giuseppe Caire)
This focus of this talk was on the benefits of using joint source channel coding for more “practical” systems. The main motivation was the catastrophic behavior of excess channel errors on source reconstruction. If you optimally compress an image and send it over a noisy channel, the reconstruction becomes terrible if the channel is a bit noisier than you expected. Since wireless channel variations are a real problem, this can have a real impact. Can we have a graceful degradation in the reconstruction quality via JSSC? He also gave two more examples, one from beamforming on the MIMO broadcast channel, and one about multicasting a Gaussian source over a Gaussian MAC using a Costa-like layering construction.
- The Case for Structured Random Codes in Network Communication Theorems (Bobak Nazer and Michael Gastpar)
This paper focused on giving some motivations for exploiting a structured approach to random codes, rather than the standard information theoretic constructions we all know and love. For problems such as computing over a multiple-access channel, linear network coding, or interference networks, separating source and channel coding can result in performance losses. A simple scheme based on lattices can get better performance in these cases.
- On Separation in the Presence of Feedback (Haim Permuter and Tsachy Weissman)
It is well-known that separation is optimal for point-to-point channels and that the directed mutual information gives the capacity of channels with feedback. For ergodic channels, they show that for construction with distortion D the condition CFB > R(D) is necessary and sufficient, which shows that separation is optimal in this setting as well. For certain MACs which can be decomposed into a “multiplexer plus point-to-point” channel, they can show a similar result. In the talk they also looked at a binary MAC with stationary ergodic noise and feedback, and showed that separation is optimal there. Their results extend to mod-additive MACs.
- Equivalence of LP Relaxation and Max-Product for Weighted Matching in General Graphs (Sujay Sanghavi)
The problem of maximum weight matching in a graph with edge weights is to find the largest weight set of disjoint edges. One algorithm for doing this is max-product, which is an iterative message passing algorithm. The matching problem can also be posed as an integer program, which has a linear programming relaxation. Sujay shows that the performance of max-product is tied to the tightness of the LP — if the LP is tight then MP converges to the correct answer, and if the LP is loose then MP will not converge. This provides a nice guarantee of when MP will converge that ties together the iterative and linear-programming approaches to optimization on graphs. Clearly this has relations to decoding, but I’ll let Alex comment on those.
- On the Duality and Difference Between Slepian-Wolf Coding and Channel Coding (Jun Chen, Da-ke He, Ashish Jagmohan, and Luis A. Lastras-Montano)
This paper explored the similarities between channel coding and Slepian-Wolf coding. The comment made by Jun Chen at the talk was that the “difference” mentioned in the title is actually another kind of duality. By duality, they are referring to a kind of mirror pairing of the error exponents for source and channel coding. Using relations between linear coding strategies shows that nonlinear Slepian Wolf codes can strictly outperform linear codes.
- On Network Interference Management (Aleksandar Jovicic, Hua Wang, and Pramod Viswanath)
This paper looked at two basic interference patterns — one long range link causing problems for many short range links, and one long range link experiencing interference from many short links. The question is this — is orthogonalization/degree-of-freedom sharing/reuse optimal? They characterize the high SNR degree-of-freedom range for both scenarios using a multi-layer superposition coding scheme and use genie-aided converse arguments to get a “within one bit” characterization in the spirit of the Etkin/Tse/Wang approach to interference channels. In the latter interference model they also show that power control can get similar performance.
- Capacity for Classes of Broadcast Channels with Receiver Side Information (Gerhard Kramer and Shlomo Shamai (Shitz))
The channel model here is a broadcast channel where each user has as side information the other user’s message. The coding schemes used are nested lattice codes, and they show that they can get all capacity points — furthermore they can generalize their scheme to cost-constrained channels and coding for degraded message sets. At the end of the talk there was some very recent work by a student working with Gerhard (Wei Kang, I believe?) who showed a similar result for the case of general correlated sources with a certain kind of degraded side information at one receiver.
- Successive Refinement of Vector Sources under Individual Distortion Criteria (Jayanth Nayak, Ertem Tuncel, Deniz Gunduz, and Elza Erkip)
The problem here is to make a successive refinement code (if you get more of the codeword you can decode to lower distortion) for vector sources where we have a distortion constraint on each of the components of the code. If we look at a two sources (vectors of length 2) we can mark achievable distortion pairs (D1, D2) on the plane and ask “what kind of trajectory can we make with successive refinement codes? It turns out the answer is rather complicated — for Gaussian sources we get three regions and there are rules about what paths we can take.
- On Cognitive Interference Networks (Amos Lapidoth, Shlomo Shamai (Shitz), and Michele A. Wigger)
This is a follow-up to the ISIT paper, which focused on characterizing the pre-log of the sum-capacity of a certain interference network that looked like a chain. The “cognitive” aspect of the problem is that the transmitters know the message of their interfering neighbors. They can use some linear encoding to zero-force their own interference and then a number of results pop out — conditions on when the pre-log for K users can be K or K-1, necessary and sufficient conditions on the network for achieving K, and the fact that pre-log can be noninteger, unlike the case where no side information is available.
- Reliable Communication in Networks with Multi-access Interference (Danail Traskov and Gerhard Kramer)
Danail presented an interesting scheme for generating correlated messages in networks (with an eye to network coding, it seemed). As a simple example, a source transmitter sends rate-constrained messages to two relays who then communicate to the destination. This is similar to the Schein-Gallager network but without the broadcast component at the beginning. The source can correlate the messages at the relays, and in some cases they can get a capacity result. More importantly, the scheme scales to more general networks of MACs and has a nice “flow” interpretation.
ITW 2007 (Tahoe) : Day One
A lot of my comments will be short since in retrospect my notes were quite poor for most of the talks I attended. I attribute it to fatigue… I also skipped out on a number of talks to think over some new problems that had come to mind. I’m going to let Alex blog about some of the coding talks.
- Plenary Talk: On Modeling Distributions over Large Alphabets (Alon Orlitsky)
Orlitsky started with the non-paradoxical nature of the “birthday paradox.” After querying people in the audience one by one, he finally ended with two people who had the same birthday (9/3), namely Shlomo Shamai and Rüdiger Urbanke. This was all a set-up of course, but I honestly thought Urbanke was just sick of the process and tried to end it. One way of turning the birthday paradox on its head is to ask “if we saw that it took K queries to find two people with the same birthday, how should we estimate the length of the year?” This is precisely the problem of large alphabet size estimation with small sample size. He then went on to tour the techniques of patterns, profiles, and the relation to capture-recapture statistics. I had seen a lot of this work before, but it was nice to see it again and a bit more slowly.
- Power Allocation for Discrete-Input Non-Ergodic Block-Fading Channels (Khoa D. Nguyen, Albert Guillén i Fàbregas, and Lars K. Rasmussen)
This work was about finding the SNR exponent of the outage probability for power allocation strategies over a particular class of fading channels (Nakagami). They differentiated between long-term and short-term power allocation and found a relationship between the short-term optimal scheme and a Singleton bound. The more interesting part is that suboptimal, lower complexity schemes are nearly optimal in this outage-exponent sense.
- Probabilistic Capacity and Optimal Coding for Asynchronous Channel (Ning Cai, Siu-Wai Ho, and Raymond W. Yeung)
This paper looked at time jitter in a continuous time model, and showed that run-length limited (RLL) codes are suboptimal for this kind of asynchrony. When the jitter is bounded in some sense they found 0-error codes, and showed that “interval codes” are better than RLL codes. They used an approximation for the rate loss involving characteristic functions in the case where we demand a bound on the time that the encoder and decoder can be out of synch. They come up with a channel model and capacity formulation and design new capacity achieving codes.
- The Effect of Finite Memory on Throughput of Wireline Packet Networks (Badri N. Vellambi, Nazanin Rahnavard, and Faramarz Fekri)
In a packet erasure network on a line, with finite buffer sizes at the nodes, what are the benefits of network coding? The authors came up with a large Markov chain model and have to find the steady state distribution to get the capacity of the network. Unfortunately, the chain is a bit unwieldy so they come up with approximations for the distribution. Some simulations are also provided.
you say routing, I say routing…
Let’s call the whole thing off. One might say the enemy army was “routed,” but do we ever use the word “routing” in that sense? It sounds wrong to me — does it only appear as a participle?
In the networking context though, people say “routing” either to rhyme with “pouting” or with “tooting.” I’d use the latter for “route 66” but I usually use the former for networking.
In case these linguistic musings bore people, fear not — I will write about other things soon.
Guest Blogger : Alex Dimakis
My esteemed colleague Alexandros Dimakis will be helping blog the 2007 Information Theory Workshop in Lake Tahoe (and maybe MMSP and other exciting things in the future). This should be exciting!
sequences and series
I think that learning math has colored my ideas about the connotations of words. In particular, the words “sequence” and “series” have the meanings “a set of related things that follow each other in a particular order” and “a number of things of a similar nature coming one after the other.” They appear to be mostly interchangeable. But consider the phrase “in a sequence/series of of papers, Csiszár and Narayan proved…” Is there a difference in meaning?
To me, “series” connotes a cumulative effect — the set of papers build upon each other, as in the summation of a series encountered in calculus. The word “sequence” is milder — these are set of related papers that follow chronologically, but may look at different angles of the same problem rather than building on each other. Clearly this difference is leaking in from the technical definitions into my writing. Does this happen to anyone else?
SSP 2007 : aftermath
I just finished up attending the 2007 Statistical Signal Processing Workshop in Madison. Normally I would have blogged about all the talks I saw, but (a) the only “talks” were tutorials and plenaries, and (b) I’m a little burned out to write much. Despite the fact that I applied to grad schools for signal processing and took the DSP prelim exam at Berkeley, I’m not really much of a signal processing guy these days. All of the submitted papers were given as posters, and despite being organized into “sessions,” all the posters were in the same room, so there were about 30-40 talks going on at the same time in parallel for 2 hours. I was a bit dubious at first, since my experience with poster presentations is that they have a large effort-to-value ratio, but this format worked for me. I was unfamiliar with about 80% of the problems that people were trying to solve, so going to talks would have made me confused. Instead, I could at least talk to someone and get the point of what they were trying to do, if not the scale of their contribution.
The one downside to the conference for me was that several of the posters that I wanted to see were in the same session as me, so I ended up missing them! Luckily I was next to “Distributed Average Consensus using Probabilistic Quantization,” which is right up my alley (from my work on gossip algorithms), but I could only listen in every once in a while. If only we could encode our talks using an erasure code — then if I listen to 7 minutes our of 10 I could interpolate the other 3…
Nerd Wisdom
I learned today in a meeting that Jonathan Yedidia has a blog. He seems more dedicated to his than I am to mine, but maybe as I enter into heavy-writing mode this semester that will change…
