On the limits of communication with low-precision analog-to-digital conversion at the receiver

11
IEEE TRANSACTIONS ON COMMUNICATIONS, VOL. 57, NO. 12, DECEMBER 2009 3629 On the Limits of Communication with Low-Precision Analog-to-Digital Conversion at the Receiver Jaspreet Singh, Student Member, IEEE, Onkar Dabeer, Member, IEEE, and Upamanyu Madhow, Fellow, IEEE Abstractโ€”As communication systems scale up in speed and bandwidth, the cost and power consumption of high-precision (e.g., 8-12 bits) analog-to-digital conversion (ADC) becomes the limiting factor in modern transceiver architectures based on digital signal processing. In this work, we explore the impact of lowering the precision of the ADC on the performance of the communication link. Speci๏ฌcally, we evaluate the communi- cation limits imposed by low-precision ADC (e.g., 1-3 bits) for transmission over the real discrete-time Additive White Gaussian Noise (AWGN) channel, under an average power constraint on the input. For an ADC with quantization bins (i.e., a precision of log 2 bits), we show that the input distribution need not have any more than +1 mass points to achieve the channel capacity. For 2-bin (1-bit) symmetric quantization, this result is tightened to show that binary antipodal signaling is optimum for any signal- to-noise ratio (SNR). For multi-bit quantization, a dual formula- tion of the channel capacity problem is used to obtain tight upper bounds on the capacity. The cutting-plane algorithm is employed to compute the capacity numerically, and the results obtained are used to make the following encouraging observations : (a) up to a moderately high SNR of 20 dB, 2-3 bit quantization results in only 10-20% reduction of spectral ef๏ฌciency compared to un- quantized observations, (b) standard equiprobable pulse ampli- tude modulated input with quantizer thresholds set to implement maximum likelihood hard decisions is asymptotically optimum at high SNR, and works well at low to moderate SNRs as well. Index Termsโ€”Channel capacity, optimum input distribution, AWGN channel, analog to digital converter, quantization. I. I NTRODUCTION D IGITAL signal processing (DSP) forms the core of modern digital communication receiver implementations, with the analog baseband signal being converted to digital form using analog-to-digital converters (ADCs) which typi- cally have fairly high (e.g., 8-12 bits) precision. Operations such as synchronization, equalization and demodulation are then performed in the digital domain, greatly enhancing the ๏ฌ‚exibility available to the designer. The continuing exponential Paper approved by F. Alajaji, the Editor for Source and Source/Channel Coding of the IEEE Communications Society. Manuscript received October 24, 2008; revised March 6, 2009. J. Singh and U. Madhow are with the Department of Electrical and Computer Engineering, University of California, Santa Barbara, CA, 93106, USA (e-mail: {jsingh, madhow}@ece.ucsb.edu). Their research was supported by the National Science Foundation under grants CCF-0729222 and CNS- 0832154. O. Dabeer is with the School of Technology and Computer Science, Tata Institute of Fundamental Research, Mumbai, 400005, India (e-mail: [email protected]). His research was supported by the Homi Bhabha Fellowship. Digital Object Identi๏ฌer 10.1109/TCOMM.2009.12.080559 ADC Quantizer Q + X Y N Fig. 1. = Q( + ) : The AWGN-Quantized Ouput channel induced by the output quantizer Q. advances in digital electronics, often summarized by Mooreโ€™s โ€œlawโ€ [1], imply that integrated circuit implementations of such DSP-centric architectures can be expected to continue scaling up in speed and down in cost. However, as the bandwidth of a communication system increases, accurate conversion of the analog received signal into digital form requires high-precision, high-speed ADC, which is costly and power-hungry [2]. One possible approach for designing such high-speed systems is to drastically reduce the number of bits of ADC precision (e.g., to 1-3 bits) as sampling rates scale up. Such a design choice has signi๏ฌcant implications for all aspects of receiver design, including carrier and timing synchronization, equalization, demodulation and decoding. However, before embarking on a comprehensive rethinking of communication system design, it is important to understand the fundamental limits on communication performance im- posed by low-precision ADC. In this paper, we take a ๏ฌrst step in this direction, investigating Shannon-theoretic performance limits for the following idealized model : linear modulation over a real baseband Additive White Gaussian Noise (AWGN) channel, with symbol rate Nyquist samples quantized by a low-precision ADC at the receiver. This induces a discrete- time memoryless AWGN-Quantized Output channel, which is depicted in Figure 1. Under an average power constraint on the input, we obtain the following results: 1) For -bin (i.e., log 2 bits) output quantization, we prove that the input distribution need not have any more than +1 mass points to achieve the channel capacity. (Numerical computation of optimal input distributions reveals that mass points are suf๏ฌcient.) An intermediate result of interest is that, when the AWGN channel output is quantized with ๏ฌnite-precision, an average power constraint on the input leads to an 0090-6778/09$25.00 c โƒ 2009 IEEE

Transcript of On the limits of communication with low-precision analog-to-digital conversion at the receiver

IEEE TRANSACTIONS ON COMMUNICATIONS, VOL. 57, NO. 12, DECEMBER 2009 3629

On the Limits of Communication withLow-Precision Analog-to-Digital

Conversion at the ReceiverJaspreet Singh, Student Member, IEEE, Onkar Dabeer, Member, IEEE, and Upamanyu Madhow, Fellow, IEEE

Abstractโ€”As communication systems scale up in speed andbandwidth, the cost and power consumption of high-precision(e.g., 8-12 bits) analog-to-digital conversion (ADC) becomes thelimiting factor in modern transceiver architectures based ondigital signal processing. In this work, we explore the impactof lowering the precision of the ADC on the performance ofthe communication link. Specifically, we evaluate the communi-cation limits imposed by low-precision ADC (e.g., 1-3 bits) fortransmission over the real discrete-time Additive White GaussianNoise (AWGN) channel, under an average power constraint onthe input. For an ADC with ๐พ quantization bins (i.e., a precisionof log2 ๐พ bits), we show that the input distribution need not haveany more than ๐พ+1 mass points to achieve the channel capacity.For 2-bin (1-bit) symmetric quantization, this result is tightenedto show that binary antipodal signaling is optimum for any signal-to-noise ratio (SNR). For multi-bit quantization, a dual formula-tion of the channel capacity problem is used to obtain tight upperbounds on the capacity. The cutting-plane algorithm is employedto compute the capacity numerically, and the results obtained areused to make the following encouraging observations : (a) up toa moderately high SNR of 20 dB, 2-3 bit quantization resultsin only 10-20% reduction of spectral efficiency compared to un-quantized observations, (b) standard equiprobable pulse ampli-tude modulated input with quantizer thresholds set to implementmaximum likelihood hard decisions is asymptotically optimumat high SNR, and works well at low to moderate SNRs as well.

Index Termsโ€”Channel capacity, optimum input distribution,AWGN channel, analog to digital converter, quantization.

I. INTRODUCTION

D IGITAL signal processing (DSP) forms the core ofmodern digital communication receiver implementations,

with the analog baseband signal being converted to digitalform using analog-to-digital converters (ADCs) which typi-cally have fairly high (e.g., 8-12 bits) precision. Operationssuch as synchronization, equalization and demodulation arethen performed in the digital domain, greatly enhancing theflexibility available to the designer. The continuing exponential

Paper approved by F. Alajaji, the Editor for Source and Source/ChannelCoding of the IEEE Communications Society. Manuscript received October24, 2008; revised March 6, 2009.

J. Singh and U. Madhow are with the Department of Electrical andComputer Engineering, University of California, Santa Barbara, CA, 93106,USA (e-mail: {jsingh, madhow}@ece.ucsb.edu). Their research was supportedby the National Science Foundation under grants CCF-0729222 and CNS-0832154.

O. Dabeer is with the School of Technology and Computer Science,Tata Institute of Fundamental Research, Mumbai, 400005, India (e-mail:[email protected]). His research was supported by the Homi BhabhaFellowship.

Digital Object Identifier 10.1109/TCOMM.2009.12.080559

ADC QuantizerQ+X Y

N

Fig. 1. ๐‘Œ = Q(๐‘‹ + ๐‘) : The AWGN-Quantized Ouput channel inducedby the output quantizer Q.

advances in digital electronics, often summarized by Mooreโ€™sโ€œlawโ€ [1], imply that integrated circuit implementations ofsuch DSP-centric architectures can be expected to continuescaling up in speed and down in cost. However, as thebandwidth of a communication system increases, accurateconversion of the analog received signal into digital formrequires high-precision, high-speed ADC, which is costly andpower-hungry [2]. One possible approach for designing suchhigh-speed systems is to drastically reduce the number ofbits of ADC precision (e.g., to 1-3 bits) as sampling ratesscale up. Such a design choice has significant implications forall aspects of receiver design, including carrier and timingsynchronization, equalization, demodulation and decoding.However, before embarking on a comprehensive rethinking ofcommunication system design, it is important to understandthe fundamental limits on communication performance im-posed by low-precision ADC. In this paper, we take a first stepin this direction, investigating Shannon-theoretic performancelimits for the following idealized model : linear modulationover a real baseband Additive White Gaussian Noise (AWGN)channel, with symbol rate Nyquist samples quantized by alow-precision ADC at the receiver. This induces a discrete-time memoryless AWGN-Quantized Output channel, which isdepicted in Figure 1. Under an average power constraint onthe input, we obtain the following results:

1) For ๐พ-bin (i.e., log2 ๐พ bits) output quantization, weprove that the input distribution need not have any morethan ๐พ+1 mass points to achieve the channel capacity.(Numerical computation of optimal input distributionsreveals that ๐พ mass points are sufficient.) Anintermediate result of interest is that, when the AWGNchannel output is quantized with finite-precision, anaverage power constraint on the input leads to an

0090-6778/09$25.00 cโƒ 2009 IEEE

3630 IEEE TRANSACTIONS ON COMMUNICATIONS, VOL. 57, NO. 12, DECEMBER 2009

implicit peak power constraint, in the sense that anoptimal input distribution must have bounded support.

2) For 1-bit symmetric quantization, the preceding resultis tightened analytically to show that binary antipodalsignaling is optimal for any signal-to-noise ratio (SNR).

3) For multi-bit quantizers, tight upper bounds on capacityare obtained using a dual formulation of the channelcapacity problem. Near-optimal input distributionsthat approach these bounds are computed using thecutting-plane algorithm [3].

4) While the preceding results optimize the inputdistribution for a fixed quantizer, comparison withan unquantized system requires optimization over thechoice of the quantizer as well. We numerically obtainoptimal 2-bit and 3-bit symmetric quantizers.

5) From our numerical results, we infer that the use of low-precision ADC incurs a relatively small loss in spectralefficiency compared to unquantized observations. Forexample, at 0 dB SNR, a receiver with 2-bit ADCachieves 95% of the spectral efficiency attained withunquantized observations. Even at a moderately highSNR of 20 dB, a receiver with 3-bit ADC achieves 85%of the spectral efficiency attained with unquantizedobservations. This indicates that DSP-centric designbased on low-precision ADC is indeed attractive ascommunication bandwidths scale up, since the small lossin spectral efficiency should be acceptable in this regime.Furthermore, we observe that a โ€œsensibleโ€ choice ofstandard equiprobable pulse amplitude modulated(PAM) input with ADC thresholds set to implementmaximum likelihood (ML) hard decisions achieves per-formance which is quite close to that obtained by numer-ical optimization of the quantizer and input distribution.

Related Work: For a discrete memoryless channel (DMC),Gallager first showed that the number of input points withnonzero probability mass need not exceed the cardinalityof the output [4, p. 96, Corollary 3]. In our setting, theinput alphabet is not a priori discrete, and there is a powerconstraint, so that the result in [4] does not apply. Our keyresult on the achievability of the capacity by a discrete inputis actually inspired by Witsenhausenโ€™s result in [5], whereDubinsโ€™ theorem [6] was used to show that the capacity ofa discrete-time memoryless channel with output cardinality๐พ , under a peak power constraint is achievable by a discreteinput with at most ๐พ points. The key to our proof is to showthat under output quantization, an average power constraintautomatically induces a peak power constraint, after whichwe can use Dubinsโ€™ theorem in a manner similar to thedevelopment in [5] to show that ๐พ + 1 mass points sufficeto achieve the average power constrained capacity.

Prior work on the effect of reduced ADC precision onchannel capacity with fixed input distribution includes [7]โ€“[9]. However, other than our own results reported earlier in[10]โ€“[12], we are not aware of an information-theoretic inves-tigation with low-precision ADC that includes optimizationof the input distribution. Another related class of problemsthat deserves mention relates to the impact of finite-precisionquantization on the information-theoretic measure of channelcut-off rate rather than channel capacity (see, e.g. [13], [14]).

Given the encouraging results here, it becomes importantto explore the impact of low-precision ADC on receiver taskssuch as synchronization and equalization, which we haveignored in our idealized model (essentially assuming that thesetasks have somehow already been accomplished). Relatedwork on estimation using low-precision samples which maybe relevant for this purpose includes the use of dither forsignal reconstruction [15]โ€“[17], frequency estimation using 1-bit ADC [18], [19], choice of quantization threshold for signalamplitude estimation [20], and signal parameter estimationusing 1-bit dithered quantization [21], [22].

Organization of the Paper: The rest of the paper isorganized as follows. In Section II, we describe the channelmodel and present results concerning the structure of theoptimal input distributions. In Section III, we discuss capacitycomputation, including duality-based upper bounds on capac-ity. Numerical results are provided in Section IV, followed bythe conclusions in Section V.

II. CHANNEL MODEL AND STRUCTURE OF OPTIMAL

INPUT DISTRIBUTIONS

We consider linear modulation over a real AWGN channel,with symbol rate Nyquist samples quantized by a ๐พ-bin quan-tizer Q at the receiver. This induces the following discrete-timememoryless AWGN-Quantized Output (AWGN-QO) channel

๐‘Œ = Q (๐‘‹ +๐‘) . (1)

Here ๐‘‹ โˆˆ โ„ is the channel input with cumulative distributionfunction ๐น (๐‘ฅ), ๐‘Œ โˆˆ {๐‘ฆ1, โ‹… โ‹… โ‹… , ๐‘ฆ๐พ} is the (discrete) channeloutput, and ๐‘ is ๐’ฉ (0, ๐œŽ2) (the Gaussian random variable withmean 0 and variance ๐œŽ2). Q maps the real valued input ๐‘‹ +๐‘ to one of the ๐พ bins, producing a discrete output ๐‘Œ . Inthis work, we only consider quantizers for which each bin isan interval of the real line. The quantizer Q with ๐พ bins istherefore characterized by the set of its (๐พ โˆ’ 1) thresholds๐‘ž๐‘ž๐‘ž := [๐‘ž1, ๐‘ž2, โ‹… โ‹… โ‹… , ๐‘ž๐พโˆ’1] โˆˆ โ„

๐พโˆ’1, such that โˆ’โˆž := ๐‘ž0 <๐‘ž1 < ๐‘ž2 < โ‹… โ‹… โ‹… < ๐‘ž๐พโˆ’1 < ๐‘ž๐พ := โˆž. The output ๐‘Œ is assignedthe value ๐‘ฆ๐‘– when the quantizer input (๐‘‹ + ๐‘) falls in the๐‘–๐‘กโ„Ž bin, which is given by the interval (๐‘ž๐‘–โˆ’1, ๐‘ž๐‘–]. The resultingtransition probability functions are

๐‘Š๐‘–(๐‘ฅ) = P(๐‘Œ = ๐‘ฆ๐‘–โˆฃ๐‘‹ = ๐‘ฅ)

= ๐‘„

(๐‘ž๐‘–โˆ’1 โˆ’ ๐‘ฅ

๐œŽ

)โˆ’๐‘„

(๐‘ž๐‘– โˆ’ ๐‘ฅ

๐œŽ

), 1 โ‰ค ๐‘– โ‰ค ๐พ,

(2)

where ๐‘„(โ‹…) is the complementary Gaussian distribution func-tion,

๐‘„(๐‘ง) =1โˆš2๐œ‹

โˆซ โˆž

๐‘ง

exp(โˆ’๐‘ก2/2)๐‘‘๐‘ก . (3)

The Probability Mass Function (PMF) of the output ๐‘Œ , cor-responding to the input distribution ๐น is

๐‘…(๐‘ฆ๐‘–;๐น ) =

โˆซ โˆž

โˆ’โˆž๐‘Š๐‘–(๐‘ฅ)๐‘‘๐น (๐‘ฅ), 1 โ‰ค ๐‘– โ‰ค ๐พ, (4)

SINGH et al.: ON THE LIMITS OF COMMUNICATION WITH LOW-PRECISION ANALOG-TO-DIGITAL CONVERSION AT THE RECEIVER 3631

and the input-output mutual information ๐ผ(๐‘‹ ;๐‘Œ ), expressedexplicitly as a function of ๐น is

๐ผ(๐น ) =

โˆซ โˆž

โˆ’โˆž

๐พโˆ‘๐‘–=1

๐‘Š๐‘–(๐‘ฅ) log๐‘Š๐‘–(๐‘ฅ)

๐‘…(๐‘ฆ๐‘–;๐น )๐‘‘๐น (๐‘ฅ) .1 (5)

Under an average power constraint ๐‘ƒ , we wish to find thecapacity of the channel (1), given by

๐ถ = sup๐นโˆˆโ„ฑ

๐ผ(๐น ), (6)

where โ„ฑ ={๐น : ๐”ผ[๐‘‹2] =

โˆซโˆžโˆ’โˆž ๐‘ฅ2๐‘‘๐น (๐‘ฅ) โ‰ค ๐‘ƒ

}, i.e., the set

of all average power constrained distributions on โ„.

Structural Properties of Optimal Inputs: We begin by em-ploying the Karush-Kuhn-Tucker (KKT) optimality conditionto show that, even though we have not imposed a peak powerconstraint on the input, it is automatically induced by theaverage power constraint. Specifically, a capacity achievingdistribution for the AWGN-QO channel (1) must have boundedsupport. 2

A. An Implicit Peak Power Constraint

Using convex optimization principles, the followingnecessary and sufficient KKT optimality condition canbe derived for our problem, in a manner similar to thedevelopment in [25], [26]. An input distribution ๐น โˆ— achievesthe capacity ๐ถ in (6) if and only if there exists ๐›พ โ‰ฅ 0 such that

๐พโˆ‘๐‘–=1

๐‘Š๐‘–(๐‘ฅ) log๐‘Š๐‘–(๐‘ฅ)

๐‘…(๐‘ฆ๐‘–;๐น โˆ—)+ ๐›พ(๐‘ƒ โˆ’ ๐‘ฅ2) โ‰ค ๐ถ (7)

for all ๐‘ฅ, with equality if ๐‘ฅ is in the support 3 of ๐น โˆ—, wherethe transition probability function ๐‘Š๐‘–(๐‘ฅ), and the output prob-ability ๐‘…(๐‘ฆ๐‘–;๐น

โˆ—) are as specified in (2) and (4), respectively.The summation on the left-hand side (LHS) of (7) is the

Kullback-Leibler divergence (or the relative entropy) betweenthe transition PMF {๐‘Š๐‘–(๐‘ฅ), ๐‘– = 1, . . . ,๐พ} and the outputPMF {๐‘…(๐‘ฆ๐‘–;๐น ), ๐‘– = 1, . . . ,๐พ}. For convenience, let us denotethis divergence function by ๐‘‘(๐‘ฅ;๐น ), that is,

๐‘‘(๐‘ฅ;๐น ) =

๐พโˆ‘๐‘–=1

๐‘Š๐‘–(๐‘ฅ) log๐‘Š๐‘–(๐‘ฅ)

๐‘…(๐‘ฆ๐‘–;๐น ). (8)

We begin by studying the behavior of this function in thelimit as ๐‘ฅ โ†’ โˆž.

Lemma 1: For the AWGN-QO channel (1), the divergencefunction ๐‘‘(๐‘ฅ;๐น ) satisfies the following properties(a) lim

๐‘ฅโ†’โˆž๐‘‘(๐‘ฅ;๐น ) = โˆ’ log๐‘…(๐‘ฆ๐พ ;๐น ).(b) There exists a finite constant ๐ด0 such thatfor ๐‘ฅ > ๐ด0, ๐‘‘(๐‘ฅ;๐น ) < โˆ’ log๐‘…(๐‘ฆ๐พ ;๐น ).4

1The logarithm is base 2 throughout the paper, so the mutual informationis measured in bits.

2That there exists a capacity achieving distribution follows by standardfunction analytic arguments [23]. For details, see our technical report [24].

3The support of a distribution ๐น (or the set of increase points of ๐น ) is theset ๐‘†๐‘‹(๐น ) = {๐‘ฅ : ๐น (๐‘ฅ+ ๐œ–)โˆ’ ๐น (๐‘ฅโˆ’ ๐œ–) > 0, โˆ€๐œ– > 0}.

4The constant ๐ด0 depends on the choice of the input ๐น . For notationalsimplicity, we do not explicitly show this dependence.

Proof: We have

๐‘‘(๐‘ฅ;๐น ) =

๐พโˆ‘๐‘–=1

๐‘Š๐‘–(๐‘ฅ) log๐‘Š๐‘–(๐‘ฅ)

๐‘…(๐‘ฆ๐‘–;๐น )

=

๐พโˆ‘๐‘–=1

๐‘Š๐‘–(๐‘ฅ) log๐‘Š๐‘–(๐‘ฅ) โˆ’๐พโˆ‘๐‘–=1

๐‘Š๐‘–(๐‘ฅ) log๐‘…(๐‘ฆ๐‘–;๐น ) .

As ๐‘ฅ โ†’ โˆž, the PMF {๐‘Š๐‘–(๐‘ฅ), ๐‘– = 1, . . . ,๐พ} โ†’ 1(๐‘– = ๐พ),where 1(โ‹…) is the indicator function. This observation,combined with the fact that the entropy of a finite alphabetrandom variable is a continuous function of its probability law,gives lim

๐‘ฅโ†’โˆž ๐‘‘(๐‘ฅ;๐น ) = 0โˆ’ log๐‘…(๐‘ฆ๐พ ;๐น ) = โˆ’ log๐‘…(๐‘ฆ๐พ ;๐น ).

Next we prove part (b). For ๐‘ฅ > ๐‘ž๐พโˆ’1, ๐‘Š๐‘–(๐‘ฅ) is a strictlydecreasing function of ๐‘ฅ for ๐‘– โ‰ค ๐พโˆ’ 1 and strictly increasingfunction of ๐‘ฅ for ๐‘– = ๐พ . Since {๐‘Š๐‘–(๐‘ฅ)} โ†’ 1(๐‘– = ๐พ) as๐‘ฅ โ†’ โˆž, it follows that there is a constant ๐ด0 such that๐‘Š๐‘–(๐ด0) < ๐‘…(๐‘ฆ๐‘–;๐น ) for ๐‘– โ‰ค ๐พโˆ’1 and ๐‘Š๐พ(๐ด0) > ๐‘…(๐‘ฆ๐พ ;๐น ).Therefore, it follows that for ๐‘ฅ > ๐ด0,

๐‘‘(๐‘ฅ;๐น ) =

๐พโˆ‘๐‘–=1

๐‘Š๐‘–(๐‘ฅ) log๐‘Š๐‘–(๐‘ฅ)

๐‘…(๐‘ฆ๐‘–;๐น )

< ๐‘Š๐พ(๐‘ฅ) log๐‘Š๐พ(๐‘ฅ)

๐‘…(๐‘ฆ๐พ ;๐น )< โˆ’ log๐‘…(๐‘ฆ๐พ ;๐น ).

The saturating nature of the divergence function for theAWGN-QO channel, as stated above, coupled with the KKTcondition, is now used to prove that a capacity achievingdistribution must have bounded support.

Proposition 1: For the average power constrained AWGN-QO channel (1), an optimal input distribution must havebounded support.

Proof: Let ๐น โˆ— be an optimal input, so that there exists๐›พ โ‰ฅ 0 such that (7) is satisfied with equality at every pointin the support of ๐น โˆ—. We exploit this necessary condition toshow that the support of ๐น โˆ— is upper bounded. Specifically,we prove that there exists a finite constant ๐ด2

โˆ— such that it isnot possible to attain equality in (7) for any ๐‘ฅ > ๐ด2

โˆ—.From Lemma 1, we get lim

๐‘ฅโ†’โˆž๐‘‘(๐‘ฅ;๐นโˆ—)=โˆ’ log๐‘…(๐‘ฆ๐พ ;๐น โˆ—)=:

๐ฟ. Also, there exists a finite constant ๐ด0 such that for๐‘ฅ > ๐ด0, ๐‘‘(๐‘ฅ;๐น

โˆ—) < ๐ฟ.We consider two possible cases.

โˆ™ Case 1: ๐›พ > 0.For ๐‘ฅ > ๐ด0, we have ๐‘‘(๐‘ฅ;๐น โˆ—) < ๐ฟ.For ๐‘ฅ >

โˆšmax{0, (๐ฟโˆ’ ๐ถ + ๐›พ๐‘ƒ )/๐›พ} =: ๐ด, we have

๐›พ(๐‘ƒ โˆ’ ๐‘ฅ2) < ๐ถ โˆ’ ๐ฟ.Defining ๐ดโˆ—

2 = max{๐ด0, ๐ด}, we get the desired result.โˆ™ Case 2: ๐›พ = 0.

Since ๐›พ = 0, the KKT condition (7) reduces to

๐‘‘(๐‘ฅ;๐น โˆ—) โ‰ค ๐ถ , โˆ€๐‘ฅ.Taking limit ๐‘ฅ โ†’ โˆž on both sides, we get๐ฟ = lim

๐‘ฅโ†’โˆž ๐‘‘(๐‘ฅ;๐น โˆ—) โ‰ค ๐ถ.Hence, choosing ๐ดโˆ—

2 = ๐ด0, for ๐‘ฅ > ๐ดโˆ—2 we get,

๐‘‘(๐‘ฅ;๐น โˆ—) < ๐ฟ โ‰ค ๐ถ, that is, ๐‘‘(๐‘ฅ;๐น โˆ—) + ๐›พ(๐‘ƒ โˆ’ ๐‘ฅ2) < ๐ถ.

Combining the two cases, we have shown that the supportof the distribution ๐น โˆ— has a finite upper bound ๐ด2

โˆ—. Usingsimilar arguments, it can easily be shown that the support of

3632 IEEE TRANSACTIONS ON COMMUNICATIONS, VOL. 57, NO. 12, DECEMBER 2009

๐น โˆ— has a finite lower bound ๐ด1โˆ— as well, which implies that

๐น โˆ— has bounded support.

B. Achievability of Capacity by a Discrete Input

In [5], Witsenhausen considered a stationary discrete-timememoryless channel, with a continuous input ๐‘‹ takingvalues on the bounded interval [๐ด1, ๐ด2] โŠ‚ โ„, and a discreteoutput ๐‘Œ of finite cardinality ๐พ . Using Dubinsโ€™ theorem[6], it was shown that if the transition probability functionsare continuous (i.e., ๐‘Š๐‘–(๐‘ฅ) is continuous in ๐‘ฅ, for each๐‘– = 1, โ‹… โ‹… โ‹… ,๐พ), then the capacity is achievable by a discreteinput distribution with at most ๐พ mass points. As statedin Proposition 2 below (proved in Appendix A), this resultcan be extended to show that, if an additional averagepower constraint is imposed on the input, the capacity is thenachievable by a discrete input with at most ๐พ+1 mass points.

Proposition 2: Consider a stationary discrete-time memo-ryless channel with a continuous input ๐‘‹ that takes val-ues in the bounded interval [๐ด1, ๐ด2], and a discrete output๐‘Œ โˆˆ {๐‘ฆ1, ๐‘ฆ2, โ‹… โ‹… โ‹… , ๐‘ฆ๐พ}. Let the channel transition probabilityfunction ๐‘Š๐‘–(๐‘ฅ) = P(๐‘Œ = ๐‘ฆ๐‘–โˆฃ๐‘‹ = ๐‘ฅ) be continuous in ๐‘ฅ foreach ๐‘–, where 1 โ‰ค ๐‘– โ‰ค ๐พ . The capacity of this channel, underan average power constraint on the input, is achievable by adiscrete input distribution with at most ๐พ + 1 mass points.

Proof: See Appendix A.Proposition 2, coupled with the implicit peak power con-

straint derived in the previous subsection (Proposition 1), givesus the following result.

Theorem 1: The capacity of the average power constrainedAWGN-QO channel (1) is achievable by a discrete inputdistribution with at most ๐พ + 1 points of support.

Proof: From Proposition 1, we know that an optimalinput ๐น โˆ— has bounded support [๐ด1

โˆ—, ๐ด2โˆ—]. Hence, to obtain

the capacity in (6), we can maximize ๐ผ(๐น ) over only thoseaverage power constrained distributions that have support in[๐ด1

โˆ—, ๐ด2โˆ—]. Since the transition functions ๐‘Š๐‘–(๐‘ฅ) are continu-

ous, Proposition 2 guarantees that this maximum is achievableby a discrete input with at most ๐พ + 1 points.

C. Symmetric Inputs for Symmetric Quantization

For our capacity computations ahead, we assume that thequantizer Q employed in (1) is symmetric, i.e., its thresholdvector ๐‘ž๐‘ž๐‘ž is symmetric about the origin. Given the symmetricnature of the AWGN noise and the power constraint, it seemsintuitively plausible that restriction to symmetric quantizersshould not be suboptimal from the point of view of optimizingover the quantizer choice in (1), although a proof of thisconjecture has eluded us. However, once we assume that thequantizer in (1) is symmetric, we can restrict attention to onlysymmetric inputs without loss of optimality, as stated in thefollowing Lemma.5

Lemma 2: If the quantizer in (1) is symmetric, then, with-out loss of optimality, we can consider only symmetric inputsfor the capacity computation in (6).Proof: Suppose we are given an input random variable ๐‘‹ (withdistribution ๐น ) that is not necessarily symmetric. Denote the

5A random variable ๐‘‹ (with distribution F) is symmetric if ๐‘‹ and โˆ’๐‘‹have the same distribution, that is, ๐น (๐‘ฅ) = 1โˆ’ ๐น (โˆ’๐‘ฅ),โˆ€ ๐‘ฅ โˆˆ โ„.

distribution of โˆ’๐‘‹ by ๐บ (so that ๐บ(๐‘ฅ) = 1โˆ’ ๐น (โˆ’๐‘ฅ), โˆ€ ๐‘ฅ โˆˆโ„). Due to the symmetric nature of the noise ๐‘ and thequantizer Q, it is easy to see that ๐‘‹ and โˆ’๐‘‹ result in thesame input-output mutual information, that is, ๐ผ(๐น ) = ๐ผ(๐บ).Consider now the following symmetric mixture distribution

๐น (๐‘ฅ) =๐น (๐‘ฅ) +๐บ(๐‘ฅ)

2.

Since the mutual information is concave in the input distri-bution, we get ๐ผ(๐น ) โ‰ฅ ๐ผ(๐น )+๐ผ(๐บ)

2 = ๐ผ(๐น ), which proves thedesired result.

III. CAPACITY COMPUTATION

In this section, we consider capacity computation for theAWGN-QO channel. We first provide an explicit capacityformula for the extreme scenario of 1-bit symmetric quanti-zation, and then discuss numerical computations for multi-bitquantization.

A. 1-bit Symmetric Quantization : Binary Antipodal Signalingis Optimal

With 1-bit symmetric quantization, the channel is

๐‘Œ = sign(๐‘‹ +๐‘). (9)

Theorem 1 (Section II-B) guarantees that the capacity of thischannel, under an average power constraint, is achievable bya discrete input distribution with at most 3 points. This resultis further tightened by the following theorem that shows theoptimality of binary antipodal signaling for all SNRs.

Theorem 2: For the 1-bit symmetric quantized channelmodel (9), the capacity is achieved by binary antipodal sig-naling and is given by

๐ถ = 1โˆ’ โ„Ž(๐‘„(โˆš

SNR))

, SNR =๐‘ƒ

๐œŽ2,

where โ„Ž(โ‹…) is the binary entropy function,

โ„Ž(๐‘) = โˆ’๐‘ log(๐‘)โˆ’ (1โˆ’ ๐‘) log(1โˆ’ ๐‘) , 0 โ‰ค ๐‘ โ‰ค 1 ,

and ๐‘„(โ‹…) is the complementary Gaussian distribution functionshown in (3).

Proof: Since ๐‘Œ is binary it is easy to see that

๐ป(๐‘Œ โˆฃ๐‘‹) = ๐”ผ

[โ„Ž

(๐‘„

(๐‘‹

๐œŽ

))],

where ๐”ผ denotes the expectation operator. Therefore

๐ผ(๐‘‹,๐‘Œ ) = ๐ป(๐‘Œ )โˆ’ ๐”ผ

[โ„Ž

(๐‘„

(๐‘‹

๐œŽ

))],

which we wish to maximize over all input distributionssatisfying ๐”ผ[๐‘‹2] โ‰ค ๐‘ƒ . Since the quantizer is symmetric, wecan restrict attention to symmetric input distributions withoutloss of optimality (cf. Lemma 2). On doing so, we obtainthat the PMF of the output ๐‘Œ is also symmetric (since thequantizer and the noise are already symmetric). Therefore,๐ป(๐‘Œ ) = 1 bit, and we get

๐ถ = 1โˆ’ min๐‘‹ symmetric๐”ผ[๐‘‹2]โ‰ค๐‘ƒ

๐”ผ

[โ„Ž

(๐‘„

(๐‘‹

๐œŽ

))].

SINGH et al.: ON THE LIMITS OF COMMUNICATION WITH LOW-PRECISION ANALOG-TO-DIGITAL CONVERSION AT THE RECEIVER 3633

Since โ„Ž(๐‘„(๐‘ง)) is an even function, we get that

๐ป(๐‘Œ โˆฃ๐‘‹) = ๐”ผ

[โ„Ž

(๐‘„

(๐‘‹

๐œŽ

))]= ๐”ผ

[โ„Ž

(๐‘„

( โˆฃ๐‘‹ โˆฃ๐œŽ

))].

In [24], we show that the function โ„Ž(๐‘„(โˆš๐‘ง)) is convex in ๐‘ง.

Jensenโ€™s inequality [27] thus implies

๐ป(๐‘Œ โˆฃ๐‘‹) โ‰ฅ โ„Ž(๐‘„(โˆš

SNR))

with equality iff ๐‘‹2 = ๐‘ƒ . Coupled with the symmetrycondition on ๐‘‹ , this implies that binary antipodal signalingachieves capacity and the capacity is

๐ถ = 1โˆ’ โ„Ž(๐‘„(โˆš

SNR))

.

B. Multi-Bit Quantization

We now consider ๐พ-bin symmetric quantization for ๐พ > 2.Every choice of the quantizer results in a unique channelmodel (1). In this section, we discuss capacity computationassuming a fixed quantizer only. Optimization over the quan-tizer choice is performed in Section IV.

1) Capacity computation using cutting-plane algorithm:Contrary to the 1-bit case, closed form expressions for optimalinput and capacity appear unlikely for multi-bit quantization,due to the complicated expression for mutual information.We therefore resort to the cutting-plane algorithm [3, Sec IV-A] to generate optimal inputs numerically. For channels withcontinuous input alphabets, the cutting-plane algorithm can,in general, be used to generate nearly optimal discrete inputdistributions. It is therefore well matched to our problem, forwhich we already know that the capacity is achievable by adiscrete input distribution.

For our simulations, we fix the noise variance ๐œŽ2 = 1, andvary the power ๐‘ƒ to obtain capacity at different SNRs. Toapply the cutting-plane algorithm, we take a fine quantizeddiscrete grid on the interval [โˆ’10

โˆš๐‘ƒ , 10

โˆš๐‘ƒ ], and optimize

the input distribution over this grid. Note that Proposition 1(Section II-A) tells us that an optimal input distribution forour problem must have bounded support, but it does not giveexplicit values that we can use directly in our simulations.However, on employing the cutting-plane algorithm over theinterval [โˆ’10

โˆš๐‘ƒ , 10

โˆš๐‘ƒ ], we find that the resulting input dis-

tributions have support sets well within this interval. Moreover,increasing the interval length further does not change theseresults.

While the cutting-plane algorithm optimizes the distributionof the channel input, a dual formulation of the channelcapacity problem, involving an optimization over the outputdistribution, can alternately be used to obtain easily com-putable tight upper bounds on the capacity. We discuss theseduality-based upper bounds next.

2) Duality-based upper bound on channel capacity:In the dual formulation of the channel capacity problem, wefocus on the distribution of the output, rather than that of theinput. Specifically, assume a channel with input alphabet ๐’ณ ,transition law ๐‘Š (๐‘ฆโˆฃ๐‘ฅ), and an average power constraint ๐‘ƒ .Then, for every choice of the output distribution ๐‘…(๐‘ฆ), wehave the following upper bound on the channel capacity ๐ถ

๐ถ โ‰ค ๐‘ˆ(๐‘…) = min๐›พโ‰ฅ0

sup๐‘ฅโˆˆ๐’ณ

[๐ท(๐‘Š (โ‹…โˆฃ๐‘ฅ)โˆฃโˆฃ๐‘…(โ‹…))+๐›พ(๐‘ƒ โˆ’๐‘ฅ2)] , (10)

05

1015

โˆ’5

0

5

0

0.1

0.2

0.3

0.4

0.5

Location

SNR (dB)

Pro

babi

lity

Fig. 2. Probability mass function of the optimal input generated by thecutting-plane algorithm [3] at various SNRs, for the 2-bit symmetric quantizerwith thresholds {โˆ’2, 0, 2}. (noise variance ๐œŽ2 = 1.)

TABLE IDUALITY-BASED UPPER BOUNDS ON CHANNEL CAPACITY, COMPARED

WITH THE MUTUAL INFORMATION (MI) ACHIEVED BY THEDISTRIBUTIONS GENERATED USING THE CUTTING-PLANE ALGORITHM.

SNR(๐‘‘๐ต) โˆ’5 0 5 10 15 20Upper Bound 0.163 0.406 0.867 1.386 1.513 1.515

๐‘€๐ผ 0.155 0.405 0.867 1.379 1.484 1.484

where ๐›พ is a Lagrange parameter, and ๐ท(๐‘Š (โ‹…โˆฃ๐‘ฅ)โˆฃโˆฃ๐‘…(โ‹…)) isthe divergence between the transition and output distributions.While [28] provides this bound for a discrete channel, its ex-tension to continuous alphabet channels has been establishedin [29]. For a more detailed perspective on duality-based upperbounds, see [30].

For an arbitrary choice of ๐‘…(๐‘ฆ), the bound (10) might bequite loose. Therefore, to obtain a tight upper bound, we mayneed to evaluate (10) for a large number of output distributionsand pick the minimum of the resulting upper bounds. Thiscould be tedious in general, especially if the output alphabetis continuous. However, for the channel model we consider,the output alphabet is discrete with small cardinality. Forexample, for 2-bit quantization, the space of all symmetricoutput distributions is characterized by a single parameter๐›ผ โˆˆ (0, 0.5). This makes the dual formulation attractive, sincewe can easily obtain a tight upper bound on capacity byevaluating the upper bound in (10) for different choices of ๐›ผ.

It remains to specify how to compute the upper bound(10) for a given output distribution ๐‘…. The structure of thedivergence function ๐ท(๐‘Š (โ‹…โˆฃ๐‘ฅ)โˆฃโˆฃ๐‘…(โ‹…)) for our problem can beexploited to develop a systematic procedure for this purpose.We refer the reader to [24] for details of this procedure, andproceed directly to the numerical results.

3) Numerical example: We compare results obtainedusing the cutting-plane algorithm with capacity upper boundsobtained using the dual formulation. We consider 2-bitquantization, and provide results for the specific choice ofquantizer having thresholds at {โˆ’2, 0, 2}.

The input distributions generated by the cutting-plane algo-rithm at various SNRs (setting ๐œŽ2 = 1) are shown in Figure 2,and the mutual information achieved by them is given in TableI. As predicted by Theorem 1 (Section II-B), the support setof the input distribution (at each SNR) has cardinality โ‰ค 5.

For upper bound computations, we evaluate (10) for differ-

3634 IEEE TRANSACTIONS ON COMMUNICATIONS, VOL. 57, NO. 12, DECEMBER 2009

0 0.5 1 1.5 2 2.5 3 3.5

0.76

0.78

0.8

0.82

0.84

0.86

0.88

x

LHS

of t

he K

KT

con

ditio

n (7

)

Fig. 3. The left-hand side of the KKT condition (7) for the input distributiongenerated by the cutting-plane algorithm (SNR = 5 dB). The KKT conditionis seen to be satisfied (up to the numerical precision of our computations).

ent symmetric output distributions. For 2-bit quantization, theset of symmetric outputs is characterized by just one parameter๐›ผ โˆˆ (0, 0.5), with the probability distribution on the outputbeing {0.5โˆ’๐›ผ, ๐›ผ, ๐›ผ, 0.5โˆ’๐›ผ}. We vary ๐›ผ over a fine discretegrid on (0, 0.5), and compute the upper bound for each valueof ๐›ผ. The least upper bound achieved thus, at a number ofdifferent SNRs, is shown in Table I. The small gap between theupper bound and the mutual information (at every SNR) showsthe tightness of the obtained upper bounds, and also confirmsthe near-optimality of the input distributions generated by thecutting-plane algorithm.

It is insightful to verify that the preceding near-optimalinput distributions satisfy the KKT condition (7). For instance,consider an SNR of 5 dB, for which the input distributiongenerated by the cutting-plane algorithm has support set{โˆ’2.86,โˆ’0.52, 0.52, 2.86}. Figure 3 plots, as a function of๐‘ฅ, the LHS of (7) for this input distribution. (The value of ๐›พused in the plot was obtained by equating the LHS of (7) tothe capacity value of 0.867, at ๐‘ฅ = 0.52 .) The KKT conditionis seen to be satisfied (up to the numerical precision of ourcomputations), as the LHS of (7) equals the capacity at pointsin the support set of the input, and is less than the capacityeverywhere else. Note that we show the plot for ๐‘ฅ โ‰ฅ 0 onlybecause it is symmetric in ๐‘ฅ.

IV. QUANTIZER OPTIMIZATION AND NUMERICAL

RESULTS

Until now, we have addressed the problem of optimizing theinput distribution for a fixed output quantizer. In this section,we optimize over the choice of the quantizer, and presentnumerical results for 2-bit and 3-bit symmetric quantization.

A Simple Benchmark Input-Quantizer Pair: While an op-timal quantizer, along with a corresponding optimal inputdistribution, provides the absolute communication limits forour model, we do not have a simple analytical characteriza-tion of their dependence on SNR. From a system designerโ€™s

0 1 2 3 4 5 6 70

0.5

1

1.5

2

Quantizer threshold โ€™qโ€™

Cap

acity

(bi

ts /

chan

nel u

se)

โˆ’5 dB

0 dB

3 dB

7 dB

10 dB

15 dB

Fig. 4. 2-bit symmetric quantization : channel capacity (in bits per channeluse) as a function of the quantizer threshold ๐‘ž. (noise variance ๐œŽ2 = 1.)

perspective, therefore, it is of interest to also examine subop-timal choices that are easy to adapt as a function of SNR,as long as the penalty relative to the optimal solution isnot excessive. Specifically, we take the following input andquantizer pair to be our benchmark strategy : for a ๐พ-binquantizer, consider equiprobable, equispaced ๐พ-PAM (pulseamplitude modulated) input, with quantizer thresholds chosento be the mid-points of the input mass point locations. Thatis, the quantizer thresholds correspond to the ML (maximumlikelihood) hard decision boundaries. Both the input masspoints and the quantizer thresholds have a simple, well-defineddependence on SNR, and can therefore be adapted easily atthe receiver based on the measured SNR. With our K-pointuniform PAM input, we have the entropy ๐ป(๐‘‹) = log2 ๐พbits for any SNR. Also, it is easy to see that as SNR โ†’ โˆž,๐ป(๐‘‹ โˆฃ๐‘Œ ) โ†’ 0 for the benchmark input-quantizer pair. Thisimplies that the benchmark scheme is near-optimal if weoperate at high SNR. The issue to investigate therefore is howmuch gain an optimal quantizer and input pair provides overthis benchmark at low to moderate SNR. Note that, for 1-bitsymmetric quantization, the benchmark input corresponds tobinary antipodal signaling, which has already been shown tobe optimal for all SNRs.

As before, we set the noise variance ๐œŽ2 = 1 for conve-nience. Of course, the results are scale-invariant, in the sensethat if both ๐‘ƒ and ๐œŽ2 are scaled by the same factor ๐‘… (thuskeeping the SNR unchanged), then there is an equivalentquantizer (obtained by scaling the thresholds by

โˆš๐‘…) that

gives identical performance.

A. 2-Bit Symmetric Quantization

A 2-bit symmetric quantizer is characterized by a singleparameter ๐‘ž, with the quantizer thresholds being {โˆ’๐‘ž, 0, ๐‘ž}.We therefore employ a brute force search over ๐‘ž to find anoptimal 2-bit symmetric quantizer. In Figure 4, we plot thevariation of the channel capacity (computed using the cutting-plane algorithm) as a function of the parameter ๐‘ž at various

SINGH et al.: ON THE LIMITS OF COMMUNICATION WITH LOW-PRECISION ANALOG-TO-DIGITAL CONVERSION AT THE RECEIVER 3635

SNRs. Based on our simulations, we make the followingobservations:

โˆ™ For any SNR, there is an optimal choice of ๐‘ž whichmaximizes capacity. For the benchmark quantizer (whichis optimal at high SNR), ๐‘ž scales as

โˆšSNR, hence it

is not surprising to note that the optimal value of ๐‘ž weobtain increases monotonically with SNR at high SNR.

โˆ™ For low SNRs, the variation in the capacity as a functionof ๐‘ž is quite small, whereas the variation becomes appre-ciable as the SNR increases. A practical implication ofthis observation is that imperfections in Automatic GainControl (AGC) have more severe consequences at higherSNRs.

โˆ™ For any SNR, as ๐‘ž โ†’ 0 or ๐‘ž โ†’ โˆž, we approachthe same capacity as with 1-bit symmetric quantization(not shown for ๐‘ž โ†’ โˆž in the plots for 10 and 15dB in Figure 4). This conforms to intuition: ๐‘ž = 0reduces the 2-bit quantizer to a 1-bit quantizer, while๐‘ž โ†’ โˆž renders the thresholds at โˆ’๐‘ž and ๐‘ž ineffective indistinguishing between two finite valued inputs, so thatonly the comparison with the quantizer threshold at 0yields useful information.

Comparison with the Benchmark: In Table II, we comparethe performance of the preceding optimal solutions with thebenchmark scheme (see the relevant columns for 2-bit ADC).The corresponding plots are shown in Figure 6. In addition tobeing nearly optimal at high SNR, the benchmark scheme isseen to perform fairly well at low to moderate SNR as well.For instance, even at -10 dB SNR, which might correspond to awideband system designed for very low bandwidth efficiency,it achieves 86% of the capacity achieved with optimal choiceof 2-bit quantizer and input distribution. On the other hand, forSNR of 0 dB or above, the capacity is better than 95% of theoptimal. These results are encouraging from a practical stand-point, given the ease of implementing the benchmark scheme.

Optimal Input Distributions: It is interesting to examinethe optimal input distributions (given by the cutting-planealgorithm) corresponding to the optimal quantizers obtainedabove. Figure 5 shows these distributions, along with optimalquantizer thresholds, for different SNRs. The solid verticallines show the locations of the input distribution points andtheir probabilities, while the quantizer thresholds are depictedby the dashed vertical lines. As expected, binary signalingis found to be optimal for low SNR, since it would bedifficult for the receiver to distinguish between multiple inputpoints located close to each other. The number of mass pointsincreases as SNR is increased, with a new point emergingat 0. On increasing SNR further, we see that the non zeroconstellation points (and also the quantizer thresholds) movefarther apart, resulting in increased capacity. When the SNRbecomes enough that four input points can be disambiguated,the point at 0 disappears, and we get two new points, resultingin a 4-point constellation. The eventual convergence of this 4-point constellation to uniform PAM with mid-point quantizerthresholds (i.e., the benchmark scheme) is to be expected,since the benchmark scheme approaches the capacity bound oftwo bits at high SNR. It is worth noting that the optimal inputswe obtained all have at most four points, even though Theorem

0 4 7 10 15 20

โˆ’10

0

10

0

0.1

0.2

0.3

0.4

0.5

Location

SNR (dB)

Pro

ba

bili

ty

Fig. 5. 2-bit symmetric quantization : optimal input distribution (solid verticallines) and quantizer thresholds (dashed vertical lines) at various SNRs.

1 (Section II-B) is looser, guaranteeing the achievability ofcapacity by at most five points.

B. 3-bit Symmetric Quantization

For 3-bit symmetric quantization, we need to optimize overa space of 3 parameters : {0 < ๐‘ž1 < ๐‘ž2 < ๐‘ž3}, with the quan-tizer thresholds being {0,ยฑ๐‘ž1,ยฑ๐‘ž2,ยฑ๐‘ž3}. Since brute forcesearch is computationally complex, we investigate an alternateiterative optimization procedure for joint optimization of theinput and the quantizer in this case. Specifically, we beginwith an initial quantizer choice ๐‘„1, and then iterate as follows(starting at ๐‘– = 1)

โˆ™ For the quantizer ๐‘„๐‘–, find an optimal input. Call this input๐น๐‘–.

โˆ™ For the input ๐น๐‘–, find a locally optimal quantizer, initial-izing the search at ๐‘„๐‘–. Call the resulting quantizer ๐‘„๐‘–+1.

โˆ™ Repeat the first two steps with ๐‘– = ๐‘–+ 1.We terminate the process when the capacity gain betweenconsecutive iterations becomes less than a small threshold ๐œ–.

Although the input-output mutual information is a concavefunctional of the input distribution (for a fixed quantizer), itis not guaranteed to be concave jointly over the input and thequantizer. Hence, the iterative procedure is not guaranteed toprovide an optimal input-quantizer pair in general. A goodchoice of the initial quantizer ๐‘„1 is crucial to enhance thelikelihood that it does converge to an optimal solution. Wediscuss this next.

High SNR Regime: For high SNRs, we know that uniformPAM with mid-point quantizer thresholds (i.e., the benchmarkscheme) is nearly optimal. Hence, this quantizer is a goodchoice for initialization at high SNRs. The results we obtainindeed demonstrate that this initialization works well at highSNRs. This is seen by comparing the results of the iterativeprocedure with the results of a brute force search over thequantizer choice (similar to the 2-bit case considered earlier),as both of them provide almost identical capacity values.

Lower SNRs: For lower SNRs, one possibility is totry different initializations ๐‘„1. However, on trying thebenchmark initialization at some lower SNRs as well, we findthat the iterative procedure still provides us with near-optimal

3636 IEEE TRANSACTIONS ON COMMUNICATIONS, VOL. 57, NO. 12, DECEMBER 2009

TABLE IIPERFORMANCE COMPARISON : FOR 1, 2, AND 3โˆ’BIT ADC, THE TABLE SHOWS THE MUTUAL INFORMATION (IN BITS PER CHANNEL USE) ACHIEVED BY

THE OPTIMAL SOLUTIONS, AS WELL AS THE BENCHMARK SOLUTIONS. ALSO SHOWN ARE THE CAPACITY ESTIMATES OBTAINED BY ASSUMING THE

ADDITIVE QUANTIZATION NOISE MODEL (AQNM). (NOTE THAT FOR 1-BIT ADC, THE BENCHMARK SOLUTION COINCIDES WITH THE OPTIMAL

SOLUTION, AND HENCE IS NOT SHOWN SEPARATELY.)

SNR 1-bit ADC 2-bit ADC SNR 3-bit ADC Unquantized(in dB) Optimal AQNM Optimal Benchmark AQNM (in dB) Optimal Benchmark AQNM

-20 0.005 0.007 0.006 0.005 0.007 -20 0.007 0.005 0.007 0.007-10 0.045 0.067 0.061 0.053 0.068 -10 0.067 0.056 0.069 0.069-5 0.135 0.185 0.179 0.166 0.195 -5 0.193 0.177 0.197 0.1980 0.369 0.424 0.455 0.440 0.479 0 0.482 0.471 0.494 0.5003 0.602 0.610 0.693 0.687 0.736 3 0.759 0.744 0.777 0.7915 0.769 0.733 0.889 0.869 0.931 5 0.975 0.955 1.002 1.0297 0.903 0.843 1.098 1.064 1.133 7 1.215 1.180 1.248 1.29410 0.991 0.972 1.473 1.409 1.417 10 1.584 1.533 1.634 1.73012 0.992 1.032 1.703 1.655 1.579 12 1.846 1.766 1.886 2.03715 1.000 1.091 1.930 1.921 1.765 15 2.253 2.138 2.232 2.51417 1.000 1.115 1.987 1.987 1.853 17 2.508 2.423 2.427 2.83820 1.000 1.136 1.999 1.999 1.938 20 2.837 2.808 2.655 3.329

0 5 10 15 200

0.5

1

1.5

2

2.5

3

3.5

4

SNR (dB)

Capacity (

bits/c

hannel use)

1โˆ’bit ADC(optimal)

2โˆ’bit ADC(optimal and PAM / ML)

3โˆ’bit ADC(optimal and PAM / ML)

Unquantized

Fig. 6. Capacity plots for different ADC precisions. For 2 and 3-bit ADC,solid curves correspond to optimal solutions, while dashed curves show theperformance of the benchmark scheme (PAM input with ML quantization).

solutions (again verified by comparing with brute forceoptimization results).

While our results show that the iterative procedure (withbenchmark initialization) has provided (near) optimal solutionsat different SNRs, we leave the question of whether it willconverge to an optimal solution in general as an open problem.

Comparison with the Benchmark: The efficacy of thebenchmark initialization at lower SNRs suggests that theperformance of the benchmark scheme should not be too farfrom optimal at small SNRs as well. This is indeed the case,as seen from the data values in Table II and the correspondingplots in Figure 6. At 0 dB SNR, for instance, the benchmarkscheme achieves 98% of the capacity achievable with anoptimal input-quantizer pair.

Optimal Input Distributions: Although not depicted here, weagain observe (as for the 2-bit case) that the optimal inputsobtained all have at most K points (๐พ = 8 in this case),while Theorem 1 guarantees the achievability of capacity by

at most ๐พ+1 points. Of course, Theorem 1 is applicable to anyquantizer choice (and not just optimal symmetric quantizers).Thus, it is possible that there might exist a ๐พ-bin quantizerfor which the capacity is indeed achieved by exactly ๐พ + 1points. We leave open, therefore, the question of whether ornot the result in Theorem 1 can be tightened to guaranteethe achievability of capacity with at most ๐พ points for theAWGN-QO channel.

C. Comparison with Unquantized Observations

We now compare the capacity results for different quantizerprecisions against the capacity with unquantized observations.Again, the plots are shown in Figure 6 and the data values aregiven in Table II. We observe that at low SNR, the performancedegradation due to low-precision quantization is small. For in-stance, at -5 dB SNR, 1-bit receiver quantization achieves 68%of the capacity achievable without any quantization, whilewith 2-bit quantization, we can get as much as 90% of theunquantized capacity. Even at moderately high SNRs, the lossdue to low-precision quantization remains quite acceptable.For example, 2-bit quantization achieves 85% of the capacityattained using unquantized observations at 10 dB SNR, while3-bit quantization achieves 85% of the unquantized capacity at20 dB SNR. For the specific case of binary antipodal signaling,[7] has earlier shown that a large fraction of the capacity canbe obtained by 2-bit quantization.

On the other hand, if we fix the spectral efficiency to thatattained by an unquantized system at 10 dB (which is 1.73bits/channel use), then 2-bit quantization incurs a loss of 2.30dB (see Table III). For wideband systems, this penalty inpower maybe more significant compared to the 15% loss inspectral efficiency on using 2-bit quantization at 10 dB SNR.This suggests, for example, that in order to weather the impactof low-precision ADC, a moderate reduction in the spectralefficiency might be a better design choice than an increase inthe transmit power.

D. Additive Quantization Noise Model (AQNM)

It is common to model the quantization noise as independentadditive noise [31, pp. 122]. Next, we compare this approx-imation with our exact capacity calculations. In this model

SINGH et al.: ON THE LIMITS OF COMMUNICATION WITH LOW-PRECISION ANALOG-TO-DIGITAL CONVERSION AT THE RECEIVER 3637

TABLE IIISNR (IN DB) REQUIRED TO ACHIEVE A SPECIFIED SPECTRAL EFFICIENCY

WITH DIFFERENT ADC PRECISIONS.

Spectral Efficiency (bits per channel use)0.25 0.5 1.0 1.73 2.5

1-bit ADC โˆ’2.04 1.79 โˆ’ โˆ’ โˆ’2-bit ADC โˆ’3.32 0.59 6.13 12.30 โˆ’3-bit ADC โˆ’3.67 0.23 5.19 11.04 16.90

Unquantized โˆ’3.83 0.00 4.77 10.00 14.91

๐‘Œ = ๐‘‹ + ๐‘ + ๐‘๐‘„, where the quantization noise ๐‘๐‘„ isassumed to be uniformly distributed, and independent of ๐‘‹,๐‘ .The signal to quantization noise ratio ๐‘ƒ

๐”ผ(๐‘๐‘„2)

is assumed tobe 6 log2 ๐พ dB for ๐พ-bin quantization [31, pp. 122]. AsSNR โ†’ 0, the distribution of ๐‘ + ๐‘๐‘„ approaches that ofa Gaussian, and hence we expect

1

2log

(1 +

๐‘ƒ

๐œŽ2 + ๐”ผ(๐‘๐‘„2)

)

to be a good approximation of the capacity at low SNR. TableII shows that this approximation can be useful in terms of pro-viding a quick estimate, although it can either underestimate oroverestimate the actual capacity, depending on the parameters.

V. CONCLUSIONS

Our Shannon-theoretic investigation indicates that the useof low-precision ADC may be a feasible option for designingfuture high-bandwidth communication systems. The availabil-ity of a large amount of bandwidth encourages power-efficientcommunication using small constellations, so that the symbolrate, and hence the sampling rate, for a given bit rate must behigh. This forces us towards using ADCs with lower precision,but fortunately, this is consistent with the use of small constel-lations in the first place for power-efficient design. Thus, if weplan on operating at low to moderate SNR, the small reductionin spectral efficiency due to low-precision ADC is acceptablein such systems, given that the bandwidth is plentiful.

There are several unresolved technical issues that we leaveas open problems. While we show that at most ๐พ + 1 pointsare needed to achieve capacity for ๐พ-bin output quantizationof the AWGN channel, our numerical results reveal that ๐พmass points are sufficient. Can this be proven analytically,at least for symmetric quantizers ? Are symmetric quantizersoptimal ? Does our iterative procedure (with the benchmarkinitialization, or some other judicious initialization) for jointoptimization of the input and the quantizer converge to anoptimal solution in general ?

A technical assumption worth revisiting is that of Nyquistsampling (which induces the discrete-time memorylessAWGN-Quantized Output channel model considered in thiswork). While symbol rate Nyquist sampling is optimal forunquantized systems in which the transmit and receive filtersare square root Nyquist and the channel is ideal, for quan-tized samples, we have obtained numerical results that showthat fractionally spaced samples can actually lead to smallperformance gains. A detailed study quantifying such gains isimportant in understanding the tradeoffs between ADC speedand precision. However, we do not expect oversampling toplay a significant role at low to moderate SNR, given the

small degradation in our Nyquist sampled system relativeto unquantized observations (for which Nyquist sampling isindeed optimal) in these regimes. Of course, oversampling inconjunction with hybrid analog/digital processing (e.g., usingideas analogous to delta-sigma quantization) could producebigger performance gains, but this falls outside the scope ofthe present model.

While our focus in this paper was on non-spread systems, itis known that low-precision ADC is often employed in spreadspectrum systems for low cost implementations [32]. In ourprior examination of Shannon limits for direct sequence spreadspectrum systems with 1-bit ADC [10], we demonstratedthat binary signaling was suboptimal, but did not completelycharacterize an optimal input distribution. The approach in thepresent paper implies that, for a spreading gain ๐บ, a discreteinput distribution with at most ๐บ + 2 points can achievecapacity (although in practice, much smaller constellationswould probably work well).

Finally, we would like to emphasize that the Shannon-theoretic perspective provided in this paper is but a firststep towards the design of communication systems with low-precision ADC. Major technical challenges include the designof ADC-constrained methods for receiver tasks such as carrierand timing synchronization, channel estimation and equaliza-tion, demodulation and decoding.

APPENDIX APROOF OF PROPOSITION 2

Denote by ๐’ฎ the set of probability distributions on [๐ด1, ๐ด2]satisfying the average power constraint ๐‘ƒ . We need to provethat the maximum of ๐ผ(๐‘‹ ;๐‘Œ ) over ๐’ฎ is achievable by adiscrete input with at most ๐พ+1 points. The proof has twoparts: 1) showing that maximum occurs at an extreme point ofa suitably defined constraint set; 2) showing that every extremepoint of this constraint set has at most ๐พ+1 support points.The first part of the proof relies on the fact that ๐ผ(๐‘‹ ;๐‘Œ ) islinear given the output constraints (as in [5]), while the secondpart is an application of Dubinsโ€™ theorem [6] along the linesof [5].

The existence of an optimal distribution ๐น โˆ— โˆˆ ๐’ฎ followsfrom standard function analytic arguments [23] (see [24] formore details). Let ๐‘…โˆ— = [๐‘โˆ—1, ..., ๐‘โˆ—๐พ ] be the correspondingoutput distribution. Consider the (convex) set

๐’ฐ = {๐น โˆˆ ๐’ฎโˆฃ๐‘…(๐‘ฆ;๐น ) = ๐‘…โˆ— and ๐ธ(๐‘‹2) = ๐‘ƒ0} , (11)

where ๐‘ƒ0 โ‰ค ๐‘ƒ is the average power of ๐‘‹ under ๐น โˆ—. On thisset, ๐ผ(๐‘‹ ;๐‘Œ ) is a linear functional of the distribution of ๐‘‹(that is, it is a linear functional of ๐น ):

๐ผ(๐‘‹ ;๐‘Œ ) = ๐ป(๐‘Œ )โˆ’๐ป(๐‘Œ โˆฃ๐‘‹)

= โˆ’๐พโˆ‘๐‘–=1

๐‘๐‘–โˆ— log ๐‘๐‘–โˆ— +

โˆซ ๐ด2

๐ด1

๐‘‘๐น (๐‘ฅ)

๐พโˆ‘๐‘–=1

๐‘Š๐‘–(๐‘ฅ) log๐‘Š๐‘–(๐‘ฅ).

It follows that ๐ผ(๐‘‹ ;๐‘Œ ) attains its maximum over ๐’ฐ (andhence over ๐’ฎ) at an extreme point of ๐’ฐ .

To complete the proof we next show that the extreme pointsof ๐’ฐ have at most ๐พ + 1 support points. We can view the

3638 IEEE TRANSACTIONS ON COMMUNICATIONS, VOL. 57, NO. 12, DECEMBER 2009

set ๐’ฐ as the intersection of the (convex) set of probabilitydistributions over [๐ด1, ๐ด2] (say โ„ณ) and the ๐พ hyperplanes:

๐ป๐‘– :

โˆซ ๐ด2

๐ด1

๐‘Š๐‘–(๐‘ฅ)๐‘‘๐น (๐‘ฅ) = ๐‘๐‘–โˆ— , 1 โ‰ค ๐‘– โ‰ค ๐พ โˆ’ 1 , (12)

and

๐ป๐พ :

โˆซ ๐ด2

๐ด1

๐‘ฅ2๐‘‘๐น (๐‘ฅ) = ๐‘ƒ0 . (13)

Since the extreme points of โ„ณ are precisely the set of unitmasses {๐›ฟ๐‘ฅ, ๐‘ฅ โˆˆ [๐ด1, ๐ด2]}, from Dubinsโ€™ theorem [6] we getthat the extreme points of ๐’ฐ have at most ๐พ+1 support points.Remark: The details of verifying that the technical conditionsrequired to apply Dubinsโ€™ theorem are indeed satisfied in thepreceding setting can be found in our technical report [24].

ACKNOWLEDGMENTS

The authors thank an anonymous reviewer for suggestingthe analysis of the additive quantization noise model in SectionIV-D. The first author also thanks Prof. Janos Englander (De-partment of Statistics and Applied Probability, UCSB), Prof.Shiv Chandrasekaran (Department of Electrical and ComputerEngineering, UCSB) and Mr. Vinay Melkote (Signal Compres-sion Laboratory, UCSB) for several useful discussions.

REFERENCES

[1] R. Hiremane, โ€œFrom Mooreโ€™s law to Intel innovation-prediction toreality," Technology Intel Mag., Apr. 2005.

[2] R. Walden, โ€œAnalog-to-digital converter survey and analysis," IEEE J.Sel. Areas Commun., vol. 17, no. 4, pp. 539-550, Apr. 1999.

[3] J. Huang and S. P. Meyn, โ€œCharacterization and computation of optimaldistributions for channel coding," IEEE Trans. Inf. Theory, vol. 51, no.7, pp. 2336-2351, July 2005.

[4] R. G. Gallager, Information Theory and Reliable Communication. NewYork: John Wiley and Sons, 1968.

[5] H. S. Witsenhausen, โ€œSome aspects of convexity useful in informationtheory," IEEE Trans. Inf. Theory, vol. 26, no. 3, pp. 265-271, May 1980.

[6] L. E. Dubins, โ€œOn extreme points of convex sets," J. Math. Anal. Appl.,vol. 5, pp. 237-244, May 1962.

[7] N. Phamdo and F. Alajaji, โ€œSoft-decision demodulation design forCOVQ over white, colored, and ISI Gaussian channels," IEEE Trans.Commun., vol. 48, no. 9, pp. 1499-1506, Sep. 2000.

[8] F. Behnamfar, F. Alajaji, and T. Linder, โ€œChannel-optimized quantizationwith soft-decision demodulation for space-time orthogonal block-codedchannels," IEEE Trans. Signal Process., vol. 54, no. 10, pp. 3935-3946,Oct. 2006.

[9] J. A. Nossek and M. T. Ivrlac, โ€œCapacity and coding for quantizedMIMO systems," in Proc. Intl. Conf. Wireless Commun. Mobile Com-puting (IWCMC), 2006, Vancouver, Canada.

[10] O. Dabeer, J. Singh, and U. Madhow, โ€œOn the limits of communicationperformance with one-bit analog-to-digital conversion," in Proc. IEEEWorkshop Signal Proc. Advances Wireless Commun. (SPAWC), 2006,Cannes, France.

[11] J. Singh, O. Dabeer, and U. Madhow, โ€œCommunication limits with lowprecision analog-to-digital conversion at the receiver," in Proc. IEEEIntl. Conf. Commun. (ICC), 2007, Glasgow, Scotland.

[12] J. Singh, O. Dabeer, and U. Madhow, โ€œCapacity of the discrete-timeAWGN Channel under output quantization," in Proc. IEEE Intl. Symp.Inform. Theory (ISIT), 2008, Toronto, Canada.

[13] L. N. Lee, โ€œOn optimal soft decision demodulation," IEEE Trans. Inf.Theory, vol. 22, pp. 437-444, July 1976.

[14] J. Salz and E. Zehavi, โ€œDecoding under integer metrics constraints,"IEEE Trans. Commun., vol. 43, no. 3, pp. 307-317, Mar. 1995.

[15] E. Masry, โ€œThe reconstruction of analog signals from the sign of theirnoisy samples," IEEE Trans. Inf. Theory, vol. 27, pp. 735-745, Nov.1981.

[16] Z. Cvetkovic and I. Daubechies, โ€œSingle-bit oversampled A/D conver-sion with exponential accuracy in the bit-rate," in Proc. Data Compres-sion Conf. (DCC), 2000, Utah, USA.

[17] P. Ishwar, A. Kumar, and K. Ramachandran, โ€œDistributed sampling fordense sensor networks: a bit-conservation principal," in Proc. Inform.Processing Sensor Networks (IPSN), 2003, Palo Alto, USA.

[18] A. Host-Madsen and P. Handel, โ€œEffects of sampling and quantizationon single-tone frequency estimation," IEEE Trans. Signal Process., vol.48, pp. 650-662, Mar. 2000.

[19] T. Andersson, M. Skoglund, and P. Handel, โ€œFrequency estimation by1-bit quantization and table look-up processing," in Proc. Euro. Sig.Proc. Conf. (EUSIPCO), 2000, Tampere, Finland.

[20] D. Rousseau, G. V. Anand, and F. Chapeau-Blondeau, โ€œNonlinearestimation from quantized signals: quantizer optimization and stochasticresonance," in Proc. Third Intl. Symp. Physics Signal Image Proc.(PSIP), 2003, Grenoble, France.

[21] O. Dabeer and A. Karnik, โ€œSignal parameter estimation using 1-bitdithered quantization," IEEE Trans. Inf. Theory, vol. 52, no. 12, pp.5389-5405, Dec. 2006.

[22] O. Dabeer and E. Masry, โ€œMultivariate signal parameter estimationunder dependent noise from 1-bit dithered quantized data," IEEE Trans.Inf. Theory, vol. 54, no. 4, pp. 1637-1654, Apr. 2008.

[23] M. Fozunbal, S. W. McLaughlin, and R. W. Schafer, โ€œCapacity analysisfor continuous-alphabet channels with side information, part I: a generalframework," IEEE Trans. Inf. Theory, vol. 51, no. 9, pp. 3075-3085, Sep.2005.

[24] J. Singh, O. Dabeer, and U. Madhow, โ€œOn the limits of commu-nication with low-precision analog-to-digital conversion at the re-ceiver," WCSL Technical Report, UCSB. [Online]. Available: โ„Ž๐‘ก๐‘ก๐‘ ://๐‘ค๐‘ค๐‘ค.๐‘’๐‘๐‘’.๐‘ข๐‘๐‘ ๐‘.๐‘’๐‘‘๐‘ข/๐‘ค๐‘๐‘ ๐‘™/๐‘๐‘ข๐‘๐‘™๐‘–๐‘๐‘Ž๐‘ก๐‘–๐‘œ๐‘›๐‘ .โ„Ž๐‘ก๐‘š๐‘™

[25] J. G. Smith, โ€œOn the information capacity of peak and average powerconstrained Gaussian channels," Ph.D. dissertation, Univ. of California,Berkeley, Dec. 1969.

[26] I. C. Abou-Faycal, M. D. Trott, and S. Shamai, โ€œThe capacity ofdiscrete-time memoryless Rayleigh fading channels," IEEE Trans. Inf.Theory, vol. 47, no. 4, pp. 1290-1301, May 2001.

[27] T. Cover and J. Thomas, Elements of Information Theory. New York:Wiley Series in Telecommunications, 2nd Edition.

[28] I. Csiszar and J. Korner, Information Theory: Coding Theorems forDiscrete Memoryless Systems. Academic Press, 1981.

[29] M. Chiang and S. Boyd, โ€œGeometric programming duals of channelcapacity and rate distortion," IEEE Trans. Inf. Theory, vol. 50, no. 2,pp. 245-258, Feb. 2004.

[30] A. Lapidoth and S. M. Moser, โ€œCapacity bounds via duality withapplications to multiple-antenna systems on flat fading channels," IEEETrans. Inf. Theory, vol. 49, no. 10, pp. 2426-2467, Oct. 2003.

[31] J. G. Proakis, Digital Communications. McGraw Hill, 4th Edition.[32] M. S. Braasch and A. J. van Dierendonck, โ€œGPS receiver architectures

and measurements," Proc. IEEE, vol. 87, no. 1, pp. 48-64, Jan. 1999.

Jaspreet Singh received the B.Tech. degree inelectrical engineering from the Indian Institute ofTechnology, Delhi, in 2004, and the M.S. degreein electrical and computer engineering from theUniversity of California, Santa Barbara (UCSB) in2005. He is currently pursuing the Ph.D. degreein electrical and computer engineering at UCSB.He has been a summer intern at the Swiss FederalInstitute of Technology (EPFL), Lausanne, and atQualcomm Flarion Technologies, New Jersey, dur-ing 2003 and 2008 respectively. He is a recipient of

the National Talent Scholarship (1998-2004) awarded by the Government ofIndia.

Onkar Dabeer (Sโ€™99-Mโ€™02) was born in Solapur,India, on September 15, 1974. He received theB.Tech. and M.Tech. degrees in Electrical Engi-neering from the Indian Institute of Technology,Bombay, in 1996 and 1998 respectively, and, thePh.D. degree in Electrical Engineering from theUniversity of California at San Diego in June 2002.He was a postdoctoral researcher at University ofCalifornia, Santa Barbara from July 2002 to July2003. From August 2003 to August 2004, he servedas a Senior Engineer at Qualcomm Inc., San Diego.

Since September 2004 he has been on the faculty at the School of Technologyand Computer Science, Tata Institute of Fundamental Research, Mumbai(Bombay), India. His research interests include wireless communicationnetworks, sensor networks, low-precision signal processing, and informationtheory.

SINGH et al.: ON THE LIMITS OF COMMUNICATION WITH LOW-PRECISION ANALOG-TO-DIGITAL CONVERSION AT THE RECEIVER 3639

Upamanyu Madhow received his bachelorโ€™s degreein electrical engineering from the Indian Institute ofTechnology, Kanpur, in 1985. He received the M. S.and Ph. D. degrees in electrical engineering from theUniversity of Illinois, Urbana-Champaign in 1987and 1990, respectively.

From 1990 to 1991, he was a Visiting AssistantProfessor at the University of Illinois. From 1991 to1994, he was a research scientist at Bell Commu-nications Research, Morristown, NJ. From 1994 to1999, he was on the faculty of the Department of

Electrical and Computer Engineering at the University of Illinois, Urbana-

Champaign. Since December 1999, he has been with the Department ofElectrical and Computer Engineering at the University of California, SantaBarbara, where he is currently a Professor. His research interests are incommunication systems and networking, with current emphasis on wirelesscommunication, sensor networks and multimedia security.

Dr. Madhow is a recipient of the NSF CAREER award. He has servedas Associate Editor for the IEEE TRANSACTIONS ON COMMUNICATIONS,IEEE TRANSACTIONS ON INFORMATION THEORY, and IEEE TRANSAC-TIONS ON INFORMATION FORENSICS AND SECURITY. He is the author ofthe graduate textbook, Fundamentals of Digital Communication, published byCambridge University Press.