On the limits of communication with low-precision analog-to-digital conversion at the receiver
-
Upload
independent -
Category
Documents
-
view
4 -
download
0
Transcript of On the limits of communication with low-precision analog-to-digital conversion at the receiver
IEEE TRANSACTIONS ON COMMUNICATIONS, VOL. 57, NO. 12, DECEMBER 2009 3629
On the Limits of Communication withLow-Precision Analog-to-Digital
Conversion at the ReceiverJaspreet Singh, Student Member, IEEE, Onkar Dabeer, Member, IEEE, and Upamanyu Madhow, Fellow, IEEE
AbstractโAs communication systems scale up in speed andbandwidth, the cost and power consumption of high-precision(e.g., 8-12 bits) analog-to-digital conversion (ADC) becomes thelimiting factor in modern transceiver architectures based ondigital signal processing. In this work, we explore the impactof lowering the precision of the ADC on the performance ofthe communication link. Specifically, we evaluate the communi-cation limits imposed by low-precision ADC (e.g., 1-3 bits) fortransmission over the real discrete-time Additive White GaussianNoise (AWGN) channel, under an average power constraint onthe input. For an ADC with ๐พ quantization bins (i.e., a precisionof log2 ๐พ bits), we show that the input distribution need not haveany more than ๐พ+1 mass points to achieve the channel capacity.For 2-bin (1-bit) symmetric quantization, this result is tightenedto show that binary antipodal signaling is optimum for any signal-to-noise ratio (SNR). For multi-bit quantization, a dual formula-tion of the channel capacity problem is used to obtain tight upperbounds on the capacity. The cutting-plane algorithm is employedto compute the capacity numerically, and the results obtained areused to make the following encouraging observations : (a) up toa moderately high SNR of 20 dB, 2-3 bit quantization resultsin only 10-20% reduction of spectral efficiency compared to un-quantized observations, (b) standard equiprobable pulse ampli-tude modulated input with quantizer thresholds set to implementmaximum likelihood hard decisions is asymptotically optimumat high SNR, and works well at low to moderate SNRs as well.
Index TermsโChannel capacity, optimum input distribution,AWGN channel, analog to digital converter, quantization.
I. INTRODUCTION
D IGITAL signal processing (DSP) forms the core ofmodern digital communication receiver implementations,
with the analog baseband signal being converted to digitalform using analog-to-digital converters (ADCs) which typi-cally have fairly high (e.g., 8-12 bits) precision. Operationssuch as synchronization, equalization and demodulation arethen performed in the digital domain, greatly enhancing theflexibility available to the designer. The continuing exponential
Paper approved by F. Alajaji, the Editor for Source and Source/ChannelCoding of the IEEE Communications Society. Manuscript received October24, 2008; revised March 6, 2009.
J. Singh and U. Madhow are with the Department of Electrical andComputer Engineering, University of California, Santa Barbara, CA, 93106,USA (e-mail: {jsingh, madhow}@ece.ucsb.edu). Their research was supportedby the National Science Foundation under grants CCF-0729222 and CNS-0832154.
O. Dabeer is with the School of Technology and Computer Science,Tata Institute of Fundamental Research, Mumbai, 400005, India (e-mail:[email protected]). His research was supported by the Homi BhabhaFellowship.
Digital Object Identifier 10.1109/TCOMM.2009.12.080559
ADC QuantizerQ+X Y
N
Fig. 1. ๐ = Q(๐ + ๐) : The AWGN-Quantized Ouput channel inducedby the output quantizer Q.
advances in digital electronics, often summarized by Mooreโsโlawโ [1], imply that integrated circuit implementations ofsuch DSP-centric architectures can be expected to continuescaling up in speed and down in cost. However, as thebandwidth of a communication system increases, accurateconversion of the analog received signal into digital formrequires high-precision, high-speed ADC, which is costly andpower-hungry [2]. One possible approach for designing suchhigh-speed systems is to drastically reduce the number ofbits of ADC precision (e.g., to 1-3 bits) as sampling ratesscale up. Such a design choice has significant implications forall aspects of receiver design, including carrier and timingsynchronization, equalization, demodulation and decoding.However, before embarking on a comprehensive rethinking ofcommunication system design, it is important to understandthe fundamental limits on communication performance im-posed by low-precision ADC. In this paper, we take a first stepin this direction, investigating Shannon-theoretic performancelimits for the following idealized model : linear modulationover a real baseband Additive White Gaussian Noise (AWGN)channel, with symbol rate Nyquist samples quantized by alow-precision ADC at the receiver. This induces a discrete-time memoryless AWGN-Quantized Output channel, which isdepicted in Figure 1. Under an average power constraint onthe input, we obtain the following results:
1) For ๐พ-bin (i.e., log2 ๐พ bits) output quantization, weprove that the input distribution need not have any morethan ๐พ+1 mass points to achieve the channel capacity.(Numerical computation of optimal input distributionsreveals that ๐พ mass points are sufficient.) Anintermediate result of interest is that, when the AWGNchannel output is quantized with finite-precision, anaverage power constraint on the input leads to an
0090-6778/09$25.00 cโ 2009 IEEE
3630 IEEE TRANSACTIONS ON COMMUNICATIONS, VOL. 57, NO. 12, DECEMBER 2009
implicit peak power constraint, in the sense that anoptimal input distribution must have bounded support.
2) For 1-bit symmetric quantization, the preceding resultis tightened analytically to show that binary antipodalsignaling is optimal for any signal-to-noise ratio (SNR).
3) For multi-bit quantizers, tight upper bounds on capacityare obtained using a dual formulation of the channelcapacity problem. Near-optimal input distributionsthat approach these bounds are computed using thecutting-plane algorithm [3].
4) While the preceding results optimize the inputdistribution for a fixed quantizer, comparison withan unquantized system requires optimization over thechoice of the quantizer as well. We numerically obtainoptimal 2-bit and 3-bit symmetric quantizers.
5) From our numerical results, we infer that the use of low-precision ADC incurs a relatively small loss in spectralefficiency compared to unquantized observations. Forexample, at 0 dB SNR, a receiver with 2-bit ADCachieves 95% of the spectral efficiency attained withunquantized observations. Even at a moderately highSNR of 20 dB, a receiver with 3-bit ADC achieves 85%of the spectral efficiency attained with unquantizedobservations. This indicates that DSP-centric designbased on low-precision ADC is indeed attractive ascommunication bandwidths scale up, since the small lossin spectral efficiency should be acceptable in this regime.Furthermore, we observe that a โsensibleโ choice ofstandard equiprobable pulse amplitude modulated(PAM) input with ADC thresholds set to implementmaximum likelihood (ML) hard decisions achieves per-formance which is quite close to that obtained by numer-ical optimization of the quantizer and input distribution.
Related Work: For a discrete memoryless channel (DMC),Gallager first showed that the number of input points withnonzero probability mass need not exceed the cardinalityof the output [4, p. 96, Corollary 3]. In our setting, theinput alphabet is not a priori discrete, and there is a powerconstraint, so that the result in [4] does not apply. Our keyresult on the achievability of the capacity by a discrete inputis actually inspired by Witsenhausenโs result in [5], whereDubinsโ theorem [6] was used to show that the capacity ofa discrete-time memoryless channel with output cardinality๐พ , under a peak power constraint is achievable by a discreteinput with at most ๐พ points. The key to our proof is to showthat under output quantization, an average power constraintautomatically induces a peak power constraint, after whichwe can use Dubinsโ theorem in a manner similar to thedevelopment in [5] to show that ๐พ + 1 mass points sufficeto achieve the average power constrained capacity.
Prior work on the effect of reduced ADC precision onchannel capacity with fixed input distribution includes [7]โ[9]. However, other than our own results reported earlier in[10]โ[12], we are not aware of an information-theoretic inves-tigation with low-precision ADC that includes optimizationof the input distribution. Another related class of problemsthat deserves mention relates to the impact of finite-precisionquantization on the information-theoretic measure of channelcut-off rate rather than channel capacity (see, e.g. [13], [14]).
Given the encouraging results here, it becomes importantto explore the impact of low-precision ADC on receiver taskssuch as synchronization and equalization, which we haveignored in our idealized model (essentially assuming that thesetasks have somehow already been accomplished). Relatedwork on estimation using low-precision samples which maybe relevant for this purpose includes the use of dither forsignal reconstruction [15]โ[17], frequency estimation using 1-bit ADC [18], [19], choice of quantization threshold for signalamplitude estimation [20], and signal parameter estimationusing 1-bit dithered quantization [21], [22].
Organization of the Paper: The rest of the paper isorganized as follows. In Section II, we describe the channelmodel and present results concerning the structure of theoptimal input distributions. In Section III, we discuss capacitycomputation, including duality-based upper bounds on capac-ity. Numerical results are provided in Section IV, followed bythe conclusions in Section V.
II. CHANNEL MODEL AND STRUCTURE OF OPTIMAL
INPUT DISTRIBUTIONS
We consider linear modulation over a real AWGN channel,with symbol rate Nyquist samples quantized by a ๐พ-bin quan-tizer Q at the receiver. This induces the following discrete-timememoryless AWGN-Quantized Output (AWGN-QO) channel
๐ = Q (๐ +๐) . (1)
Here ๐ โ โ is the channel input with cumulative distributionfunction ๐น (๐ฅ), ๐ โ {๐ฆ1, โ โ โ , ๐ฆ๐พ} is the (discrete) channeloutput, and ๐ is ๐ฉ (0, ๐2) (the Gaussian random variable withmean 0 and variance ๐2). Q maps the real valued input ๐ +๐ to one of the ๐พ bins, producing a discrete output ๐ . Inthis work, we only consider quantizers for which each bin isan interval of the real line. The quantizer Q with ๐พ bins istherefore characterized by the set of its (๐พ โ 1) thresholds๐๐๐ := [๐1, ๐2, โ โ โ , ๐๐พโ1] โ โ
๐พโ1, such that โโ := ๐0 <๐1 < ๐2 < โ โ โ < ๐๐พโ1 < ๐๐พ := โ. The output ๐ is assignedthe value ๐ฆ๐ when the quantizer input (๐ + ๐) falls in the๐๐กโ bin, which is given by the interval (๐๐โ1, ๐๐]. The resultingtransition probability functions are
๐๐(๐ฅ) = P(๐ = ๐ฆ๐โฃ๐ = ๐ฅ)
= ๐
(๐๐โ1 โ ๐ฅ
๐
)โ๐
(๐๐ โ ๐ฅ
๐
), 1 โค ๐ โค ๐พ,
(2)
where ๐(โ ) is the complementary Gaussian distribution func-tion,
๐(๐ง) =1โ2๐
โซ โ
๐ง
exp(โ๐ก2/2)๐๐ก . (3)
The Probability Mass Function (PMF) of the output ๐ , cor-responding to the input distribution ๐น is
๐ (๐ฆ๐;๐น ) =
โซ โ
โโ๐๐(๐ฅ)๐๐น (๐ฅ), 1 โค ๐ โค ๐พ, (4)
SINGH et al.: ON THE LIMITS OF COMMUNICATION WITH LOW-PRECISION ANALOG-TO-DIGITAL CONVERSION AT THE RECEIVER 3631
and the input-output mutual information ๐ผ(๐ ;๐ ), expressedexplicitly as a function of ๐น is
๐ผ(๐น ) =
โซ โ
โโ
๐พโ๐=1
๐๐(๐ฅ) log๐๐(๐ฅ)
๐ (๐ฆ๐;๐น )๐๐น (๐ฅ) .1 (5)
Under an average power constraint ๐ , we wish to find thecapacity of the channel (1), given by
๐ถ = sup๐นโโฑ
๐ผ(๐น ), (6)
where โฑ ={๐น : ๐ผ[๐2] =
โซโโโ ๐ฅ2๐๐น (๐ฅ) โค ๐
}, i.e., the set
of all average power constrained distributions on โ.
Structural Properties of Optimal Inputs: We begin by em-ploying the Karush-Kuhn-Tucker (KKT) optimality conditionto show that, even though we have not imposed a peak powerconstraint on the input, it is automatically induced by theaverage power constraint. Specifically, a capacity achievingdistribution for the AWGN-QO channel (1) must have boundedsupport. 2
A. An Implicit Peak Power Constraint
Using convex optimization principles, the followingnecessary and sufficient KKT optimality condition canbe derived for our problem, in a manner similar to thedevelopment in [25], [26]. An input distribution ๐น โ achievesthe capacity ๐ถ in (6) if and only if there exists ๐พ โฅ 0 such that
๐พโ๐=1
๐๐(๐ฅ) log๐๐(๐ฅ)
๐ (๐ฆ๐;๐น โ)+ ๐พ(๐ โ ๐ฅ2) โค ๐ถ (7)
for all ๐ฅ, with equality if ๐ฅ is in the support 3 of ๐น โ, wherethe transition probability function ๐๐(๐ฅ), and the output prob-ability ๐ (๐ฆ๐;๐น
โ) are as specified in (2) and (4), respectively.The summation on the left-hand side (LHS) of (7) is the
Kullback-Leibler divergence (or the relative entropy) betweenthe transition PMF {๐๐(๐ฅ), ๐ = 1, . . . ,๐พ} and the outputPMF {๐ (๐ฆ๐;๐น ), ๐ = 1, . . . ,๐พ}. For convenience, let us denotethis divergence function by ๐(๐ฅ;๐น ), that is,
๐(๐ฅ;๐น ) =
๐พโ๐=1
๐๐(๐ฅ) log๐๐(๐ฅ)
๐ (๐ฆ๐;๐น ). (8)
We begin by studying the behavior of this function in thelimit as ๐ฅ โ โ.
Lemma 1: For the AWGN-QO channel (1), the divergencefunction ๐(๐ฅ;๐น ) satisfies the following properties(a) lim
๐ฅโโ๐(๐ฅ;๐น ) = โ log๐ (๐ฆ๐พ ;๐น ).(b) There exists a finite constant ๐ด0 such thatfor ๐ฅ > ๐ด0, ๐(๐ฅ;๐น ) < โ log๐ (๐ฆ๐พ ;๐น ).4
1The logarithm is base 2 throughout the paper, so the mutual informationis measured in bits.
2That there exists a capacity achieving distribution follows by standardfunction analytic arguments [23]. For details, see our technical report [24].
3The support of a distribution ๐น (or the set of increase points of ๐น ) is theset ๐๐(๐น ) = {๐ฅ : ๐น (๐ฅ+ ๐)โ ๐น (๐ฅโ ๐) > 0, โ๐ > 0}.
4The constant ๐ด0 depends on the choice of the input ๐น . For notationalsimplicity, we do not explicitly show this dependence.
Proof: We have
๐(๐ฅ;๐น ) =
๐พโ๐=1
๐๐(๐ฅ) log๐๐(๐ฅ)
๐ (๐ฆ๐;๐น )
=
๐พโ๐=1
๐๐(๐ฅ) log๐๐(๐ฅ) โ๐พโ๐=1
๐๐(๐ฅ) log๐ (๐ฆ๐;๐น ) .
As ๐ฅ โ โ, the PMF {๐๐(๐ฅ), ๐ = 1, . . . ,๐พ} โ 1(๐ = ๐พ),where 1(โ ) is the indicator function. This observation,combined with the fact that the entropy of a finite alphabetrandom variable is a continuous function of its probability law,gives lim
๐ฅโโ ๐(๐ฅ;๐น ) = 0โ log๐ (๐ฆ๐พ ;๐น ) = โ log๐ (๐ฆ๐พ ;๐น ).
Next we prove part (b). For ๐ฅ > ๐๐พโ1, ๐๐(๐ฅ) is a strictlydecreasing function of ๐ฅ for ๐ โค ๐พโ 1 and strictly increasingfunction of ๐ฅ for ๐ = ๐พ . Since {๐๐(๐ฅ)} โ 1(๐ = ๐พ) as๐ฅ โ โ, it follows that there is a constant ๐ด0 such that๐๐(๐ด0) < ๐ (๐ฆ๐;๐น ) for ๐ โค ๐พโ1 and ๐๐พ(๐ด0) > ๐ (๐ฆ๐พ ;๐น ).Therefore, it follows that for ๐ฅ > ๐ด0,
๐(๐ฅ;๐น ) =
๐พโ๐=1
๐๐(๐ฅ) log๐๐(๐ฅ)
๐ (๐ฆ๐;๐น )
< ๐๐พ(๐ฅ) log๐๐พ(๐ฅ)
๐ (๐ฆ๐พ ;๐น )< โ log๐ (๐ฆ๐พ ;๐น ).
The saturating nature of the divergence function for theAWGN-QO channel, as stated above, coupled with the KKTcondition, is now used to prove that a capacity achievingdistribution must have bounded support.
Proposition 1: For the average power constrained AWGN-QO channel (1), an optimal input distribution must havebounded support.
Proof: Let ๐น โ be an optimal input, so that there exists๐พ โฅ 0 such that (7) is satisfied with equality at every pointin the support of ๐น โ. We exploit this necessary condition toshow that the support of ๐น โ is upper bounded. Specifically,we prove that there exists a finite constant ๐ด2
โ such that it isnot possible to attain equality in (7) for any ๐ฅ > ๐ด2
โ.From Lemma 1, we get lim
๐ฅโโ๐(๐ฅ;๐นโ)=โ log๐ (๐ฆ๐พ ;๐น โ)=:
๐ฟ. Also, there exists a finite constant ๐ด0 such that for๐ฅ > ๐ด0, ๐(๐ฅ;๐น
โ) < ๐ฟ.We consider two possible cases.
โ Case 1: ๐พ > 0.For ๐ฅ > ๐ด0, we have ๐(๐ฅ;๐น โ) < ๐ฟ.For ๐ฅ >
โmax{0, (๐ฟโ ๐ถ + ๐พ๐ )/๐พ} =: ๐ด, we have
๐พ(๐ โ ๐ฅ2) < ๐ถ โ ๐ฟ.Defining ๐ดโ
2 = max{๐ด0, ๐ด}, we get the desired result.โ Case 2: ๐พ = 0.
Since ๐พ = 0, the KKT condition (7) reduces to
๐(๐ฅ;๐น โ) โค ๐ถ , โ๐ฅ.Taking limit ๐ฅ โ โ on both sides, we get๐ฟ = lim
๐ฅโโ ๐(๐ฅ;๐น โ) โค ๐ถ.Hence, choosing ๐ดโ
2 = ๐ด0, for ๐ฅ > ๐ดโ2 we get,
๐(๐ฅ;๐น โ) < ๐ฟ โค ๐ถ, that is, ๐(๐ฅ;๐น โ) + ๐พ(๐ โ ๐ฅ2) < ๐ถ.
Combining the two cases, we have shown that the supportof the distribution ๐น โ has a finite upper bound ๐ด2
โ. Usingsimilar arguments, it can easily be shown that the support of
3632 IEEE TRANSACTIONS ON COMMUNICATIONS, VOL. 57, NO. 12, DECEMBER 2009
๐น โ has a finite lower bound ๐ด1โ as well, which implies that
๐น โ has bounded support.
B. Achievability of Capacity by a Discrete Input
In [5], Witsenhausen considered a stationary discrete-timememoryless channel, with a continuous input ๐ takingvalues on the bounded interval [๐ด1, ๐ด2] โ โ, and a discreteoutput ๐ of finite cardinality ๐พ . Using Dubinsโ theorem[6], it was shown that if the transition probability functionsare continuous (i.e., ๐๐(๐ฅ) is continuous in ๐ฅ, for each๐ = 1, โ โ โ ,๐พ), then the capacity is achievable by a discreteinput distribution with at most ๐พ mass points. As statedin Proposition 2 below (proved in Appendix A), this resultcan be extended to show that, if an additional averagepower constraint is imposed on the input, the capacity is thenachievable by a discrete input with at most ๐พ+1 mass points.
Proposition 2: Consider a stationary discrete-time memo-ryless channel with a continuous input ๐ that takes val-ues in the bounded interval [๐ด1, ๐ด2], and a discrete output๐ โ {๐ฆ1, ๐ฆ2, โ โ โ , ๐ฆ๐พ}. Let the channel transition probabilityfunction ๐๐(๐ฅ) = P(๐ = ๐ฆ๐โฃ๐ = ๐ฅ) be continuous in ๐ฅ foreach ๐, where 1 โค ๐ โค ๐พ . The capacity of this channel, underan average power constraint on the input, is achievable by adiscrete input distribution with at most ๐พ + 1 mass points.
Proof: See Appendix A.Proposition 2, coupled with the implicit peak power con-
straint derived in the previous subsection (Proposition 1), givesus the following result.
Theorem 1: The capacity of the average power constrainedAWGN-QO channel (1) is achievable by a discrete inputdistribution with at most ๐พ + 1 points of support.
Proof: From Proposition 1, we know that an optimalinput ๐น โ has bounded support [๐ด1
โ, ๐ด2โ]. Hence, to obtain
the capacity in (6), we can maximize ๐ผ(๐น ) over only thoseaverage power constrained distributions that have support in[๐ด1
โ, ๐ด2โ]. Since the transition functions ๐๐(๐ฅ) are continu-
ous, Proposition 2 guarantees that this maximum is achievableby a discrete input with at most ๐พ + 1 points.
C. Symmetric Inputs for Symmetric Quantization
For our capacity computations ahead, we assume that thequantizer Q employed in (1) is symmetric, i.e., its thresholdvector ๐๐๐ is symmetric about the origin. Given the symmetricnature of the AWGN noise and the power constraint, it seemsintuitively plausible that restriction to symmetric quantizersshould not be suboptimal from the point of view of optimizingover the quantizer choice in (1), although a proof of thisconjecture has eluded us. However, once we assume that thequantizer in (1) is symmetric, we can restrict attention to onlysymmetric inputs without loss of optimality, as stated in thefollowing Lemma.5
Lemma 2: If the quantizer in (1) is symmetric, then, with-out loss of optimality, we can consider only symmetric inputsfor the capacity computation in (6).Proof: Suppose we are given an input random variable ๐ (withdistribution ๐น ) that is not necessarily symmetric. Denote the
5A random variable ๐ (with distribution F) is symmetric if ๐ and โ๐have the same distribution, that is, ๐น (๐ฅ) = 1โ ๐น (โ๐ฅ),โ ๐ฅ โ โ.
distribution of โ๐ by ๐บ (so that ๐บ(๐ฅ) = 1โ ๐น (โ๐ฅ), โ ๐ฅ โโ). Due to the symmetric nature of the noise ๐ and thequantizer Q, it is easy to see that ๐ and โ๐ result in thesame input-output mutual information, that is, ๐ผ(๐น ) = ๐ผ(๐บ).Consider now the following symmetric mixture distribution
๐น (๐ฅ) =๐น (๐ฅ) +๐บ(๐ฅ)
2.
Since the mutual information is concave in the input distri-bution, we get ๐ผ(๐น ) โฅ ๐ผ(๐น )+๐ผ(๐บ)
2 = ๐ผ(๐น ), which proves thedesired result.
III. CAPACITY COMPUTATION
In this section, we consider capacity computation for theAWGN-QO channel. We first provide an explicit capacityformula for the extreme scenario of 1-bit symmetric quanti-zation, and then discuss numerical computations for multi-bitquantization.
A. 1-bit Symmetric Quantization : Binary Antipodal Signalingis Optimal
With 1-bit symmetric quantization, the channel is
๐ = sign(๐ +๐). (9)
Theorem 1 (Section II-B) guarantees that the capacity of thischannel, under an average power constraint, is achievable bya discrete input distribution with at most 3 points. This resultis further tightened by the following theorem that shows theoptimality of binary antipodal signaling for all SNRs.
Theorem 2: For the 1-bit symmetric quantized channelmodel (9), the capacity is achieved by binary antipodal sig-naling and is given by
๐ถ = 1โ โ(๐(โ
SNR))
, SNR =๐
๐2,
where โ(โ ) is the binary entropy function,
โ(๐) = โ๐ log(๐)โ (1โ ๐) log(1โ ๐) , 0 โค ๐ โค 1 ,
and ๐(โ ) is the complementary Gaussian distribution functionshown in (3).
Proof: Since ๐ is binary it is easy to see that
๐ป(๐ โฃ๐) = ๐ผ
[โ
(๐
(๐
๐
))],
where ๐ผ denotes the expectation operator. Therefore
๐ผ(๐,๐ ) = ๐ป(๐ )โ ๐ผ
[โ
(๐
(๐
๐
))],
which we wish to maximize over all input distributionssatisfying ๐ผ[๐2] โค ๐ . Since the quantizer is symmetric, wecan restrict attention to symmetric input distributions withoutloss of optimality (cf. Lemma 2). On doing so, we obtainthat the PMF of the output ๐ is also symmetric (since thequantizer and the noise are already symmetric). Therefore,๐ป(๐ ) = 1 bit, and we get
๐ถ = 1โ min๐ symmetric๐ผ[๐2]โค๐
๐ผ
[โ
(๐
(๐
๐
))].
SINGH et al.: ON THE LIMITS OF COMMUNICATION WITH LOW-PRECISION ANALOG-TO-DIGITAL CONVERSION AT THE RECEIVER 3633
Since โ(๐(๐ง)) is an even function, we get that
๐ป(๐ โฃ๐) = ๐ผ
[โ
(๐
(๐
๐
))]= ๐ผ
[โ
(๐
( โฃ๐ โฃ๐
))].
In [24], we show that the function โ(๐(โ๐ง)) is convex in ๐ง.
Jensenโs inequality [27] thus implies
๐ป(๐ โฃ๐) โฅ โ(๐(โ
SNR))
with equality iff ๐2 = ๐ . Coupled with the symmetrycondition on ๐ , this implies that binary antipodal signalingachieves capacity and the capacity is
๐ถ = 1โ โ(๐(โ
SNR))
.
B. Multi-Bit Quantization
We now consider ๐พ-bin symmetric quantization for ๐พ > 2.Every choice of the quantizer results in a unique channelmodel (1). In this section, we discuss capacity computationassuming a fixed quantizer only. Optimization over the quan-tizer choice is performed in Section IV.
1) Capacity computation using cutting-plane algorithm:Contrary to the 1-bit case, closed form expressions for optimalinput and capacity appear unlikely for multi-bit quantization,due to the complicated expression for mutual information.We therefore resort to the cutting-plane algorithm [3, Sec IV-A] to generate optimal inputs numerically. For channels withcontinuous input alphabets, the cutting-plane algorithm can,in general, be used to generate nearly optimal discrete inputdistributions. It is therefore well matched to our problem, forwhich we already know that the capacity is achievable by adiscrete input distribution.
For our simulations, we fix the noise variance ๐2 = 1, andvary the power ๐ to obtain capacity at different SNRs. Toapply the cutting-plane algorithm, we take a fine quantizeddiscrete grid on the interval [โ10
โ๐ , 10
โ๐ ], and optimize
the input distribution over this grid. Note that Proposition 1(Section II-A) tells us that an optimal input distribution forour problem must have bounded support, but it does not giveexplicit values that we can use directly in our simulations.However, on employing the cutting-plane algorithm over theinterval [โ10
โ๐ , 10
โ๐ ], we find that the resulting input dis-
tributions have support sets well within this interval. Moreover,increasing the interval length further does not change theseresults.
While the cutting-plane algorithm optimizes the distributionof the channel input, a dual formulation of the channelcapacity problem, involving an optimization over the outputdistribution, can alternately be used to obtain easily com-putable tight upper bounds on the capacity. We discuss theseduality-based upper bounds next.
2) Duality-based upper bound on channel capacity:In the dual formulation of the channel capacity problem, wefocus on the distribution of the output, rather than that of theinput. Specifically, assume a channel with input alphabet ๐ณ ,transition law ๐ (๐ฆโฃ๐ฅ), and an average power constraint ๐ .Then, for every choice of the output distribution ๐ (๐ฆ), wehave the following upper bound on the channel capacity ๐ถ
๐ถ โค ๐(๐ ) = min๐พโฅ0
sup๐ฅโ๐ณ
[๐ท(๐ (โ โฃ๐ฅ)โฃโฃ๐ (โ ))+๐พ(๐ โ๐ฅ2)] , (10)
05
1015
โ5
0
5
0
0.1
0.2
0.3
0.4
0.5
Location
SNR (dB)
Pro
babi
lity
Fig. 2. Probability mass function of the optimal input generated by thecutting-plane algorithm [3] at various SNRs, for the 2-bit symmetric quantizerwith thresholds {โ2, 0, 2}. (noise variance ๐2 = 1.)
TABLE IDUALITY-BASED UPPER BOUNDS ON CHANNEL CAPACITY, COMPARED
WITH THE MUTUAL INFORMATION (MI) ACHIEVED BY THEDISTRIBUTIONS GENERATED USING THE CUTTING-PLANE ALGORITHM.
SNR(๐๐ต) โ5 0 5 10 15 20Upper Bound 0.163 0.406 0.867 1.386 1.513 1.515
๐๐ผ 0.155 0.405 0.867 1.379 1.484 1.484
where ๐พ is a Lagrange parameter, and ๐ท(๐ (โ โฃ๐ฅ)โฃโฃ๐ (โ )) isthe divergence between the transition and output distributions.While [28] provides this bound for a discrete channel, its ex-tension to continuous alphabet channels has been establishedin [29]. For a more detailed perspective on duality-based upperbounds, see [30].
For an arbitrary choice of ๐ (๐ฆ), the bound (10) might bequite loose. Therefore, to obtain a tight upper bound, we mayneed to evaluate (10) for a large number of output distributionsand pick the minimum of the resulting upper bounds. Thiscould be tedious in general, especially if the output alphabetis continuous. However, for the channel model we consider,the output alphabet is discrete with small cardinality. Forexample, for 2-bit quantization, the space of all symmetricoutput distributions is characterized by a single parameter๐ผ โ (0, 0.5). This makes the dual formulation attractive, sincewe can easily obtain a tight upper bound on capacity byevaluating the upper bound in (10) for different choices of ๐ผ.
It remains to specify how to compute the upper bound(10) for a given output distribution ๐ . The structure of thedivergence function ๐ท(๐ (โ โฃ๐ฅ)โฃโฃ๐ (โ )) for our problem can beexploited to develop a systematic procedure for this purpose.We refer the reader to [24] for details of this procedure, andproceed directly to the numerical results.
3) Numerical example: We compare results obtainedusing the cutting-plane algorithm with capacity upper boundsobtained using the dual formulation. We consider 2-bitquantization, and provide results for the specific choice ofquantizer having thresholds at {โ2, 0, 2}.
The input distributions generated by the cutting-plane algo-rithm at various SNRs (setting ๐2 = 1) are shown in Figure 2,and the mutual information achieved by them is given in TableI. As predicted by Theorem 1 (Section II-B), the support setof the input distribution (at each SNR) has cardinality โค 5.
For upper bound computations, we evaluate (10) for differ-
3634 IEEE TRANSACTIONS ON COMMUNICATIONS, VOL. 57, NO. 12, DECEMBER 2009
0 0.5 1 1.5 2 2.5 3 3.5
0.76
0.78
0.8
0.82
0.84
0.86
0.88
x
LHS
of t
he K
KT
con
ditio
n (7
)
Fig. 3. The left-hand side of the KKT condition (7) for the input distributiongenerated by the cutting-plane algorithm (SNR = 5 dB). The KKT conditionis seen to be satisfied (up to the numerical precision of our computations).
ent symmetric output distributions. For 2-bit quantization, theset of symmetric outputs is characterized by just one parameter๐ผ โ (0, 0.5), with the probability distribution on the outputbeing {0.5โ๐ผ, ๐ผ, ๐ผ, 0.5โ๐ผ}. We vary ๐ผ over a fine discretegrid on (0, 0.5), and compute the upper bound for each valueof ๐ผ. The least upper bound achieved thus, at a number ofdifferent SNRs, is shown in Table I. The small gap between theupper bound and the mutual information (at every SNR) showsthe tightness of the obtained upper bounds, and also confirmsthe near-optimality of the input distributions generated by thecutting-plane algorithm.
It is insightful to verify that the preceding near-optimalinput distributions satisfy the KKT condition (7). For instance,consider an SNR of 5 dB, for which the input distributiongenerated by the cutting-plane algorithm has support set{โ2.86,โ0.52, 0.52, 2.86}. Figure 3 plots, as a function of๐ฅ, the LHS of (7) for this input distribution. (The value of ๐พused in the plot was obtained by equating the LHS of (7) tothe capacity value of 0.867, at ๐ฅ = 0.52 .) The KKT conditionis seen to be satisfied (up to the numerical precision of ourcomputations), as the LHS of (7) equals the capacity at pointsin the support set of the input, and is less than the capacityeverywhere else. Note that we show the plot for ๐ฅ โฅ 0 onlybecause it is symmetric in ๐ฅ.
IV. QUANTIZER OPTIMIZATION AND NUMERICAL
RESULTS
Until now, we have addressed the problem of optimizing theinput distribution for a fixed output quantizer. In this section,we optimize over the choice of the quantizer, and presentnumerical results for 2-bit and 3-bit symmetric quantization.
A Simple Benchmark Input-Quantizer Pair: While an op-timal quantizer, along with a corresponding optimal inputdistribution, provides the absolute communication limits forour model, we do not have a simple analytical characteriza-tion of their dependence on SNR. From a system designerโs
0 1 2 3 4 5 6 70
0.5
1
1.5
2
Quantizer threshold โqโ
Cap
acity
(bi
ts /
chan
nel u
se)
โ5 dB
0 dB
3 dB
7 dB
10 dB
15 dB
Fig. 4. 2-bit symmetric quantization : channel capacity (in bits per channeluse) as a function of the quantizer threshold ๐. (noise variance ๐2 = 1.)
perspective, therefore, it is of interest to also examine subop-timal choices that are easy to adapt as a function of SNR,as long as the penalty relative to the optimal solution isnot excessive. Specifically, we take the following input andquantizer pair to be our benchmark strategy : for a ๐พ-binquantizer, consider equiprobable, equispaced ๐พ-PAM (pulseamplitude modulated) input, with quantizer thresholds chosento be the mid-points of the input mass point locations. Thatis, the quantizer thresholds correspond to the ML (maximumlikelihood) hard decision boundaries. Both the input masspoints and the quantizer thresholds have a simple, well-defineddependence on SNR, and can therefore be adapted easily atthe receiver based on the measured SNR. With our K-pointuniform PAM input, we have the entropy ๐ป(๐) = log2 ๐พbits for any SNR. Also, it is easy to see that as SNR โ โ,๐ป(๐ โฃ๐ ) โ 0 for the benchmark input-quantizer pair. Thisimplies that the benchmark scheme is near-optimal if weoperate at high SNR. The issue to investigate therefore is howmuch gain an optimal quantizer and input pair provides overthis benchmark at low to moderate SNR. Note that, for 1-bitsymmetric quantization, the benchmark input corresponds tobinary antipodal signaling, which has already been shown tobe optimal for all SNRs.
As before, we set the noise variance ๐2 = 1 for conve-nience. Of course, the results are scale-invariant, in the sensethat if both ๐ and ๐2 are scaled by the same factor ๐ (thuskeeping the SNR unchanged), then there is an equivalentquantizer (obtained by scaling the thresholds by
โ๐ ) that
gives identical performance.
A. 2-Bit Symmetric Quantization
A 2-bit symmetric quantizer is characterized by a singleparameter ๐, with the quantizer thresholds being {โ๐, 0, ๐}.We therefore employ a brute force search over ๐ to find anoptimal 2-bit symmetric quantizer. In Figure 4, we plot thevariation of the channel capacity (computed using the cutting-plane algorithm) as a function of the parameter ๐ at various
SINGH et al.: ON THE LIMITS OF COMMUNICATION WITH LOW-PRECISION ANALOG-TO-DIGITAL CONVERSION AT THE RECEIVER 3635
SNRs. Based on our simulations, we make the followingobservations:
โ For any SNR, there is an optimal choice of ๐ whichmaximizes capacity. For the benchmark quantizer (whichis optimal at high SNR), ๐ scales as
โSNR, hence it
is not surprising to note that the optimal value of ๐ weobtain increases monotonically with SNR at high SNR.
โ For low SNRs, the variation in the capacity as a functionof ๐ is quite small, whereas the variation becomes appre-ciable as the SNR increases. A practical implication ofthis observation is that imperfections in Automatic GainControl (AGC) have more severe consequences at higherSNRs.
โ For any SNR, as ๐ โ 0 or ๐ โ โ, we approachthe same capacity as with 1-bit symmetric quantization(not shown for ๐ โ โ in the plots for 10 and 15dB in Figure 4). This conforms to intuition: ๐ = 0reduces the 2-bit quantizer to a 1-bit quantizer, while๐ โ โ renders the thresholds at โ๐ and ๐ ineffective indistinguishing between two finite valued inputs, so thatonly the comparison with the quantizer threshold at 0yields useful information.
Comparison with the Benchmark: In Table II, we comparethe performance of the preceding optimal solutions with thebenchmark scheme (see the relevant columns for 2-bit ADC).The corresponding plots are shown in Figure 6. In addition tobeing nearly optimal at high SNR, the benchmark scheme isseen to perform fairly well at low to moderate SNR as well.For instance, even at -10 dB SNR, which might correspond to awideband system designed for very low bandwidth efficiency,it achieves 86% of the capacity achieved with optimal choiceof 2-bit quantizer and input distribution. On the other hand, forSNR of 0 dB or above, the capacity is better than 95% of theoptimal. These results are encouraging from a practical stand-point, given the ease of implementing the benchmark scheme.
Optimal Input Distributions: It is interesting to examinethe optimal input distributions (given by the cutting-planealgorithm) corresponding to the optimal quantizers obtainedabove. Figure 5 shows these distributions, along with optimalquantizer thresholds, for different SNRs. The solid verticallines show the locations of the input distribution points andtheir probabilities, while the quantizer thresholds are depictedby the dashed vertical lines. As expected, binary signalingis found to be optimal for low SNR, since it would bedifficult for the receiver to distinguish between multiple inputpoints located close to each other. The number of mass pointsincreases as SNR is increased, with a new point emergingat 0. On increasing SNR further, we see that the non zeroconstellation points (and also the quantizer thresholds) movefarther apart, resulting in increased capacity. When the SNRbecomes enough that four input points can be disambiguated,the point at 0 disappears, and we get two new points, resultingin a 4-point constellation. The eventual convergence of this 4-point constellation to uniform PAM with mid-point quantizerthresholds (i.e., the benchmark scheme) is to be expected,since the benchmark scheme approaches the capacity bound oftwo bits at high SNR. It is worth noting that the optimal inputswe obtained all have at most four points, even though Theorem
0 4 7 10 15 20
โ10
0
10
0
0.1
0.2
0.3
0.4
0.5
Location
SNR (dB)
Pro
ba
bili
ty
Fig. 5. 2-bit symmetric quantization : optimal input distribution (solid verticallines) and quantizer thresholds (dashed vertical lines) at various SNRs.
1 (Section II-B) is looser, guaranteeing the achievability ofcapacity by at most five points.
B. 3-bit Symmetric Quantization
For 3-bit symmetric quantization, we need to optimize overa space of 3 parameters : {0 < ๐1 < ๐2 < ๐3}, with the quan-tizer thresholds being {0,ยฑ๐1,ยฑ๐2,ยฑ๐3}. Since brute forcesearch is computationally complex, we investigate an alternateiterative optimization procedure for joint optimization of theinput and the quantizer in this case. Specifically, we beginwith an initial quantizer choice ๐1, and then iterate as follows(starting at ๐ = 1)
โ For the quantizer ๐๐, find an optimal input. Call this input๐น๐.
โ For the input ๐น๐, find a locally optimal quantizer, initial-izing the search at ๐๐. Call the resulting quantizer ๐๐+1.
โ Repeat the first two steps with ๐ = ๐+ 1.We terminate the process when the capacity gain betweenconsecutive iterations becomes less than a small threshold ๐.
Although the input-output mutual information is a concavefunctional of the input distribution (for a fixed quantizer), itis not guaranteed to be concave jointly over the input and thequantizer. Hence, the iterative procedure is not guaranteed toprovide an optimal input-quantizer pair in general. A goodchoice of the initial quantizer ๐1 is crucial to enhance thelikelihood that it does converge to an optimal solution. Wediscuss this next.
High SNR Regime: For high SNRs, we know that uniformPAM with mid-point quantizer thresholds (i.e., the benchmarkscheme) is nearly optimal. Hence, this quantizer is a goodchoice for initialization at high SNRs. The results we obtainindeed demonstrate that this initialization works well at highSNRs. This is seen by comparing the results of the iterativeprocedure with the results of a brute force search over thequantizer choice (similar to the 2-bit case considered earlier),as both of them provide almost identical capacity values.
Lower SNRs: For lower SNRs, one possibility is totry different initializations ๐1. However, on trying thebenchmark initialization at some lower SNRs as well, we findthat the iterative procedure still provides us with near-optimal
3636 IEEE TRANSACTIONS ON COMMUNICATIONS, VOL. 57, NO. 12, DECEMBER 2009
TABLE IIPERFORMANCE COMPARISON : FOR 1, 2, AND 3โBIT ADC, THE TABLE SHOWS THE MUTUAL INFORMATION (IN BITS PER CHANNEL USE) ACHIEVED BY
THE OPTIMAL SOLUTIONS, AS WELL AS THE BENCHMARK SOLUTIONS. ALSO SHOWN ARE THE CAPACITY ESTIMATES OBTAINED BY ASSUMING THE
ADDITIVE QUANTIZATION NOISE MODEL (AQNM). (NOTE THAT FOR 1-BIT ADC, THE BENCHMARK SOLUTION COINCIDES WITH THE OPTIMAL
SOLUTION, AND HENCE IS NOT SHOWN SEPARATELY.)
SNR 1-bit ADC 2-bit ADC SNR 3-bit ADC Unquantized(in dB) Optimal AQNM Optimal Benchmark AQNM (in dB) Optimal Benchmark AQNM
-20 0.005 0.007 0.006 0.005 0.007 -20 0.007 0.005 0.007 0.007-10 0.045 0.067 0.061 0.053 0.068 -10 0.067 0.056 0.069 0.069-5 0.135 0.185 0.179 0.166 0.195 -5 0.193 0.177 0.197 0.1980 0.369 0.424 0.455 0.440 0.479 0 0.482 0.471 0.494 0.5003 0.602 0.610 0.693 0.687 0.736 3 0.759 0.744 0.777 0.7915 0.769 0.733 0.889 0.869 0.931 5 0.975 0.955 1.002 1.0297 0.903 0.843 1.098 1.064 1.133 7 1.215 1.180 1.248 1.29410 0.991 0.972 1.473 1.409 1.417 10 1.584 1.533 1.634 1.73012 0.992 1.032 1.703 1.655 1.579 12 1.846 1.766 1.886 2.03715 1.000 1.091 1.930 1.921 1.765 15 2.253 2.138 2.232 2.51417 1.000 1.115 1.987 1.987 1.853 17 2.508 2.423 2.427 2.83820 1.000 1.136 1.999 1.999 1.938 20 2.837 2.808 2.655 3.329
0 5 10 15 200
0.5
1
1.5
2
2.5
3
3.5
4
SNR (dB)
Capacity (
bits/c
hannel use)
1โbit ADC(optimal)
2โbit ADC(optimal and PAM / ML)
3โbit ADC(optimal and PAM / ML)
Unquantized
Fig. 6. Capacity plots for different ADC precisions. For 2 and 3-bit ADC,solid curves correspond to optimal solutions, while dashed curves show theperformance of the benchmark scheme (PAM input with ML quantization).
solutions (again verified by comparing with brute forceoptimization results).
While our results show that the iterative procedure (withbenchmark initialization) has provided (near) optimal solutionsat different SNRs, we leave the question of whether it willconverge to an optimal solution in general as an open problem.
Comparison with the Benchmark: The efficacy of thebenchmark initialization at lower SNRs suggests that theperformance of the benchmark scheme should not be too farfrom optimal at small SNRs as well. This is indeed the case,as seen from the data values in Table II and the correspondingplots in Figure 6. At 0 dB SNR, for instance, the benchmarkscheme achieves 98% of the capacity achievable with anoptimal input-quantizer pair.
Optimal Input Distributions: Although not depicted here, weagain observe (as for the 2-bit case) that the optimal inputsobtained all have at most K points (๐พ = 8 in this case),while Theorem 1 guarantees the achievability of capacity by
at most ๐พ+1 points. Of course, Theorem 1 is applicable to anyquantizer choice (and not just optimal symmetric quantizers).Thus, it is possible that there might exist a ๐พ-bin quantizerfor which the capacity is indeed achieved by exactly ๐พ + 1points. We leave open, therefore, the question of whether ornot the result in Theorem 1 can be tightened to guaranteethe achievability of capacity with at most ๐พ points for theAWGN-QO channel.
C. Comparison with Unquantized Observations
We now compare the capacity results for different quantizerprecisions against the capacity with unquantized observations.Again, the plots are shown in Figure 6 and the data values aregiven in Table II. We observe that at low SNR, the performancedegradation due to low-precision quantization is small. For in-stance, at -5 dB SNR, 1-bit receiver quantization achieves 68%of the capacity achievable without any quantization, whilewith 2-bit quantization, we can get as much as 90% of theunquantized capacity. Even at moderately high SNRs, the lossdue to low-precision quantization remains quite acceptable.For example, 2-bit quantization achieves 85% of the capacityattained using unquantized observations at 10 dB SNR, while3-bit quantization achieves 85% of the unquantized capacity at20 dB SNR. For the specific case of binary antipodal signaling,[7] has earlier shown that a large fraction of the capacity canbe obtained by 2-bit quantization.
On the other hand, if we fix the spectral efficiency to thatattained by an unquantized system at 10 dB (which is 1.73bits/channel use), then 2-bit quantization incurs a loss of 2.30dB (see Table III). For wideband systems, this penalty inpower maybe more significant compared to the 15% loss inspectral efficiency on using 2-bit quantization at 10 dB SNR.This suggests, for example, that in order to weather the impactof low-precision ADC, a moderate reduction in the spectralefficiency might be a better design choice than an increase inthe transmit power.
D. Additive Quantization Noise Model (AQNM)
It is common to model the quantization noise as independentadditive noise [31, pp. 122]. Next, we compare this approx-imation with our exact capacity calculations. In this model
SINGH et al.: ON THE LIMITS OF COMMUNICATION WITH LOW-PRECISION ANALOG-TO-DIGITAL CONVERSION AT THE RECEIVER 3637
TABLE IIISNR (IN DB) REQUIRED TO ACHIEVE A SPECIFIED SPECTRAL EFFICIENCY
WITH DIFFERENT ADC PRECISIONS.
Spectral Efficiency (bits per channel use)0.25 0.5 1.0 1.73 2.5
1-bit ADC โ2.04 1.79 โ โ โ2-bit ADC โ3.32 0.59 6.13 12.30 โ3-bit ADC โ3.67 0.23 5.19 11.04 16.90
Unquantized โ3.83 0.00 4.77 10.00 14.91
๐ = ๐ + ๐ + ๐๐, where the quantization noise ๐๐ isassumed to be uniformly distributed, and independent of ๐,๐ .The signal to quantization noise ratio ๐
๐ผ(๐๐2)
is assumed tobe 6 log2 ๐พ dB for ๐พ-bin quantization [31, pp. 122]. AsSNR โ 0, the distribution of ๐ + ๐๐ approaches that ofa Gaussian, and hence we expect
1
2log
(1 +
๐
๐2 + ๐ผ(๐๐2)
)
to be a good approximation of the capacity at low SNR. TableII shows that this approximation can be useful in terms of pro-viding a quick estimate, although it can either underestimate oroverestimate the actual capacity, depending on the parameters.
V. CONCLUSIONS
Our Shannon-theoretic investigation indicates that the useof low-precision ADC may be a feasible option for designingfuture high-bandwidth communication systems. The availabil-ity of a large amount of bandwidth encourages power-efficientcommunication using small constellations, so that the symbolrate, and hence the sampling rate, for a given bit rate must behigh. This forces us towards using ADCs with lower precision,but fortunately, this is consistent with the use of small constel-lations in the first place for power-efficient design. Thus, if weplan on operating at low to moderate SNR, the small reductionin spectral efficiency due to low-precision ADC is acceptablein such systems, given that the bandwidth is plentiful.
There are several unresolved technical issues that we leaveas open problems. While we show that at most ๐พ + 1 pointsare needed to achieve capacity for ๐พ-bin output quantizationof the AWGN channel, our numerical results reveal that ๐พmass points are sufficient. Can this be proven analytically,at least for symmetric quantizers ? Are symmetric quantizersoptimal ? Does our iterative procedure (with the benchmarkinitialization, or some other judicious initialization) for jointoptimization of the input and the quantizer converge to anoptimal solution in general ?
A technical assumption worth revisiting is that of Nyquistsampling (which induces the discrete-time memorylessAWGN-Quantized Output channel model considered in thiswork). While symbol rate Nyquist sampling is optimal forunquantized systems in which the transmit and receive filtersare square root Nyquist and the channel is ideal, for quan-tized samples, we have obtained numerical results that showthat fractionally spaced samples can actually lead to smallperformance gains. A detailed study quantifying such gains isimportant in understanding the tradeoffs between ADC speedand precision. However, we do not expect oversampling toplay a significant role at low to moderate SNR, given the
small degradation in our Nyquist sampled system relativeto unquantized observations (for which Nyquist sampling isindeed optimal) in these regimes. Of course, oversampling inconjunction with hybrid analog/digital processing (e.g., usingideas analogous to delta-sigma quantization) could producebigger performance gains, but this falls outside the scope ofthe present model.
While our focus in this paper was on non-spread systems, itis known that low-precision ADC is often employed in spreadspectrum systems for low cost implementations [32]. In ourprior examination of Shannon limits for direct sequence spreadspectrum systems with 1-bit ADC [10], we demonstratedthat binary signaling was suboptimal, but did not completelycharacterize an optimal input distribution. The approach in thepresent paper implies that, for a spreading gain ๐บ, a discreteinput distribution with at most ๐บ + 2 points can achievecapacity (although in practice, much smaller constellationswould probably work well).
Finally, we would like to emphasize that the Shannon-theoretic perspective provided in this paper is but a firststep towards the design of communication systems with low-precision ADC. Major technical challenges include the designof ADC-constrained methods for receiver tasks such as carrierand timing synchronization, channel estimation and equaliza-tion, demodulation and decoding.
APPENDIX APROOF OF PROPOSITION 2
Denote by ๐ฎ the set of probability distributions on [๐ด1, ๐ด2]satisfying the average power constraint ๐ . We need to provethat the maximum of ๐ผ(๐ ;๐ ) over ๐ฎ is achievable by adiscrete input with at most ๐พ+1 points. The proof has twoparts: 1) showing that maximum occurs at an extreme point ofa suitably defined constraint set; 2) showing that every extremepoint of this constraint set has at most ๐พ+1 support points.The first part of the proof relies on the fact that ๐ผ(๐ ;๐ ) islinear given the output constraints (as in [5]), while the secondpart is an application of Dubinsโ theorem [6] along the linesof [5].
The existence of an optimal distribution ๐น โ โ ๐ฎ followsfrom standard function analytic arguments [23] (see [24] formore details). Let ๐ โ = [๐โ1, ..., ๐โ๐พ ] be the correspondingoutput distribution. Consider the (convex) set
๐ฐ = {๐น โ ๐ฎโฃ๐ (๐ฆ;๐น ) = ๐ โ and ๐ธ(๐2) = ๐0} , (11)
where ๐0 โค ๐ is the average power of ๐ under ๐น โ. On thisset, ๐ผ(๐ ;๐ ) is a linear functional of the distribution of ๐(that is, it is a linear functional of ๐น ):
๐ผ(๐ ;๐ ) = ๐ป(๐ )โ๐ป(๐ โฃ๐)
= โ๐พโ๐=1
๐๐โ log ๐๐โ +
โซ ๐ด2
๐ด1
๐๐น (๐ฅ)
๐พโ๐=1
๐๐(๐ฅ) log๐๐(๐ฅ).
It follows that ๐ผ(๐ ;๐ ) attains its maximum over ๐ฐ (andhence over ๐ฎ) at an extreme point of ๐ฐ .
To complete the proof we next show that the extreme pointsof ๐ฐ have at most ๐พ + 1 support points. We can view the
3638 IEEE TRANSACTIONS ON COMMUNICATIONS, VOL. 57, NO. 12, DECEMBER 2009
set ๐ฐ as the intersection of the (convex) set of probabilitydistributions over [๐ด1, ๐ด2] (say โณ) and the ๐พ hyperplanes:
๐ป๐ :
โซ ๐ด2
๐ด1
๐๐(๐ฅ)๐๐น (๐ฅ) = ๐๐โ , 1 โค ๐ โค ๐พ โ 1 , (12)
and
๐ป๐พ :
โซ ๐ด2
๐ด1
๐ฅ2๐๐น (๐ฅ) = ๐0 . (13)
Since the extreme points of โณ are precisely the set of unitmasses {๐ฟ๐ฅ, ๐ฅ โ [๐ด1, ๐ด2]}, from Dubinsโ theorem [6] we getthat the extreme points of ๐ฐ have at most ๐พ+1 support points.Remark: The details of verifying that the technical conditionsrequired to apply Dubinsโ theorem are indeed satisfied in thepreceding setting can be found in our technical report [24].
ACKNOWLEDGMENTS
The authors thank an anonymous reviewer for suggestingthe analysis of the additive quantization noise model in SectionIV-D. The first author also thanks Prof. Janos Englander (De-partment of Statistics and Applied Probability, UCSB), Prof.Shiv Chandrasekaran (Department of Electrical and ComputerEngineering, UCSB) and Mr. Vinay Melkote (Signal Compres-sion Laboratory, UCSB) for several useful discussions.
REFERENCES
[1] R. Hiremane, โFrom Mooreโs law to Intel innovation-prediction toreality," Technology Intel Mag., Apr. 2005.
[2] R. Walden, โAnalog-to-digital converter survey and analysis," IEEE J.Sel. Areas Commun., vol. 17, no. 4, pp. 539-550, Apr. 1999.
[3] J. Huang and S. P. Meyn, โCharacterization and computation of optimaldistributions for channel coding," IEEE Trans. Inf. Theory, vol. 51, no.7, pp. 2336-2351, July 2005.
[4] R. G. Gallager, Information Theory and Reliable Communication. NewYork: John Wiley and Sons, 1968.
[5] H. S. Witsenhausen, โSome aspects of convexity useful in informationtheory," IEEE Trans. Inf. Theory, vol. 26, no. 3, pp. 265-271, May 1980.
[6] L. E. Dubins, โOn extreme points of convex sets," J. Math. Anal. Appl.,vol. 5, pp. 237-244, May 1962.
[7] N. Phamdo and F. Alajaji, โSoft-decision demodulation design forCOVQ over white, colored, and ISI Gaussian channels," IEEE Trans.Commun., vol. 48, no. 9, pp. 1499-1506, Sep. 2000.
[8] F. Behnamfar, F. Alajaji, and T. Linder, โChannel-optimized quantizationwith soft-decision demodulation for space-time orthogonal block-codedchannels," IEEE Trans. Signal Process., vol. 54, no. 10, pp. 3935-3946,Oct. 2006.
[9] J. A. Nossek and M. T. Ivrlac, โCapacity and coding for quantizedMIMO systems," in Proc. Intl. Conf. Wireless Commun. Mobile Com-puting (IWCMC), 2006, Vancouver, Canada.
[10] O. Dabeer, J. Singh, and U. Madhow, โOn the limits of communicationperformance with one-bit analog-to-digital conversion," in Proc. IEEEWorkshop Signal Proc. Advances Wireless Commun. (SPAWC), 2006,Cannes, France.
[11] J. Singh, O. Dabeer, and U. Madhow, โCommunication limits with lowprecision analog-to-digital conversion at the receiver," in Proc. IEEEIntl. Conf. Commun. (ICC), 2007, Glasgow, Scotland.
[12] J. Singh, O. Dabeer, and U. Madhow, โCapacity of the discrete-timeAWGN Channel under output quantization," in Proc. IEEE Intl. Symp.Inform. Theory (ISIT), 2008, Toronto, Canada.
[13] L. N. Lee, โOn optimal soft decision demodulation," IEEE Trans. Inf.Theory, vol. 22, pp. 437-444, July 1976.
[14] J. Salz and E. Zehavi, โDecoding under integer metrics constraints,"IEEE Trans. Commun., vol. 43, no. 3, pp. 307-317, Mar. 1995.
[15] E. Masry, โThe reconstruction of analog signals from the sign of theirnoisy samples," IEEE Trans. Inf. Theory, vol. 27, pp. 735-745, Nov.1981.
[16] Z. Cvetkovic and I. Daubechies, โSingle-bit oversampled A/D conver-sion with exponential accuracy in the bit-rate," in Proc. Data Compres-sion Conf. (DCC), 2000, Utah, USA.
[17] P. Ishwar, A. Kumar, and K. Ramachandran, โDistributed sampling fordense sensor networks: a bit-conservation principal," in Proc. Inform.Processing Sensor Networks (IPSN), 2003, Palo Alto, USA.
[18] A. Host-Madsen and P. Handel, โEffects of sampling and quantizationon single-tone frequency estimation," IEEE Trans. Signal Process., vol.48, pp. 650-662, Mar. 2000.
[19] T. Andersson, M. Skoglund, and P. Handel, โFrequency estimation by1-bit quantization and table look-up processing," in Proc. Euro. Sig.Proc. Conf. (EUSIPCO), 2000, Tampere, Finland.
[20] D. Rousseau, G. V. Anand, and F. Chapeau-Blondeau, โNonlinearestimation from quantized signals: quantizer optimization and stochasticresonance," in Proc. Third Intl. Symp. Physics Signal Image Proc.(PSIP), 2003, Grenoble, France.
[21] O. Dabeer and A. Karnik, โSignal parameter estimation using 1-bitdithered quantization," IEEE Trans. Inf. Theory, vol. 52, no. 12, pp.5389-5405, Dec. 2006.
[22] O. Dabeer and E. Masry, โMultivariate signal parameter estimationunder dependent noise from 1-bit dithered quantized data," IEEE Trans.Inf. Theory, vol. 54, no. 4, pp. 1637-1654, Apr. 2008.
[23] M. Fozunbal, S. W. McLaughlin, and R. W. Schafer, โCapacity analysisfor continuous-alphabet channels with side information, part I: a generalframework," IEEE Trans. Inf. Theory, vol. 51, no. 9, pp. 3075-3085, Sep.2005.
[24] J. Singh, O. Dabeer, and U. Madhow, โOn the limits of commu-nication with low-precision analog-to-digital conversion at the re-ceiver," WCSL Technical Report, UCSB. [Online]. Available: โ๐ก๐ก๐ ://๐ค๐ค๐ค.๐๐๐.๐ข๐๐ ๐.๐๐๐ข/๐ค๐๐ ๐/๐๐ข๐๐๐๐๐๐ก๐๐๐๐ .โ๐ก๐๐
[25] J. G. Smith, โOn the information capacity of peak and average powerconstrained Gaussian channels," Ph.D. dissertation, Univ. of California,Berkeley, Dec. 1969.
[26] I. C. Abou-Faycal, M. D. Trott, and S. Shamai, โThe capacity ofdiscrete-time memoryless Rayleigh fading channels," IEEE Trans. Inf.Theory, vol. 47, no. 4, pp. 1290-1301, May 2001.
[27] T. Cover and J. Thomas, Elements of Information Theory. New York:Wiley Series in Telecommunications, 2nd Edition.
[28] I. Csiszar and J. Korner, Information Theory: Coding Theorems forDiscrete Memoryless Systems. Academic Press, 1981.
[29] M. Chiang and S. Boyd, โGeometric programming duals of channelcapacity and rate distortion," IEEE Trans. Inf. Theory, vol. 50, no. 2,pp. 245-258, Feb. 2004.
[30] A. Lapidoth and S. M. Moser, โCapacity bounds via duality withapplications to multiple-antenna systems on flat fading channels," IEEETrans. Inf. Theory, vol. 49, no. 10, pp. 2426-2467, Oct. 2003.
[31] J. G. Proakis, Digital Communications. McGraw Hill, 4th Edition.[32] M. S. Braasch and A. J. van Dierendonck, โGPS receiver architectures
and measurements," Proc. IEEE, vol. 87, no. 1, pp. 48-64, Jan. 1999.
Jaspreet Singh received the B.Tech. degree inelectrical engineering from the Indian Institute ofTechnology, Delhi, in 2004, and the M.S. degreein electrical and computer engineering from theUniversity of California, Santa Barbara (UCSB) in2005. He is currently pursuing the Ph.D. degreein electrical and computer engineering at UCSB.He has been a summer intern at the Swiss FederalInstitute of Technology (EPFL), Lausanne, and atQualcomm Flarion Technologies, New Jersey, dur-ing 2003 and 2008 respectively. He is a recipient of
the National Talent Scholarship (1998-2004) awarded by the Government ofIndia.
Onkar Dabeer (Sโ99-Mโ02) was born in Solapur,India, on September 15, 1974. He received theB.Tech. and M.Tech. degrees in Electrical Engi-neering from the Indian Institute of Technology,Bombay, in 1996 and 1998 respectively, and, thePh.D. degree in Electrical Engineering from theUniversity of California at San Diego in June 2002.He was a postdoctoral researcher at University ofCalifornia, Santa Barbara from July 2002 to July2003. From August 2003 to August 2004, he servedas a Senior Engineer at Qualcomm Inc., San Diego.
Since September 2004 he has been on the faculty at the School of Technologyand Computer Science, Tata Institute of Fundamental Research, Mumbai(Bombay), India. His research interests include wireless communicationnetworks, sensor networks, low-precision signal processing, and informationtheory.
SINGH et al.: ON THE LIMITS OF COMMUNICATION WITH LOW-PRECISION ANALOG-TO-DIGITAL CONVERSION AT THE RECEIVER 3639
Upamanyu Madhow received his bachelorโs degreein electrical engineering from the Indian Institute ofTechnology, Kanpur, in 1985. He received the M. S.and Ph. D. degrees in electrical engineering from theUniversity of Illinois, Urbana-Champaign in 1987and 1990, respectively.
From 1990 to 1991, he was a Visiting AssistantProfessor at the University of Illinois. From 1991 to1994, he was a research scientist at Bell Commu-nications Research, Morristown, NJ. From 1994 to1999, he was on the faculty of the Department of
Electrical and Computer Engineering at the University of Illinois, Urbana-
Champaign. Since December 1999, he has been with the Department ofElectrical and Computer Engineering at the University of California, SantaBarbara, where he is currently a Professor. His research interests are incommunication systems and networking, with current emphasis on wirelesscommunication, sensor networks and multimedia security.
Dr. Madhow is a recipient of the NSF CAREER award. He has servedas Associate Editor for the IEEE TRANSACTIONS ON COMMUNICATIONS,IEEE TRANSACTIONS ON INFORMATION THEORY, and IEEE TRANSAC-TIONS ON INFORMATION FORENSICS AND SECURITY. He is the author ofthe graduate textbook, Fundamentals of Digital Communication, published byCambridge University Press.