Smoothing Codes and Lattices:
Systematic Study and New Bounds
Abstract.
In this article we revisit smoothing bounds in parallel between lattices and codes. Initially introduced by Micciancio and Regev, these bounds were instantiated with Gaussian distributions and were crucial for arguing the security of many lattice-based cryptosystems. Unencumbered by direct application concerns, we provide a systematic study of how these bounds are obtained for both lattices and codes, transferring techniques between both areas. We also consider multiple choices of spherically symmetric noise distribution.
We found that the best strategy for a worst-case bound combines Parseval’s Identity, the Cauchy-Schwarz inequality, and the second linear programming bound, and this holds for both codes and lattices and all noise distributions at hand. For an average-case analysis, the linear programming bound can be replaced by a tight average count.
This alone gives optimal results for spherically uniform noise over random codes and random lattices. This also improves previous Gaussian smoothing bound for worst-case lattices, but surprisingly this provides even better results with uniform ball noise than for Gaussian (or Bernoulli noise for codes).
This counter-intuitive situation can be resolved by adequate decomposition and truncation of Gaussian and Bernoulli distributions into a superposition of uniform noise, giving further improvement for those cases, and putting them on par with the uniform cases.
1. Introduction
1.1. Smoothing bounds.
In either a code or a lattice, smoothing refers to fact that, as an error distribution grows wider and wider, the associated syndrome distribution tends towards a uniform distribution. In other words, the error distribution, reduced modulo the code or the lattice, becomes essentially flat. This phenomenon is pivotal in arguing security of cryptosystems [MR07, GPV08, DST19]. In information theoretic literature, it is also sometimes referred to as flatness [LLBS14]. Informally, by a “smoothing bound” we are referring to a result which lower bounds the amount of noise which needs to be added so that the smoothed distribution “looks” flat.
To be more concrete, by a “flat distribution”, we are referring to a uniform distribution over the ambient space modulo the group of interest. For a (linear) code , this quotient space is ; for a lattice , it is . We then consider some “noise” vector distributed over the ambient space (resp. ), and attempt to prove that (resp. ) is “close” to the uniform distribution over the quotient space (resp. ). To quantify “closeness” between distributions, we will use the standard choice of statistical distance.
An important question to be addressed is the choice of distribution for the noise vector . In lattice-based cryptography (where such smoothing bounds originated [MR07]), the literature ubiquitously uses Gaussian distributions for errors, and smoothness is guaranteed for an error growing as the inverse of the minimum distance of the dual lattice. The original chain [MR07] of argument goes as follows:
-
•
Apply the Poisson summation formula (PSF);
-
•
Bound variations via the triangle inequality (TI) over all non-zero dual lattice points;
-
•
Bound the absolute sum above via the Banaszczyk tail bound [Ban93] for discrete Gaussian (BT).
An intermediate quantity called the smoothing parameter introduced by [MR07] before the last step is also often used in the lattice-based cryptographic literature. Each bounding step is potentially non-tight, and indeed more recent works have replaced the last step by the following [ADRS15]:
-
•
Bound the number of lattice points in balls of a given radius via the Linear Programming bound [Lev79] (LP) and “sum over all radii” (with care).
With this LP strategy, it is in principle possible to also compute a smoothing bound for spherically symmetric distributions of errors other than the Gaussian; however, we are not aware of prior work doing this explicitly. A very natural choice would be uniform distributions over Euclidean balls.
For codes, there are also two natural distributions of errors: Bernoulli noise, i.e. flip each bit independently with some probability (a.k.a. the binary symmetric channel), and a uniform noise over a Hamming sphere of a fixed radius. The latter is typically preferred for the design of concrete and practical cryptosystems [McE78, Ale11, MTSB13, DST19], while the former appears more convenient in theoretical works (1)(1)(1)A third choice of distribution, described as a discrete-time random walk, also made an appearance for a complexity theoretic result [BLVW19]. The expert reader may note that the Bernoulli distribution can also be treated as a continuous-time random walk, and both can be analysed via the heat kernel formalism [Chu97, Chap. 10].. Cryptographic interest for code smoothing has recently arisen [BLVW19, YZ21], but results are so far limited to codes with extreme parameters and specific “balancedness” constraints. However we note that the question is not entirely new in the coding literature (see for instance [Klø07]). In particular, an understanding of the smoothing properties of Bernoulli noise is intimately connected to the undetected error probability of a code transmitted through the .
In this light, it is interesting to revisit and systematize our understanding of smoothing bounds, unencumbered by direct application concerns. We find it enlightening to do this exploration in parallel between codes and lattices, transferring techniques back and forth between both areas whenever possible.
Furthermore, we keep our arguments agnostic to the specific choice of error distribution, allowing us to apply them with different error distributions and compare the results. To compare different (symmetric) distributions, we advocate parametrizing them by the expected weight/norm of a vector. That is, we quantify the magnitude of a noise vector by (where denotes either the Hamming weight or the Euclidean norm of the vector). Our smoothing bounds will depend on this parameter, and we consider a smoothing bound to be more effective if for the smoothed distribution to be close to uniform we require a smaller lower-bound on .
1.2. Contributions.
In this work, we collect the techniques that have been used for smoothing, both in the code and lattice contexts. We view individual steps as modular components of arguments, and consider all permissible combinations of steps, thereby determining the most effective arguments. In the following, we outline our systematization efforts, describing the various proof frameworks that we tried before settling on the most effective argument.
Code smoothing bounds.
Given the relative dearth of results concerning code smoothing, it seems natural to start by adapting the first argument (PSF+TI+BT) to codes following the proof techniques of [Ban93, MR07]. And indeed, the whole strategy translates flawlessly, with only one caveat: it leads to a very poor result, barely better than the trivial bound. Namely, smoothness is established only for Bernoulli errors with parameter very close to .
The adaptation of Banaszczyk tail bound [Ban93] to codes (together with replacing the Gaussian by a Bernoulli distribution) is rather naïve, and it is therefore not very surprising that it leads to a disappointing result. Instead, we can also follow the improved strategy for lattices from [ADRS15], and resort to linear programming bounds for codes [Bas65, MRJW77, ABL01]. Briefly, by an LP bound we are referring to a result that bounds the number of codewords (resp. lattice vectors) of a certain weight (resp. norm) in terms of the dual distance (resp. shortest dual vector) of the code (resp. lattice). In both cases, the results are obtained by considering a certain LP relaxation of the combinatorial quantities one wishes to bound, hence the name. Even more, the bounds for codes and lattices are obtained via essentially the same arguments [MRJW77, DL98, CE03]. We therefore find it natural to apply LP bounds in our effort to develop proof techniques which apply to both code- and lattice-smoothing.
The strategy (PSF+TI+LP) turns out to give a significantly better result, but it nevertheless still appears to be far from optimal. We believe that the application of the triangle inequality in the second step to bound the sum of Fourier coefficients given by the Poisson summation formula leads to the unsatisfactory bound. Indeed, a common heuristic when dealing with sums of Fourier coefficients is that, unless there is a good reason otherwise, the sum should have magnitude roughly the square-root of the order of the group (as is the case for random signs): the triangle inequality is far too crude to notice this.
Instead, we turn to another common upper-bound on a sum, namely, the Cauchy-Schwarz (CS) inequality. It is natural to subsequently apply Parseval’s Identity (PI). It turns out that this strategy yields very promising results, upon which we now elucidate. The upper-bound is described in terms of the weight distribution of a code, i.e. the number of codewords of weight for each . Unfortunately, it is quite difficult to understand the weight distribution of arbitrary codes, and the bounds that we do have are quite technical.
Random codes.
For this reason, we first apply our proof template to random codes, as it is quite simple to compute the (expected) weight distribution of a random code. Quite satisfyingly, the simple two steps arguments (PI+CS) already yields optimal results for this case, but when the error is sampled uniformly at random from a sphere! That is, we can show that the support size of the error distribution matches the obvious lower bound that applies to any distribution that successfully smooths a code: namely, for a code the support size must be at least . Using coding-theoretic terminology, the weight of the error vector that we need to smooth is given by the ubiquitous Gilbert-Varshamov bound
which characterizes the trade-off between a random code’s rate and its minimum distance. Here, is the inverse of the binary entropy function.
Moreover, as the argument is versatile enough to apply to essentially all spherical error distributions, we also tried applying it to the Bernoulli distribution, and the random walk distribution of [BLVW19]. Comparing them, we were rather surprised that our argument provided better bounds for the uniform distribution over a Hamming sphere than the other two distributions for the same average Hamming weight.
However, while the (PI+CS) sequence of arguments is more effective when the noise is sampled uniformly on the sphere, we can exploit the fact that the Hamming weight of a Bernoulli-distributed vector is tightly concentrated to recover the same smoothing bound for this distribution. In more detail, we use a “truncated” argument. First, we decompose the Bernoulli distribution into a convex combination of uniform sphere distributions. But, by Chernoff’s bound, a Bernoulli distribution is concentrated on vectors whose weight lies in a width interval around its expected weight. Therefore, outside of this interval, the contribution of the Bernoulli on the statistical distance is negligible. Then apply the (PI+CS) sequence of arguments to each constituent distribution close to the expected weight. In this way, we are able to demonstrate that Bernoulli distributions also optimally smooth random codes.
Arbitrary codes.
Next, we turn our attention to smoothing worst-case codes. Motivated by our success in smoothing random codes, we again follow the (PI+CS) sequence of arguments and combine this with LP bounds to derive smoothing bounds when the dual distance of the code is sufficiently large. Again, the sequence of arguments is most effective when the error is distributed uniformly over the sphere, with one caveat: we are also required to assume that the dual code is balanced in the sense that it also does not contain any vectors of too large weight. While this assumption has appeared in other works [BLVW19, YZ21], we find it somewhat unsatisfactory.
Fortunately, this condition is not required if the error is sampled according to the Bernoulli distribution. But then we run into the same issue that we had earlier with random codes: the (PI+CS) argument, followed by LP bounds, natively yields a lesser result when instantiated with Bernoulli noise. Fortunately, we have already seen how to resolve this issue: we pass to the truncated Bernoulli distribution and decompose it into uniform sphere distributions. This yields a best-of-both-worlds result: we obtain the strongest smoothing bound we can in terms of the noise magnitude, while requiring the weakest assumption on the code.
And back to lattices.
Random lattices.
First, just as we set our expectations for code-smoothing by first studying the random case, we analogously start here by considering random lattices. However, defining a random lattice is a non-trivial task. We actually consider two distributions. The first, which is based on the deep Minkowski-Hlwaka-Siegel (MHS) Theorem, we only abstractly describe. Thanks to the MHS Theorem, we can very easily compute the (expected value) of our upper-bound.
For the MHS distribution of lattices, we consider two natural error distributions: the Gaussian distribution (which is used ubiquitously in the literature), as well as the uniform distribution over the Euclidean ball. And again, perhaps surprisingly (although less so now thanks to our experience with the code case), we obtain a better result with the uniform distribution over the Euclidean ball. And moreover, the Euclidean ball result is optimal in the same sense that we had for codes: the support volume of the error distribution is exactly equal to the covolume of the lattice (2)(2)(2)That is, for a lattice , the volume of the torus . We will denote this quantity by from now on. . We view the value such that the volume of the -ball of radius is equal to the covolume of a lattice (which is half the quantity that appears in Minkowski bound) as being the lattice-theoretic analogue of the Gilbert-Varshamov quantity:
However, as Gaussian vectors satisfy many pleasing properties that are often exploited in lattice-theoretic literature, we would like to obtain the same smoothing bound for this error distribution. Fortunately, our experience with codes also tells us how to recover the result for Gaussian noise from the Euclidean ball noise smoothing bound: we decompose the Gaussian distribution appropriately into a convex combination of Euclidean ball distributions. Together with a basic tail bound, we recover the same smoothing bound for Gaussian noise that we had for the uniform ball noise.
We also study random -ary lattices, which are more concretely defined: following the traditional lattice-theoretic terminology, they are obtained by applying Construction A to a random code. This does lead to a slight increase in the technicality of the argument – in particular, we need to apply a certain “summing over annuli” trick – but the computations are still relatively elementary. Again, we find that the argument naturally works better when the errors are distributed uniformly over a ball, but we can still transfer the bound to the Gaussian noise.
Interestingly, the same optimal bound has been recovered in a concurrent work [LLB22, Theorem 1.] for Gaussian distributions. Their arguments are quite unlike ours: [LLB22] uses the Kullback–Leibler divergence in combination with other information-theoretic arguments. However, contrary to our bounds obtained via the (PI + CS) sequence of arguments, [LLB22, Theorem 1] only holds for random -ary lattices.
Arbitrary lattices.
Next, we address the challenge of smoothing arbitrary lattices. And again, we follow the (PI+CS) sequence of arguments, and subsequently use the Kabatiansky and Levenshtein bound [KL78] to obtain a smoothing bound in terms of the minimum distance of the dual lattice. The Kabatiansky and Levenshtein bound is the lattice-analogue of the second LP bound from coding theory. We can directly apply the arguments with both of our error distributions of interest, and again, the uniform ball distribution wins. But the decomposition and tail-bound trick again applies to yield the same result for the Gaussian distribution that we had for the uniform ball distribution.
Comparison.
We summarize how our work improves on the state of the art in Table 1 for lattices, and in Table 2 and Figure 1 for codes. For this discussion, we let (resp. )) denote the uniform distribution over (resp. , and let denote the statistical distance.
In the case of lattices (Table 1), we fix the smoothing bound target to exponentially small, that is we state the minimal value of such that the bound over the statistical distance implies when the error follows the prescribed distribution and of an average Euclidean length of .(3)(3)(3)In fact, the values in this table guarantee exponentially small statistical distance from the uniform distribution.
Distribution | Proof strategy | smoothing factor | General statement |
---|---|---|---|
Gaussian | PSF+TI+BT | Lemma 3.2 [MR07] | |
Gaussian | PSF+TI+LP | Lemma 6.1 [ADRS15] | |
Gaussian | PI+CS+LP | Theorem 4.18 (this work) | |
Unif. Euclidean ball | PI+CS+LP | Theorem 4.17 (this work) | |
Gaussian | via Unif. + Trunc. | Theorem 4.19 (this work) |
In the case of codes we also fix the smoothing bound target to negligible,(4)(4)(4)Again, it is the same if we insist the statistical distance to uniform is exponentially small. but we compare two cases: smoothing bounds for random codes (in average) and for a fixed code (worst case). In Figure 1 we compare the minimal value such that when the error follows the prescribed distribution and with an expectation that is taken over codes of rate . In Table 1 we make the same comparison but to reach for a fixed code such that the minimum distance of its dual is known.

Distribution | smoothing factor | Balanced-code | General statement |
---|---|---|---|
Bernoulli | NO | Eq. (17), Prop. 3.11, 3.12 | |
Discrete Rand. Walk | YES | Theorem 3.14 | |
Unif. Hamming sphere | YES | Theorem 3.14 | |
Bernoulli + Trunc. | NO | Theorem 3.16 |
2. Preliminaries: Notations and Fourier Analysis over Locally Compact Abelian Group
2.1. General Notation.
The notation means that is defined as being equal to . Given a set , its indicator function will be denoted . For a finite set , we will denote by its cardinality. Vectors will be written with bold letters (such as ). Furthermore, we denotes by the set of integers .
The statistical distance between two discrete probability distributions and over a same space is defined as:
Similarly, for two continuous probability density functions and over a same measure space , the statistical distance is defined as
2.2. Codes and Lattices
We give here some basic definitions and notation about linear codes and lattices.
Linear codes. In the whole paper, we will deal exclusively with binary linear codes, namely subspaces of for some positive integer . The space will be embedded with the Hamming weight , namely
We will denote by the sphere with center and radius ; its size is given by and we have where denotes the binary-entropy, namely .
An -code is defined as a dimension subspace of . The rate of is . Its minimal distance is given by
The number of codewords of of weight will be denoted by , namely
The dual of a code is defined as where denotes the standard inner product on .
Lattices. We will consider lattices of which is embedded with the Euclidean norm , namely
We will denote by the ball with center and radius ; its volume is given by
An -dimension lattice is defined as a discrete subgroup of . The covolume of is the volume of any fundamental parallelotope. The minimal distance of is given by The number of lattice points of of weight will be denoted by , namely
2.3. Fourier Analysis
We give here a brief introduction to Fourier analysis over arbitrary locally compact Abelian groups. Our general treatment will allow us to apply directly some basic results in a code and lattice context, obviating the need in each case to introduce essentially the same definitions and to provide the same proofs.
Corollary 2.4 at the end of this subsection is the starting point of our smoothing bounds: all of our results are obtained by using different facts
to bound the right hand side of the inequality.
Groups and Their Duals. In what follows will denote a locally compact Abelian group. Such a group admits a Haar measure . For instance with the Lebesgue measure , or with the counting measure .
The dual group is given by the continuous group homomorphisms from into the multiplicative group of complex numbers of absolute value , and it is again a locally compact Abelian group. In Figure 2 we give groups, their duals as well as their associated Haar measures that will be considered in this work.
.
It is important to note that if is a closed subgroup, then and are also locally compact groups. Furthermore, has a dual group that satisfies the following isomorphism
Norms and Fourier Transforms. For any , will denote the space of measurable functions (up to functions which agree almost everywhere) with finite norm which is defined as
The Fourier transform of is defined as
We omitted here the dependence on . It will be clear from the context.
Theorem 2.1 (Parseval’s Identity).
Let , then with appropriate normalization of the Haar measure
Poisson Formula. Given and any function , its restriction over is defined as . We define its periodization as follows.
Definition 2.2 (Periodization).
Let be a closed subgroup of and . We define the -periodization of as
where denotes any choice of the Haar measure for .
There always exists a Haar measure such that for any continuous function with compact support the quotient integral formula holds
(1) |
Theorem 2.3 (Poisson Formula).
Let be a closed subgroup and , then with appropriate normalization of the Haar measures,
The following corollary is a simple consequence of the Cauchy-Schwarz inequality, Parseval identity and the Poisson formula. Our results on smoothing bounds are all based on this corollary.
Corollary 2.4.
Let be a closed subgroup of . Let and such that . Then with appropriate normalization of the Haar measure,(5)(5)(5)We choose the Haar measures , , and for which both the Poisson formula and Parseval’s Identity hold.
where denotes the identity element of .
Proof.
We have
(2) | ||||
where in Equation (2) we used the following equalities:
and
which concludes the proof. ∎
In this work we will choose and or and . Haar measures associated to and for which the corollary holds are given in Figure 2. Furthermore, we will use Fourier transforms over and . We describe in Figure 3 these dual groups that we will consider.
3. Smoothing Bounds: Code Case
Given a binary linear code of length , the aim of a smoothing bound is to quantify at which condition on the noise is statistically close to the uniform distribution over when is uniformly drawn from and sampled according to some noise distribution . Equivalently, we want to understand when is close to the uniform distribution. We will focus on the case where the distribution of is radial, meaning that it only depends on the Hamming weight of .
Notation 3.1.
We will use throughout this section the following notation.
-
•
The uniform probability distribution over the quotient space will frequently recur and for this reason we just denote it by . The uniform distribution over the whole space is denoted by and the uniform distribution over the codewords of is denoted by .
-
•
We also use the uniform distribution over the sphere which we denote by .
-
•
For two probability distributions and over we denote by the convolution over : .
It will be more convenient to work in the quotient space and for this we use the following proposition.
Proposition 3.2.
Let be a probability distribution over and be an -code. We have
Proof.
Let and be distributed according to and . We have the following computation:
(3) | ||||
where in Equation (3) we used that each term of the sum is constant on . ∎
As a rewriting of Corollary 2.4 we get the following proposition that upper-bounds , namely:
Proposition 3.3.
Let be an -code and be a radial distribution on . We have
where by abuse of notation we denote by the common value of on vectors of weight .
Proof.
We have that is a closed subgroup of with associated Haar measures:
for which we can apply Corollary 2.4. Let and . First, it is clear that and that
where we used that is a distribution. Therefore we can apply Corollary 2.4 with functions and . Furthermore, by definition of . We get the following computation:
(4) |
To conclude the proof it remains to apply Corollary 2.4 with Equation (4) and then to use that is radial and therefore also . ∎
Our upper-bound of Proposition 3.3 involves the weight distribution of the code , namely . To understand how our bound behaves for a given distribution , we will start (in the following subsection) with the case of random codes. The expected value for is well known in this case. This will lead us to estimate our bound on almost all codes and gives us some hints about the best distribution to choose for our smoothing bound in the worst case (which is the case that we treat in Subsection 3.2).
3.1. Smoothing Random Codes.
The probabilistic model that we use for our random code of length is defined by sampling uniformly at random a generator matrix for it, i.e.
It is straightforward to check that the expected number of codewords of weight in the dual is given by:
Fact 3.4.
For chosen according to
This estimation combined with Proposition 3.3 enables us to upper-bound .
Proposition 3.5.
We have:
(5) |
Proof.
It remains now to choose the distribution . A natural choice in code-based cryptography is the uniform distribution over the sphere of radius centered around .
Uniform Distribution over a Sphere. The Fourier transform of is intimately connected to Krawtchouk polynomials. The Krawtchouk polynomial of order and degree is defined as
To simplify notation, since is clear here from context, we will drop the dependency on and simply write . The following fact allows to relate with (see for instance [vL99, Lem. 3.5.1, §3.5])
Fact 3.6.
For any ,
(6) |
The above sum can be upper-bounded by observing that is an orthonormal basis of functions for the inner product . It can be viewed as the standard inner product between radial functions over . In particular, [Lev95, Corollary 2.3]. Therefore, for random codes we obtain the following proposition
Proposition 3.7.
We have for random chosen according to
(8) |
In other words, if one wants to smooth a random code with target distance via the uniform distribution over a sphere, one has to choose its radius such that . It is readily seen that for fixed code rate , choosing any fixed ratio such that is enough, where corresponds to the asymptotic relative Gilbert-Varshamov (GV) bound
with being the inverse of the binary entropy function . The GV bound appears ubiquitously in the coding-theoretic literature: amongst other contexts, it arises as the (expected) relative minimum distance of a random code of dimension , or as the maximum relative minimum error weight for which decoding over the binary symmetric channel can be successful with non-vanishing error probability.
This value of radius is optimal: clearly, the support size of an error distribution smoothing a code must exceed . Thus, we cannot expect to smooth a code with errors in the sphere if its volume is smaller than .
Therefore the uniform distribution over a sphere is optimal for random codes. By this, we mean that it leads to the smallest amount of possible noise (when it is concentrated on a ball) to smooth a random code. Notice that we obtained this result after applying the chain of arguments Cauchy-Schwarz, Parseval and Poisson to bound the statistical distance.
About the original chain of arguments of Micciancio and Regev. It can be verified that by coming back to the original steps of [MR07, ADRS15], namely the Poisson summation formula and then the triangle inequality, we would obtain
(9) |
By using that (when ) we see that our bound (Proposition 3.3) is sharper. It turns out that our bound is exponentially sharper for random codes (and even in the worst case) when choosing as the uniform distribution over a sphere of radius , namely . In this case the Micciancio-Regev argument yields the following computation
(10) |
To carefully estimate this upper-bound (and to compare with (8)) we are going to use the following proposition, which gives the asymptotic behaviour of (see for instance [IS98, DT17]).
Proposition 3.8.
Let and be three positive integers. We set , and . We assume . Let where . In the case ,
In the case , is negative, and
We let,
In Figure 4 we compare the asymptotic values of and as functions of . Notice that . We see that is undefined for a rate . In other words, it is impossible to show that with the standard approach of [MR07, ADRS15] when . Furthermore, for larger rates (and sufficiently large ), is much smaller than .

Bernoulli Distribution. Another natural distribution to consider when dealing with codes is the so-called “Bernoulli” distribution , which is defined for as
This choice leads to simpler computations compared to the uniform distribution over a sphere. For instance we have . By plugging this in Equation (5) of Proposition 3.5 we obtain
(11) |
Thus, if one wants to smooth a random code at target distance with the Bernoulli distribution, the above argument says that one has to choose where . As , it is meaningful to compare and . It is readily seen that . In other words, this time the upper-bound given by Proposition 3.5 does not give what would be optimal, namely the Gilbert-Varshamov relative distance , but a quantity which is bigger. However, it is expected that the average amount of noise to smooth a random code is the same in both cases, since a Bernoulli distribution of parameter is extremely concentrated over words of Hamming weight and that therefore . This suggests that Proposition 3.5 is not tight in this case. This is indeed the case, we can prove that we can smooth a random code with the Bernoulli noise as soon as . This follows from the following proposition.
Proposition 3.9.
Let and . Then,
Proof.
See Appendix A. ∎
This proposition shows that if one wants it is enough to have for any . This can be achieved by choosing and such that .
To summarize this subsection we have the following theorem
Theorem 3.10.
Let be a random code chosen according to , . Let (resp. ) be the uniform distribution over (resp. ) and be the Bernoulli distribution over of parameter . We have,
In particular, and for any fixed .
3.2. Smoothing a Fixed Code.
Our upper-bound on given in Proposition 3.3 involves the weight distribution of the dual of , namely the ’s. To derive smoothing bounds on a fixed code our strategy will simply consist in using the best known upper bounds on the ’s. Roughly speaking, these bounds show that for some constant which is function of .
Notation. Let and ,
(12) |
where the maximum is taken over all codes of length and minimum distance .
We recall (or slightly extend) results taken from [ABL01]:
Proposition 3.11.
Proof.
See Appendix B. ∎
Proposition 3.12 ([ABL01, Proposition 4]).
Let and
where
For any
(14) |
Both of these bounds are derived from “linear programming arguments” which were initially used to upper-bound the size of a code given its minimum distance. Proposition 3.11 is an extension of [ABL01, Theorem 3] in the case of linear codes, in particular we give an upper-bound for any (and not for only ). The proof is in the appendix. The second bound is usually called the the second linear programming bound. In terms of and , Proposition 3.11 and 3.12 are among the best (known) upper-bounds on . In the case where , Proposition 3.12 leads to better smoothing bounds compared to Proposition 3.11.
Remark 3.13.
There exist many other bounds on , like [ACKL05, Theorem 8] which holds only for linear codes or [ACKL05, Theorem 7]. However for our smoothing bounds, Propositions 3.11 and 3.12 lead to the best results, partly because these are the best bounds on the number of codewords of Hamming weight close to the minimum distance of the code.
We draw in Figures 5 and 6 the bounds of Propositions 3.11 and 3.12 as function of for a couple values of .


Equipped with these bounds we are ready to give our smoothing bounds for codes in the worst case, namely for a fixed code. Our study with random codes gave a hint that the choice of the uniform distribution over a sphere could give better results than the Bernoulli distribution. However, as we will show now, the distribution on a sphere forces us to assume that no codewords of large weight belong to the dual when we want to smooth . It corresponds to the hypothesis of balanced-codes made in [BLVW19] to obtain a worst-to-average case reduction. We would like to avoid making this assumption as nothing forbids large weight vectors from belonging to a fixed code. Fortunately, as we will later show, we can avoid making this hypothesis while still keeping the advantages of the uniform distribution over a sphere.
Impossibility to smooth a code whose dual is not balanced with the uniform distribution over a sphere.
It is readily seen that in the case where the dual code is not balanced, meaning that it contains the all-one vector (and therefore that the dual weight distribution is symmetric: for any when the codelength is ), then it is impossible to smooth it with the uniform distribution over a sphere. Indeed, this implies that all codewords of have an even Hamming weight (they have to be orthogonal to the all-one vector). The parity of the Hamming weights of vectors in a coset (i.e. in the class of representatives of some element in ) will be the same. Therefore, half of the cosets cannot be reached when periodizing over .
Difficulty of using Proposition 3.3 for proving smoothness of the uniform distribution if the dual has large weight codewords. Even in the case where the dual is balanced, difficulties can arise if we want to use Proposition 3.3 for proving smoothness of the uniform distribution over a sphere when the dual has large weight codewords. First of all, the fact that it contains
the all-one codeword also reflects in the upper-bound of Proposition 3.3. Recall that and that we have (see Fact 3.6). Therefore, when the full weight vector belongs to , our upper-bound on of Proposition 3.3 cannot be smaller than . Furthermore, even if the dual does not contain the all-one codeword, codewords of weight say also give a non-negligible contribution to the upper-bound of Proposition 3.3: the contribution is a polynomial .
Difficulty of using Proposition 3.3 for proving smoothness of the “discrete walk distribution” if the dual has large weight codewords. Other meaningful distributions in the cryptographic context display the same problem as the uniform distribution concerning the difficulty of applying Proposition 3.3 to them if the dual contains large weight codewords. This applies to the discrete time random walk distribution introduced in [BLVW19] for worst-to-average case reductions. The authors were only able to prove smoothness of this distribution if the dual code has no small and no large weight codewords. This distribution is given by
where the ’s are independently and uniformly drawn at random in and denotes the -th canonical basis vector. Recall that [BLVW19]
Therefore, when , as for the Fourier transform of the uniform distribution over a sphere, showing that cannot smooth a code when the full weight vector belongs to its dual. In summary, a direct application of Proposition 3.3 is quite unsatisfactory for these distributions and . If we are willing to also make an assumption on the largest weight of a codeword, then certainly a direct application of Proposition 3.3 is able to provide meaningful smoothing bounds for them. Indeed, the following theorem is obtained by just combining Propositions 3.3, 3.11 and 3.12.
Theorem 3.14.
Let be a binary linear code of length and . Suppose that and that has no element of Hamming weight for some . We have
Avoiding making an assumption on the largest dual codeword: the case of the Bernoulli distribution. Even if the Bernoulli distribution has some drawbacks compared to the uniform distribution over a sphere, when applying Proposition 3.3 with random codes, it has however a nice property concerning the large weight codewords: the large weight dual codewords have a negligible contribution in the upper-bound of Proposition 3.3. To see this let us first recall that
(15) |
Therefore, by Proposition 3.3 we have
(16) |
On the other hand, we have the following lemma which shows that large weight codewords can only have an exponentially small contribution to the above upper-bound.
Lemma 3.15.
Let be a linear code of length and let . There is at most one codeword of weight .
Proof.
Suppose by contradiction that there exists two distinct codewords of Hamming weight . By using the triangle inequality we obtain (where denotes the all-one vector)
which contradicts the fact that has minimum distance . ∎
In other words, large weight dual codewords (if they exist) have only an exponentially small contribution to our smoothing bound with the Bernoulli distribution. In principle, we could plug in Equation (17) bounds on the ’s given in Propositions 3.11 and 3.12.
We will improve on the bounds obtained in this way by truncating the Bernoulli distribution, then
-
prove that by appropriately truncating both distributions have the same smoothness property,
-
show that the truncated distribution has the same nice properties with respect to large weights,
-
show that we can apply Proposition 3.3 to the truncated distribution and get appropriate smoothness properties.
We obtain in this way:
Theorem 3.16.
Proof.
See Appendix C. ∎
Let and be the smallest that enables to reach with
-
•
Theorem 3.16 when ,
- •
In Figure 7 we compare the smallest that enables one to reach with Equation (17) and with Theorem 3.16. As we can see Theorem 3.16 leads to significantly better bounds. Furthermore, it turns out that is roughly equal to the smallest radius such that if we had supposed that no codewords of weight belong to . In other words, our proof using the tweak of truncating the Bernoulli enables us to obtain a smoothing bound without the hypothesis of no dual codewords of large Hamming weight which is as good as with the uniform distribution over a sphere if we had made this assumption.

4. Smoothing Bounds: Lattice Case
Given an -dimensional lattice the aim of smoothing bounds is to give a non-trivial model of noise for (namely the reduction of modulo ) to be uniformly distributed. Following Micciancio and Regev [MR07], the standard choice of noise is given by the Gaussian distribution, defined via
The parametrization is chosen such that is the standard deviation of . Micciancio and Regev showed that when is distributed according to , choosing large enough enables to be statistically close to the uniform distribution.
However, following the intuition from the case of codes we will first analyze the case where is sampled uniformly from a Euclidean ball. Interestingly, just as with codes where our methodology led to stronger bounds when the uniform distribution over a sphere was used to smooth rather than the Bernoulli distribution, we will obtain better results when we work with the uniform distribution over a ball. Fortunately, using concentration of the Gaussian measure one can translate results from the case where is uniformly distributed over a ball to the case that it is sampled according to ; see Proposition 4.5. This is analogous to the translation from results for the uniform distribution over a sphere to the Bernoulli distribution for codes elucidated in Proposition 3.9.
For either choice of noise, to obtain a smoothing bound we are required to bound the statistical distance between the distribution of if has density , and the uniform distribution over . It is readily seen that has density which is defined as (see Definition 2.2 with the choice of Haar measures given in Table 2)
Notation. For any ,
In the following proposition we specialize Corollary 2.4 to the case of lattices.
Proposition 4.1.
Let be an -dimensional lattice. Let be some density function on and be the density of the uniform distribution over . We have
We will restrict our instantiations to functions whose Fourier transforms are radial, that is, depends only on the Euclidean norm of , namely .
4.1. Smoothing Random Lattices
As with codes, we begin our investigation of smoothing lattices by considering the random case. However, defining a “random lattice” is much more involved than the analogous notion of random codes. Fortunately for us, we can apply the Siegel version of the Minkowski-Hlawka theorem to conclude that there exists a random lattice model which behaves very nicely from the perspective of “test functions”. We first state the technical theorem that we require.
Theorem 4.2 (Minkowski-Hlawka-Siegel).
On the set of all the lattices of covolume in there exists a probability measure such that, for any Riemann integrable function which vanishes outside some bounded region,(6)(6)(6)This statement holds for a larger class of functions. In particular it holds for our instantiation with the Gaussian distribution.
As intuition for the above theorem, consider the case that is the indicator function for a bounded, measurable subset . Then, Theorem 4.2 promises that the expected number of lattice points (other than the origin(7)(7)(7)Note that as with certainty, there is really no “randomness” for this event.) in is equal to the volume of over the covolume of the lattice.
Uniform Distribution over a Ball. Let
be the density of the uniform distribution over the Euclidean ball of radius . Let us recall that denotes the volume of any ball of radius . From Theorem 4.2, we may obtain the following proposition. This should be compared with Proposition 3.7.
Proposition 4.3.
On the set of all lattices of covolume in there exists a probability measure such that, for any
In particular, defining
if we have
Proof.
We define to be the procedure that samples a lattice according to of covolume , then outputs its dual. In the following chain, we first apply Proposition 4.1; then, Jensen’s inequality; then, the Minkowski-Hlawka-Siegel (MHS) Theorem (Theorem 4.2) to the function ; and, lastly, Parseval’s Identity (Theorem 2.1). This yields:
For the “in particular” part of the proposition, we use Stirling’s estimate to derive
from which it follows that if
we have
which concludes the proof. ∎
It is easily verified that the value of defined in Proposition 4.4 corresponds to the so-called Gaussian heuristic. We view this condition on as the equivalent of the Gilbert-Varshamov bound for codes as we discussed just below Proposition 3.7. In particular, as we need the support of the noise to have volume at least if we hope to smooth a lattice of covolume , we see that the uniform distribution over a ball is optimal for smoothing random lattices, just as the uniform distribution over a sphere was optimal for smoothing random codes.
Gaussian Noise. We now turn to the case of Gaussian noise. Following the proof of Proposition 4.3 to the point where we apply Parseval’s identity, but replacing by , we obtain that
To conclude, one uses the following routine computation
Thus, we obtain:
Proposition 4.4.
On the set of all the lattices of covolume in there exists a probability measure such that, for any ,
In particular, if , we have
To compare Propositions 4.3 and 4.4, we note that a random vector sampled according to has an expected Euclidean norm given by . So, it is fair to compare the effectiveness of smoothing with a parameter Gaussian distribution and the uniform distribution over a ball of radius . We note that, if is as in Proposition 4.4 and is the radius of the so-called Gaussian heuristic, then
Thus, we conclude that the parameter from Proposition 4.4 is larger than what we could hope by a factor .
4.2. Connecting Uniform Ball Distribution to Gaussian
However, recall that in the code-case we argued that, as the Hamming weight of a vector sampled according to the Bernoulli distribution is tightly concentrated, we could obtain the same smoothing bound for the Bernoulli distribution as we did for the uniform sphere distribution, essentially by showing that we can approximate a Bernoulli distribution by a convex combination of uniform sphere distributions. Similarly, we can relate the Gaussian distribution to the uniform distribution over a ball, and thereby remove this additional factor.
We state a general proposition that allows us to translate smoothing bounds for the uniform ball distribution to the Gaussian distribution. It guarantees that if the uniform ball distribution smooths whenever , the Gaussian distribution smooths whenever . While the intuition for the argument is the same as that which we used in the code-case, the argument is itself a bit more sophisticated.
Proposition 4.5.
Let be a random lattice of covolume and let be the uniform distribution over its cosets. Suppose that for all there is a function such that
Let . Then, for all , defining , we have
Proof.
See Appendix D. ∎
Combining the above proposition with Theorem 4.2, setting , we obtain the following theorem.
Theorem 4.6.
Let be a random lattice of covolume sampled according to , let be the uniform distribution over its cosets, and let
Then, for any , setting , we have
4.3. Smoothing Random -ary Lattices
While the method of sampling lattices promised by the Minkowski-Hlawka-Siegel Theorem (Theorem 4.2) is indeed very convenient for computations, it does not tell us much about how to explicitly sample from the distribution. Furthermore it is not very relevant if one is interested in the random lattices that are used in cryptography.
For a more concrete sampling procedure that is relevant to cryptography, we can consider the randomized Construction A (or, more precisely, its dual), which gives a very popular random model of lattices which are easily constructed from random codes. Specifically, for a prime and a linear code we obtain a lattice as follows. First, we “lift” the codewords to vectors in in the natural way by identifying with the set ; denote the lifted vector as . Then, we can define the following lattice
In other words: consists of all vectors in the integer lattice whose reductions modulo give an element of .
Fix integers , a prime and a desired covolume . We sample a random lattice as follows
-
•
First, sample a random linear code of dimension (recall this means that we sample a random matrix and define ),
-
•
Then, we scale by ,
-
•
Lastly, we output the dual of .
Notice that the scaling is chosen so that, as long as is of full rank, the lattice we output has the desired covolume . We denote this procedure of sampling by (the dependence on , and is left implicit).
The important fact is that, up to an error term (which decreases as increases), the expected number of lattice points from in a Euclidean ball of radius is roughly , as one would hope.
Proposition 4.7 ([Zam14, Lemma 7.9.2]).
For every , and prime power , for the expected number of lattice points from in a Euclidean ball of radius satisfies
We now turn to bounding the expected statistical distance between and , where and is the radius of the Euclidean ball from which the noise is uniformly sampled. First, we state an explicit formula for the Fourier transform of , the indicator function of a Euclidean ball of radius , in terms of Bessel functions.
Notation 4.8.
For a positive real number , we denote by the Bessel function of the first kind of order .
The important fact concerning Bessel functions that we will use is the following.
Fact 4.9.
We have
(18) |
We will refrain from providing an explicit formula for Bessel functions, and instead use the following upper-bound as a black-box.
Proposition 4.10 ([Kra06]).
For any we have
Using this proposition, we first prove a technical lemma that will be reused when we discuss smoothing arbitrary lattices. In order to state the lemma, we introduce the following auxiliary function.
Notation 4.11.
For a real , we define via
where is any vector in of norm . Note that as depends only on , this is indeed well-defined.
The following lemma leverages Proposition 4.10 to upper-bound on a closed interval.
Lemma 4.12.
For any and any and we have, for some constant
Proof.
First, we notice that for all
for some constant . We now use Proposition 4.10 to derive
for an appropriate constant which concludes the proof. ∎
We now provide the main theorem of this section. It demonstrates that to smooth our ensemble of random -ary codes (in expectation) with the uniform distribution over the ball of radius , it still suffices to choose , assuming is not too small.
Theorem 4.13.
Let and . Let be a prime and set . Let . For some constant , we have
In particular, if , we have
Proof.
Let for and
Now, we apply Proposition 4.1 and the above definitions to obtain
By Proposition 4.7, we may upper-bound
(19) |
Now, recalling we have for any
Thus, we conclude
Now, by Lemma 4.12 we have for all . Hence,
for an appropriate constant . Thus, putting everything together we derive
for some constant . The “in particular” part of the Theorem follows analogously to the corresponding argumentation (Stirling’s estimate) used in the proof of Proposition 4.3. ∎
Next, turning to Gaussian noise, we could again prove a smoothing bound “directly,” but this will lose the same factor of as we had earlier. Instead, we apply Proposition 4.5 with the function to conclude the following.
Theorem 4.14.
Let and . Let be a prime and set . Let be a random -ary lattice sampled according to , let be the uniform distribution over its cosets, and let
Then, for any , setting , we have
4.4. Smoothing Arbitrary Lattices
We now turn our attention to the task of smoothing arbitrary lattices.
Analogously to how we used the minimum distance of the dual code to give our smoothing bound for worst-case codes, we will use the shortest vector of the dual lattice in order to provide our smoothing bound for worst-case lattices. The lemma that we will apply is the following where
Lemma 4.15 ([PS09, Lemma 3]).
For any -dimensional lattice ,
Remark 4.16.
We begin by considering the effectiveness of smoothing with noise uniformly sampled from the ball. The following theorem is proved using similar techniques to those we used for Theorem 4.13, although instead of using Proposition 4.7 to bound the ’s, we use Lemma 4.15.
Theorem 4.17.
Let be an -dimensional lattice and be the uniform distribution over its cosets. Then, it holds that
In particular, setting
for all , it holds that
Proof.
Define
where we recall the definition of with (see Notation 4.11). We also define
With this notation and Proposition 4.1 we have
(20) |
By Lemma 4.12, for some constant , we obtain
Combining this with the upper-bound on provided by Lemma 4.15 (note that for all ), we find
In the above, all necessary constants were absorbed into the term. Combining this with (20), we obtain the first part of the theorem. The “in particular” part again follows using Stirling’s approximation. ∎
Next, we can consider the effectiveness of smoothing with the Gaussian distribution. As usual, we could follow the steps of the proof of Theorem 4.17 and obtain the same result, but with an additional multiplicative factor of . That is, we obtain
Theorem 4.18.
Let be an -dimensional lattice and be the uniform distribution over its cosets. Then, it holds.
In particular, setting
it holds for any that .
However, as usual it is more effective to combine the bound for the uniform ball distribution and decompose the Gaussian as a convex combination of uniform ball distributions, i.e. to apply Proposition 4.5. In this way, we can obtain the following theorem, improving the smoothing bound by another factor. In the following theorem, we are setting the function of Proposition 4.5 with the term in the bound of Theorem 4.17.
Theorem 4.19.
Let be an -dimensional lattice, the uniform distribution over its cosets, and
Then, for any and letting , it holds that
5. Acknowledgement
We would like to thank Iosif Pinelis for help with the proof of Proposition 4.5.
References
- [ABL01] Alexei E. Ashikhmin, Alexander Barg, and Simon Litsyn. Estimates of the distance distribution of codes and designs. Electron. Notes Discret. Math., 6:4–14, 2001.
- [ACKL05] Alexei E. Ashikhmin, Gérard D. Cohen, Michael Krivelevich, and Simon Litsyn. Bounds on distance distributions in codes of known size. IEEE Trans. Inf. Theory, 51(1):250–258, 2005.
- [ADRS15] Divesh Aggarwal, Daniel Dadush, Oded Regev, and Noah Stephens-Davidowitz. Solving the shortest vector problem in time using discrete Gaussian sampling. In Proceedings of the forty-seventh annual ACM symposium on Theory of computing, pages 733–742, 2015.
- [Ale11] Michael Alekhnovich. More on average case vs approximation complexity. Computational Complexity, 20(4):755–786, 2011.
- [Ban93] Wojciech Banaszczyk. New bounds in some transference theorems in the geometry of numbers. Mathematische Annalen, 296(1):625–635, 1993.
- [Bas65] LA Bassalygo. New upper bounds for codes correcting errors. Probl. Peredachi Inform, 1(4):41–44, 1965.
- [BLVW19] Zvika Brakerski, Vadim Lyubashevsky, Vinod Vaikuntanathan, and Daniel Wichs. Worst-case hardness for LPN and cryptographic hashing via code smoothing. In Annual international conference on the theory and applications of cryptographic techniques, pages 619–635. Springer, 2019.
- [CE03] Henry Cohn and Noam Elkies. New upper bounds on sphere packings I. Ann. of Math, (157-2):689–714, 2003.
- [Chu97] Fan R. K. Chung. Spectral graph theory, volume 92 of CBMS Regional Conference Series in Mathematics. American Mathematical Society, 1997.
- [DL98] Philippe Delsarte and Vladimir Iossifovitch Levenshtein. Association schemes and coding theory. IEEE Trans. Inform. Theory, 44(6):2477–2504, 1998.
- [DST19] Thomas Debris-Alazard, Nicolas Sendrier, and Jean-Pierre Tillich. Wave: A new family of trapdoor one-way preimage sampleable functions based on codes. In Advances in Cryptology - ASIACRYPT 2019, LNCS, Kobe, Japan, December 2019.
- [DT17] Thomas Debris-Alazard and Jean-Pierre Tillich. Statistical decoding. preprint, January 2017. arXiv:1701.07416.
- [GPV08] Craig Gentry, Chris Peikert, and Vinod Vaikuntanathan. Trapdoors for hard lattices and new cryptographic constructions. In Proceedings of the fortieth annual ACM symposium on Theory of computing, pages 197–206, 2008.
- [IS98] Mourad E.H. Ismail and Plamen Simeonov. Strong asymptotics for Krawtchouk polynomials. Journal of Computational and Applied Mathematics, pages 121–144, 1998.
- [KL78] Grigory Kabatiansky and Vladimir I. Levenshtein. Bounds for packings on a sphere and in space. Problems of Information Transmission, (14):1–17, 1978.
- [Klø07] Torleiv Kløve. Codes for Error Detection, volume 2 of Series on Coding Theory and Cryptology. WorldScientific, 2007.
- [Kra06] Ilia Krasikov. Uniform bounds for bessel functions. Journal of Applied Analysis, 12:83–91, 06 2006.
- [Lev79] Vladimir I. Levenshtein. On bounds for packings in -dimensional euclidean space. Dokl. Akad. Nauk SSSR, 245:1299–1303, 1979.
- [Lev95] Vladimir I. Levenshtein. Krawtchouk polynomials and universal bounds for codes and designs in hamming spaces. IEEE Trans. Inf. Theory, 41(5):1303–1321, 1995.
- [LLB22] Laura Luzzi, Cong Ling, and Matthieu R. Bloch. Secret key generation from Gaussian sources using lattice-based extractors. CoRR, abs/2206.10443, 2022.
- [LLBS14] Cong Ling, Laura Luzzi, Jean-Claude Belfiore, and Damien Stehlé. Semantically secure lattice codes for the Gaussian wiretap channel. IEEE Transactions on Information Theory, 60(10):6399–6416, 2014.
- [McE78] Robert J. McEliece. A Public-Key System Based on Algebraic Coding Theory, pages 114–116. Jet Propulsion Lab, 1978. DSN Progress Report 44.
- [MR07] Daniele Micciancio and Oded Regev. Worst-case to average-case reductions based on Gaussian measures. SIAM Journal on Computing, 37(1):267–302, 2007.
- [MRJW77] Robert J. McEliece, Eugene R. Rodemich, Howard Rumsey Jr., and Lloyd R. Welch. New upper bounds on the rate of a code via the Delsarte-MacWilliams inequalities. IEEE Trans. Inf. Theory, 23(2):157–166, 1977.
- [MTSB13] Rafael Misoczki, Jean-Pierre Tillich, Nicolas Sendrier, and Paulo S. L. M. Barreto. MDPC-McEliece: New McEliece variants from moderate density parity-check codes. In Proc. IEEE Int. Symposium Inf. Theory - ISIT, pages 2069–2073, 2013.
- [PS09] Xavier Pujol and Damien Stehlé. Solving the shortest lattice vector problem in time 2. IACR Cryptol. ePrint Arch., 2009:605, 2009.
- [vL99] Jacobus Hendricus van Lint. Introduction to coding theory. Graduate texts in mathematics. Springer, 3rd edition edition, 1999.
- [Wai19] Martin J Wainwright. High-dimensional statistics: A non-asymptotic viewpoint, volume 48. Cambridge University Press, 2019.
- [YZ21] Yu Yu and Jiang Zhang. Smoothing out binary linear codes and worst-case sub-exponential hardness for LPN. In Tal Malkin and Chris Peikert, editors, Advances in Cryptology - CRYPTO 2021 - 41st Annual International Cryptology Conference, CRYPTO 2021, Virtual Event, August 16-20, 2021, Proceedings, Part III, volume 12827 of Lecture Notes in Computer Science, pages 473–501. Springer, 2021.
- [Zam14] Ram Zamir. Lattice Coding for Signals and Networks: A Structured Coding Approach to Quantization, Modulation and Multiuser Information Theory. Cambridge University Press, 2014.
Appendix A Proof of Proposition 3.9
Our aim in this section is to prove the following proposition
See 3.9
Roughly speaking, this proposition is a consequence of the fact that a Bernoulli distribution concentrates Hamming weights over a small number of slices close to the expected weight (here ) and, on each slice the Bernoulli distribution is uniform. Let us introduce the truncated Bernoulli distribution over words of Hamming weight for some , namely
(21) |
where
(22) |
is the probability normalizing constant.
Proposition 3.9 is a consequence of the following lemmas.
Lemma A.1.
Let . We have
Proof.
By Chernoff’s bound
(23) |
Therefore for any ,
(24) |
We have now the following computation:
where in the last line we used that is a probability distribution. ∎
Lemma A.2.
We have
Proof.
By the triangle inequality,
Focusing on the second term now
which concludes the proof by Lemma A.1. ∎
The following lemma is a basic property of the statistical distance.
Lemma A.3.
For any distribution and we have
where the ’s are positive and sum to one.
We are now ready to prove Proposition 3.9.
Appendix B Proof of Proposition 3.11
Our aim in this section is to prove the following proposition which is an extension of [ABL01, Theorem 3] for ([ABL01, Theorem 3] only applied for .)
See 3.11
Our proof is mainly a rewriting of the proof of [ABL01, Theorem 3] which relies on the following proposition.
Proposition B.1 ([ABL01, Proposition with ]).
Let be a binary code of length such that . Let and be such that
where denotes the first root of the Krawtchouk polynomial of order , namely .
When , we have
(27) |
The approach is to optimize on the choice of in Proposition B.1 to give an upper-bound on . More precisely we observe that
(28) |
and then choose to minimize .
Proof of Proposition 3.11.
It will be helpful to bring in the following map:
It can be verified that this application is an involution, is symmetric and decreasing on .
Let be a binary code of length such that where and be defined as in Proposition B.1. Let and . Then by Proposition B.1 we have (see Equation (28))
(29) |
Case 1: .
It is optimal to choose in this case such that where
and as tends to infinity. Let us first notice that implies that which together with implies that which in turn is equivalent to the condition for being able to apply Proposition B.1. Moreover also implies and by using Proposition 3.8 we obtain
Therefore
Appendix C Proof of Theorem 3.16
Our aim in this appendix is to prove the following theorem.
See 3.16
Sketch of proof. We will use the following proof strategy
-
1.
By Lemma A.2 we know that on one hand
(30) This is actually a consequence of Chernoff’s bound. This argument can also be used to show that the Fourier transforms are also close to each other pointwise
(31) - 2.
- 3.
Lemma C.1.
We have
Proof.
Recall that where by Chernoff’s bound, we have
(33) |
Notice now that,
Let . Notice that . By linearity of the Fourier transform we obtain the following computation:
(34) |
where in the last line we used Equation (33). Recall now that by definition of the Fourier transform for functions over we have:
By plugging this in Equation (34) we get:
which concludes the proof. ∎
Proof of Step 2. This corresponds to proving the following lemma.
Lemma C.2.
Proof.
By applying Proposition 3.3 to we obtain
(35) |
where denotes the common value of the radial function on vectors of Hamming weight . Recall now that and by Lemma C.1 that . Therefore,
By plugging this in Equation (35) we obtain (as there is at most one dual codeword of weight for each , see Lemma 3.15)
(36) |
which concludes the proof. ∎
Appendix D Proof of Proposition 4.5
Our aim in this section is to prove the following proposition.
See 4.5
It will be a consequence of the following lemmas. We begin with the following result decomposing the Gaussian as a convex combination of balls.
Lemma D.1.
The Gaussian distribution in dimension of parameter is the following convex combination of uniform distributions over balls:
where . Furthermore, we have .
Proof.
First, let (i.e. the value the probability density function takes on vectors of weight ) and denote . For any , setting , as we have
Above, we denoted by the function which takes value on input if , and otherwise. To conclude, note that and recall .
For the “furthermore” part of the lemma, we compute
(37) |
We make the substitution , which means . Also, we recall . Thus,
which concludes the proof. ∎
We now quote the following bound, which makes precise the intuition that it is exponentially unlikely that a random Gaussian vector has norm factor smaller than its expected norm. This result provides the analogy for the Chernoff bound that we used for the code-case.
Lemma D.2 ([Wai19, Example 2.5]).
Let be a random Gaussian vector of dimension and parameter . Let . Then
This lemma allows us to prove the following lemma bounding when .
Lemma D.3.
Let and . Then
Proof.
Let . By Lemma D.2, if denotes a random Gaussian vector of dimension and parameter , we have
(38) |
To compute this last integral, note that
(39) |
where denotes the Euclidean sphere of radius and is the area element. If denotes the surface area of , then and thus
(40) |
Further, it is known that . Therefore, plugging Equations (39) and (40) into (38) leads to
(41) |
Now, we look at the left-hand side of the inequality we wish to prove. We begin by making the substitution . So then . Moreover, let and note that when we have .
Plugging this last inequality with (41) yields
To conclude the proof, note that and therefore
It concludes the proof. ∎
We are now ready to prove Proposition 4.5.
Proof of Proposition 4.5..
By Lemma D.1, is a convex combination of uniform distribution over balls, namely . Therefore (we use here the analogue of Lemma A.3 in the context of the statistical distance between two probability density functions)
We split the integral in two parts at radius . For the first part , we use the trivial bound which gives:
We then apply Lemma D.3, which bounds this part by .
For the second part , we use the trivial bound and, noting
we may apply the assumption of the proposition, yielding
Adding these bounds yields the proposition. ∎