An enchancment to Bennett’s inequality for the Poisson distribution
[ad_1]
If , a Poisson random variable with imply is a random variable taking values within the pure numbers with likelihood distribution
One is commonly keen on bounding higher tail possibilities
for , or decrease tail possibilities
for . A regular instrument for that is Bennett’s inequality:
Proposition 1 (Bennett’s inequality) One has
for and
for , the place
From the Taylor growth for we conclude Gaussian sort tail bounds within the regime (and particularly when (within the spirit of the Chernoff, Bernstein, and Hoeffding inequalities). however within the regime the place is giant and constructive one obtains a slight acquire over these different classical bounds (of sort, fairly than ).
Proof: We use the exponential second methodology. For any , we’ve from Markov’s inequality that
A regular computation reveals that the second producing perform of the Poisson distribution is given by
and therefore
For , it seems that the right-hand facet is optimized by setting , wherein case the right-hand facet simplifies to . This proves the primary inequality; the second inequality is confirmed equally (however now and are non-positive fairly than non-negative).
Comment 2 Bennett’s inequality additionally applies for (suitably normalized) sums of bounded impartial random variables. In some circumstances there are direct comparability inequalities obtainable to narrate these variables to the Poisson case. As an illustration, suppose is the sum of impartial Boolean variables of whole imply and with for some . Then for any pure quantity , we’ve
As such, for small, one can effectively management the tail possibilities of when it comes to the tail likelihood of a Poisson random variable of imply near ; that is in fact very carefully associated to the well-known undeniable fact that the Poisson distribution emerges because the restrict of sums of many impartial boolean variables, every of which is non-zero with small likelihood. See this paper of Bentkus and this paper of Pinelis for some additional helpful (and fewer apparent) comparability inequalities of this sort.
On this be aware I needed to file the commentary that one can enhance the Bennett certain by a small polynomial issue as soon as one leaves the Gaussian regime , particularly gaining an element of when . This commentary will not be troublesome and is implicitly within the literature (one can extract it as an example from the rather more normal outcomes of this paper of Talagrand, and the essential concept already seems in this paper of Glynn), however I used to be not capable of finding a clear model of this assertion within the literature, so I’m inserting it right here on my weblog. (But when a reader is aware of of a reference that mainly incorporates the certain beneath, I might be pleased to know of it.)
Proposition 3 (Improved Bennett’s inequality) One has
for and
for .
Proof: We start with the primary inequality. We could assume that , since in any other case the declare follows from the same old Bennett inequality. We broaden out the left-hand facet as
Observe that for that
Thus the sum is dominated by the primary time period occasions a geometrical sequence . We are able to thus certain the left-hand facet by
By the Stirling approximation, that is
The expression contained in the supremum is lowering in for , thus we are able to certain it by
which simplifies to
after a routine calculation.
Now we flip to the second inequality. As earlier than we could assume that . We first eliminate a degenerate case wherein . Right here the left-hand facet is simply
and the right-hand facet is akin to
Since is adverse and , we see that the right-hand facet is , and the estimate holds on this case.
It stays to think about the regime the place and . The left-hand facet expands as
The sum is dominated by the primary time period occasions a geometrical sequence . The maximal is akin to , so we are able to certain the left-hand facet by
Utilizing the Stirling approximation as earlier than we are able to certain this by
which simplifies to
after a routine calculation.
The identical evaluation will be reversed to point out that the bounds given above are mainly sharp as much as constants, a minimum of when (and ) are giant.
[ad_2]