SlideShare a Scribd company logo
1 of 47
Download to read offline
Nonparametric Estimation of Distribution and Density
Functions Using q-Bernstein Polynomials
Yogendra P. Chaubey
Department of Mathematics and Statistics
Concordia University, Montreal, Canada H3G 1M8
E-mail:yogen.chaubey@concordia.ca
Talk to be presented at the 6th International Conference
of IMBIC, Kolkata, India, December 21-23, 2016
Acknowledgment: The computations were performed by Qi Zhang as
part of his BSc (Honours) project.
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 1 / 47
Abstract
Here we consider the smooth density estimation for the random variables
that have a compact support. Without loss of generality, we restrict the
support to the closed interval [0, 1], for which Bernstein polynomial
estimator, originally proposed by Vitale (1973), is known to have some
good properties, such as being a genuine density estimator and being free
from boundary bias (see Babu, Canty and Chaubey (2002) and Kakizawa
(2004)). Generalized version of these estimators were proposed by Prakasa
Rao (2005) and Kakizawa (2011) by considering the generalized Bernstein
polynomials introduced in Cao (1997). Some work on improving
convergence rate of the Bernstein polynomial density estimator were
carried out in Leblanc (2009, 2012) and Igarshi and Kakizawa (2014). In
this paper, we consider yet another generalization of the Bernstein
polynomials introduced by Phillips (1997), that may have some advantage
over the usual Bernstein polynomial estimator.
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 2 / 47
1 Introduction
2 Preliminaries
3 Smooth Estimators of the Distribution and Density Functions
4 Asymptotic Properties of Estimators
Asymptotic Properties of ˜Fn,m,q
Asymptotic Properties of ˜fn,m,q
5 Numerical Studies
Cross Validation Method to Determine the Smoothing Parameter q
Likelihood Based Cross Validation
Integrated Square Error Cross Validation
Least Square Cross Validation
Illustration for Distribution Function Estimator
Illustration for Density Estimator
Comparing Estimators by Average Squared Error
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 3 / 47
1. Introduction
Let h be a bounded continuous function defined on the interval [0, 1],
then the corresponding Bernstein polynomial of degree m is defined as
Bm(x; h) =
m
k=0
h
k
m
bk(x, m), (1.1)
where
bk(x, m) =
m
k
xk
(1 − x)m−k
, k = 0, 1, .... (1.2)
This polynomial was proposed as an uniform approximation to
bounded continuous functions defined on [0, 1] by S.N. Bernstein
(1912), that provided an existential proof of Weierstrass
approximation theorem.
The monograph by Lorentz (1986) makes an excellent reference
regarding the properties of the Bernstein polynomials.
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 4 / 47
1. Introduction
Assuming that the function h has a continuous derivative, the
derivative Bm(x; h) given by
Bm(x; h) = m
m−1
k=0
h
k + 1
m
− h
k
m
bk(x, m − 1) (1.3)
uniformly approximates the derivative h (x).
Given a random sample {X1, ..., Xn} from an absolutely continuous
distribution function F, we can estimate it by the empirical
distribution function
Fn(x) =
1
n
n
i=1
I{Ti ≤ x}. (1.4)
Since the above function is a step function, it is not appropriate for
finding the density function f = F , thus smoothing of this function
is desired.
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 5 / 47
1. Introduction
Replacing h by the empirical distribution function Fn, the Bernstein
polynomial approximation mentioned earlier motivates the following
smooth estimator of the distribution function
˜Fn,m(x) =
m
k=0
Fn
k
m
bk(x, m). (1.5)
This estimator is a genuine distribution function and therefore its
derivative serves as a smooth density estimator given by
˜fn,m(x) = Bm(x; Fn)
= m
m−1
k=0
Fn
k + 1
m
− Fn
k
m
bk(x, m − 1)
(1.6)
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 6 / 47
1. Introduction
Bernstein polynomial density estimator: Originally proposed by Vitale
(1973) and later investigated by many authors
Gowronski and Stadm¨uler (1981): Smoothing of histogram using
Feller’s Theorem
Stadtm¨uller, (1983): Convergence of the density estimator
Petrone (1999): Nonparametric Bayesian prior
Ghosal, 2001: Convergence of nonparametric posteriors
Babu, Canty and Chaubey (2002): Further studies
Leblanc (2009, 2012): Smooth distribution function estimator.
Prakasa Rao (2005) and Kakizawa (2011): Generalized Bernstein
polynomials introduced by Cao (1997).
Another generalization is introduced by Phillips (1997) that is the
subject of this talk.
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 7 / 47
2. Preliminaries
Let q > 0, then for any k ∈ N ∪ {0}, the q−integer [k]q and
q−factorial [k]q! are respectively defined by
[k]q =
(1 − qk)/(1 − q), for q = 1
k, for q = 1,
(2.1)
and
[k]q! =
[k]q[k − 1]q...[1]q, for k ≥ 1
1, for k = 0.
(2.2)
For integers 0 ≤ k ≤ m, the q−binomial coefficients are defined by
m
k q
=
[m]q!
[k]q![m − k]q!
(2.3)
For q = 1, we clearly have [m]q = m, [m]q! = m! and
m
k q
=
m
k
,
q−factorials and q−binomial coefficients reduce to the usual factorials
and binomial coefficients, respectively.
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 8 / 47
2. Preliminaries
Definition
(Phillips 1997). Let h ∈ C[0, 1]. The generalized Bernstein polynomial
based on the q−integers is
Bm,q(x; h) =
m
k=0
h
[m]q
[k]q
bk,q(x, m) (2.4)
where
bk,q(x, m) =
m
k q
xk
m−k−1
s=0
(1 − qs
x), m = 1, 2, ...
and an empty product is taken to be equal to 1.
The polynomials defined in (2.4) reduce to the usual Bernstein
polynomials for q = 1, and hence they are referred to as the
generalized Bernstein polynomials, however, in the sequel we will refer
them as the q−Bernstein polynomials.
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 9 / 47
2. Preliminaries
The following analog of Bernstein’s Theorem for q−Bernstein
polynomials was proved by Phillips (1997).
Theorem
Let a sequence {qm}∞
m=1 satisfy 0 < qm < 1 and qm → 1 as m → ∞.
Then for any function h ∈ C[0, 1],
Bm,qm (x; h) → h(x) [x ∈ [0, 1]; m → ∞].
Here the expression gm(x) → g(x)[x ∈ [0, 1]; m → ∞] denotes
uniform convergence of gm to g with respect to x ∈ [0, 1].
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 10 / 47
2. Preliminaries
The following theorem from Phillips (1996) gives convergence results
of the derivative of the q−Bernstein polynomial
Theorem
Let h ∈ C1[0, 1] and let the sequence (qm) be chosen so that the sequence
( m) converges to zero from above faster than 1/3m, where
m =
m
1 + qm + q2
m + ... + qm−1
m
− 1.
Then the sequence of derivatives of the q−Bernstein polynomials
converges uniformly on [0, 1] to h (x).
Phillips (1996) showed that the condition on ( m) can be satisfied by
choosing (qm) such that
1 −
a−m
m
≤ qm < 1 with a > 3.
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 11 / 47
3. Estimators of the Distribution and Density Functions
Similar to the estimator given in (1.5) using the Bernstein polynomial,
the q− Bernstein polynomial may be used to define the following
non-parametric smooth estimator of F. Thus, the q−Bernstein
polynomial estimator of the distribution function F is given by
˜Fn;m,q(x) =
m
k=0
Fn
[k]q
[m]q
bk,q(x, m), x ∈ [0, 1], (3.1)
where
bk,q(x, m) =
m
k q
xk
m−k−1
s=0
(1 − qs
x). (3.2)
For q = 1, the above estimator reduces to the usual Bernstein
polynomial smooth estimator of the distribution function studied in
Babu, Canty and Chaubey (2004).
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 12 / 47
2. Estimators of the Distribution and Density Functions
By Euler identity (see Philips 1997) the product in (3.2) can be
written as a sum and thus taking the derivative of ˜Fn;m,q, the
following smooth estimator of f is obtained that will be called
q−Bernstein density estimator:
˜fn;m,q(x) =
m
k=0
m−k
s=0
Fn
[k]q
[m]q
ωs,k(x; m, q) (3.3)
where
ωs,k(x; m, q) =
m
k q
(−1)s
q
s(s−1)
2
m − k
s q
(k + s)xk+s−1
.
This paper presents a comparative study of the resulting smooth
estimators with Bernstein polynomial estimators while choosing the
value of q by cross validation and the value of m = n/ log(n) as
suggested in Babu, Canty and Chaubey (2004).
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 13 / 47
2. Estimators of the Distribution and Density Functions
Since, the approximation improves as m → ∞, one may propose the
use of limiting q−Bernstein polynomials, that was studies in Il’inskii
and Ostrovska (2002). They noted that as m → ∞ for 0 < q < 1
lim
m→∞
[k]q
[m]q
= 1 − qk
lim
m→∞
bk,q(x, m) =
xk
(1 − qk)[k]q!
∞
s=0
(1 − qs
x) =: bk,∞(x, q)
the limiting q−Bernstein polynomial becomes
B∞,q(x, h) =
∞
k=0 h(1 − qk)bk,∞(x, q), if x ∈ [0, 1),
h(1) if x = 1.
Theorem
For any f ∈ C[0, 1],
B∞,q(x, h) → f(x)[x ∈ [0, 1]; q ↑ 1].
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 14 / 47
2. Estimators of the Distribution and Density Functions
Theorem
For any f ∈ C[0, 1],
B∞,q(x, h) → f(x)[x ∈ [0, 1]; q ↑ 1].
Theorem
Let 0 < α < 1. Then for any f ∈ C[0, 1],
Bm,q(x, h) → h(x)[x ∈ [0, 1]; q ∈ [α, 1]; m → ∞].
The above theorems basically recast Phillip’s results by explicitly
exhibiting the limiting form of Bm,q(x; h) as m → ∞. So, for general
approximation, the problem of determining the constants m and q
reduces to determining q only. In our experience, the corresponding
smooth estimator of F is meaningful in practice, however, the
corresponding density estimator may not be very smooth.
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 15 / 47
2. Estimators of the Distribution and Density Functions
In our experience, the corresponding smooth estimator of F is
meaningful in practice, however, the corresponding density estimator
may not be very smooth.
Thus, for numerical studies, we propose to consider m = n/ log(n)
[the choice suggested in Babu, Canty and Chaubey (2004)] and
determine q by cross-validation.
In the next section we provide the almost sure convergence results for
the smooth estimators similar to those obtained in Babu, Canty and
Chaubey (2004) using Bernstein polynomials. But, first some pictures
of the estimators.
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 16 / 47
Distribution Function Estimator for Beta(3, 3) density
0.0 0.2 0.4 0.6 0.8 1.0
0.00.20.40.60.81.0
q−Bernstein Smooth Distribution
Sample from Beta(3,3) Density, n=50
x
Fn(x)
Bernstein
q−Bernstein,q=.99
Figure: 1. Distribution Function Estimators for Beta(3, 3) density; n = 50.
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 17 / 47
Distribution Function Estimator for Beta(3, 3) density
0.0 0.2 0.4 0.6 0.8 1.0
0.00.20.40.60.81.0
q−Bernstein Smooth Distribution
Sample from Beta(3,3) Density, n=50
x
Fn(x)
Bernstein
q−Bernstein,q=.99,m=200
Figure: 2. Distribution Function Estimators for Beta(3, 3) density; n = 50.
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 18 / 47
Density Function Estimator for Beta(3, 3) density
q−Bernstein Smooth Desnity estimator
Sample from Beta(3,3) Density, n=50
x
fn(x)
0.0 0.2 0.4 0.6 0.8 1.0
0.00.51.01.52.0
Bernstein
q−Bernstein,q=.99
Figure: 3. Density Function Estimators for Beta(3, 3) density; n = 50.
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 19 / 47
Density Function Estimator for Beta(3, 3) density
q−Bernstein Smooth Desnity estimator
Sample from Beta(3,3) Density, n=50
x
fn(x)
0.0 0.2 0.4 0.6 0.8 1.0
0.00.51.01.52.0
Bernstein
q−Bernstein,q=.99,m=50
Figure: 4. Density Function Estimators for Beta(3, 3) density; n = 50.
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 20 / 47
Density Function Estimator for Beta(3, 3) density
q−Bernstein Smooth Desnity estimator
Sample from Beta(3,3) Density, n=50
x
fn(x)
0.0 0.2 0.4 0.6 0.8 1.0
0.00.51.01.52.0
Bernstein
q−Bernstein,q=.99,m=25
Figure: 5. Density Function Estimators for Beta(3, 3) density; n = 50.
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 21 / 47
4. Asymptotic Properties of Estimators
Throughout this paper we use the notation
G = sup
x∈[0,1]
|G(x)|,
for a bounded function G on [0, 1],
an = (n−1
log n)1/2
, and bn,m = (n−1
log n)1/2
(m−1
log m)1/4
.
(4.1)
The following theorem shows that ˜Fn,m is strongly consistent.
Theorem
If m, n → ∞, then ˜Fn,m,q − F → 0 a.s. as q → 1.
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 22 / 47
4.1 Asymptotic Properties of Distribution Function
Estimator
Next theorem gives a result depicting the closeness of the smooth
estimator with the empirical distribution function.
Theorem
Let F be continuous and differentiable on the interval [0, 1] with density f.
If f is Lipschitz of order 1, then for n2/3 ≤ m ≤ (n/ log n)2, we have a.s.
as n → ∞ ,
lim
q↑1
˜Fn,m − Fn = O (n−1
log n)1/2
(m−1
log m)1/4
. (4.2)
For m = n, we have
lim
q↑1
˜Fn,m − Fn = O(n−3/4
(log n)3/4
) a.s.. (4.3)
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 23 / 47
4.2. Asymptotic Properties of Density Function Estimator
We now establish the strong convergence of ˜fn(t) similar to that of
˜Fn(t).
Theorem
Let F be continuous on the interval [0, 1] with a continuous density f.
Then for 2 ≤ m ≤ (n/ log n), we have a.s. as n → ∞ and q → 1,
˜fn,m,q − f = O(m1/2
an) + O( F∗
m − f ), (4.4)
a.s. as n → ∞ , where F∗
m denotes the derivative of F∗
m. Consequently, if
m = o(n/ log n), then ˜fn,m,q − f → 0 a.s. as m, n → ∞.
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 24 / 47
4.2. Asymptotic Properties of Density Function Estimator
The following theorem establishes the asymptotic normality of the
˜fn,m,q(x).
Theorem
Under conditions in theorem 3.1, if f(x) > 0 then as q → 1
n1/2
m1/4
( ˜fn,m,q(x) − f(x))
D
→ N 0,
f(x)
2 πx(1 − x)
(4.5)
as m, n → ∞ such that m ≤ (n/ log n).
Remark: If mn−2/5 → δ−2 > 0, then as q → 1,
n(2/5)
( ˜fn,m,q(x) − f(x))
D
→ N 0,
δf(x)
2 πx(1 − x)
(4.6)
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 25 / 47
4.3. Estimation for Other Support
The case of supports other than [0, 1] may be handled by
transformation.
1. For a finite support [a, b], transform the sample
values to Y1, ...., Yn, where Yi = (Xi − a)/(b − a).
Denoting the ˜Gn(y) as the smooth distribution function
of the transformed sample, the smooth distribution
˜Fn(x) is given by
˜Fn(x) = ˜Gn(y) where y = (x − a)/(b − a).
2.For the interval (−∞, ∞), the transformation
Yi = (1/2) + (1/π) tan−1 Xi is useful. In this case
˜Fn(x) = ˜Gn(y) where y = (1/2)(1/π) tan−1
x.
3. For the non-negative support [0, ∞) :, we may use
the transformation: y = x/(1 + x)
˜Fn(x) = ˜Gn(y) where y = x/(1 + x).
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 26 / 47
5. Numerical Studies
Here we illustrate the results for the estimation of distribution and
density for the following three cases: (i) Beta (3,3) distribution, (ii)
Beta(2,4) distribution and (iii) 0.65×Beta(2,10)+0.35×Beta(10,2)
distribution.
The three cases are chosen to depict three scenarios, namely that of
symmetric unimodal distribution, asymmetric unimodal distribution
and a bimodal distribution. We investigate the possibility of choosing
q different than 1, for the choice of m = n/ log(n), a choice
recommended in Babu, Canty and Chaubey (2002) for Bernstein
polynomials. We would also like to investigate the optimal choice of q
using the limiting form of the q−Bernstein polynomials, however, this
is not pursued here. For a given m, the choice of q is obtained by
cross-validation that is explained below.
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 27 / 47
5.1 Cross Validation
For the q−Bernstein polynomial CDF and PDF estimators, in order to
choose an appropriate parameter q, we introduce a numerical method
called Cross Validation, detail of relative equations are referenced
from Chaubey and Sen (2009).
Kullback-Liebler divergence between the estimated density ˜fn and the
true density f is given by
KL( ˜fn; f) = E log
f(x)
˜fn(x)
dF(x)
In practice, the optimum cross-validation method estimates such
divergence from the data for a given smoothing parameter and
chooses one which gives the smallest estimated divergence. Bowman
(1984) shows that this procedure is equivalent to the minimization of
the negative likelihood,
CVKL(q) = − log
n
i=1
˜fn(Xi; Di) = −
n
i=1
log( ˜fn(Xi; Di)) (5.1)
where D denotes data with X removed from D, which is the wholeYogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 28 / 47
5.1 ISE Cross Validation
According to this criterion we determine λn that minimizes the
criterion related to the mean integrated squared error,
MISE(q) = E ( ˜fn(x) − f(x))2
dx
Estimating this from the data and minimizing it is equivalent to the
minimization of (see Silverman (1986),
CVISE(q) = ˜f2
n(x; q, D)dx −
2
n
n
i=1
˜fn−1(Xi; q, Di) (5.2)
The first part can be obtain by numerical integration, say composite
Simpson Method. And again, Di denotes data with Xi removed from
D.
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 29 / 47
5.1 LS Cross Validation
The above cross validation works well if f is a density estimator. For
CDF, it is better to use least square cross validation (see H¨ardle,
1991).
CVLS(q) =
n
i=1
( ˜Fn(Xi; q, D) − ˜Fn(Xi; q, Di))2
(5.3)
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 30 / 47
Distribution Function-Illustration
For a given sample of size n we start with m = n and determine the
optimum value of the smoothing parameter as illustrated in Figures
6-8 for a random sample generated from each of the distributions.
We also investigate the effect of increasing the value of m. This is
done by selecting a grid of size n on the interval [0, 1], and we let M
denote the value of m that gives the average error of estimation
between the successive values of M and M + 1 less than ε, in order
to minimize the amount of computation. [In order to control the
individual errors, M will depend on the value of x also.].
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 31 / 47
0.0 0.2 0.4 0.6 0.8 1.0
0.050.100.150.200.25
q
SumofSquareErros(SSE)
(a) n = 20, q = 0.94, m = 20
0.2 0.4 0.6 0.8
0.00.20.40.60.81.0
x
cdf
(b) M = 147, ε = 9.96 × 10−5
0.0 0.2 0.4 0.6 0.8 1.0
0.00.20.40.60.8
q
SumofSquareErros(SSE)
(c) n = 50, q = 1, m = 50
0.2 0.4 0.6 0.8
0.00.20.40.60.81.0
x
cdf
(d) M = 1024, ε = 5.31 × 10−5
0.0 0.2 0.4 0.6 0.8 1.0
0.00.20.40.60.81.01.2
q
SumofSquareErros(SSE)
(e) n = 100; q = 1, m = 100
0.2 0.4 0.6 0.8 1.0
0.00.20.40.60.81.0
x
cdf
(f) M = 801, ε = 7.56 × 10−5
Figure 6: LSCV plot for q−Bernstein estimator (Left Panel); CDF Estima-
tor for Beta(3, 3) Samples (Right Panel).
Remark: red - true density; dash - q−Bernstein with m = n; blue -
q−Bernstein with parameter q, M
12
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 32 / 47
0.0 0.2 0.4 0.6 0.8 1.0
0.20.40.60.8
q
SumofSquareErros(SSE)
(a) n = 20, q = 0.96, m = 20
0.1 0.2 0.3 0.4 0.5 0.6 0.7
0.00.20.40.60.81.0
x
cdf
(b) M = 206, ε = 9.60 × 10−5
0.0 0.2 0.4 0.6 0.8 1.0
0.00.20.40.60.81.01.21.4
q
SumofSquareErros(SSE)
(c) n = 50, q = 0.99, m = 50
0.2 0.4 0.6 0.8
0.00.20.40.60.81.0
x
cdf
(d) M = 105, ε = 8.28 × 10−5
0.0 0.2 0.4 0.6 0.8 1.0
0.00.51.01.52.02.53.03.5
q
SumofSquareErros(SSE)
(e) n = 100; q = 1, m = 100
0.0 0.2 0.4 0.6 0.8
0.00.20.40.60.81.0
x
cdf
(f) M = 611, ε = 6.64 × 10−5
Figure 7: Smooth Bernstein(Left) and q−Bernstein(Right) CDF Estimator
for Beta(2, 4) Samples.
Remark: red - true density; dash - q−Bernstein with m = n; blue -
q−Bernstein with parameter q, M
13
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 33 / 47
0.0 0.2 0.4 0.6 0.8 1.0
0.20.40.60.81.01.2
q
SumofSquareErros(SSE)
(a) n = 20, q = 0.99, m = 20
0.0 0.2 0.4 0.6 0.8 1.0
0.00.20.40.60.81.0
x
cdf
(b) M = 73, ε = 6.88 × 10−5
0.0 0.2 0.4 0.6 0.8 1.0
0.00.51.01.52.02.5
q
SumofSquareErros(SSE)
(c) n = 50, q = 1, m = 50
0.0 0.2 0.4 0.6 0.8 1.0
0.00.20.40.60.81.0
x
cdf
(d) M = 355, ε = 4.41 × 10−5
0.0 0.2 0.4 0.6 0.8 1.0
012345
q
SumofSquareErros(SSE)
(e) n = 100; q = 1, m = 100
0.0 0.2 0.4 0.6 0.8 1.0
0.00.20.40.60.81.0
x
cdf
(f) M = 371, ε = 8.98 × 10−5
Figure 8: Smooth Bernstein(Left) and q−Bernstein(Right) CDF Estimator
for 0.65 · Beta(2, 10) + 0.35 · Beta(10, 2) Samples.
Remark: red - true density; dash - q−Bernstein with m = n; blue -
q−Bernstein with parameter q, M
14
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 34 / 47
0.0 0.2 0.4 0.6 0.8 1.0
0.00.20.40.60.81.0
x
cdf
(a) n = 20, q = 0.94, m = 7, M = 147
0.0 0.2 0.4 0.6 0.8 1.0
0.00.20.40.60.81.0
x
cdf
(b) n = 20, q = 0.96, m = 7, M = 206
0.0 0.2 0.4 0.6 0.8 1.0
0.00.20.40.60.81.0
x
cdf
(c) n = 50, q = 1, m = 13, M = 1024
0.0 0.2 0.4 0.6 0.8 1.0
0.00.20.40.60.81.0
x
cdf
(d) n = 50, q = 0.99, m = 13, M = 105
0.0 0.2 0.4 0.6 0.8 1.0
0.00.20.40.60.81.0
x
cdf
(e) n = 100, q = 1, m = 22, M = 801
0.0 0.2 0.4 0.6 0.8 1.0
0.00.20.40.60.81.0
x
cdf
(f) n = 100, q = 1, m = 22, M = 611
Figure 9: Smooth Bernstein and q−Bernstein CDF Estimator for
Beta(3, 3)(Left) and Beta(2, 4)(Right) Samples.
Remark: red is theoretical, black is Bernstein ,dash is q−Bernstein with
M = m, blue is q−Bernstein with parameter q, M
15
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 35 / 47
0.0 0.2 0.4 0.6 0.8 1.0
0.00.20.40.60.81.0
x
cdf
(a) n = 20, q = 0.99, m = 7, M = 73
0.0 0.2 0.4 0.6 0.8 1.0
0.00.20.40.60.81.0
x
cdf
(b) n = 50, q = 1, m = 13, M = 355
0.0 0.2 0.4 0.6 0.8 1.0
0.00.20.40.60.81.0
x
cdf
(c) n = 100, q = 1, m = 22, M = 371
Figure 10: Smooth Bernstein(Left) and q−Bernstein(Right) CDF Estimator
for 0.65 · Beta(2, 10) + 0.35 · Beta(10, 2) Samples.
Remark: red is theoretical, dash is q−Bernstein with M = m, blue is
q−Bernstein with parameter q, M
16
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 36 / 47
Density Function-Illustration
In Figures 11-12, we present the plots of density estimators corresponding
to the three distributions based on a random sample for sample sizes
n = 20, 50, 100, where m = n/ log(n) for different values of q ∈ [.85, 1].
We will use the following colors to highlight specific q’s.
We will use red for the theoretical density, while use green3 for q = 0.85,
lightseagreen for q = 0.9, olivedrab for q = 0.95 and darkorange3 for
optimized q and blue for the Bernstein density estimator. For optimized
q ∈ {0.85, 0.9, 0.95}, we will display it in darkorange3. Note that we have
not considered q−Bernstein polynomial density estimator for the M, q
case, as in this situation the density shape is very rough. These are
illustrated in Figures 9-10.
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 37 / 47
0.0 0.2 0.4 0.6 0.8 1.0
0.00.51.01.5
x
density
(a) n = 20, m = 7, q = 0.85
0.0 0.2 0.4 0.6 0.8 1.0
0.00.51.01.52.0
x
density
(b) n = 20, m = 7, q = 0.85
0.0 0.2 0.4 0.6 0.8 1.0
0.00.51.01.5
x
density
(c) n = 50, m = 13, q = 0.96
0.0 0.2 0.4 0.6 0.8 1.0
0.00.51.01.52.0
x
density
(d) n = 50, m = 13, q = 0.96
0.0 0.2 0.4 0.6 0.8 1.0
0.00.51.01.5
x
density
(e) n = 100, m = 22, q = 0.99
0.0 0.2 0.4 0.6 0.8 1.0
0.00.51.01.52.0
x
density
(f) n = 100, m = 22, q = 0.99
Figure 11: Comparison of Bernstein and q−Bernstein PDF estimator with
q ∈ [0.85, 1] for Beta(3, 3)(left) and Beta(2, 4)(right)
18
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 38 / 47
0.0 0.2 0.4 0.6 0.8 1.0
0.00.51.01.52.02.5
xcdf
(a) n = 20, m = 7, q = 0.95
0.0 0.2 0.4 0.6 0.8 1.0
0.00.51.01.52.02.5
x
cdf
(b) n = 50, m = 13, q = 0.99
0.0 0.2 0.4 0.6 0.8 1.0
0.00.51.01.52.02.5
x
cdf
(c) n = 100, m = 22, q = 1
Figure 12: Comparison of Bernstein and q−Bernstein PDF estimator with
q ∈ [0.85, 1] for 0.65 · Beta(2, 10) + 0.35 · Beta(10, 2)
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 39 / 47
Average Squared Error Comparison
The following tables display the average squared errors (ASE) based on
B = 40 comparisons, where ASE(g) = 1
B (˜g(xi) − g(xi))2, where g is
the true distribution or density, and ˜g is its nonparametric estimator.
Distribution Sample size Bernstein q−Bernstein
20 0.0020684 0.0004371
Beta(3,3) 50 0.0037048 0.0043542
100 0.0003380 0.0001946
20 0.0036687 0.0021716
Beta(2,4) 50 0.0049030 0.0041611
100 0.0013453 0.0006329
20 0.0014412 0.0036309
Mixture 50 0.0012856 0.0032519
100 0.0041389 0.0046857
Table: Average Square Error for Bernstein and q−Bernstein CDF estimator with
optimized q
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 40 / 47
Average Squared Error Comparison
Distribution Sample size Bernstein q−Bernstein
20 0.1388213 0.0760121
Beta(3,3) 50 0.0278798 0.0669346
100 0.0182203 0.0185668
20 0.1669993 0.1340892
Beta(2,4) 50 0.0488355 0.0780358
100 0.0221844 0.0219737
20 0.4132526 0.5370016
Mixture 50 0.1991615 0.2336240
100 0.2113997 0.1938492
Table: Average Square Error for Bernstein and q−Bernstein PDF estimator with
optimized q
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 41 / 47
By looking at the figures, we see that as the sample size increase, the ASE
has a decreasing trend. The distribution function estimator may have
smaller errors specially for small sample sizes. For density estimators, q−
Bernstein polynomial may have slight advantage over the Bernstein
polynomial that too for small samples and unimodal distributions. Overall,
we can say that the classical Bernstein polynomial estimator for estimating
density may still be considered superior over the general q− Bernstein
polynomial estimator, however, the q−Bernstein polynomial estimator for
the distribution function might be potentially better. This requires further
investigation, especially in the manner in which we have selected M. The
typical question is that of finding optimal q when m = ∞?
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 42 / 47
References
1 Babu, G.J., Canty, A.J. and Chaubey, Y.P. (2002). Application of
Bernstein polynomials for smooth estimation of a distribution and
density function. Journal of Statistical Planning and Inference 105,
377-392.
2 Bernstein, S.N. (1912). D´emonstration du th´eor`eme de Weierstrass,
fonde´e sur le calcul des probabilit´es. Commun. Soc. Math. Kharkov,
No.1. Series XIII, 1–2.
3 Bowman, A.W. (1984). An alternative method of cross-validation for
the smoothing of density estimates.Biometrika 71, 353-360.
4 Cao, J.D. (1997). A generalization of the Bernstein polynomials.
Journal Of Mathematical Analysis And Applications 209, 140–146.
5 Chaubey, Y. P., & Sen, P. K. (2009). On the selection of the
smoothing parameter in poisson smoothing of histogram estimator:
Computational aspects. Pakistan Journal of Statistics 25, 385-401.
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 43 / 47
References
6 Gawronski,W. and StadtM¨uller, U. (1981). Smoothing histograms by
means of lattice- and continuous distributions. Metrika 28, 155164.
7 Ghosal, S. (2001). Convergence rates for density estimation with
bernstein polynomials. The Annals of Statistics 29, 1264–1280.
8 H¨ardle, W. (1991). Smoothing Techniques with Implementation in S.
Springer-Verlag: New York.
9 Igarashi, G. and Yoshihide Kakizawa, Y. (2014). On improving
convergence rate of Bernstein polynomial density estimator. Journal
of Nonparametric Statistics 26, 6184.
10 Il’inskii, A. and Ostrovska, S. (2002). Convergence of generalized
Bernstein polynomials. Journal of Approximation Theory 116,
100–112.
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 44 / 47
References
11 Kakizawa, Y. (2004). Bernstein polynomial probability density
estimation. Journal of Nonparametric Statistics 16, 709–729.
12 Kakizawa, Y. (2011). A note on generalized bernstein polynomial
density estimators. Statistical Methodology 8, 136153.
13 Leblanc, A. (2009). ChungSmirnov property for Bernstein estimators
of distribution functions. Journal of Nonparametric Statistics 21,
133142.
14 Leblanc, A. (2012). On estimating distribution functions using
Bernstein polynomials. Ann Inst Stat Math 64, 919943.
15 Lorentz, G. G. (1986). Bernstein Polynomials (2nd ed.) New York:
Chelsea Publishing.
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 45 / 47
References
16 Phillips, G. M. (1997). Bernstein polynomials based on the q-integers.
The heritage of P. L. Chebyshev: a Festschrift in honor of the 70th
birthday of T. J. Rivlin. Ann. Numer. Math. 14, 511–518.
17 Prakasa Rao, B. L. S. (2005). Estimation of distribution and density
functions by generalized Bernstein polynomials. Indian Journal of
Pure and Applied Mathematics 36, 63–88.
18 Silverman, B. W. (1986). Density estimation for statistics and data
analysis, Chapman and Hall, London.
19 StadtM¨uller, U. (1983). Asymptotic distributions of smoothed
histograms. Metrika 30, 145–158.
20 Vitale, R.A. (1973). A Bernstein polynomial approach to density
estimation. Comm. Statist. 2, 493–506.
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 46 / 47
Talk slides will be available on SlideShare:
www.slideshare.net/ychaubey/talk-slides-msast2016-70014046
THANKS!!
Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 47 / 47

More Related Content

What's hot

Compatible Mapping and Common Fixed Point Theorem
Compatible Mapping and Common Fixed Point TheoremCompatible Mapping and Common Fixed Point Theorem
Compatible Mapping and Common Fixed Point TheoremIOSR Journals
 
A common fixed point theorem in cone metric spaces
A common fixed point theorem in cone metric spacesA common fixed point theorem in cone metric spaces
A common fixed point theorem in cone metric spacesAlexander Decker
 
Fixed points of contractive and Geraghty contraction mappings under the influ...
Fixed points of contractive and Geraghty contraction mappings under the influ...Fixed points of contractive and Geraghty contraction mappings under the influ...
Fixed points of contractive and Geraghty contraction mappings under the influ...IJERA Editor
 
Chaubey seminarslides2017
Chaubey seminarslides2017Chaubey seminarslides2017
Chaubey seminarslides2017ychaubey
 
A Note on “   Geraghty contraction type mappings”
A Note on “   Geraghty contraction type mappings”A Note on “   Geraghty contraction type mappings”
A Note on “   Geraghty contraction type mappings”IOSRJM
 
Kumaraswamy distributin:
Kumaraswamy distributin:Kumaraswamy distributin:
Kumaraswamy distributin:Pankaj Das
 
Common fixed point theorems using faintly compatible
Common fixed point theorems using faintly compatibleCommon fixed point theorems using faintly compatible
Common fixed point theorems using faintly compatibleAlexander Decker
 
Structured Regularization for conditional Gaussian graphical model
Structured Regularization for conditional Gaussian graphical modelStructured Regularization for conditional Gaussian graphical model
Structured Regularization for conditional Gaussian graphical modelLaboratoire Statistique et génome
 
11.common fixed points of weakly reciprocally continuous maps using a gauge f...
11.common fixed points of weakly reciprocally continuous maps using a gauge f...11.common fixed points of weakly reciprocally continuous maps using a gauge f...
11.common fixed points of weakly reciprocally continuous maps using a gauge f...Alexander Decker
 
Common fixed points of weakly reciprocally continuous maps using a gauge func...
Common fixed points of weakly reciprocally continuous maps using a gauge func...Common fixed points of weakly reciprocally continuous maps using a gauge func...
Common fixed points of weakly reciprocally continuous maps using a gauge func...Alexander Decker
 
Delayed acceptance for Metropolis-Hastings algorithms
Delayed acceptance for Metropolis-Hastings algorithmsDelayed acceptance for Metropolis-Hastings algorithms
Delayed acceptance for Metropolis-Hastings algorithmsChristian Robert
 
Kernel estimation(ref)
Kernel estimation(ref)Kernel estimation(ref)
Kernel estimation(ref)Zahra Amini
 
The existence of common fixed point theorems of generalized contractive mappi...
The existence of common fixed point theorems of generalized contractive mappi...The existence of common fixed point theorems of generalized contractive mappi...
The existence of common fixed point theorems of generalized contractive mappi...Alexander Decker
 
Generalized fixed point theorems for compatible mapping in fuzzy 2 metric spa...
Generalized fixed point theorems for compatible mapping in fuzzy 2 metric spa...Generalized fixed point theorems for compatible mapping in fuzzy 2 metric spa...
Generalized fixed point theorems for compatible mapping in fuzzy 2 metric spa...Alexander Decker
 

What's hot (20)

Compatible Mapping and Common Fixed Point Theorem
Compatible Mapping and Common Fixed Point TheoremCompatible Mapping and Common Fixed Point Theorem
Compatible Mapping and Common Fixed Point Theorem
 
A common fixed point theorem in cone metric spaces
A common fixed point theorem in cone metric spacesA common fixed point theorem in cone metric spaces
A common fixed point theorem in cone metric spaces
 
Fixed points of contractive and Geraghty contraction mappings under the influ...
Fixed points of contractive and Geraghty contraction mappings under the influ...Fixed points of contractive and Geraghty contraction mappings under the influ...
Fixed points of contractive and Geraghty contraction mappings under the influ...
 
Chaubey seminarslides2017
Chaubey seminarslides2017Chaubey seminarslides2017
Chaubey seminarslides2017
 
Cluster
ClusterCluster
Cluster
 
A Note on “   Geraghty contraction type mappings”
A Note on “   Geraghty contraction type mappings”A Note on “   Geraghty contraction type mappings”
A Note on “   Geraghty contraction type mappings”
 
Kumaraswamy distributin:
Kumaraswamy distributin:Kumaraswamy distributin:
Kumaraswamy distributin:
 
Common fixed point theorems using faintly compatible
Common fixed point theorems using faintly compatibleCommon fixed point theorems using faintly compatible
Common fixed point theorems using faintly compatible
 
ABC in Venezia
ABC in VeneziaABC in Venezia
ABC in Venezia
 
Structured Regularization for conditional Gaussian graphical model
Structured Regularization for conditional Gaussian graphical modelStructured Regularization for conditional Gaussian graphical model
Structured Regularization for conditional Gaussian graphical model
 
11.common fixed points of weakly reciprocally continuous maps using a gauge f...
11.common fixed points of weakly reciprocally continuous maps using a gauge f...11.common fixed points of weakly reciprocally continuous maps using a gauge f...
11.common fixed points of weakly reciprocally continuous maps using a gauge f...
 
Common fixed points of weakly reciprocally continuous maps using a gauge func...
Common fixed points of weakly reciprocally continuous maps using a gauge func...Common fixed points of weakly reciprocally continuous maps using a gauge func...
Common fixed points of weakly reciprocally continuous maps using a gauge func...
 
Distributed perceptron
Distributed perceptronDistributed perceptron
Distributed perceptron
 
Delayed acceptance for Metropolis-Hastings algorithms
Delayed acceptance for Metropolis-Hastings algorithmsDelayed acceptance for Metropolis-Hastings algorithms
Delayed acceptance for Metropolis-Hastings algorithms
 
41-50
41-5041-50
41-50
 
quantum gravity
quantum gravityquantum gravity
quantum gravity
 
Kernel estimation(ref)
Kernel estimation(ref)Kernel estimation(ref)
Kernel estimation(ref)
 
The existence of common fixed point theorems of generalized contractive mappi...
The existence of common fixed point theorems of generalized contractive mappi...The existence of common fixed point theorems of generalized contractive mappi...
The existence of common fixed point theorems of generalized contractive mappi...
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
Generalized fixed point theorems for compatible mapping in fuzzy 2 metric spa...
Generalized fixed point theorems for compatible mapping in fuzzy 2 metric spa...Generalized fixed point theorems for compatible mapping in fuzzy 2 metric spa...
Generalized fixed point theorems for compatible mapping in fuzzy 2 metric spa...
 

Similar to Talk slides msast2016

TS-IASSL2014
TS-IASSL2014TS-IASSL2014
TS-IASSL2014ychaubey
 
Introduction to Evidential Neural Networks
Introduction to Evidential Neural NetworksIntroduction to Evidential Neural Networks
Introduction to Evidential Neural NetworksFederico Cerutti
 
Theory of Relational Calculus and its Formalization
Theory of Relational Calculus and its FormalizationTheory of Relational Calculus and its Formalization
Theory of Relational Calculus and its FormalizationYoshihiro Mizoguchi
 
Talk at 2013 WSC, ISI Conference in Hong Kong, August 26, 2013
Talk at 2013 WSC, ISI Conference in Hong Kong, August 26, 2013Talk at 2013 WSC, ISI Conference in Hong Kong, August 26, 2013
Talk at 2013 WSC, ISI Conference in Hong Kong, August 26, 2013Christian Robert
 
Hso201 a solutions to practice problem set 3-2015-16-ii
Hso201 a solutions to practice problem set 3-2015-16-iiHso201 a solutions to practice problem set 3-2015-16-ii
Hso201 a solutions to practice problem set 3-2015-16-iiOm Prakash Bankolia
 
Probability cheatsheet
Probability cheatsheetProbability cheatsheet
Probability cheatsheetJoachim Gwoke
 
Slides csm
Slides csmSlides csm
Slides csmychaubey
 
Welcome to International Journal of Engineering Research and Development (IJERD)
Welcome to International Journal of Engineering Research and Development (IJERD)Welcome to International Journal of Engineering Research and Development (IJERD)
Welcome to International Journal of Engineering Research and Development (IJERD)IJERD Editor
 
Existence results for fractional q-differential equations with integral and m...
Existence results for fractional q-differential equations with integral and m...Existence results for fractional q-differential equations with integral and m...
Existence results for fractional q-differential equations with integral and m...IJRTEMJOURNAL
 
A Family Of Extragradient Methods For Solving Equilibrium Problems
A Family Of Extragradient Methods For Solving Equilibrium ProblemsA Family Of Extragradient Methods For Solving Equilibrium Problems
A Family Of Extragradient Methods For Solving Equilibrium ProblemsYasmine Anino
 
bau-22-1.pdf
bau-22-1.pdfbau-22-1.pdf
bau-22-1.pdfShengBau
 
Fixed point theorem in fuzzy metric space with e.a property
Fixed point theorem in fuzzy metric space with e.a propertyFixed point theorem in fuzzy metric space with e.a property
Fixed point theorem in fuzzy metric space with e.a propertyAlexander Decker
 
Dimension Reduction Introduction & PCA.pptx
Dimension Reduction Introduction & PCA.pptxDimension Reduction Introduction & PCA.pptx
Dimension Reduction Introduction & PCA.pptxRohanBorgalli
 
Markov chain Monte Carlo methods and some attempts at parallelizing them
Markov chain Monte Carlo methods and some attempts at parallelizing themMarkov chain Monte Carlo methods and some attempts at parallelizing them
Markov chain Monte Carlo methods and some attempts at parallelizing themPierre Jacob
 
Different kind of distance and Statistical Distance
Different kind of distance and Statistical DistanceDifferent kind of distance and Statistical Distance
Different kind of distance and Statistical DistanceKhulna University
 
Probability and Statistics
Probability and StatisticsProbability and Statistics
Probability and StatisticsMalik Sb
 
Vector measures and classical disjointification methods
Vector measures and classical disjointification methodsVector measures and classical disjointification methods
Vector measures and classical disjointification methodsesasancpe
 
Vector measures and classical disjointification methods...
Vector measures and classical disjointification methods...Vector measures and classical disjointification methods...
Vector measures and classical disjointification methods...esasancpe
 

Similar to Talk slides msast2016 (20)

TS-IASSL2014
TS-IASSL2014TS-IASSL2014
TS-IASSL2014
 
Introduction to Evidential Neural Networks
Introduction to Evidential Neural NetworksIntroduction to Evidential Neural Networks
Introduction to Evidential Neural Networks
 
Theory of Relational Calculus and its Formalization
Theory of Relational Calculus and its FormalizationTheory of Relational Calculus and its Formalization
Theory of Relational Calculus and its Formalization
 
Talk at 2013 WSC, ISI Conference in Hong Kong, August 26, 2013
Talk at 2013 WSC, ISI Conference in Hong Kong, August 26, 2013Talk at 2013 WSC, ISI Conference in Hong Kong, August 26, 2013
Talk at 2013 WSC, ISI Conference in Hong Kong, August 26, 2013
 
Hso201 a solutions to practice problem set 3-2015-16-ii
Hso201 a solutions to practice problem set 3-2015-16-iiHso201 a solutions to practice problem set 3-2015-16-ii
Hso201 a solutions to practice problem set 3-2015-16-ii
 
Probability cheatsheet
Probability cheatsheetProbability cheatsheet
Probability cheatsheet
 
Slides csm
Slides csmSlides csm
Slides csm
 
Welcome to International Journal of Engineering Research and Development (IJERD)
Welcome to International Journal of Engineering Research and Development (IJERD)Welcome to International Journal of Engineering Research and Development (IJERD)
Welcome to International Journal of Engineering Research and Development (IJERD)
 
2018 MUMS Fall Course - Problem Set 1 (Attachment for Class Presentation) - J...
2018 MUMS Fall Course - Problem Set 1 (Attachment for Class Presentation) - J...2018 MUMS Fall Course - Problem Set 1 (Attachment for Class Presentation) - J...
2018 MUMS Fall Course - Problem Set 1 (Attachment for Class Presentation) - J...
 
Existence results for fractional q-differential equations with integral and m...
Existence results for fractional q-differential equations with integral and m...Existence results for fractional q-differential equations with integral and m...
Existence results for fractional q-differential equations with integral and m...
 
A Family Of Extragradient Methods For Solving Equilibrium Problems
A Family Of Extragradient Methods For Solving Equilibrium ProblemsA Family Of Extragradient Methods For Solving Equilibrium Problems
A Family Of Extragradient Methods For Solving Equilibrium Problems
 
bau-22-1.pdf
bau-22-1.pdfbau-22-1.pdf
bau-22-1.pdf
 
Fixed point theorem in fuzzy metric space with e.a property
Fixed point theorem in fuzzy metric space with e.a propertyFixed point theorem in fuzzy metric space with e.a property
Fixed point theorem in fuzzy metric space with e.a property
 
Dimension Reduction Introduction & PCA.pptx
Dimension Reduction Introduction & PCA.pptxDimension Reduction Introduction & PCA.pptx
Dimension Reduction Introduction & PCA.pptx
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
Markov chain Monte Carlo methods and some attempts at parallelizing them
Markov chain Monte Carlo methods and some attempts at parallelizing themMarkov chain Monte Carlo methods and some attempts at parallelizing them
Markov chain Monte Carlo methods and some attempts at parallelizing them
 
Different kind of distance and Statistical Distance
Different kind of distance and Statistical DistanceDifferent kind of distance and Statistical Distance
Different kind of distance and Statistical Distance
 
Probability and Statistics
Probability and StatisticsProbability and Statistics
Probability and Statistics
 
Vector measures and classical disjointification methods
Vector measures and classical disjointification methodsVector measures and classical disjointification methods
Vector measures and classical disjointification methods
 
Vector measures and classical disjointification methods...
Vector measures and classical disjointification methods...Vector measures and classical disjointification methods...
Vector measures and classical disjointification methods...
 

Recently uploaded

NATIONAL ANTHEMS OF AFRICA (National Anthems of Africa)
NATIONAL ANTHEMS OF AFRICA (National Anthems of Africa)NATIONAL ANTHEMS OF AFRICA (National Anthems of Africa)
NATIONAL ANTHEMS OF AFRICA (National Anthems of Africa)Basil Achie
 
LANDMARKS AND MONUMENTS IN NIGERIA.pptx
LANDMARKS  AND MONUMENTS IN NIGERIA.pptxLANDMARKS  AND MONUMENTS IN NIGERIA.pptx
LANDMARKS AND MONUMENTS IN NIGERIA.pptxBasil Achie
 
Presentation for the Strategic Dialogue on the Future of Agriculture, Brussel...
Presentation for the Strategic Dialogue on the Future of Agriculture, Brussel...Presentation for the Strategic Dialogue on the Future of Agriculture, Brussel...
Presentation for the Strategic Dialogue on the Future of Agriculture, Brussel...Krijn Poppe
 
CTAC 2024 Valencia - Sven Zoelle - Most Crucial Invest to Digitalisation_slid...
CTAC 2024 Valencia - Sven Zoelle - Most Crucial Invest to Digitalisation_slid...CTAC 2024 Valencia - Sven Zoelle - Most Crucial Invest to Digitalisation_slid...
CTAC 2024 Valencia - Sven Zoelle - Most Crucial Invest to Digitalisation_slid...henrik385807
 
Simulation-based Testing of Unmanned Aerial Vehicles with Aerialist
Simulation-based Testing of Unmanned Aerial Vehicles with AerialistSimulation-based Testing of Unmanned Aerial Vehicles with Aerialist
Simulation-based Testing of Unmanned Aerial Vehicles with AerialistSebastiano Panichella
 
Open Source Strategy in Logistics 2015_Henrik Hankedvz-d-nl-log-conference.pdf
Open Source Strategy in Logistics 2015_Henrik Hankedvz-d-nl-log-conference.pdfOpen Source Strategy in Logistics 2015_Henrik Hankedvz-d-nl-log-conference.pdf
Open Source Strategy in Logistics 2015_Henrik Hankedvz-d-nl-log-conference.pdfhenrik385807
 
Call Girls in Rohini Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Rohini Delhi 💯Call Us 🔝8264348440🔝Call Girls in Rohini Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Rohini Delhi 💯Call Us 🔝8264348440🔝soniya singh
 
Gaps, Issues and Challenges in the Implementation of Mother Tongue Based-Mult...
Gaps, Issues and Challenges in the Implementation of Mother Tongue Based-Mult...Gaps, Issues and Challenges in the Implementation of Mother Tongue Based-Mult...
Gaps, Issues and Challenges in the Implementation of Mother Tongue Based-Mult...marjmae69
 
Mathan flower ppt.pptx slide orchids ✨🌸
Mathan flower ppt.pptx slide orchids ✨🌸Mathan flower ppt.pptx slide orchids ✨🌸
Mathan flower ppt.pptx slide orchids ✨🌸mathanramanathan2005
 
Philippine History cavite Mutiny Report.ppt
Philippine History cavite Mutiny Report.pptPhilippine History cavite Mutiny Report.ppt
Philippine History cavite Mutiny Report.pptssuser319dad
 
SBFT Tool Competition 2024 -- Python Test Case Generation Track
SBFT Tool Competition 2024 -- Python Test Case Generation TrackSBFT Tool Competition 2024 -- Python Test Case Generation Track
SBFT Tool Competition 2024 -- Python Test Case Generation TrackSebastiano Panichella
 
Open Source Camp Kubernetes 2024 | Running WebAssembly on Kubernetes by Alex ...
Open Source Camp Kubernetes 2024 | Running WebAssembly on Kubernetes by Alex ...Open Source Camp Kubernetes 2024 | Running WebAssembly on Kubernetes by Alex ...
Open Source Camp Kubernetes 2024 | Running WebAssembly on Kubernetes by Alex ...NETWAYS
 
Genesis part 2 Isaiah Scudder 04-24-2024.pptx
Genesis part 2 Isaiah Scudder 04-24-2024.pptxGenesis part 2 Isaiah Scudder 04-24-2024.pptx
Genesis part 2 Isaiah Scudder 04-24-2024.pptxFamilyWorshipCenterD
 
OSCamp Kubernetes 2024 | Zero-Touch OS-Infrastruktur für Container und Kubern...
OSCamp Kubernetes 2024 | Zero-Touch OS-Infrastruktur für Container und Kubern...OSCamp Kubernetes 2024 | Zero-Touch OS-Infrastruktur für Container und Kubern...
OSCamp Kubernetes 2024 | Zero-Touch OS-Infrastruktur für Container und Kubern...NETWAYS
 
call girls in delhi malviya nagar @9811711561@
call girls in delhi malviya nagar @9811711561@call girls in delhi malviya nagar @9811711561@
call girls in delhi malviya nagar @9811711561@vikas rana
 
The 3rd Intl. Workshop on NL-based Software Engineering
The 3rd Intl. Workshop on NL-based Software EngineeringThe 3rd Intl. Workshop on NL-based Software Engineering
The 3rd Intl. Workshop on NL-based Software EngineeringSebastiano Panichella
 
OSCamp Kubernetes 2024 | SRE Challenges in Monolith to Microservices Shift at...
OSCamp Kubernetes 2024 | SRE Challenges in Monolith to Microservices Shift at...OSCamp Kubernetes 2024 | SRE Challenges in Monolith to Microservices Shift at...
OSCamp Kubernetes 2024 | SRE Challenges in Monolith to Microservices Shift at...NETWAYS
 
Genshin Impact PPT Template by EaTemp.pptx
Genshin Impact PPT Template by EaTemp.pptxGenshin Impact PPT Template by EaTemp.pptx
Genshin Impact PPT Template by EaTemp.pptxJohnree4
 
The Ten Facts About People With Autism Presentation
The Ten Facts About People With Autism PresentationThe Ten Facts About People With Autism Presentation
The Ten Facts About People With Autism PresentationNathan Young
 
Work Remotely with Confluence ACE 2.pptx
Work Remotely with Confluence ACE 2.pptxWork Remotely with Confluence ACE 2.pptx
Work Remotely with Confluence ACE 2.pptxmavinoikein
 

Recently uploaded (20)

NATIONAL ANTHEMS OF AFRICA (National Anthems of Africa)
NATIONAL ANTHEMS OF AFRICA (National Anthems of Africa)NATIONAL ANTHEMS OF AFRICA (National Anthems of Africa)
NATIONAL ANTHEMS OF AFRICA (National Anthems of Africa)
 
LANDMARKS AND MONUMENTS IN NIGERIA.pptx
LANDMARKS  AND MONUMENTS IN NIGERIA.pptxLANDMARKS  AND MONUMENTS IN NIGERIA.pptx
LANDMARKS AND MONUMENTS IN NIGERIA.pptx
 
Presentation for the Strategic Dialogue on the Future of Agriculture, Brussel...
Presentation for the Strategic Dialogue on the Future of Agriculture, Brussel...Presentation for the Strategic Dialogue on the Future of Agriculture, Brussel...
Presentation for the Strategic Dialogue on the Future of Agriculture, Brussel...
 
CTAC 2024 Valencia - Sven Zoelle - Most Crucial Invest to Digitalisation_slid...
CTAC 2024 Valencia - Sven Zoelle - Most Crucial Invest to Digitalisation_slid...CTAC 2024 Valencia - Sven Zoelle - Most Crucial Invest to Digitalisation_slid...
CTAC 2024 Valencia - Sven Zoelle - Most Crucial Invest to Digitalisation_slid...
 
Simulation-based Testing of Unmanned Aerial Vehicles with Aerialist
Simulation-based Testing of Unmanned Aerial Vehicles with AerialistSimulation-based Testing of Unmanned Aerial Vehicles with Aerialist
Simulation-based Testing of Unmanned Aerial Vehicles with Aerialist
 
Open Source Strategy in Logistics 2015_Henrik Hankedvz-d-nl-log-conference.pdf
Open Source Strategy in Logistics 2015_Henrik Hankedvz-d-nl-log-conference.pdfOpen Source Strategy in Logistics 2015_Henrik Hankedvz-d-nl-log-conference.pdf
Open Source Strategy in Logistics 2015_Henrik Hankedvz-d-nl-log-conference.pdf
 
Call Girls in Rohini Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Rohini Delhi 💯Call Us 🔝8264348440🔝Call Girls in Rohini Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Rohini Delhi 💯Call Us 🔝8264348440🔝
 
Gaps, Issues and Challenges in the Implementation of Mother Tongue Based-Mult...
Gaps, Issues and Challenges in the Implementation of Mother Tongue Based-Mult...Gaps, Issues and Challenges in the Implementation of Mother Tongue Based-Mult...
Gaps, Issues and Challenges in the Implementation of Mother Tongue Based-Mult...
 
Mathan flower ppt.pptx slide orchids ✨🌸
Mathan flower ppt.pptx slide orchids ✨🌸Mathan flower ppt.pptx slide orchids ✨🌸
Mathan flower ppt.pptx slide orchids ✨🌸
 
Philippine History cavite Mutiny Report.ppt
Philippine History cavite Mutiny Report.pptPhilippine History cavite Mutiny Report.ppt
Philippine History cavite Mutiny Report.ppt
 
SBFT Tool Competition 2024 -- Python Test Case Generation Track
SBFT Tool Competition 2024 -- Python Test Case Generation TrackSBFT Tool Competition 2024 -- Python Test Case Generation Track
SBFT Tool Competition 2024 -- Python Test Case Generation Track
 
Open Source Camp Kubernetes 2024 | Running WebAssembly on Kubernetes by Alex ...
Open Source Camp Kubernetes 2024 | Running WebAssembly on Kubernetes by Alex ...Open Source Camp Kubernetes 2024 | Running WebAssembly on Kubernetes by Alex ...
Open Source Camp Kubernetes 2024 | Running WebAssembly on Kubernetes by Alex ...
 
Genesis part 2 Isaiah Scudder 04-24-2024.pptx
Genesis part 2 Isaiah Scudder 04-24-2024.pptxGenesis part 2 Isaiah Scudder 04-24-2024.pptx
Genesis part 2 Isaiah Scudder 04-24-2024.pptx
 
OSCamp Kubernetes 2024 | Zero-Touch OS-Infrastruktur für Container und Kubern...
OSCamp Kubernetes 2024 | Zero-Touch OS-Infrastruktur für Container und Kubern...OSCamp Kubernetes 2024 | Zero-Touch OS-Infrastruktur für Container und Kubern...
OSCamp Kubernetes 2024 | Zero-Touch OS-Infrastruktur für Container und Kubern...
 
call girls in delhi malviya nagar @9811711561@
call girls in delhi malviya nagar @9811711561@call girls in delhi malviya nagar @9811711561@
call girls in delhi malviya nagar @9811711561@
 
The 3rd Intl. Workshop on NL-based Software Engineering
The 3rd Intl. Workshop on NL-based Software EngineeringThe 3rd Intl. Workshop on NL-based Software Engineering
The 3rd Intl. Workshop on NL-based Software Engineering
 
OSCamp Kubernetes 2024 | SRE Challenges in Monolith to Microservices Shift at...
OSCamp Kubernetes 2024 | SRE Challenges in Monolith to Microservices Shift at...OSCamp Kubernetes 2024 | SRE Challenges in Monolith to Microservices Shift at...
OSCamp Kubernetes 2024 | SRE Challenges in Monolith to Microservices Shift at...
 
Genshin Impact PPT Template by EaTemp.pptx
Genshin Impact PPT Template by EaTemp.pptxGenshin Impact PPT Template by EaTemp.pptx
Genshin Impact PPT Template by EaTemp.pptx
 
The Ten Facts About People With Autism Presentation
The Ten Facts About People With Autism PresentationThe Ten Facts About People With Autism Presentation
The Ten Facts About People With Autism Presentation
 
Work Remotely with Confluence ACE 2.pptx
Work Remotely with Confluence ACE 2.pptxWork Remotely with Confluence ACE 2.pptx
Work Remotely with Confluence ACE 2.pptx
 

Talk slides msast2016

  • 1. Nonparametric Estimation of Distribution and Density Functions Using q-Bernstein Polynomials Yogendra P. Chaubey Department of Mathematics and Statistics Concordia University, Montreal, Canada H3G 1M8 E-mail:yogen.chaubey@concordia.ca Talk to be presented at the 6th International Conference of IMBIC, Kolkata, India, December 21-23, 2016 Acknowledgment: The computations were performed by Qi Zhang as part of his BSc (Honours) project. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 1 / 47
  • 2. Abstract Here we consider the smooth density estimation for the random variables that have a compact support. Without loss of generality, we restrict the support to the closed interval [0, 1], for which Bernstein polynomial estimator, originally proposed by Vitale (1973), is known to have some good properties, such as being a genuine density estimator and being free from boundary bias (see Babu, Canty and Chaubey (2002) and Kakizawa (2004)). Generalized version of these estimators were proposed by Prakasa Rao (2005) and Kakizawa (2011) by considering the generalized Bernstein polynomials introduced in Cao (1997). Some work on improving convergence rate of the Bernstein polynomial density estimator were carried out in Leblanc (2009, 2012) and Igarshi and Kakizawa (2014). In this paper, we consider yet another generalization of the Bernstein polynomials introduced by Phillips (1997), that may have some advantage over the usual Bernstein polynomial estimator. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 2 / 47
  • 3. 1 Introduction 2 Preliminaries 3 Smooth Estimators of the Distribution and Density Functions 4 Asymptotic Properties of Estimators Asymptotic Properties of ˜Fn,m,q Asymptotic Properties of ˜fn,m,q 5 Numerical Studies Cross Validation Method to Determine the Smoothing Parameter q Likelihood Based Cross Validation Integrated Square Error Cross Validation Least Square Cross Validation Illustration for Distribution Function Estimator Illustration for Density Estimator Comparing Estimators by Average Squared Error Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 3 / 47
  • 4. 1. Introduction Let h be a bounded continuous function defined on the interval [0, 1], then the corresponding Bernstein polynomial of degree m is defined as Bm(x; h) = m k=0 h k m bk(x, m), (1.1) where bk(x, m) = m k xk (1 − x)m−k , k = 0, 1, .... (1.2) This polynomial was proposed as an uniform approximation to bounded continuous functions defined on [0, 1] by S.N. Bernstein (1912), that provided an existential proof of Weierstrass approximation theorem. The monograph by Lorentz (1986) makes an excellent reference regarding the properties of the Bernstein polynomials. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 4 / 47
  • 5. 1. Introduction Assuming that the function h has a continuous derivative, the derivative Bm(x; h) given by Bm(x; h) = m m−1 k=0 h k + 1 m − h k m bk(x, m − 1) (1.3) uniformly approximates the derivative h (x). Given a random sample {X1, ..., Xn} from an absolutely continuous distribution function F, we can estimate it by the empirical distribution function Fn(x) = 1 n n i=1 I{Ti ≤ x}. (1.4) Since the above function is a step function, it is not appropriate for finding the density function f = F , thus smoothing of this function is desired. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 5 / 47
  • 6. 1. Introduction Replacing h by the empirical distribution function Fn, the Bernstein polynomial approximation mentioned earlier motivates the following smooth estimator of the distribution function ˜Fn,m(x) = m k=0 Fn k m bk(x, m). (1.5) This estimator is a genuine distribution function and therefore its derivative serves as a smooth density estimator given by ˜fn,m(x) = Bm(x; Fn) = m m−1 k=0 Fn k + 1 m − Fn k m bk(x, m − 1) (1.6) Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 6 / 47
  • 7. 1. Introduction Bernstein polynomial density estimator: Originally proposed by Vitale (1973) and later investigated by many authors Gowronski and Stadm¨uler (1981): Smoothing of histogram using Feller’s Theorem Stadtm¨uller, (1983): Convergence of the density estimator Petrone (1999): Nonparametric Bayesian prior Ghosal, 2001: Convergence of nonparametric posteriors Babu, Canty and Chaubey (2002): Further studies Leblanc (2009, 2012): Smooth distribution function estimator. Prakasa Rao (2005) and Kakizawa (2011): Generalized Bernstein polynomials introduced by Cao (1997). Another generalization is introduced by Phillips (1997) that is the subject of this talk. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 7 / 47
  • 8. 2. Preliminaries Let q > 0, then for any k ∈ N ∪ {0}, the q−integer [k]q and q−factorial [k]q! are respectively defined by [k]q = (1 − qk)/(1 − q), for q = 1 k, for q = 1, (2.1) and [k]q! = [k]q[k − 1]q...[1]q, for k ≥ 1 1, for k = 0. (2.2) For integers 0 ≤ k ≤ m, the q−binomial coefficients are defined by m k q = [m]q! [k]q![m − k]q! (2.3) For q = 1, we clearly have [m]q = m, [m]q! = m! and m k q = m k , q−factorials and q−binomial coefficients reduce to the usual factorials and binomial coefficients, respectively. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 8 / 47
  • 9. 2. Preliminaries Definition (Phillips 1997). Let h ∈ C[0, 1]. The generalized Bernstein polynomial based on the q−integers is Bm,q(x; h) = m k=0 h [m]q [k]q bk,q(x, m) (2.4) where bk,q(x, m) = m k q xk m−k−1 s=0 (1 − qs x), m = 1, 2, ... and an empty product is taken to be equal to 1. The polynomials defined in (2.4) reduce to the usual Bernstein polynomials for q = 1, and hence they are referred to as the generalized Bernstein polynomials, however, in the sequel we will refer them as the q−Bernstein polynomials. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 9 / 47
  • 10. 2. Preliminaries The following analog of Bernstein’s Theorem for q−Bernstein polynomials was proved by Phillips (1997). Theorem Let a sequence {qm}∞ m=1 satisfy 0 < qm < 1 and qm → 1 as m → ∞. Then for any function h ∈ C[0, 1], Bm,qm (x; h) → h(x) [x ∈ [0, 1]; m → ∞]. Here the expression gm(x) → g(x)[x ∈ [0, 1]; m → ∞] denotes uniform convergence of gm to g with respect to x ∈ [0, 1]. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 10 / 47
  • 11. 2. Preliminaries The following theorem from Phillips (1996) gives convergence results of the derivative of the q−Bernstein polynomial Theorem Let h ∈ C1[0, 1] and let the sequence (qm) be chosen so that the sequence ( m) converges to zero from above faster than 1/3m, where m = m 1 + qm + q2 m + ... + qm−1 m − 1. Then the sequence of derivatives of the q−Bernstein polynomials converges uniformly on [0, 1] to h (x). Phillips (1996) showed that the condition on ( m) can be satisfied by choosing (qm) such that 1 − a−m m ≤ qm < 1 with a > 3. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 11 / 47
  • 12. 3. Estimators of the Distribution and Density Functions Similar to the estimator given in (1.5) using the Bernstein polynomial, the q− Bernstein polynomial may be used to define the following non-parametric smooth estimator of F. Thus, the q−Bernstein polynomial estimator of the distribution function F is given by ˜Fn;m,q(x) = m k=0 Fn [k]q [m]q bk,q(x, m), x ∈ [0, 1], (3.1) where bk,q(x, m) = m k q xk m−k−1 s=0 (1 − qs x). (3.2) For q = 1, the above estimator reduces to the usual Bernstein polynomial smooth estimator of the distribution function studied in Babu, Canty and Chaubey (2004). Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 12 / 47
  • 13. 2. Estimators of the Distribution and Density Functions By Euler identity (see Philips 1997) the product in (3.2) can be written as a sum and thus taking the derivative of ˜Fn;m,q, the following smooth estimator of f is obtained that will be called q−Bernstein density estimator: ˜fn;m,q(x) = m k=0 m−k s=0 Fn [k]q [m]q ωs,k(x; m, q) (3.3) where ωs,k(x; m, q) = m k q (−1)s q s(s−1) 2 m − k s q (k + s)xk+s−1 . This paper presents a comparative study of the resulting smooth estimators with Bernstein polynomial estimators while choosing the value of q by cross validation and the value of m = n/ log(n) as suggested in Babu, Canty and Chaubey (2004). Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 13 / 47
  • 14. 2. Estimators of the Distribution and Density Functions Since, the approximation improves as m → ∞, one may propose the use of limiting q−Bernstein polynomials, that was studies in Il’inskii and Ostrovska (2002). They noted that as m → ∞ for 0 < q < 1 lim m→∞ [k]q [m]q = 1 − qk lim m→∞ bk,q(x, m) = xk (1 − qk)[k]q! ∞ s=0 (1 − qs x) =: bk,∞(x, q) the limiting q−Bernstein polynomial becomes B∞,q(x, h) = ∞ k=0 h(1 − qk)bk,∞(x, q), if x ∈ [0, 1), h(1) if x = 1. Theorem For any f ∈ C[0, 1], B∞,q(x, h) → f(x)[x ∈ [0, 1]; q ↑ 1]. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 14 / 47
  • 15. 2. Estimators of the Distribution and Density Functions Theorem For any f ∈ C[0, 1], B∞,q(x, h) → f(x)[x ∈ [0, 1]; q ↑ 1]. Theorem Let 0 < α < 1. Then for any f ∈ C[0, 1], Bm,q(x, h) → h(x)[x ∈ [0, 1]; q ∈ [α, 1]; m → ∞]. The above theorems basically recast Phillip’s results by explicitly exhibiting the limiting form of Bm,q(x; h) as m → ∞. So, for general approximation, the problem of determining the constants m and q reduces to determining q only. In our experience, the corresponding smooth estimator of F is meaningful in practice, however, the corresponding density estimator may not be very smooth. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 15 / 47
  • 16. 2. Estimators of the Distribution and Density Functions In our experience, the corresponding smooth estimator of F is meaningful in practice, however, the corresponding density estimator may not be very smooth. Thus, for numerical studies, we propose to consider m = n/ log(n) [the choice suggested in Babu, Canty and Chaubey (2004)] and determine q by cross-validation. In the next section we provide the almost sure convergence results for the smooth estimators similar to those obtained in Babu, Canty and Chaubey (2004) using Bernstein polynomials. But, first some pictures of the estimators. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 16 / 47
  • 17. Distribution Function Estimator for Beta(3, 3) density 0.0 0.2 0.4 0.6 0.8 1.0 0.00.20.40.60.81.0 q−Bernstein Smooth Distribution Sample from Beta(3,3) Density, n=50 x Fn(x) Bernstein q−Bernstein,q=.99 Figure: 1. Distribution Function Estimators for Beta(3, 3) density; n = 50. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 17 / 47
  • 18. Distribution Function Estimator for Beta(3, 3) density 0.0 0.2 0.4 0.6 0.8 1.0 0.00.20.40.60.81.0 q−Bernstein Smooth Distribution Sample from Beta(3,3) Density, n=50 x Fn(x) Bernstein q−Bernstein,q=.99,m=200 Figure: 2. Distribution Function Estimators for Beta(3, 3) density; n = 50. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 18 / 47
  • 19. Density Function Estimator for Beta(3, 3) density q−Bernstein Smooth Desnity estimator Sample from Beta(3,3) Density, n=50 x fn(x) 0.0 0.2 0.4 0.6 0.8 1.0 0.00.51.01.52.0 Bernstein q−Bernstein,q=.99 Figure: 3. Density Function Estimators for Beta(3, 3) density; n = 50. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 19 / 47
  • 20. Density Function Estimator for Beta(3, 3) density q−Bernstein Smooth Desnity estimator Sample from Beta(3,3) Density, n=50 x fn(x) 0.0 0.2 0.4 0.6 0.8 1.0 0.00.51.01.52.0 Bernstein q−Bernstein,q=.99,m=50 Figure: 4. Density Function Estimators for Beta(3, 3) density; n = 50. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 20 / 47
  • 21. Density Function Estimator for Beta(3, 3) density q−Bernstein Smooth Desnity estimator Sample from Beta(3,3) Density, n=50 x fn(x) 0.0 0.2 0.4 0.6 0.8 1.0 0.00.51.01.52.0 Bernstein q−Bernstein,q=.99,m=25 Figure: 5. Density Function Estimators for Beta(3, 3) density; n = 50. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 21 / 47
  • 22. 4. Asymptotic Properties of Estimators Throughout this paper we use the notation G = sup x∈[0,1] |G(x)|, for a bounded function G on [0, 1], an = (n−1 log n)1/2 , and bn,m = (n−1 log n)1/2 (m−1 log m)1/4 . (4.1) The following theorem shows that ˜Fn,m is strongly consistent. Theorem If m, n → ∞, then ˜Fn,m,q − F → 0 a.s. as q → 1. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 22 / 47
  • 23. 4.1 Asymptotic Properties of Distribution Function Estimator Next theorem gives a result depicting the closeness of the smooth estimator with the empirical distribution function. Theorem Let F be continuous and differentiable on the interval [0, 1] with density f. If f is Lipschitz of order 1, then for n2/3 ≤ m ≤ (n/ log n)2, we have a.s. as n → ∞ , lim q↑1 ˜Fn,m − Fn = O (n−1 log n)1/2 (m−1 log m)1/4 . (4.2) For m = n, we have lim q↑1 ˜Fn,m − Fn = O(n−3/4 (log n)3/4 ) a.s.. (4.3) Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 23 / 47
  • 24. 4.2. Asymptotic Properties of Density Function Estimator We now establish the strong convergence of ˜fn(t) similar to that of ˜Fn(t). Theorem Let F be continuous on the interval [0, 1] with a continuous density f. Then for 2 ≤ m ≤ (n/ log n), we have a.s. as n → ∞ and q → 1, ˜fn,m,q − f = O(m1/2 an) + O( F∗ m − f ), (4.4) a.s. as n → ∞ , where F∗ m denotes the derivative of F∗ m. Consequently, if m = o(n/ log n), then ˜fn,m,q − f → 0 a.s. as m, n → ∞. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 24 / 47
  • 25. 4.2. Asymptotic Properties of Density Function Estimator The following theorem establishes the asymptotic normality of the ˜fn,m,q(x). Theorem Under conditions in theorem 3.1, if f(x) > 0 then as q → 1 n1/2 m1/4 ( ˜fn,m,q(x) − f(x)) D → N 0, f(x) 2 πx(1 − x) (4.5) as m, n → ∞ such that m ≤ (n/ log n). Remark: If mn−2/5 → δ−2 > 0, then as q → 1, n(2/5) ( ˜fn,m,q(x) − f(x)) D → N 0, δf(x) 2 πx(1 − x) (4.6) Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 25 / 47
  • 26. 4.3. Estimation for Other Support The case of supports other than [0, 1] may be handled by transformation. 1. For a finite support [a, b], transform the sample values to Y1, ...., Yn, where Yi = (Xi − a)/(b − a). Denoting the ˜Gn(y) as the smooth distribution function of the transformed sample, the smooth distribution ˜Fn(x) is given by ˜Fn(x) = ˜Gn(y) where y = (x − a)/(b − a). 2.For the interval (−∞, ∞), the transformation Yi = (1/2) + (1/π) tan−1 Xi is useful. In this case ˜Fn(x) = ˜Gn(y) where y = (1/2)(1/π) tan−1 x. 3. For the non-negative support [0, ∞) :, we may use the transformation: y = x/(1 + x) ˜Fn(x) = ˜Gn(y) where y = x/(1 + x). Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 26 / 47
  • 27. 5. Numerical Studies Here we illustrate the results for the estimation of distribution and density for the following three cases: (i) Beta (3,3) distribution, (ii) Beta(2,4) distribution and (iii) 0.65×Beta(2,10)+0.35×Beta(10,2) distribution. The three cases are chosen to depict three scenarios, namely that of symmetric unimodal distribution, asymmetric unimodal distribution and a bimodal distribution. We investigate the possibility of choosing q different than 1, for the choice of m = n/ log(n), a choice recommended in Babu, Canty and Chaubey (2002) for Bernstein polynomials. We would also like to investigate the optimal choice of q using the limiting form of the q−Bernstein polynomials, however, this is not pursued here. For a given m, the choice of q is obtained by cross-validation that is explained below. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 27 / 47
  • 28. 5.1 Cross Validation For the q−Bernstein polynomial CDF and PDF estimators, in order to choose an appropriate parameter q, we introduce a numerical method called Cross Validation, detail of relative equations are referenced from Chaubey and Sen (2009). Kullback-Liebler divergence between the estimated density ˜fn and the true density f is given by KL( ˜fn; f) = E log f(x) ˜fn(x) dF(x) In practice, the optimum cross-validation method estimates such divergence from the data for a given smoothing parameter and chooses one which gives the smallest estimated divergence. Bowman (1984) shows that this procedure is equivalent to the minimization of the negative likelihood, CVKL(q) = − log n i=1 ˜fn(Xi; Di) = − n i=1 log( ˜fn(Xi; Di)) (5.1) where D denotes data with X removed from D, which is the wholeYogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 28 / 47
  • 29. 5.1 ISE Cross Validation According to this criterion we determine λn that minimizes the criterion related to the mean integrated squared error, MISE(q) = E ( ˜fn(x) − f(x))2 dx Estimating this from the data and minimizing it is equivalent to the minimization of (see Silverman (1986), CVISE(q) = ˜f2 n(x; q, D)dx − 2 n n i=1 ˜fn−1(Xi; q, Di) (5.2) The first part can be obtain by numerical integration, say composite Simpson Method. And again, Di denotes data with Xi removed from D. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 29 / 47
  • 30. 5.1 LS Cross Validation The above cross validation works well if f is a density estimator. For CDF, it is better to use least square cross validation (see H¨ardle, 1991). CVLS(q) = n i=1 ( ˜Fn(Xi; q, D) − ˜Fn(Xi; q, Di))2 (5.3) Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 30 / 47
  • 31. Distribution Function-Illustration For a given sample of size n we start with m = n and determine the optimum value of the smoothing parameter as illustrated in Figures 6-8 for a random sample generated from each of the distributions. We also investigate the effect of increasing the value of m. This is done by selecting a grid of size n on the interval [0, 1], and we let M denote the value of m that gives the average error of estimation between the successive values of M and M + 1 less than ε, in order to minimize the amount of computation. [In order to control the individual errors, M will depend on the value of x also.]. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 31 / 47
  • 32. 0.0 0.2 0.4 0.6 0.8 1.0 0.050.100.150.200.25 q SumofSquareErros(SSE) (a) n = 20, q = 0.94, m = 20 0.2 0.4 0.6 0.8 0.00.20.40.60.81.0 x cdf (b) M = 147, ε = 9.96 × 10−5 0.0 0.2 0.4 0.6 0.8 1.0 0.00.20.40.60.8 q SumofSquareErros(SSE) (c) n = 50, q = 1, m = 50 0.2 0.4 0.6 0.8 0.00.20.40.60.81.0 x cdf (d) M = 1024, ε = 5.31 × 10−5 0.0 0.2 0.4 0.6 0.8 1.0 0.00.20.40.60.81.01.2 q SumofSquareErros(SSE) (e) n = 100; q = 1, m = 100 0.2 0.4 0.6 0.8 1.0 0.00.20.40.60.81.0 x cdf (f) M = 801, ε = 7.56 × 10−5 Figure 6: LSCV plot for q−Bernstein estimator (Left Panel); CDF Estima- tor for Beta(3, 3) Samples (Right Panel). Remark: red - true density; dash - q−Bernstein with m = n; blue - q−Bernstein with parameter q, M 12 Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 32 / 47
  • 33. 0.0 0.2 0.4 0.6 0.8 1.0 0.20.40.60.8 q SumofSquareErros(SSE) (a) n = 20, q = 0.96, m = 20 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.00.20.40.60.81.0 x cdf (b) M = 206, ε = 9.60 × 10−5 0.0 0.2 0.4 0.6 0.8 1.0 0.00.20.40.60.81.01.21.4 q SumofSquareErros(SSE) (c) n = 50, q = 0.99, m = 50 0.2 0.4 0.6 0.8 0.00.20.40.60.81.0 x cdf (d) M = 105, ε = 8.28 × 10−5 0.0 0.2 0.4 0.6 0.8 1.0 0.00.51.01.52.02.53.03.5 q SumofSquareErros(SSE) (e) n = 100; q = 1, m = 100 0.0 0.2 0.4 0.6 0.8 0.00.20.40.60.81.0 x cdf (f) M = 611, ε = 6.64 × 10−5 Figure 7: Smooth Bernstein(Left) and q−Bernstein(Right) CDF Estimator for Beta(2, 4) Samples. Remark: red - true density; dash - q−Bernstein with m = n; blue - q−Bernstein with parameter q, M 13 Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 33 / 47
  • 34. 0.0 0.2 0.4 0.6 0.8 1.0 0.20.40.60.81.01.2 q SumofSquareErros(SSE) (a) n = 20, q = 0.99, m = 20 0.0 0.2 0.4 0.6 0.8 1.0 0.00.20.40.60.81.0 x cdf (b) M = 73, ε = 6.88 × 10−5 0.0 0.2 0.4 0.6 0.8 1.0 0.00.51.01.52.02.5 q SumofSquareErros(SSE) (c) n = 50, q = 1, m = 50 0.0 0.2 0.4 0.6 0.8 1.0 0.00.20.40.60.81.0 x cdf (d) M = 355, ε = 4.41 × 10−5 0.0 0.2 0.4 0.6 0.8 1.0 012345 q SumofSquareErros(SSE) (e) n = 100; q = 1, m = 100 0.0 0.2 0.4 0.6 0.8 1.0 0.00.20.40.60.81.0 x cdf (f) M = 371, ε = 8.98 × 10−5 Figure 8: Smooth Bernstein(Left) and q−Bernstein(Right) CDF Estimator for 0.65 · Beta(2, 10) + 0.35 · Beta(10, 2) Samples. Remark: red - true density; dash - q−Bernstein with m = n; blue - q−Bernstein with parameter q, M 14 Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 34 / 47
  • 35. 0.0 0.2 0.4 0.6 0.8 1.0 0.00.20.40.60.81.0 x cdf (a) n = 20, q = 0.94, m = 7, M = 147 0.0 0.2 0.4 0.6 0.8 1.0 0.00.20.40.60.81.0 x cdf (b) n = 20, q = 0.96, m = 7, M = 206 0.0 0.2 0.4 0.6 0.8 1.0 0.00.20.40.60.81.0 x cdf (c) n = 50, q = 1, m = 13, M = 1024 0.0 0.2 0.4 0.6 0.8 1.0 0.00.20.40.60.81.0 x cdf (d) n = 50, q = 0.99, m = 13, M = 105 0.0 0.2 0.4 0.6 0.8 1.0 0.00.20.40.60.81.0 x cdf (e) n = 100, q = 1, m = 22, M = 801 0.0 0.2 0.4 0.6 0.8 1.0 0.00.20.40.60.81.0 x cdf (f) n = 100, q = 1, m = 22, M = 611 Figure 9: Smooth Bernstein and q−Bernstein CDF Estimator for Beta(3, 3)(Left) and Beta(2, 4)(Right) Samples. Remark: red is theoretical, black is Bernstein ,dash is q−Bernstein with M = m, blue is q−Bernstein with parameter q, M 15 Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 35 / 47
  • 36. 0.0 0.2 0.4 0.6 0.8 1.0 0.00.20.40.60.81.0 x cdf (a) n = 20, q = 0.99, m = 7, M = 73 0.0 0.2 0.4 0.6 0.8 1.0 0.00.20.40.60.81.0 x cdf (b) n = 50, q = 1, m = 13, M = 355 0.0 0.2 0.4 0.6 0.8 1.0 0.00.20.40.60.81.0 x cdf (c) n = 100, q = 1, m = 22, M = 371 Figure 10: Smooth Bernstein(Left) and q−Bernstein(Right) CDF Estimator for 0.65 · Beta(2, 10) + 0.35 · Beta(10, 2) Samples. Remark: red is theoretical, dash is q−Bernstein with M = m, blue is q−Bernstein with parameter q, M 16 Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 36 / 47
  • 37. Density Function-Illustration In Figures 11-12, we present the plots of density estimators corresponding to the three distributions based on a random sample for sample sizes n = 20, 50, 100, where m = n/ log(n) for different values of q ∈ [.85, 1]. We will use the following colors to highlight specific q’s. We will use red for the theoretical density, while use green3 for q = 0.85, lightseagreen for q = 0.9, olivedrab for q = 0.95 and darkorange3 for optimized q and blue for the Bernstein density estimator. For optimized q ∈ {0.85, 0.9, 0.95}, we will display it in darkorange3. Note that we have not considered q−Bernstein polynomial density estimator for the M, q case, as in this situation the density shape is very rough. These are illustrated in Figures 9-10. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 37 / 47
  • 38. 0.0 0.2 0.4 0.6 0.8 1.0 0.00.51.01.5 x density (a) n = 20, m = 7, q = 0.85 0.0 0.2 0.4 0.6 0.8 1.0 0.00.51.01.52.0 x density (b) n = 20, m = 7, q = 0.85 0.0 0.2 0.4 0.6 0.8 1.0 0.00.51.01.5 x density (c) n = 50, m = 13, q = 0.96 0.0 0.2 0.4 0.6 0.8 1.0 0.00.51.01.52.0 x density (d) n = 50, m = 13, q = 0.96 0.0 0.2 0.4 0.6 0.8 1.0 0.00.51.01.5 x density (e) n = 100, m = 22, q = 0.99 0.0 0.2 0.4 0.6 0.8 1.0 0.00.51.01.52.0 x density (f) n = 100, m = 22, q = 0.99 Figure 11: Comparison of Bernstein and q−Bernstein PDF estimator with q ∈ [0.85, 1] for Beta(3, 3)(left) and Beta(2, 4)(right) 18 Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 38 / 47
  • 39. 0.0 0.2 0.4 0.6 0.8 1.0 0.00.51.01.52.02.5 xcdf (a) n = 20, m = 7, q = 0.95 0.0 0.2 0.4 0.6 0.8 1.0 0.00.51.01.52.02.5 x cdf (b) n = 50, m = 13, q = 0.99 0.0 0.2 0.4 0.6 0.8 1.0 0.00.51.01.52.02.5 x cdf (c) n = 100, m = 22, q = 1 Figure 12: Comparison of Bernstein and q−Bernstein PDF estimator with q ∈ [0.85, 1] for 0.65 · Beta(2, 10) + 0.35 · Beta(10, 2) Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 39 / 47
  • 40. Average Squared Error Comparison The following tables display the average squared errors (ASE) based on B = 40 comparisons, where ASE(g) = 1 B (˜g(xi) − g(xi))2, where g is the true distribution or density, and ˜g is its nonparametric estimator. Distribution Sample size Bernstein q−Bernstein 20 0.0020684 0.0004371 Beta(3,3) 50 0.0037048 0.0043542 100 0.0003380 0.0001946 20 0.0036687 0.0021716 Beta(2,4) 50 0.0049030 0.0041611 100 0.0013453 0.0006329 20 0.0014412 0.0036309 Mixture 50 0.0012856 0.0032519 100 0.0041389 0.0046857 Table: Average Square Error for Bernstein and q−Bernstein CDF estimator with optimized q Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 40 / 47
  • 41. Average Squared Error Comparison Distribution Sample size Bernstein q−Bernstein 20 0.1388213 0.0760121 Beta(3,3) 50 0.0278798 0.0669346 100 0.0182203 0.0185668 20 0.1669993 0.1340892 Beta(2,4) 50 0.0488355 0.0780358 100 0.0221844 0.0219737 20 0.4132526 0.5370016 Mixture 50 0.1991615 0.2336240 100 0.2113997 0.1938492 Table: Average Square Error for Bernstein and q−Bernstein PDF estimator with optimized q Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 41 / 47
  • 42. By looking at the figures, we see that as the sample size increase, the ASE has a decreasing trend. The distribution function estimator may have smaller errors specially for small sample sizes. For density estimators, q− Bernstein polynomial may have slight advantage over the Bernstein polynomial that too for small samples and unimodal distributions. Overall, we can say that the classical Bernstein polynomial estimator for estimating density may still be considered superior over the general q− Bernstein polynomial estimator, however, the q−Bernstein polynomial estimator for the distribution function might be potentially better. This requires further investigation, especially in the manner in which we have selected M. The typical question is that of finding optimal q when m = ∞? Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 42 / 47
  • 43. References 1 Babu, G.J., Canty, A.J. and Chaubey, Y.P. (2002). Application of Bernstein polynomials for smooth estimation of a distribution and density function. Journal of Statistical Planning and Inference 105, 377-392. 2 Bernstein, S.N. (1912). D´emonstration du th´eor`eme de Weierstrass, fonde´e sur le calcul des probabilit´es. Commun. Soc. Math. Kharkov, No.1. Series XIII, 1–2. 3 Bowman, A.W. (1984). An alternative method of cross-validation for the smoothing of density estimates.Biometrika 71, 353-360. 4 Cao, J.D. (1997). A generalization of the Bernstein polynomials. Journal Of Mathematical Analysis And Applications 209, 140–146. 5 Chaubey, Y. P., & Sen, P. K. (2009). On the selection of the smoothing parameter in poisson smoothing of histogram estimator: Computational aspects. Pakistan Journal of Statistics 25, 385-401. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 43 / 47
  • 44. References 6 Gawronski,W. and StadtM¨uller, U. (1981). Smoothing histograms by means of lattice- and continuous distributions. Metrika 28, 155164. 7 Ghosal, S. (2001). Convergence rates for density estimation with bernstein polynomials. The Annals of Statistics 29, 1264–1280. 8 H¨ardle, W. (1991). Smoothing Techniques with Implementation in S. Springer-Verlag: New York. 9 Igarashi, G. and Yoshihide Kakizawa, Y. (2014). On improving convergence rate of Bernstein polynomial density estimator. Journal of Nonparametric Statistics 26, 6184. 10 Il’inskii, A. and Ostrovska, S. (2002). Convergence of generalized Bernstein polynomials. Journal of Approximation Theory 116, 100–112. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 44 / 47
  • 45. References 11 Kakizawa, Y. (2004). Bernstein polynomial probability density estimation. Journal of Nonparametric Statistics 16, 709–729. 12 Kakizawa, Y. (2011). A note on generalized bernstein polynomial density estimators. Statistical Methodology 8, 136153. 13 Leblanc, A. (2009). ChungSmirnov property for Bernstein estimators of distribution functions. Journal of Nonparametric Statistics 21, 133142. 14 Leblanc, A. (2012). On estimating distribution functions using Bernstein polynomials. Ann Inst Stat Math 64, 919943. 15 Lorentz, G. G. (1986). Bernstein Polynomials (2nd ed.) New York: Chelsea Publishing. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 45 / 47
  • 46. References 16 Phillips, G. M. (1997). Bernstein polynomials based on the q-integers. The heritage of P. L. Chebyshev: a Festschrift in honor of the 70th birthday of T. J. Rivlin. Ann. Numer. Math. 14, 511–518. 17 Prakasa Rao, B. L. S. (2005). Estimation of distribution and density functions by generalized Bernstein polynomials. Indian Journal of Pure and Applied Mathematics 36, 63–88. 18 Silverman, B. W. (1986). Density estimation for statistics and data analysis, Chapman and Hall, London. 19 StadtM¨uller, U. (1983). Asymptotic distributions of smoothed histograms. Metrika 30, 145–158. 20 Vitale, R.A. (1973). A Bernstein polynomial approach to density estimation. Comm. Statist. 2, 493–506. Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 46 / 47
  • 47. Talk slides will be available on SlideShare: www.slideshare.net/ychaubey/talk-slides-msast2016-70014046 THANKS!! Yogendra Chaubey (Concordia University) Department of Mathematics & Statistics December 21-23, 2016 47 / 47