Journal of Statistical Theory and Applications

Volume 17, Issue 3, September 2018, Pages 419 - 438

Shannon Information in K-records for Pareto-type Distributions

Authors
Zohreh Zamaniz.zamani.62@gmail.com
Department of Statistics, Faculty of Mathematics and Computer, Shahid Bahonar University of Kerman Kerman, Iran
Mohsen Madadimhnmadadi@gmail.com
Department of Statistics, Faculty of Mathematics and Computer, Shahid Bahonar University of Kerman Kerman, Iran
Received 10 July 2017, Accepted 13 November 2017, Available Online 30 September 2018.
DOI
10.2991/jsta.2018.17.3.3How to use a DOI?
Keywords
K-record values; K-record times; Shannon information; Digamma function; Zeta function; Differential entropy; Record data
Abstract

Pareto distributions provides models for many applications in the social, natural and physical sciences. In this paper, we derive the Shannon information contained in upper (lower) k-record values and associated k-record times of Pareto-type distributions for a finite sample of fixed size and for an inverse sampling plan. Properties of the Shannon information of the k-record values associated with Pareto-type distributions are investigated, both analytically and numerically.

Copyright
© 2018, the Authors. Published by Atlantis Press.
Open Access
This is an open access article under the CC BY-NC license (http://creativecommons.org/licences/by-nc/4.0/).

1. Introduction

The origin of the term entropy goes back to the works of Clausius (1864) and Boltzmann (1872) in thermodynamics. The idea of information-theoretic entropy was first introduced by Shannon (1948). Park (1995), Ebrahimi et al. 2004 and Oluyede (2006) obtained various results on the information properties of order statistics. Yari and Borzadaran (2010) calculated the Shannon entropy for Paretotype distributions and their order statistics. Baratpour et al. 2007 derived some results related to the Shannon entropy and Rényi entropy for record values.

In 1976, Dzubdziela and Kopociński introduced the concept of k-records which were further studied by Grudzieén and Szynal (1985), Raqab and Amin (1997) and called Type 2 k-records by Arnold et al. 1998. For k = 1, the usual records are obtained.

Madadi and Tata (2011) presented results on the Shannon information contained in upper (lower) record values and associated record times in a sequence of independent identically distributed continuous random variables. They also considered the Shannon information contained in record data from an inverse sampling plan. In 2014, they generalized these results to k-records. Afhami and Madadi (2013) derived the exact analytical expressions for the Shannon entropy of generalized order statistics from the Pareto (IV) distribution and related distributions.

In this paper, we obtain the Shannon information contained in upper (lower) k-record values and k-record times of a random sample of size n (nk) and of an inverse sampling plan for Pareto-type distributions. The paper is organized as follows: In Section 2, we present some preliminary results. Section 3 contains the main results of the paper. In this section, we derive the Shannon information contained in the data consisting of all upper and lower k-record values and associated k-record times of a random sample of size n and of an inverse sampling plan for Pareto-type distributions. We also present some results of the differential entropy for a finite sample of fixed size and for an inverse sampling plan. Finally, Section 4 contains a conclusion.

2. Preliminaries

In this section, we review some basic notations and definitions concerning Pareto-type distributions, k-records and entropy which will be needed in the next section.

2.1. Pareto-type distributions

A general version of Pareto-type distributions, called the Pareto (IV) distribution, is discussed in chapter 3 of Arnold (1983). The cumulative distribution function of this family is

FX(x)=1[1+(xμθ)1γ]α,x>μ,
where −∞ < µ < ∞, θ > 0 and γ > 0 are location, scale and inequality parameters, respectively and α > 0 is the shape parameter which characterizes the tail of the distribution. This distribution is denoted by Pareto (IV)(µ, θ, γ, α), and its density function is as follows:
fX(x)=α(xμθ)1γ1θγ[1+(xμθ)1γ]α+1,x>μ.
  • Setting (α = 1), (γ = 1) and (γ = 1, µ = θ) in relation (2.1) and (2.2), one at a time, leads to the cumulative distribution and probability density functions of Pareto (III), Pareto (II) and Pareto (I) distributions, respectively.

  • The Burr (XII) distribution is a special case of the Pareto (IV) distribution in which µ = 0, γ1γ. If X ∼Pareto(IV)(µ, θ, γ, α), and Zα=1+(xμθ)1γ, then

    FZα(z)=1zα,z>1,
    and
    fZα(z)=αzα+1,z>1,
    where α > 0. We denote this distribution by Pareto (α) and note that Zα has the same distribution as Z11α.

2.2. Shannon information

Consider a discrete random variable X such that P{X = xi} = pi, i = 1,2,··· and i=1Pi=1. Then

H(X)=i=1pilnpi,
is known as the “entropy” or Shannon information (SI) of the random variable X. If X is a random variable having an absolutely continuous cumulative distribution function (cdf) F(x) and probability density function (pdf) f(x), then SI is defined as
H(X)=+f(x)lnf(x)dx.
Suppose that Z be a random variable with cdf FZ, pdf fZ and Shannon information H(Z). The following lemma is well-known and easy to check.

Lemma 2.1.

If X = θZγ + µ, then FX(x)=FZ((xμθ)1γ).

So,

H(X)=ln(θγ)+(γ1)E(lnZ)+H(Z),θ,γ>0,μ.

The Shannon entropy of the random variable X is a mathematical measure of information which measures the average reduction of uncertainty of X. Because of its descriptive character, analytical expressions for univariate distributions have been obtained, among others, by Cover and Thomas (2006).

2.3. k-record data

Let {Xn : n = 1,2,…} be a sequence of independent and identically distributed (i.i.d) random variables with absolutely continuous cumulative distribution function F(x; θ) and probability density function f(x; θ). We are interested in the Shannon information (SI) contained in the sequence of k-records. A k-record is basically the k-th largest observation in a partial sample. More precisely, let Xi:n denote the i-th order statistic from a random sample of size n. We define upper k-record times Tn,k and upper k-record values Rn,k as follows :

  1. (i)

    T1,k = k and R1,k = X1:k,

  2. (ii)

    Tn,k = min{j : j > Tn−1,k, Xj > XTn−1,k−k+1:Tn−1,k} and Rn,k = XTn,kk + 1 : Tn,k, (n ≥ 2).

Suppose Nn,k be the number of k-record values in X1,…,Xn. Lower k-record values, lower k-record times and the number of lower k-records are similarly defined (Arnold et al., 1998).

In the case of record and k-record data from a random sample of size n, Madadi and Tata (2011, 2014) obtained the SI contained in the last record (maximum), upper k-record values with associated k-record times, denoted by HMU(n) and HRTU(n,k), respectively. The SI for the lower records and k-records are denoted by HML(n) and HRTL(n,k), respectively.

In an inverse sampling plan (ISP), one takes observations until a fixed number m of k-records is reached. Denote the SI contained in all upper k-record values, the last upper k-record and all upper k-record values together with associated k-record times by hRU(m,k), hMU(m,k) and hRTU(m,k), respectively. The corresponding SI for lower k-record data are denoted by hRL(m,k), hML(m,k) and hRTL(m,k), respectively.

Madadi and Tata (2011) showed that the SI contained in the last upper record is

HMU(n)=lnn+n1nnϕfU(n1,1),
where
ϕfU(i,k)=+(lnf(x))(1F(x))k1Fi(x)f(x)dx.
For lower record values, we have
HML(n)=lnn+n1nnϕfL(n1,1),
where ϕfL(i,k) is obtained from ϕfU(i,k) by replacing F by 1 −F.

Madadi and Tata (2014) also proved that the SI contained in the data consisting of all upper k-record values and k-record times of a random sample of size n (nk) is given by

HRTU(n,k) =(nk+22)lnkki=0nk(i+k1k1)ϕfU(i,k)+ki=1nkj=0k1(i+k1k)(k1j)(1)j(i+j+1)2
and
HRTU(n,1)=ψ(n+1)+γ*i=1n1i2i=0n1ϕfU(i,1),
where γ* = 0.57721566 is the Euler constant and ψ is the digamma function where
ψ(x)=1Γ(x)0yx1ex(lny)dy.
The HRTL(n,k) and HRTL(n,1) contained in corresponding lower k-record statistics are obtained by replacing ϕfU(i,k) and ϕfU(i,1) by ϕfL(i,k) and ϕfL(i,1) in (2.8) and (2.9) , respectively.

They also showed that the SI contained in all of k-record values, last upper k-record and all of upper k-record values and times of an ISP are respectively obtained as

hRU(m,k)=mk(km+12klnk)i=1mψfU(i,k),
hMU(m,k)=lnΓ(m)+m(k1)klnk(m1)ψ(m)ψfU(m,k),
and for k > 1,
hRTU(m,k)=mlnk+m(k1)k+(1γ*ψ(k+1))(m+k1km(k1)m1)+j=2m1((m+kj)kj1km(k1)mj)ζ(j,k+1)i=1mψfU(i,k),
where
ψfU(i,k)=E(lnf(Ri,k))=kiΓ(i)+(lnf(x))(ln(1F(x)))i1(1F(x))k1f(x)dx.
Here ζ(i,k)=j=k1ji is the incomplete Zeta function.

Note that for k = 1, i.e. for ordinary records,

hRTU(m,1)=hRU(m,1)+m(m1)i=2m(mi+1)ζ(i).
Similar formulas hold for lower k-records.

3. Shannon information in k-records for Pareto-type distributions

In this section, we compute the SI in k-records of Pareto-type distributions for a finite sample of fixed size as well as for an inverse sampling plan.

3.1. Shannon information in a finite sample

Since the number of observations n is fixed (and the number of records is random), the last upper record is the maximum in the sample. From (2.5) and (2.6) , we have

HMU(n,Zα)=lnn+n1n+α+1α(ψ(n+1)+γ*)lnα,
and
HML(n,Zα)=lnn+n1n+α+1nαlnα.
Therefore the SI contained in the last record for the Pareto (IV) are
HMU(n,X)= HMU(n,Zα)+ln(θγ)1γα(ψ(n+1)+γ*)+nα(1γ)j=0n1l=1(n1j)(1)jl(l+jα+α),
and
HML(n,X)=HML(n,Zα)+ln(θγ)n(1γ)[1n2ααl=11l(l+nα)].
When α is integer,
l=11l(l+jα+α)=1jα+α(ψ(jα+α+1)+γ*),
and
l=11l(l+nα)=1nα(ψ(nα+1)+γ*).
Let
ΔHMU(n,Zα)=HMU(n+1,Zα)HMU(n,Zα),
and
ΔHML(n,Zα)=HML(n+1,Zα)HML(n,Zα),
denote the n-th differential of the entropy, that is, the change in the entropy when a new observation occurs. From (3.1) and (3.2) , we obtain
ΔHMU(n,Zα)=lnnn+1+1n+1(n+1)α,
and
ΔHML(n,Zα)=lnnn+11n(n+1)α.
Fig. 1. represent behaviours of HMU(n,Z1) and HML(n,Z1) with respect to n.

Fig. 1.

(a) and (b) represent HMU(n,Z1) and HML(n,Z1), respectively.

One can easily see that

  1. (i)

    limnΔHMU(n,Zα)=limnΔHML(n,Zα)=0.

  2. (ii)

    ΔHMU(n,Zα) and ΔHML(n,Zα) are respectively always nonnegative and negative functions i.e. HMU(n,Zα) and HML(n,Zα) are increasing and decreasing functions with respect to sample size for each α. Also, ΔHMU(n,Zα) and ΔHML(n,Zα) are decreasing and increasing functions of n for each α.

  3. (iii)

    0ΔHMU(n,Zα)<ln2+1+12α and ln212α<ΔHML(n,Zα)0.

  4. (iv)

    For each fixed n, ΔHMU(n,Zα) and ΔHML(n,Zα) are decreasing convex and increasing concave functions of α, respectively.

  5. (v)

    HMU(n,Zα) and HML(n,Zα) are decreasing convex functions of α for each n.

Theorem 3.1.

For Pareto (1), the SI contained in all the upper and lower k-record values together with k-record times of a random sample of size n (n ≥ k) are equal to

HRTU(n,k,Z1)=(nk+22)lnk+2k+i=1nk(i+k1k1)(2kI(k:k+i)+iI(i+1:k+i)),
and
HRTL(n,k,Z1)=(nk+22)lnk+2kI(1:k)+i=1nk(i+k1k1)(2k+i)I(i+1:k+i),
where I(j:n)=i=0nj(nji)(1)i(i+j)2,0<jn.

Also for k = 1, we obtain

HRTU(n,1,Z1)=(1+2γ*)(ψ(n+1)+γ*)i=1n1i2+2i=1nψ(i+1)i,
and
HRTL(n,1,Z1)=(ψ(n+1)+γ*)+i=1n1i2.
From (2.3), (2.4) and (2.6), we get
ϕfU(i,k,Z1)=01(2lnu)uk1(1u)idu=2j=0i(ij)(1)j(j+k)2=2I(k:k+i).
Substituting this expression in (2.8) , we obtain
HRTU(n,k,Z1)=(nk+22)lnk+2ki=0nk(i+k1k1)I(k:k+i)+ki=1nk(i+k1k)I(i+1:k+i)=(nk+22)lnk+2k1k2+ki=1nk[2(i+k1k1)I(k:k+i)+(i+k1k)I(i+1:k+i)],
hence, we conclude (3.3). Similarly,
ϕfL(i,k,Z1)=2I(i+1:k+i),
therefore (3.4) is obtained. From (2.6), and recalling that ln(1t)=q=1tqq, we have
ϕfU(i,1,Z1)=2(ψ(i+2)+γ*)i+1,
and
ϕfL(i,1,Z1)=2(i+1)2,
thus (3.5) and (3.6) are achieved.

Corollary 3.1.

For Zα=Z11α, we obtain

HRTU(n,k,Zα)=HRTU(n,k,Z1)k(lnα)(ψ(n+1)ψ(k))+k(1α1)i=0nk(i+k1k1)I(k:k+i),
and
HRTL(n,k,Zα)=HRTL(n,k,Z1)k(lnα)(ψ(n+1)ψ(k))+k(1α1)i=0nk(i+k1k1)I(i+1:k+i).
Also,
HRTU(n,1,Zα)=HRTU(n,1,Z1)(lnα)(ψ(n+1)+γ*)+(1α1)i=0n1I(1:i+1),
and
HRTL(n,1,Zα)=HRTL(n,1,Z1)(lnα)(ψ(n+1)+γ)+(1α1)i=1n1i2.

Corollary 3.2.

The SI contained in all k-record values and k-record times of a random sample of size n (n ≥ k) from the Pareto (IV) distribution are

HRTU(n,k,X)=HRTU(n,k,Zα)+k(ln(θγ))(ψ(n+1)ψ(k))k(1γ)αi=0nk(i+k1k1)I(k:i+k)+kα(1γ)i=0nkj=0iq=1(i+k1k1)(ij)(1)jq(q+kα+jα),
and
HRTL(n,k,X)=HRTL(n,k,Zα)+k(ln(θγ))(ψ(n+1)ψ(k))+kα(1γ)i=0nkj=0k1q=1(i+k1k1)(k1j)(1)jq(q+iα+jα+α)k(1γ)αi=0nkj=0k1(i+k1k1)(k1j)(1)j(i+j+k)2,
respectively.

For ordinary records, these information can be written as

HRTU(n,1,X)=HRTU(n,1,Zα)+(ln(θγ)γ*(1γ)α)(ψ(n+1)+γ)1γαi=1nψ(i+1)i+α(1γ)i=0n1j=0iq=1(ij)(1)jq(q+jα+α),
and
HRTL(n,1,X)=HRTL(n,1,Zα)+(ln(θγ))(ψ(n+1)+γ*)1γαi=1n1i2+α(1γ)i=0n1q=11q(q+iα+α).
When α is integer, we have
q=11q(q+iα+α)=1iα+α(ψ(iα+α+1)+γ*).
For n = k, k + 1,···, let
ΔHRTU(n,k)=HRTU(n+1,k)HRTU(n,k),
and
ΔHRTL(n,k)=HRTL(n+1,k)HRTL(n,k),
denote the n-th SI differential corresponding to entropies in equations (3.3) and (3.4), then it is easy to see that
ΔHRTU(n,k,Z1)=(nk+2)lnk+(nk1)[2kI(k:n+1)+(nk+1)I(nk+2:n+1)],
and
ΔHRTL(n,k,Z1)=(nk+2)lnk+(nk1)(n+k+1)I(nk+2:n+1).
Fig. 2. displays HRTU(n,k,Z1) and HRTU(n,1,Z1) for several values of k and n. Behaviours of HRTL(n,k,Z1) and HRTL(n,1,Z1) are shown in Fig. 3. Our numerical computations suggest that when k ≥ 2 is fixed, the differential entropy is negative and decreasing in n. Also, for k = 1, we have
ΔHRTU(n,1,Z1)=n+2(n+1)2+2n+1(ψ(n+1)+γ*),
and
ΔHRTL(n,1,Z1)=n+2(n+1)2.
The following properties are easy to check and confirm via Figs. 2. and 3.
  1. (i)

    limnΔHRTU(n,1,Z1)=limnΔHRTL(n,1,Z1)=0.

  2. (ii)

    ΔHRTU(n,1,Z1) and ΔHRTL(n,1,Z1) are non-negative decreasing functions of n.

  3. (iii)

    0ΔHKTL(n,1,Z1)<34.

We note that ψ(n+1)=ψ(n)+1n and limnψ(n)n=0.

Fig. 2.

(a) and (b) represent HRTU(n,k,Z1) and HRTU(n,1,Z1), respectively.

Fig. 3.

(a) and (b) represent HRTL(n,k,Z1) and HRTL(n,1,Z1), respectively.

3.2. Shannon information in an inverse sampling plan

In an inverse sampling plan, we take observations until a fixed number m of k-records is reached. Here, we compute the SI of k-records for an inverse sampling plan from Pareto-type distributions and discuss their properties.

Theorem 3.2.

The SI contained in the m-th upper and m-th lower k-record values from Z1 are given by

hMU(m,k,Z1)=lnΓ(m)+m(k+1)klnk(m1)ψ(m),m1,
and
hML(m,k,Z1)=lnΓ(m)+m(k1)klnk(m1)ψ(m)+2(ψ(k+1)+γ*j=2mkj1ζ(j,k+1)),m2.
It is easy to check (3.7). Now from (2.13), replacing F by 1 − F, we obtain
ψfL(i,k,Z1)=2kiΓ(i)01(ln(1y))(lny)i1yk1dy.
Since ln(1y)=l=1yll, the above expression can be written as
ψfL(i,k,Z1)=2kil=k+11(lk)li=2ki[1kij=1k1jζ(2,k+1)ki1ζ(i,k+1)k]=2[ψ(k+1)+γ*j=2ikj1ζ(j,k+1)].
It is clear that i = 1, leads to
ψfL(1,k,Z1)=2kj=k+11j(jk)=2[j=11jj=k+11j]=2j=1k1j=2(ψ(k+1)+γ*).
Substituting (3.9) and (3.10) in (2.11) , the proof is complete.

Remark 3.1.

For m = 1, we have

hML(1,k,Z1)=hRL(1,k,Z1)=hRTL(1,k,Z1)=k1klnk+2(ψ(k+1)+γ*).

Corollary 3.3.

For Zα, we obtain

hMU(m,k,Zα)=hMU(m,k,Z1)lnα+(1α1)mk,
and
hML(m,k,Zα)=hML(m,k,Z1)lnα+(1α1)(ψ(k+1)+γ*j=2mkj1ζ(j,k+1)).
Also, hML(1,k,Zα)=hML(1,k,Z1)lnα+(1α1)(ψ(k+1)+γ*).

Corollary 3.4.

The SI contained in the m-th upper and m-th lower k-record values from the Pareto (IV) are

hMU(m,k,X)=hMU(m,k,Zα)+ln(θγ)+(kα)m(1γ)q=kα+11(qkα)qm(1γ)mkα,
and
hML(m,k,X)=hML(m,k,Zα)+ln(θγ)+1γα(ψ(k+1)+γ*)(1γ)αj=2mkj1ζ(j,k+1)+(1γ)kmj=1q=0(jαq)(1)qj(q+k)m,
respectively. When α is integer, one obtains
hMU(m,k,X)=hMU(m,k,Zα)+ln(θγ)+(1γ)(ψ(kα+1)+γ*)(1γ)j=2m(kα)j1ζ(j,kα+1)(1γ)mkα.

Let

ΔhMU(m,k)=hMU(m+1,k)hMU(m,k),
and
ΔhML(m,k)=hML(m+1,k)hML(m,k),
denote the change in entropy in the last upper and last lower k-record values, respectively when the new k-record is observed. Then, from (3.7) and (3.8) , we obtain
ΔhMU(m,k,Z1)=lnmψ(m)+1k,
and
ΔhML(m,k,Z1)=lnmψ(m)1k2kmζ(m+1,k+1).
It is easy to see that
  1. (i)

    ΔhMU(m,k,Z1) is a positive function of m for each fixed k. Hence the sequence of functions {hMU(m,k,Z1)} is increasing with respect to m. Also, ΔhML(m,k,Z1) is a negative function of m, for each fixed k. Therefore {hML(m,k,Z1)} is a sequence of decreasing functions of m.

  2. (ii)

    for each fixed k, ΔhMU(m,k,Z1) and ΔhML(m,k,Z1) are decreasing and increasing functions of m, respectively.

  3. (iii)

    limmΔhMU(m,k,Z1)=1k and limmΔhML(m,k,Z1)=1k.

  4. (iv)

    12m+1k<ΔhMU(m,k,Z1)<1m+1k and 1k2kζ(2,k+1)+γ*<ΔhML(m,k,Z1)<1k, for each m ≥ 2.

We notice that 12m<lnmψ(m)<1m, for m > 0, ψ(1) = −γ* and limm→∞ kmζ (m + 1, k + 1) = 0. For k = 1, we obtain
ΔhMU(m,1,Z1)=lnmψ(m)+1,
and
ΔhML(m,1,Z1)=lnmψ(m)2ζ(m+1)+1.
It is easy to prove that
  1. (i)

    limmΔhMU(m,1,Z1)=1 and limmΔhML(m,1,Z1)=1.

  2. (ii)

    12m+1<ΔhMU(m,1,Z1)<1m+1, for each m ≥ 2 and ΔhML(m,1,Z1)>12mπ23+1.

  3. (iii)

    ΔhMU(m,1,Z1) is a positive decreasing function of m, therefore hMU(m,1,Z1) is an increasing function of m.

Fig. 4. displays a graphical representation of hMU(m,k,Z1) and hML(m,k,Z1) when k = 1, 2, 3, 5, 10.

Let

Δ*hMU(m,k)=hMU(m,k+1)hMU(m,k),
and
Δ*hML(m,k)=hML(m,k+1)hML(m,k),
denote the SI difference of (k + 1)-record and k-record in the m-th upper and m-th lower cases, respectively. Then, from (3.7) and (3.8) , it is easy to check
Δ*hMU(m,k,Z1)=lnkk+1mk(k+1),
and
Δ*hML(m,k,Z1)=m+2kk(k+1)+lnkk+1+2kj=2m(kk+1)j+2j=2m(kj1(k+1)j1)ζ(j,k+2).
We note that ζ(j,k+1)=1(k+1)j+ζ(j,k+2). Hence
  1. (i)

    Δ*hMU(m,k,Z1) is a negative increasing function of k, for each fixed m. Hence, hMU(m,k,Z1) is a monotone decreasing function of k. Similarly, Δ*hML(m,k,Z1) is a positive decreasing function of k and therefore hML(m,k,Z1) is a monotone increasing function of k for each m.

  2. (ii)

    limkΔ*hMU(m,k,Z1)=limkΔ*hML(m,k,Z1)=0.

  3. (iii)

    ln2m2<Δ*hMU(m,k,Z1)<0 and 0<Δ*hML(m,k,Z1)<m2+2ln2(12)m1+j=2m(22j)ζ(j,3).

Fig. 4.

(a) and (b) represent hMU(m,k,Z1) and hML(m,k,Z1), respectively.

Remark 3.2.

For m = 1, we have

Δ*hML(1,k,Z1)=lnkk+1+2k+1k(k+1).
Therefore
  1. (i)

    limkΔ*hML(1,k,Z1)=0 and 0<Δ*hML(1,k,Z1)<ln2+32.

  2. (ii)

    Δ*hML(1,k,Z1) is positive monotone decreasing function of k for each m. So hML(1,k,Z1) is a monotone increasing function of k.

Theorem 3.3.

The SI contained in all of the first m upper and m lower k-record values of an ISP are respectively given by

hRU(m,k,Z1)=mk(k+m+12klnk),
and
hRL(m,k,Z1)=mk(km+12klnk)+2m(ψ(k+1)+γ*)2j=2m(mj+1)kj1ζ(j,k+1).

Corollary 3.5.

For Pareto (I) with parameter α, we have

hRU(m,k,Zα)=hRU(m,k,Z1)mlnα+(1α1)m(m+1)2k,
and
hRL(m,k,Zα)=hRL(m,k,Z1)mlnα+m(1α1)(ψ(k+1)+γ*)+(11α)j=2m(mj+1)kj1ζ(j,k+1).

Corollary 3.6.

For the Pareto (IV), one can obtains

hRU(m,k,X)=hRU(m,k,Zα)+mln(θγ)+(1γ)j=1mq=kα+1(kα)j(qkα)qjm(m+1)(1γ)2kα,
and
hRL(m,k,X)=hRL(m,k,Zα)+mln(θγ)+m(1γ)α(ψ(k+1)+γ*)1γαj=2m(mj+1)kj1ζ(j,k+1)+(1γ)j=1ml=1q=0(lαq)(1)qkjl(q+k)j.
When α is integer,
hRU(m,k,X)=hRU(m,k,Zα)+mln(θγ)+m(1γ)(ψ(kα+1)+γ*)(1γ)j=2m(mj+1)(kα)j1ζ(j,kα+1)m(m+1)(1γ)2kα.
Let
ΔhRU(m,k)=hRU(m+1,k)hRU(m,k),
and
ΔhRL(m,k)=hRL(m+1,k)hRL(m,k),
denotes the change in entropy in observing the k-record values from the m-th to the (m + 1)-th k-record values. Therefore,
ΔhRU(m,k,Z1)=m+1klnk+1,
and
ΔhRL(m,k,Z1)=2(ψ(k+1)+γ*)2j=2m+1kj1ζ(j,k+1)+1m+1klnk.
It can be easily shown that
  1. (i)

    ΔhRL(m,k,Z1) is a monotone decreasing in m, for each k > 1.

  2. (ii)

    ΔhRU(m,k,Z1) is a positive function for m > klnkk − 1 and otherwise is a negative function.

  3. (iii)

    ΔhRL(m,k,Z1)<2(ψ(k)+γ*kζ(2,k+1))+1lnk, ∀m.

(see Fig. 5. for illustration.)

Fig. 5.

(a) and (b) represent hRU(m,k,Z1) and hRL(m,k,Z1), respectively.

Remark 3.3.

Note that for k = 1, i.e. for ordinary records, it is easy to see that the differential entropy has the following properties

  1. (i)

    ΔhRU(m,1,Z1)=m+2 and ΔhRL(m,1,Z1)=m+22j=2m+1ζ(j).

  2. (ii)

    The sequence {hRU(m,1,Z1)} contains positive and increasing functions in m.

  3. (iii)

    ΔhRU(m,1,Z1) is a positive monotone increasing function of m.

  4. (iv)

    hRU(m,1,Z1)>2 and ΔhRU(m,1,Z1)>3, for each m.

  5. (v)

    m ≥ 7, |ΔhRL(m,1,Z1)(m1)|<0.01.

Let
Δ*hRU(m,k)=hRU(m,k+1)hRU(m,k),
and
Δ*hRL(m,k)=hRL(m,k+1)hRL(m,k).
It can be shown that
Δ*hRU(m,k,Z1)=mlnkk+1m(m+1)2k(k+1),
and
Δ*hRL(m,k,Z1)=m(m+12k(k+1)+2k+1+lnkk+1)+2kj=2m(mj+1)(kk+1)j+2j=2m(mj+1)(kj1(k+1)j1)ζ(j,k+2).
Therefore
  1. (i)

    Δ*hRU(m,k,Z1) is a negative increasing function of k for each m. Hence hRU(m,k,Z1) is a decreasing function of k for each m.

  2. (ii)

    limkΔ*hRU(m,k,Z1)=0 and m(ln2+m+14)<Δ*hRU(m,k,Z1)<0.

Theorem 3.4.

The SI contained respectively in the upper and lower k-record values and associated k-record times for k > 1 are obtained as

hRTU(m,k,Z1)=mlnk+mk(m+k)+(1γ*ψ(k+1))(m+k1km(k1)m1)+j=2m1((m+kj)kj1km(k1)mj)ζ(j,k+1),
and
hRTL(m,k,Z1)=mlnk+m(k1)k+(1γ*ψ(k+1))(m+k1km(k1)m1)+2m(ψ(k+1)+γ*)+j=2m1((k+jm2)kj1km(k1)mj)ζ(j,k+1)2km1ζ(m,k+1).

Remark 3.4.

Note that for k = 1, (3.13) and (3.14) lead to

hRTU(m,1,Z1)=m(3m+1)2j=2m(mj+1)ζ(j),
and
hRTL(m,1,Z1)=m2+m3j=2m(mj+1)ζ(j).

Corollary 3.7.

For Zα, we have

hRTU(m,k,Zα)=hRTU(m,k,Z1)mlnα+(1α1)m(m+1)2k,
and
hRTL(m,k,Zα)=hRTL(m,k,Z1)mlnα+m(1α1)(ψ(k+1)+γ*)+(11α)j=2m(mj+1)kj1ζ(j,k+1),
and for k = 1,
hRTU(m,1,Zα)=hRTU(m,1,Z1)mlnα+(1α1)m(m1)2,
and
hRTL(m,1,Zα)=hRTL(m,1,Z1)mlnα+(1α1)m(m1)2(1α1)j=2m1(mj)ζ(j).

Corollary 3.8.

For the Pareto (IV), we have

hRTU(m,k,X)=hRTU(m,k,Zα)+mln(θγ)m(m+1)(1γ)2kα+(1γ)j=1mq=kα+1(kα)j(qkα)qj,
and
hRTL(m,k,X)=hRTL(m,k,Zα)+mln(θγ)+m(1γ)α(ψ(k+1)+γ*)(1γ)αj=2m(mj+1)kj1ζ(j,k+1)+(1γ)j=1ml=1q=0(lαq)(1)qkjl(q+k)j.
When α is integer,
hRTU(m,k,X)=hRTU(m,k,Zα)+mln(θγ)+m(1γ)(ψ(kα+1)+γ*)(1γ)j=2m(mj+1)(kα)j1ζ(j,kα+1)m(m+1)(1γ)2kα.
We note that for ordinary records,
hRTU(m,1,X)=hRTU(m,1,Zα)+mln(θγ)+(1γ)i=1mj=α+1αi(jα)jim(m+1)(1γ)2α,hRTL(m,1,X)=hRTL(m,1,Zα)+mln(θγ)m(m+1)(1γ)2α+1γαj=2m(mj+1)ζ(j)+(1γ)i=1mj=1q=0(jαq)(1)qj(q+1)i,
and when α is integer
hRTU(m,1,X)=hRTU(m,1,Zα)+mln(θγ)+m(1γ)(ψ(α+1)+γ*)(1γ)j=2m(mj+1)αj1ζ(j,k+1)m(m+1)(1γ)2α.
Let
ΔhRTU(m,k)=hRTU(m+1,k)hRTU(m,k),
and
ΔhRTL(m,k)=hRTL(m+1,k)hRTL(m,k).
Then,
ΔhRTU(m,k,Z1)=lnk+2m+1k+1+(1γ*ψ(k+1))(1km(k1)m)+j=2m1(kj1km(k1)mj+1)ζ(j,k+1)km1k1ζ(m,k+1),
and
ΔhRTL(m,k,Z1)=lnk+k1k+(1γ*ψ(k+1))(1km(k1)m)+2(ψ(k+1)+γ*)2kmζ(m+1,k+1)+km12kmk1ζ(m,k+1)j=2m1(km(k1)mj+1+kj1)ζ(j,k+1).
respectively. We note that for k = 1,
ΔhRTU(m,1,Z1)=3m+2j=2m+1ζ(j),
and
ΔhRTL(m,1,Z1)=2m+23j=2m+1ζ(j),m=1,2,3,
The following properties are easily obtained:
  1. (i)

    ΔhRTL(m,1,Z1) is a negative function of m, therefore hRTL(m,1,Z1) is a decreasing function of m. Furthermore, ΔhRTL(m,1,Z1) is an increasing function of m.

  2. (ii)

    For each m ≥ 8, we have |ΔhRTL(m,1,Z1)(m1)|<1100 and |ΔhRTU(m,1,Z1)(2m+1)|<1100, for each m ≥ 6.

(see Figs. 6. and 7. for some more notes.)

Fig. 6.

(a) and (b) represent hRTU(m,k,Z1) and hRTU(m,1,Z1), respectively.

Fig. 7.

(a) and (b) represent hRTL(m,k,Z1) and hRTL(m,1,Z1), respectively.

4. Conclusions

In this paper, we obtained the Shannon information in k-record in a sample of fixed size as well as in an inverse sampling plan for Pareto-type distributions. Properties of entropies for k-record values for Pareto-type distributions are also investigated.

Acknowledgments

The authors are grateful to the editorial board of Journal of Statistical Theory and Applications (JSTA), in particular, Professor Hamedani and would like to thank two anonymous referees for useful comments and suggestions that lead to this improved version of the paper.

References

[1]B Afhami and M Madadi, Shannon entropy in generalized order statistics from Pareto-type distributions, Int. J. Nonlinear. Anal. Appl, Vol. 4, 2013, pp. 79-91.
[2]BC Arnold, Pareto Distribution, International Cooperative Publishing House, Fairland, Maryland, 1983.
[5]L Boltzmann, Neitere Studien uber das Warmegleichgewicht unter Gasmolekulen, K. Akad. Wiss. (Wein) Sitzb, Vol. 66, 1872, pp. 275.
[6]R Clausius, Abhaudlungen uber die mechanische Warmetheorie Friedrich, Vieweg, Braunschweig, 1864.
[10]Z Grudzien and D Szynal, On the expected values of k-th record values and associated characterizations of distributions, Probab. Stat. Decis. Theory Ser. A, Reidel, Dordrecht, 1985, pp. 119-127.
[14]S Park, The entropy of consecutive order statistics, IEEE Trans. Inform. Theory, Vol. 1, 1995, pp. 2003-2007.
Journal
Journal of Statistical Theory and Applications
Volume-Issue
17 - 3
Pages
419 - 438
Publication Date
2018/09/30
ISSN (Online)
2214-1766
ISSN (Print)
1538-7887
DOI
10.2991/jsta.2018.17.3.3How to use a DOI?
Copyright
© 2018, the Authors. Published by Atlantis Press.
Open Access
This is an open access article under the CC BY-NC license (http://creativecommons.org/licences/by-nc/4.0/).

Cite this article

TY  - JOUR
AU  - Zohreh Zamani
AU  - Mohsen Madadi
PY  - 2018
DA  - 2018/09/30
TI  - Shannon Information in K-records for Pareto-type Distributions
JO  - Journal of Statistical Theory and Applications
SP  - 419
EP  - 438
VL  - 17
IS  - 3
SN  - 2214-1766
UR  - https://doi.org/10.2991/jsta.2018.17.3.3
DO  - 10.2991/jsta.2018.17.3.3
ID  - Zamani2018
ER  -