1. Introduction
Recently, generalized entropies have been studied from the mathematical point of view. The typical generalizations of Shannon entropy [
1] are Rényi entropy [
2] and Tsallis entropy [
3]. The recent comprehensive book [
4] and the review [
5] support to understand the Tsallis statistics for the readers. Rényi entropy and Tsallis entropy are defined by
(1.1)
for a given information source
X = {
x1, …,
xn} with the probability
pj ≡ Pr(
X =
xj). Both entropies recover Shannon entropy
(1.2)
in the limit
q → 1. The uniqueness theorem for Tsallis entropy was firstly given in [
6] and improved in [
7].
Throughout this paper, we call a parametric extended entropy, such as Rényi entropy and Tsallis entropy, a generalized entropy. If we take n = 2 in (1.2), we have the so-called binary entropy sb(x) = −xlog x − (1 − x)log (1 − x). Also we take n = 1 in (1.2), and we have the Shannon′s entropy function f(x) = −xlog x. In this paper, we treat the entropy function with two parameters. We note that we can produce the relative entropic function −yf(x/y) = x(log x − log y) by the use of the Shannon′s entropy function f(x).
We note that Rényi entropy has the additivity
(1.3)
but Tsallis entropy has the nonadditivity
(1.4)
where
X ×
Y means that
X and
Y are independent random variables. Therefore, we have a definitive difference for these entropies although we have the simple relation between them
(1.5)
where
q-exponential function exp
q(
x) ≡ {1+(1−
q)
x}
1/(1−q) is defined if 1 + (1 −
q)
x ≥ 0. Note that we have
.
Tsallis entropy is rewritten by
(1.6)
where
q-logarithmic function (which is an inverse function of exp
q(·)) is defined by
(1.7)
which converges to log
x in the limit
q → 1.
Since Shannon entropy can be regarded as the expectation value for each value −log
pj, we may consider that Tsallis entropy can be regarded as the
q-expectation value for each value −ln
qpj, as an analogy to the Shannon entropy, where
q-expectation value
Eq is defined by
(1.8)
However, the
q-expectation value
Eq lacks the fundamental property such as
E(1) = 1, so that it was considered to be inadequate to adopt as a generalized definition of the usual expectation value. Then the normalized
q-expectation value was introduced
(1.9)
and by using this, the normalized Tsallis entropy was defined by
(1.10)
We easily find that we have the following nonadditivity relation for the normalized Tsallis entropy:
(1.11)
As for the details on the mathematical properties of the normalized Tsallis entropy, see [8], for example. See also [9] for the role of Tsallis entropy and the normalized Tsallis entropy in statistical physics. The difference between two non-additivity relations (1.4) and (1.11) is the signature of the coefficient 1 − q in the third term of the right-hand sides.
We note that Tsallis entropy is also rewritten by
(1.12)
so that we may regard it as the expectation value such as
Sq(
X) =
E1[ln
q1/
pj],
where
E1 means the usual expectation value
. However, if we adopt this formulation in the definition of Tsallis conditional entropy, we do not have an important property such as a chain rule (see [
10] for details). Therefore, we often adopt the formulation using the
q-expectation value.
As a further generalization, a two-parameter extended entropy
(1.13)
was recently introduced in [
11,
12] and systematically studied with the generalized exponential function and the generalized logarithmic function ln
κ,r(
x) ≡
xr((
xκ −
x−κ)/2
κ). In the present paper, we treat a two-parameter extended entropy defined in the following form:
(1.14)
for two positive numbers α and β. This form can be obtained by putting α = 1 + r − κ and β = 1 + r + κ in (1.13), and it coincides with the two-parameter extended entropy studied in [13]. In addition, the two-parameter extended entropy (1.14) was axiomatically characterized in [14]. Furthermore, a two-parameter extended relative entropy was also axiomatically characterized in [15].
In the paper [
16], a characterization of Tsallis entropy function was proven by using the functional equation. In the present paper, we will show that the two-parameter extended entropy function
(1.15)
can be characterized by the simple functional equation.
2. A Review of the Characterization of Tsallis Entropy Function by the Functional Equation
The following proposition was originally given in [16] by the simple and elegant proof. Here, we give the alternative proof along to the proof given in [17].
Proposition 2.1 (see [16].)If the differentiable nonnegative function fq with positive parameter q ∈ ℝ satisfies the following functional equation:
(2.1)
then the function
fq is uniquely given by
(2.2)
where
cq is a nonnegative constant depending only on the parameter
q.
Proof. If we put y = 1 in (2.1), then we have fq(1) = 0. From here, we assume that y ≠ 1. We also put gq(t) ≡ fq(t)/t then we have
(2.3)
Putting
x = 1/2 in (
2.3), we have
(2.4)
Substituting
y/2 into
y, we have
(2.5)
By repeating similar substitutions, we have
(2.6)
Then, we have
(2.7)
due to
q > 0. Differentiating (
2.3) by
y, we have
(2.8)
Putting
y = 1 in the above equation, we have
(2.9)
where
.
By integrating (2.3) from 2−N to 1 with respect to y and performing the conversion of the variables, we have
(2.10)
By differentiating the above equation with respect to
x, we have
(2.11)
Taking the limit
N →
∞ in the above, we have
(2.12)
thanks to (
2.7). From (
2.9) and (
2.12), we have the following differential equation:
(2.13)
This differential equation has the following general solution:
(2.14)
where
dq is an integral constant depending on
q. From
gq(1) = 0, we have
dq =
cq/(1 −
q). Thus, we have
(2.15)
Finally, we have
(2.16)
From
fq(
x) ≥ 0, we have
cq ≥ 0.
If we take the limit as q → 1 in Proposition 2.1, we have the following corollary.
Corollary 2.2 (see [17].)If the differentiable nonnegative function f satisfies the following functional equation:
(2.17)
then the function
f is uniquely given by
(2.18)
where
c is a nonnegative constant.
3. Main Results
In this section, we give a characterization of a two-parameter extended entropy function by the functional equation. Before we give our main theorem, we review the following result given by Kannappan [18, 19].
Proposition 3.1 (see [18], [19].)Let two probability distributions (p1, …, pn) and (q1, …, qm). If the measureable function f : (0,1) → ℝ satisfies
(3.1)
for all (
p1, …,
pn) and (
q1, …,
qm) with fixed
m,
n ≥ 3, then the function
f is given by
(3.2)
where
c and
b are arbitrary constants.
Here, we review a two-parameter generalized Shannon additivity, [
14, equation (30)]
(3.3)
where
sα,β is a component of the trace form of the two-parameter entropy [
14, equation (26)]
(3.4)
Equation (3.3) was used to prove the uniqueness theorem for two-parameter extended entropy in [14]. As for (3.3), a tree-graphical interpretation was given in [14]. The condition (3.1) can be read as the independent case (p(j∣i) = pj) in (3.3).
Here, we consider the nontrivial simplest case for (
3.3). Take
pij = {
q1,
q2,
q3},
p1 =
q1 +
q2, and
p2 =
q3. then we have
p(1∣1) =
q1/(
q1 +
q2),
p(2∣1) =
q2/(
q1 +
q2),
p(1∣2) = 1, and
p(2∣2) = 0, then (
3.3) is written by
(3.5)
If
sα,β is an entropic function, then it vanishes at 0 or 1, since the entropy has no informational quantity for the deterministic cases, then the above identity is reduced in the following:
(3.6)
In the following theorem, we adopt a simpler condition than (3.1).
Theorem 3.2. If the differentiable nonnegative function fα,β with two positive parameters α, β ∈ ℝ satisfies the following functional equation:
(3.7)
then the function
fα,β is uniquely given by
(3.8)
where
cα,β and
cα are nonnegative constants depending only on the parameters
α (and
β).
Proof. If we put y = 1, then we have fα,β(1) = 0 due to x > 0. By differentiating (3.7) with respect to y, we have
(3.9)
Putting
y = 1 in (
3.9), we have the following differential equation:
(3.10)
where we put
. Equation (
3.10) can be deformed as follows:
(3.11)
that is, we have
(3.12)
Integrating both sides on the above equation with respect to
x, we have
(3.13)
where
dα,β is a integral constant depending on
α and
β. Therefore, we have
(3.14)
By
fα,β(1) = 0, we have
dα,β =
cα,β/(
α −
β). Thus, we have
(3.15)
Also by
fα,β(
x) ≥ 0, we have
cα,β ≥ 0.
As for the case of α = β, we can prove by the similar way.
Remark 3.3. We can derive (3.6) from our condition (3.7). Firstly, we easily have fα,β(0) = fα,β(1) = 0 from our condition equation (3.7). In addition, we have for q = q1 + q2,
(3.16)
Thus, we may interpret that our condition (
3.7) contains an essential part of the two-parameter generalized Shannon additivity.
Note that we can reproduce the two-parameter entropic function by the use of fα,β as
(3.17)
with
cα,β = 1 for simplicity. This leads to two-parameter extended relative entropy [
15]
(3.18)
See also [
20] on the first appearance of the Tsallis relative entopy (generalized Kullback-Leibler information).
If we take α = q, β = 1 or α = 1, β = q in Theorem 3.2, we have the following corollary.
Corollary 3.4. If the differentiable nonnegative function fq with a positive parameter q ∈ ℝ satisfies the following functional equation:
(3.19)
then the function
fq is uniquely given by
(3.20)
where
cq is a nonnegative constant depending only on the parameter
q.
Here, we give an interpretation of the functional equation (3.19) from the view of Tsallis statistics.
Remark 3.5. We assume that we have the following two functional equations for 0 < x, y ≤ 1:
(3.21)
These equations lead to the following equations for 0 <
xi,
yj ≤ 1:
(3.22)
where
i = 1, …,
n and
j = 1, …,
m. Taking the summation on
i and
j in both sides, we have
(3.23)
(3.24)
under the condition
. If the function
fq(
x) is given by (
3.20), then two above functional equations coincide with two nonadditivity relations given in (
1.4) and (
1.11).
On the other hand, we have the following equation from (23) and (3.21):
(3.25)
By a similar way to the proof of Theorem
3.2, we can show that the functional equation (
3.25) uniquely determines the function
fq by the form given in (
3.20). Therefore, we can conclude that two functional equations (23) and (
3.21), which correspond to the non-additivity relations (
1.4) and (
1.11), also characterize Tsallis entropy function.
If we again take the limit as q → 1 in Corollary 3.4, we have the following corollary.
Corollary 3.6. If the differentiable nonnegative function f satisfies the following functional equation:
(3.26)
then the function
f is uniquely given by
(3.27)
where
c is a nonnegative constant.
4. Conclusion
As we have seen, the two-parameter extended entropy function can be uniquely determined by a simple functional equation. Also an interpretation related to a tree-graphical structure was given as a remark.
Recently, the extensive behaviours of generalized entropies were studied in [21–23]. Our condition given in (3.7) may be seen as extensive form. However, I have not yet found any relation between our functional (3.7) and the extensive behaviours of the generalized entropies. This problem is not the purpose of the present paper, but it is quite interesting to study this problem as a future work.
Acknowledgments
This paper is dedicated to Professor Kenjiro Yanagi on his 60th birthday. The author would like to thank the anonymous reviewers for providing valuable comments to improve the paper. The author was partially supported by the Japanese Ministry of Education, Science, Sports and Culture, Grant-in-Aid for Encouragement of Young Scientists (B) 20740067.