Abstract
This article proposes a new refinement of the celebrated Jensen inequality. Some refinements have been obtained for quasi-arithmetic means, Hölder and Hermite–Hadamard inequalities. Several applications are given in information theory. A more general refinement of Jensen inequality is presented associated to n functions.
Similar content being viewed by others
1 Introduction
The celebrated Jensen inequality states that: If I is an interval in \(\mathbb{R}\) and \(g,p:[a,b]\rightarrow \mathbb{R}\) are integrable functions such that \(g(\varrho )\in I, p(\varrho )>0 \ \forall \varrho \in [a,b]\). Also, if \(\psi:I\rightarrow \mathbb{R}\) is convex function and \((\psi \circ g).p\) is integrable on \([a,b]\). Then
Jensen’s inequality is one of the fundamental inequalities in mathematics and it underlies many vital statistical concepts and proofs. Some important applications involve derivation of the AM-GM inequality, estimations for Zipf–Mandelbrot and Shannon entropies, the convergence property of the expectation maximization algorithm, and positivity of Kullback–Leibler divergence [1–7]. Also, this inequality has been utilized to solve several problems in many areas of science and technology e.g. physics, engineering, financial economics and computer science.
There are several classical important inequalities which may be deduced from (1), for example Hölder, Levinson’s, and Ky Fan and Young’s inequalities. Due to the great importance of this inequality, several researchers have focused on this inequality and derived many improvements, refinements and extensions of the Jensen inequality. The Jensen inequality also has been given for some other generalized convex functions such as s-convex, preinvex, h-convex and η-convex functions. For some recent results concerning the Jensen inequality see [1–3, 5, 8–20].
In this article first of all we establish an interesting refinement of the Jensen inequality associated to two functions whose sum is equal to unity. Using this refinement, we derive refinements of Hölder, power mean, quasi-arithmetic mean and Hermite–Hadamard inequalities. We also focus on deducing bounds for Csiszár-divergence, Kullback–Leibler divergence, Shannon entropy and variational distance etc. We present a more general refinement of Jensen inequality concerning n functions whose sums are equal to unity.
2 Main results
We start to derive a new refinement of the Jensen inequality associated to two functions whose sum is equal to unity.
Theorem 1
Let\(\psi: I \rightarrow \mathbb{R}\)be a convex function defined on the intervalI. Let\(p,u,v,g:[a,b]\rightarrow \mathbb{R}\)be integrable functions such that\(g(\varrho )\in I, u(\varrho ), v(\varrho ), p(\varrho )\in \mathbb{R}^{+}\)for all\(\varrho \in [a,b]\)and\(v(\varrho )+u(\varrho )=1\), \(P=\int _{a}^{b}p(\varrho )\,d\varrho \). Then
If the functionψis concave then the reverse inequalities hold in (2).
Proof
Since \(u(\varrho )+v(\varrho )=1\), so we have
Applying the integral Jensen inequality on both terms on the right side of (3) we obtain
□
As a consequence of the above theorem we deduce the following refinement of the Hölder inequality.
Corollary 1
Let\(r_{1},r_{2}>1\)be such that\(\frac{1}{r_{1}}+\frac{1}{r_{2}}=1\). If\(u,v,\tau,g_{1}\)and\(g_{2}\)are non-negative functions defined on\([a,b]\)such that\(\tau g_{1}^{r_{1}},\tau g_{2}^{r_{2}},u\tau g_{2}^{r_{2}}, v\tau g_{2}^{r_{2}}, u\tau g_{1}g_{2},v\tau g_{1}g_{2},\tau g_{1}g_{2}\in L^{1}([a,b])\)and\(u(\varrho )+v(\varrho )=1\)for all\(\varrho \in [a,b]\), then
In the case when\(0< r_{1}<1\)and\(r_{2}=\frac{r_{1}}{r_{1}-1}\)with\(\int _{a}^{b}\tau (\varrho )g^{r_{2}}_{2}(\varrho )\,d\varrho >0\)or\(r_{1}<0\)and\(\int _{a}^{b}\tau (\varrho )g^{r_{1}}_{1}(\varrho )\,d\varrho >0\), then we have
Proof
If \(\int _{a}^{b}\tau (\varrho )g^{r_{2}}_{2}(\varrho )\,d\varrho >0\), then by using Theorem 1 for \(\psi (\varrho )=\varrho ^{r_{1}}, \varrho >0, r_{1}>1\), \(p(\varrho )=\tau (\varrho )g_{2}^{r_{2}}(\varrho ), g(\varrho )=g_{1}( \varrho )g_{2}^{\frac{-r_{2}}{r_{1}}}(\varrho )\), we obtain (5). If \(\int _{a}^{b}\tau (\varrho )g^{r_{1}}_{1}(\varrho )\,d\varrho >0\), then applying the same procedure but taking \(r_{1}, r_{2}, g_{1},g_{2}\) instead of \(r_{2}, r_{1}, g_{2},g_{1}\), we obtain (5).
Set \(\int _{a}^{b}\tau (\varrho )g^{r_{2}}_{2}(\varrho )\,d\varrho =0\) and \(\int _{a}^{b}\tau (\varrho )g^{r_{1}}_{1}(\varrho )\,d\varrho =0\). We know that
Therefore taking the integral and then using the given conditions we have \(\int _{a}^{b}\tau (\varrho )g_{1}(\varrho )\times g_{2}(\varrho )\,d\varrho =0\).
For the case \(r_{1}>1\), the proof is completed.
For the case when \(0< r_{1}<1\), \(M=\frac{1}{r_{1}}>1\) and applying (5) for M and \(N=(1-r_{1})^{-1}, \overline{g}_{1}= (g_{1}g_{2})^{r_{1}}, \overline{g}_{2}=g^{-r_{1}}_{2}\) instead of \(r_{1},r_{2},g_{1},g_{2}\).
Finally, if \(r_{1}<0\) then \(0< r_{2}<1\) and we may apply similar arguments with \(r_{1}, r_{2}, g_{1},g_{2}\) replaced by \(r_{2}, r_{1}, g_{2},g_{1} \) provided that \(\int _{a}^{b}\tau (\varrho )g^{r_{1}}_{1}(\varrho )\,d\varrho >0\). □
Another refinement of the Hölder inequality presented in the following corollary.
Corollary 2
Let\(r_{1}>1, r_{2}=\frac{r_{1}}{r_{1}-1}\). If\(u,v,\tau,g_{1}\)and\(g_{2}\)are non-negative functions defined on\([a,b]\)such that\(\tau g_{1}^{r_{1}},\tau g_{2}^{r_{2}},u\tau g_{2}^{r_{2}}, v\tau g_{2}^{r_{2}}, \tau g_{1}g_{2}\in L^{1}([a,b])\)and\(u(\varrho )+v(\varrho )=1\)for all\(\varrho \in [a,b]\), also assuming that\(\int _{a}^{b}\tau (\varrho )g^{r_{2}}_{2}(\varrho )>0\), then
In the case when\(0< r_{1}<1\)and\(r_{2}=\frac{r_{1}}{r_{1}-1}\)with\(\int _{a}^{b}\tau (\varrho )g^{r_{2}}_{2}(\varrho )\,d\varrho >0\)or\(r_{1}<0\)and\(\int _{a}^{b}\tau (\varrho )g^{r_{1}}_{1}(\varrho )\,d\varrho >0\), then we have
Proof
Assume that \(\int _{a}^{b}\tau (\varrho )g^{r_{2}}_{2}(\varrho )\,d\varrho >0\). Let \(\psi (\varrho )=\varrho ^{\frac{1}{r_{1}}}\), \(\varrho >0,r_{1}>1 \). Then clearly the function ψ is concave. Therefore applying Theorem 1 for \(\psi (\varrho )=\varrho ^{\frac{1}{r_{1}}}, p=\tau g^{r_{2}}_{2}, g=g^{r_{1}}_{1}g^{-r_{2}}_{2}\), we obtain (8). If \(\int _{a}^{b}\tau (\varrho )g^{r_{1}}_{1}(\varrho )\,d\varrho >0\), then applying the same procedure but taking \(r_{1}, r_{2}, g_{1},g_{2}\) instead of \(r_{2}, r_{1}, g_{2},g_{1}\), we obtain (8).
If \(\int _{a}^{b}\tau (\varrho )g^{r_{2}}_{2}(\varrho )\,d\varrho =0\) and \(\int _{a}^{b}\tau (\varrho )g^{r_{1}}_{1}(\varrho )\,d\varrho =0\), then since as we know that
Therefore taking the integral and then using the given conditions we have \(\int _{a}^{b}\tau (\varrho )\times g_{1}(\varrho )g_{2}(\varrho )\,d\varrho =0\).
In the case when \(0< r_{1}<1\), \(M=\frac{1}{r_{1}}>1\) and applying (8) for M and \(N=(1-r_{1})^{-1}, \overline{g}_{1}= (g_{1}g_{2})^{r_{1}}, \overline{g}_{2}=g^{-r_{1}}_{2}\) instead of \(r_{1},r_{2},g_{1},g_{2}\), we get (9).
Finally, if \(r_{1}<0\) then \(0< r_{2}<1\) and we may apply similar arguments with \(r_{1}, r_{2}, g_{1},g_{2}\) replaced by \(r_{2}, r_{1}, g_{2},g_{1} \) provided that \(\int _{a}^{b}\tau (\varrho )g^{r_{1}}_{1}(\varrho )\,d\varrho >0\). □
Remark 1
If we put \(u(\varrho )=\frac{b-\varrho }{b-a},v(\varrho )= \frac{\varrho -a}{b-a}\) in (8), then we deduce the inequalities which have been obtained by Işcan in [21].
Let p and g be positive integrable functions defined on \([a,b]\). Then the integral power means of order \(r\in \mathbb{R}\) are defined as follows:
In the following corollary we deduce inequalities for power means.
Corollary 3
Let\(p,u,v\)andgbe positive integrable functions defined on\([a,b]\)with\(u(\varrho )+v(\varrho )=1\)for all\(\varrho \in [a,b]\). Let\(s,t\in \mathbb{R}\)such that\(s\leq t\). Then
Proof
If \(s,t\in \mathbb{R}\) and \(s,t\neq 0\), then using (2) for \(\psi (\varrho )=\varrho ^{\frac{t}{s}}\), \(\varrho >0\), \(g\rightarrow g^{s}\) and then taking the power \(\frac{1}{t}\) we get (12). For the case \(t=0\), taking the limit \(t\rightarrow 0 \) in (12) we obtain (13). We have the same for \(s=0\) taking the limit.
Similarly taking (2) for \(\psi (\varrho )=\varrho ^{\frac{s}{t}}\), \(\varrho >0,s,t\neq 0\), \(g\rightarrow g^{t}\) and then taking the power \(\frac{1}{s}\) we get (14). For \(s=0\) or \(t=0\) we take the limit as above. □
Let p be positive integrable function defined on \([a,b]\) and g be any integrable function defined on \([a,b]\). Then, for a strictly monotone continuous function h whose domain belongs to the image of g, the quasi-arithmetic mean is defined as follows:
We give inequalities for the quasi-arithmetic mean.
Corollary 4
Let\(u,v,p\)be positive integrable functions defined on\([a,b]\)such that\(u(\varrho )+v(\varrho )=1\)for all\(\varrho \in [a,b]\)andgbe any integrable function defined on\([a,b]\). Also assume thathis a strictly monotone continuous function whose domain belongs to the image ofg. If\(f\circ h^{-1}\)is convex function then
If the function\(f\circ h^{-1}\)is concave then the reverse inequalities hold in (17).
Proof
The required inequalities may be deduced by using (2) for \(g\rightarrow h\circ g\) and \(\psi \rightarrow f\circ h^{-1}\). □
The following refinement of the Hermite–Hadamard inequality may be given.
Corollary 5
Let\(\psi: [a,b] \rightarrow \mathbb{R}\)be a convex function defined on the interval\([a,b]\). Let\(u,v:[a,b]\rightarrow \mathbb{R}\)be integrable functions such that\(u(\varrho ), v(\varrho )\in \mathbb{R}^{+}\)for all\(\varrho \in [a,b]\)and\(u(\varrho )+v(\varrho )=1\). Then
For the concave functionψthe reverse inequalities hold in (18).
Proof
Using Theorem 1 for \(p(\varrho )=1, g(\varrho )=\varrho \) for all \(\varrho \in [a,b]\), we obtain (18). □
3 Applications in information theory
In this section, we present some important applications for different divergences and distances in information theory [22] of our main result.
Definition 1
(Csiszár divergence)
Let \(T:I\rightarrow \mathbb{R}\) be a function defined on the positive interval I. Also let \(u_{1},v_{1}:[a,b]\rightarrow (0,\infty )\) be two integrable functions such that \(\frac{u_{1}(\varrho )}{v_{1}(\varrho )}\in I\) for all \(\varrho \in [a,b]\), then the Csiszár divergence is defined as
Theorem 2
Let\(T: I \rightarrow \mathbb{R}\)be a convex function defined on the positive intervalI. Let\(u,v,u_{1},v_{1}:[a,b]\rightarrow \mathbb{R}^{+}\)be integrable functions such that\(\frac{u_{1}(\varrho )}{v_{1}(\varrho )}\in I\)and\(u(\varrho )+v(\varrho )=1\)for all\(\varrho \in [a,b]\). Then
Proof
Using Theorem 1 for \(\psi =T\), \(g=\frac{u_{1}}{v_{1}}\) and \(p=v_{1}\), we obtain (19). □
Definition 2
(Shannon entropy)
If \(v_{1}(\varrho )\) is positive probability density function defined on \([a,b]\), then the Shannon entropy is defined by
Corollary 6
Let\(u,v,v_{1}:[a,b]\rightarrow \mathbb{R}^{+}\)be integrable functions such that\(v_{1}\)is probability density function and\(u(\varrho )+v(\varrho )=1\)for all\(\varrho \in [a,b]\). Then
Proof
Taking \(T(\varrho )=-\log \varrho, \varrho \in \mathbb{R}^{+}\), in (19), we obtain (20). □
Definition 3
(Kullback–Leibler divergence)
If \(u_{1}\) and \(v_{1}\) are two positive probability densities defined on \([a,b]\), the Kullback–Leibler divergence is defined by
Corollary 7
Let\(u,v,u_{1},v_{1}:[a,b]\rightarrow \mathbb{R}^{+}\)be integrable functions such that\(u_{1}\)and\(v_{1}\)are probability density functions and\(u(\varrho )+v(\varrho )=1\)for all\(\varrho \in [a,b]\). Then
Proof
Taking \(T(\varrho )=\varrho \log \varrho, \varrho \in \mathbb{R}^{+}\), in (19), we obtain (20). □
Definition 4
(Variational distance)
If \(u_{1}\) and \(v_{1}\) are positive probability density functions defined on \([a,b]\), then the variational distance is defined by
Corollary 8
Let\(u,v,u_{1},v_{1}\)be as stated in Corollary7. Then
Proof
Using the function \(T(\varrho )=| \varrho -1|, \varrho \in \mathbb{R}^{+}\), in (19), we obtain (22). □
Definition 5
(Jeffrey’s distance)
If \(u_{1}\) and \(v_{1}\) are two positive probability density functions defined on \([a,b]\), then the Jeffrey distance is defined by
Corollary 9
Let\(u,v,u_{1},v_{1}\)be as stated in Corollary7. Then
Proof
Using the function \(T(\varrho )=(\varrho -1)\log \varrho, \varrho \in \mathbb{R}^{+}\), in (19), we obtain (23). □
Definition 6
(Bhattacharyya coefficient)
If \(u_{1}\) and \(v_{1}\) are two positive probability density functions defined on \([a,b]\), then the Bhattacharyya coefficient is defined by
Corollary 10
Let\(u,v,u_{1},v_{1}\)be as stated in Corollary7. Then
Proof
Using the function \(T(\varrho )=-\sqrt{\varrho }, \varrho \in \mathbb{R}^{+}\), in (19), we obtain (24). □
Definition 7
(Hellinger distance)
If \(u_{1}\) and \(v_{1}\) are two positive probability density functions defined on \([a,b]\), then the Hellinger distance is defined by
Corollary 11
Let\(u,v,u_{1},v_{1}\)be as stated in Corollary7. Then
Proof
Using the function \(T(\varrho )=(\sqrt{\varrho }-1)^{2}, \varrho \in \mathbb{R}^{+}\), in (19), we obtain (25). □
Definition 8
(Triangular discrimination)
If \(u_{1}\) and \(v_{1}\) are two positive probability density functions defined on \([a,b]\), then the triangular discrimination between \(u_{1}\) and \(v_{1}\) is defined by
Corollary 12
Let\(u,v,u_{1},v_{1}\)be as stated in Corollary7. Then
Proof
Since the function \(\phi (\varrho )= \frac{(\varrho -1)^{2}}{\varrho +1}, \varrho \in \mathbb{R}^{+}\), is convex, using the function \(T(\varrho )=\phi (\varrho )\), in (19), we obtain (26). □
4 Further generalization
In the following theorem we present further refinement of the Jensen inequality concerning n functions whose sum is equal to unity.
Theorem 3
Let\(\psi: \mathtt{G} \rightarrow \mathbb{R}\)be a convex function defined on the intervalG. Let\(p,g,u_{l}\in L[a,b]\)such that\(g(\varrho )\in \mathtt{G}, p(\varrho ), u_{l}(\varrho )\in \mathbb{R}^{+}\)for all\(\varrho \in [a,b]\)\((l=1,2,\ldots,n)\)and\(\sum_{l=1}^{n}u_{l}(\varrho )=1\), \(P=\int _{a}^{b}p(\varrho )\,d\varrho \). Assume that\(L_{1}\)and\(L_{2}\)are non-empty disjoint subsets of\(\{1,2,\ldots,n\}\)such that\(L_{1}\cup L_{2}=\{1,2,\ldots,n\}\). Then
If the functionψis concave then the reverse inequalities hold in (27).
Proof
Since \(\sum_{l=1}^{n}u_{l}(\varrho )=1\), we may write
Applying integral Jensen’s inequality on both terms on the right hand side of (28) we obtain
□
Remark 2
If we take \(n=2\), in Theorem 3, we deduce Theorem 1. Also, analogously to the previous sections we may give applications of Theorem 3 for different means, the Hölder inequality and information theory.
References
Adil Khan, M., Al-Sahwi, Z.M., Ming Chu, Y.: New estimations for Shannon and Zipf–Mandelbrot entropies. Entropy 20(608), 1–10 (2018)
Adil Khan, M., Pečarić, Ð., Pečarić, J.: On Zipf–Mandelbrot entropy. J. Comput. Appl. Math. 346, 192–204 (2019)
Adil Khan, M., Pečarić, Ð., Pečarić, J.: Bounds for Shannon and Zipf–Mandelbrot entropies. Math. Methods Appl. Sci. 40(18), 7316–7322 (2017)
Dempster, A.P., Laird, N.M., Rubin, D.B.: Maximum likelihood from incomplete data via the EM algorithm. J. R. Stat. Soc., Ser. B 39, 1–38 (1977)
Adil Khan, M., Hanif, M., Khan, Z.A., Ahmad, K., Chu, Y.-M.: Association of Jensen inequality for s-convex function. J. Inequal. Appl. 2019, Article ID 162 (2019)
Adil Khan, M., Pečarić, Ð., Pečarić, J.: Bounds for Csiszár divergence and hybrid Zipf–Mandelbrot entropy. Math. Methods Appl. Sci. 42, 7411–7424 (2019)
Khan, S., Adil Khan, M., Chu, Y.-M.: Converses of the Jensen inequality derived from the Green functions with applications in information theory. Math. Methods Appl. Sci. 43(5), 2577–2587 (2020). https://doi.org/10.1002/mma.6066
Pečarić, J., Perić, J.: New improvement of the converse Jensen inequality. Math. Inequal. Appl. 21(1), 217–234 (2018)
Bakula, M.K., Nikodem, K.: Converse Jensen inequality for strongly convex set-valued maps. J. Math. Inequal. 12(2), 545–550 (2018)
Choi, D., Krnić, M., Pečarić, J.: More accurate classes of Jensen-type inequalities for convex and operator convex functions. Math. Inequal. Appl. 21(2), 301–321 (2018)
Sababheh, M.: Improved Jensen’s inequality. Math. Inequal. Appl. 20(2), 389–403 (2017)
Hot, J.M., Seo, Y.: An interpolation of Jensen’s inequality and its converses with applications to quasi-arithmetic mean inequalities. J. Math. Inequal. 12(2), 303–313 (2018)
Lu, G.: New refinements of Jensen’s inequality and entropy upper bounds. J. Math. Inequal. 12(2), 403–421 (2018)
Mikić, R., Pečarić, Ð., Pečarić, J.: Inequalities of the Jensen and Edmundson–Lah–Ribarić type for 3-convex functions with applications. J. Math. Inequal. 12(3), 677–692 (2018)
Song, Y.-Q., Adil Khan, M., Zahir Ullah, S., Ming-Chu, Y.: Integral inequalities for strongly convex functions. J. Funct. Spaces 2018, Article ID 6595921 (2018)
Khan, J., Adil Khan, M., Pečarić, J.: On Jensen’s type inequalities via generalized majorization inequalities. Filomat 32(16), 5719–5733 (2018)
Adil Khan, M., Ullah, S.Z., Chu, Y.: The concept of coordinate strongly convex functions and related inequalities. Rev. R. Acad. Cienc. Exactas Fís. Nat., Ser. A Mat. 113, 2235–2251 (2019)
Adil Khan, M., Khan, J., Pečarić, J.: Generalization of Jensen’s and Jensen–Steffensen’s inequalities by generalized majorization theorem. J. Math. Inequal. 11(4), 1049–1074 (2017)
Moradi, H.R., Omidvar, M.E., Adil Khan, M., Nikodem, K.: Around Jensen’s inequality for strongly convex functions. Aequ. Math. 92(1), 25–37 (2018)
Zaheer Ullah, S., Adil Khan, M., Chu, Y.-M.: A note on generalized convex functions. J. Inequal. Appl. 2009, Article ID 291 (2009)
Işcan, I.: New refinements for integral and sum forms of Hölder inequality. J. Inequal. Appl. 2019, Article ID 304 (2019)
Latif, N., Pečarić, Ð., Pečarić, J.: Majorization, useful Csiszar divergence and useful Zipf–Mandelbrot law. Open Math. 16, 1357–1373 (2018)
Acknowledgements
The publication was supported by the Ministry of Education and Science of the Russian Federation (Agreement number No. 02.a03.21.0008.)
Availability of data and materials
Not applicable.
Funding
There is no funding for this work.
Author information
Authors and Affiliations
Contributions
All authors contributed equally to the writing of this paper. All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare that they have no competing interests.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Adil Khan, M., Pečarić, Ð. & Pečarić, J. New refinement of the Jensen inequality associated to certain functions with applications. J Inequal Appl 2020, 76 (2020). https://doi.org/10.1186/s13660-020-02343-7
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s13660-020-02343-7