Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
125 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Sequences of Inequalities Among New Divergence Measures (1010.0412v2)

Published 3 Oct 2010 in cs.IT and math.IT

Abstract: There are three classical divergence measures exist in the literature on information theory and statistics. These are namely, Jeffryes-Kullback-Leiber J-divergence. Sibson-Burbea-Rao Jensen-Shannon divegernce and Taneja arithemtic-geometric mean divergence. These three measures bear an interesting relationship among each other and are based on logarithmic expressions. The divergence measures like Hellinger discrimination, symmetric chi-square divergence, and triangular discrimination are also known in the literature and are not based on logarithmic expressions. Past years Dragomir et al., Kumar and Johnson and Jain and Srivastava studied different kind of divergence measures. In this paper, we have presented some more new divergence measures and obtained inequalities relating these new measures and also made connections with previous ones. The idea of exponential divergence is also introduced.

Summary

We haven't generated a summary for this paper yet.