...

  • Home
  • Articles
    • Current Issue
    • Archives
  • Authors
    • Author Guidelines
    • Policies
    • Downloads
  • Editors
  • Reviewers
...

International Journal of Mathematics Trends and Technology

Research Article | Open Access | Download PDF

Volume 70 | Issue 6 | Year 2024 | Article Id. IJMTT-V70I6P101 | DOI : https://doi.org/10.14445/22315373/IJMTT-V70I6P101

Edgeworth Theoretic Characterization of Mutual Information Underneath Joint and Conditional Probability Distribution in Algorithmic Information Theory


Rohit Kumar Verma, JharanaChandrakar
Received Revised Accepted Published
06 Apr 2024 20 May 2024 09 Jun 2024 29 Jun 2024
Abstract

The science of communication started to attract a lot of attention in the early 1900s. Communication and the transmission of information was rapidly becoming a global phenomenon due to the ongoing growth of telegraph communication, the recent invention of the telephone [3], and, most importantly, the necessity for quick and dependable connections between military groups during the World War. There was an urgent need by the 1940s for a mathematical theory that would direct the development of more advanced communication technologies. 

Keywords

Entropy, Information, Information improvements, Discrete information, Continuous information. 1. Introduction The seminal work that gave rise to this theory was the groundbreaking publication A Mathematical Theory o

References

[1] Differentialentropy. [Online]. Available: https://en.wikipedia.org/wiki/Differentialentropy [2] R.V.L. Hartley, “Transmission of Information,” Bell System Technical Journal, vol. 7, no. 3, pp. 535-563, 1928.
[CrossRef] [Google Scholar] [Publisher Link]
[3] George Markowsky, Information Theory, 2024. [Online]. Available: https://www.britannica.com/science/information-theory
[4] Robert J. McEliece, The Theory of Information and Coding, Cambridge: New York, 2004.
[Google Scholar] [Publisher Link]
[5] H. Nyquist, “Certain Factors Affecting Telegraph Speed,” Journal of the A.I.E.E., vol. 43, no. 2, pp. 124-130, 1924.
[CrossRef] [Google Scholar] [Publisher Link]
[6] M.S. Pinsker, Information and Information Stability of Random Variables and Processes, San Francisco: Holden-Day, pp. 1-243, 1964.
[Google Scholar] [Publisher Link]
[7] C.E. Shannon, and W. Weaver, “The Mathematical Theory of Communication,” Urbana: University of Illinois Press, 1972.
 [8] Terence Tao, An Introduction to Measure Theory, Providence, RI: American Mathematical Society, 2013.
[Google Scholar] [Publisher Link]
[9] Rocco Tenaglia, Entropy and Information in Discrete and Continuous Information Theory, 2017.
[Google Scholar]
 [10] Norbert Wiener, Extrapolation, Interpolation and Smoothing of Stationary Time Series: With Engineering Applications, The MIT Press, 1949.
[CrossRef] [Google Scholar] [Publisher Link]
[11] Rohit Kumar Verma, and BabitaVerma, A New Approach in Mathematical Theory of Communication (A New Entropy with its Application), Lambert Academic Publishing, 2013.
[Google Scholar]
 [12] R.K. Verma, Family of Measures of Information with their Applications in Coding Theory and Channel Capacity, Lambert Academic Publishing, 2023

Citation :

Rohit Kumar Verma, JharanaChandrakar, "Edgeworth Theoretic Characterization of Mutual Information Underneath Joint and Conditional Probability Distribution in Algorithmic Information Theory," International Journal of Mathematics Trends and Technology (IJMTT), vol. 70, no. 6, pp. 1-12, 2024. Crossref, https://doi.org/10.14445/22315373/IJMTT-V70I6P101

  • PDF
  • Abstract
  • Keywords
  • References
  • Citation
Abstract Keywords References Citation
  • Home
  • Authors Guidelines
  • Paper Submission
  • APC
  • Archives
  • Downloads
  • Open Access
  • Publication Ethics
  • Copyrights Infringement
  • Journals
  • FAQ
  • Contact Us

Follow Us

Copyright © 2025 Seventh Sense Research Group® . All Rights Reserved