Pages that link to "Item:Q3245606"
From MaRDI portal
The following pages link to The Individual Ergodic Theorem of Information Theory (Q3245606):
Displaying 50 items.
- A monotone Sinai theorem (Q272941) (← links)
- The generalized entropy ergodic theorem for nonhomogeneous Markov chains (Q325891) (← links)
- On moving averages and asymptotic equipartition of information (Q343262) (← links)
- Relative complexity of random walks in random sceneries (Q693714) (← links)
- Contributions to information theory for abstract alphabets (Q775918) (← links)
- On the ergodic theory of Tanaka-Ito type \(\alpha \)-continued fractions (Q831883) (← links)
- Large deviations for non-uniformly expanding maps (Q867041) (← links)
- Some limit properties for the \(m\)th-order nonhomogeneous Markov chains indexed by an m rooted Cayley tree (Q984014) (← links)
- A limit theorem for the entropy density of nonhomogeneous Markov information source (Q1347196) (← links)
- A conversation with Leo Breiman. (Q1431203) (← links)
- The Shannon-McMillan theorem for ergodic quantum lattice systems (Q1434287) (← links)
- Optimal data compression algorithm (Q1816650) (← links)
- Some remarks concerning the individual ergodic theorem of information theory (Q1844649) (← links)
- The Markov approximation of the sequences of \(N\)-valued random variables and a class of small deviation theorems. (Q1877526) (← links)
- An extension of Shannon-McMillan theorem and some limit properties for nonhomogeneous Markov chains (Q1915828) (← links)
- On universal algorithms for classifying and predicting stationary processes (Q2039763) (← links)
- A spectral representation for the entropy of topological dynamical systems (Q2041852) (← links)
- On local metric pressure of dynamical systems (Q2043721) (← links)
- Entropy as an integral operator: erratum and modification (Q2057891) (← links)
- The Shannon-McMillan-Breiman theorem beyond amenable groups (Q2073250) (← links)
- Tree-indexed Markov chains in random environment and some of their strong limit properties (Q2082275) (← links)
- Shannon's theorem for locally compact groups (Q2119207) (← links)
- An entropy problem of the \(\alpha \)-continued fraction maps (Q2134249) (← links)
- An optimal uniform concentration inequality for discrete entropies on finite alphabets in the high-dimensional setting (Q2137047) (← links)
- Markov approximation and the generalized entropy ergodic theorem for non-null stationary process (Q2174834) (← links)
- Entropy and compression: a simple proof of an inequality of Khinchin-Ornstein-Shields (Q2190980) (← links)
- Bowen entropy of sets of generic points for fixed-point free flows (Q2202273) (← links)
- A strong limit theorem for the average of ternary functions of Markov chains in bi-infinite random environments (Q2348312) (← links)
- An information-theoretic analysis of return maximization in reinforcement learning (Q2375396) (← links)
- Singular measures and Hausdorff measures (Q2536905) (← links)
- Ergodic properties of conditional forecast functions of stable systems (Q2564223) (← links)
- The generalized entropy ergodic theorem for nonhomogeneous bifurcating Markov chains indexed by a binary tree (Q2676986) (← links)
- The strong law of large numbers and the Shannon-McMillan theorem for the mth-order nonhomogeneous Markov chains indexed by an m rooted Cayley tree (Q2811414) (← links)
- Fuzzy entropy of action of semi-groups (Q2958879) (← links)
- Strong laws of large numbers for the <i>m</i>th-order asymptotic odd–even Markov chains indexed by an <i>m</i>-rooted Cayley tree (Q2979965) (← links)
- A class of strong limit theorems for countable nonhomogeneous Markov chains on the generalized gambling system (Q3070104) (← links)
- Entropy Rate and Maximum Entropy Methods for Countable Semi-Markov Chains (Q3155281) (← links)
- An ergodic theorem for constrained sequences of functions (Q3480311) (← links)
- Nonparametric sequential prediction of time series (Q3569202) (← links)
- Some Research on Shannon–McMillan Theorem for<i>m</i>th-Order Nonhomogeneous Markov Information Source (Q3651642) (← links)
- Bounds on Data Compression Ratio with a Given Tolerable Error Probability (Q4266369) (← links)
- The Comparison between Arbitrary Information Sources and Memoryless Information Sources and its Limit Properties (Q4321113) (← links)
- Strong law of large numbers for generalized sample relative entropy of non homogeneous Markov chains (Q4563470) (← links)
- On the entropy for semi-Markov processes (Q4819515) (← links)
- The Asymptotic Equipartition Property for a Nonhomogeneous Markov Information Source (Q4950731) (← links)
- Entropy of flows, revisited (Q4950753) (← links)
- UNIVERSAL CODING AND PREDICTION ON ERGODIC RANDOM POINTS (Q5044311) (← links)
- Strong deviation theorems for general information sources (Q5045659) (← links)
- Upper Bounds on Mixing Time of Finite Markov Chains (Q5055645) (← links)
- THE STRONG LIMIT THEOREM FOR RELATIVE ENTROPY DENSITY RATES BETWEEN TWO ASYMPTOTICALLY CIRCULAR MARKOV CHAINS (Q5056609) (← links)