Comparison of two methods for approximation of probability distributions with prescribed marginals

Albert Pérez; Milan Studený

Kybernetika (2007)

  • Volume: 43, Issue: 5, page 591-618
  • ISSN: 0023-5954

Abstract

top
Let P be a discrete multidimensional probability distribution over a finite set of variables N which is only partially specified by the requirement that it has prescribed given marginals { P A ; A 𝒮 } , where 𝒮 is a class of subsets of N with 𝒮 = N . The paper deals with the problem of approximating P on the basis of those given marginals. The divergence of an approximation P ^ from P is measured by the relative entropy H ( P | P ^ ) . Two methods for approximating P are compared. One of them uses formerly introduced concept of dependence structure simplification (see Perez [Per79]). The other one is based on an explicit expression, which has to be normalized. We give examples showing that neither of these two methods is universally better than the other. If one of the considered approximations P ^ really has the prescribed marginals then it appears to be the distribution P with minimal possible multiinformation. A simple condition on the class 𝒮 implying the existence of an approximation P ^ with prescribed marginals is recalled. If the condition holds then both methods for approximating P give the same result.

How to cite

top

Pérez, Albert, and Studený, Milan. "Comparison of two methods for approximation of probability distributions with prescribed marginals." Kybernetika 43.5 (2007): 591-618. <http://eudml.org/doc/33883>.

@article{Pérez2007,
abstract = {Let $P$ be a discrete multidimensional probability distribution over a finite set of variables $N$ which is only partially specified by the requirement that it has prescribed given marginals $\lbrace P_\{A\};\ A\in \{\mathcal \{S\}\} \rbrace $, where $\{\mathcal \{S\}\}$ is a class of subsets of $N$ with $\bigcup \{\mathcal \{S\}\} = N$. The paper deals with the problem of approximating $P$ on the basis of those given marginals. The divergence of an approximation $\hat\{P\}$ from $P$ is measured by the relative entropy $H(P|\hat\{P\})$. Two methods for approximating $P$ are compared. One of them uses formerly introduced concept of dependence structure simplification (see Perez [Per79]). The other one is based on an explicit expression, which has to be normalized. We give examples showing that neither of these two methods is universally better than the other. If one of the considered approximations $\hat\{P\}$ really has the prescribed marginals then it appears to be the distribution $P$ with minimal possible multiinformation. A simple condition on the class $\{\mathcal \{S\}\}$ implying the existence of an approximation $\hat\{P\}$ with prescribed marginals is recalled. If the condition holds then both methods for approximating $P$ give the same result.},
author = {Pérez, Albert, Studený, Milan},
journal = {Kybernetika},
keywords = {marginal problem; relative entropy; dependence structure simplification; explicit expression approximation; multiinformation; decomposable model; asteroid; marginal problem; relative entropy; multiinformation; decomposable model},
language = {eng},
number = {5},
pages = {591-618},
publisher = {Institute of Information Theory and Automation AS CR},
title = {Comparison of two methods for approximation of probability distributions with prescribed marginals},
url = {http://eudml.org/doc/33883},
volume = {43},
year = {2007},
}

TY - JOUR
AU - Pérez, Albert
AU - Studený, Milan
TI - Comparison of two methods for approximation of probability distributions with prescribed marginals
JO - Kybernetika
PY - 2007
PB - Institute of Information Theory and Automation AS CR
VL - 43
IS - 5
SP - 591
EP - 618
AB - Let $P$ be a discrete multidimensional probability distribution over a finite set of variables $N$ which is only partially specified by the requirement that it has prescribed given marginals $\lbrace P_{A};\ A\in {\mathcal {S}} \rbrace $, where ${\mathcal {S}}$ is a class of subsets of $N$ with $\bigcup {\mathcal {S}} = N$. The paper deals with the problem of approximating $P$ on the basis of those given marginals. The divergence of an approximation $\hat{P}$ from $P$ is measured by the relative entropy $H(P|\hat{P})$. Two methods for approximating $P$ are compared. One of them uses formerly introduced concept of dependence structure simplification (see Perez [Per79]). The other one is based on an explicit expression, which has to be normalized. We give examples showing that neither of these two methods is universally better than the other. If one of the considered approximations $\hat{P}$ really has the prescribed marginals then it appears to be the distribution $P$ with minimal possible multiinformation. A simple condition on the class ${\mathcal {S}}$ implying the existence of an approximation $\hat{P}$ with prescribed marginals is recalled. If the condition holds then both methods for approximating $P$ give the same result.
LA - eng
KW - marginal problem; relative entropy; dependence structure simplification; explicit expression approximation; multiinformation; decomposable model; asteroid; marginal problem; relative entropy; multiinformation; decomposable model
UR - http://eudml.org/doc/33883
ER -

References

top
  1. Csiszár I., Matúš F., Information projections revisited, IEEE Trans. Inform. Theory 49 (2003), 1474–1490 Zbl1063.94016MR1984936
  2. Kellerer H. G., Verteilungsfunktionen mit gegebenem Marginalverteilungen (in German, translation: Distribution functions with given marginal distributions), Z. Wahrsch. verw. Gerbiete 3 (1964), 247–270 (1964) MR0175158
  3. Lauritzen S. L., Graphical Models, Clarendon Press, Oxford 1996 Zbl1055.62126MR1419991
  4. Perez A., ε -admissible simplifications of the dependence structure of random variables, Kybernetika 13 (1979), 439–449 (1979) MR0472224
  5. Perez A., The barycenter concept of a set of probability measures as a tool in statistical decision, In: The book of abstracts of the 4th Internat. Vilnius Conference on Probability Theory and Mathematical Statistics 1985, pp. 226–228 (1985) 
  6. Perez A., Princip maxima entropie a princip barycentra při integraci dílčích znalostí v expertních systémech (in Czech, translation: The maximum entropy principle and the barycenter principle in partial knowledge integration in expert systems), In: Metody umělé inteligence a expertní systémy III (V. Mařík and Z. Zdráhal, eds.), ČSVT – FEL ČVUT, Prague 1987, pp. 62–74 (1987) 
  7. Perez A., Explicit expression Exe – containing the same multiinformation as that in the given marginal set – for approximating probability distributions, A manuscript in Word, 2003 
  8. Studený M., Pojem multiinformace v pravděpodobnostním rozhodování (in Czech, translation: The notion of multiinformation in probabilistic decision-making), CSc Thesis, Czechoslovak Academy of Sciences, Institute of Information Theory and Automation, Prague 1987 
  9. Studený M., Probabilistic Conditional Independence Structures, Springer–Verlag, London 2005 

NotesEmbed ?

top

You must be logged in to post comments.

To embed these notes on your page include the following JavaScript code on your page where you want the notes to appear.

Only the controls for the widget will be shown in your chosen language. Notes will be shown in their authored language.

Tells the widget how many notes to show per page. You can cycle through additional notes using the next and previous controls.

    
                

Note: Best practice suggests putting the JavaScript code just before the closing </body> tag.