Cauchy-Stieltjes kernel families

Similar documents
Kernel families of probability measures. Saskatoon, October 21, 2011

Meixner matrix ensembles

Associated exponential families and elliptic functions

ULTRASPHERICAL TYPE GENERATING FUNCTIONS FOR ORTHOGONAL POLYNOMIALS

Octavio Arizmendi, Ole E. Barndorff-Nielsen and Víctor Pérez-Abreu

Failure of the Raikov Theorem for Free Random Variables

Free Meixner distributions and random matrices

TOPICS ON MEIXNER FAMILIES AND ULTRASPHERICAL TYPE GENERATING FUNCTIONS FOR ORTHOGONAL POLYNOMIALS. Marek Bozejko 1 and Nizar Demni 2

COMPUTING MOMENTS OF FREE ADDITIVE CONVOLUTION OF MEASURES

Characterizations of free Meixner distributions

On Equi-/Over-/Underdispersion. and Related Properties of Some. Classes of Probability Distributions. Vladimir Vinogradov

On Lévy measures for in nitely divisible natural exponential families

Characterizations of free Meixner distributions

Random Matrices: Beyond Wigner and Marchenko-Pastur Laws

Exponential homomorphism. non-commutative probability

Notes 9 : Infinitely divisible and stable laws

Exponential families of mixed Poisson distributions

Math 162: Calculus IIA

Measures, orthogonal polynomials, and continued fractions. Michael Anshelevich

Measures, orthogonal polynomials, and continued fractions. Michael Anshelevich

Quantum Probability and Asymptotic Spectral Analysis of Growing Roma, GraphsNovember 14, / 47

Random Toeplitz Matrices

Free Probability Theory and Random Matrices. Roland Speicher Queen s University Kingston, Canada

the convolution of f and g) given by

arxiv:math/ v3 [math.oa] 11 Jan 2006

arxiv: v2 [math.pr] 8 Feb 2016

E[X n ]= dn dt n M X(t). ). What is the mgf? Solution. Found this the other day in the Kernel matching exercise: 1 M X (t) =

Dispersion models for geometric sums

t x 1 e t dt, and simplify the answer when possible (for example, when r is a positive even number). In particular, confirm that EX 4 = 3.

Supermodular ordering of Poisson arrays

4 Invariant Statistical Decision Problems

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

FREE GENERALIZED GAMMA CONVOLUTIONS. Victor Pérez-Abreu and Noriyoshi Sakuma. Comunicación del CIMAT No I-08-19/ (PE /CIMAT)

Review of Probability Theory II

MATH115. Indeterminate Forms and Improper Integrals. Paolo Lorenzo Bautista. June 24, De La Salle University

Page Max. Possible Points Total 100

Advanced Machine Learning & Perception

Introduction to orthogonal polynomials. Michael Anshelevich

From the mesoscopic to microscopic scale in random matrix theory

Asymptotics for posterior hazards

Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao

GARCH processes probabilistic properties (Part 1)

Gaussian representation of a class of Riesz probability distributions

Lecture 1: August 28

IE 303 Discrete-Event Simulation

STAT 3610: Review of Probability Distributions

MFM Practitioner Module: Quantitative Risk Management. John Dodson. September 23, 2015

Statistics Ph.D. Qualifying Exam: Part I October 18, 2003

6.1 Moment Generating and Characteristic Functions

Jump-type Levy Processes

Empirical Comparison of ML and UMVU Estimators of the Generalized Variance for some Normal Stable Tweedie Models: a Simulation Study

Appendix B Positive definiteness

Lyapunov exponents of free operators

The random variable 1

CS Lecture 19. Exponential Families & Expectation Propagation

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Free infinite divisibility for generalized power distributions with free Poisson term

arxiv: v3 [math-ph] 21 Jun 2012

Data, Estimation and Inference

A Generalization of Wigner s Law

Lectures 6 7 : Marchenko-Pastur Law

Random matrices: Distribution of the least singular value (via Property Testing)

MIT Spring 2016

Review of Probabilities and Basic Statistics

Quadratic harnesses, q-commutations, and orthogonal martingale polynomials. Quadratic harnesses. Previous research. Notation and Motivation

Brief Review of Probability

Local law of addition of random matrices

Review for the previous lecture

Things to remember when learning probability distributions:

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable

Practice Questions for Final

Chi-square lower bounds

Method of Moments. which we usually denote by X or sometimes by X n to emphasize that there are n observations.

Exponential tail inequalities for eigenvalues of random matrices

1 Random Variable: Topics

ST5215: Advanced Statistical Theory

Asymptotics for posterior hazards

Write your Registration Number, Test Centre, Test Code and the Number of this booklet in the appropriate places on the answersheet.

1* (10 pts) Let X be a random variable with P (X = 1) = P (X = 1) = 1 2

Resistance Growth of Branching Random Networks

ON GENERALIZED VARIANCE OF NORMAL-POISSON MODEL AND POISSON VARIANCE ESTIMATION UNDER GAUSSIANITY

Approximation of Heavy-tailed distributions via infinite dimensional phase type distributions

STAT 7032 Probability. Wlodek Bryc

Free Convolution with A Free Multiplicative Analogue of The Normal Distribution

Continuous RVs. 1. Suppose a random variable X has the following probability density function: π, zero otherwise. f ( x ) = sin x, 0 < x < 2

15-388/688 - Practical Data Science: Basic probability. J. Zico Kolter Carnegie Mellon University Spring 2018

PROBABILITY MODELS OF LIFE LENGTH

Math Numerical Analysis Mid-Term Test Solutions

1 Infinitely Divisible Random Variables

Determinantal point processes and random matrix theory in a nutshell

Universality of local spectral statistics of random matrices

p. 6-1 Continuous Random Variables p. 6-2

Name: Firas Rassoul-Agha

18.175: Lecture 13 Infinite divisibility and Lévy processes

Selfadjoint Polynomials in Independent Random Matrices. Roland Speicher Universität des Saarlandes Saarbrücken

SOLUTIONS TO MATH68181 EXTREME VALUES AND FINANCIAL RISK EXAM

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Northwestern University Department of Electrical Engineering and Computer Science

Statistics for scientists and engineers

Foundations of Statistical Inference

Transcription:

Cauchy-Stieltjes kernel families Wlodek Bryc Fields Institute, July 23, 2013

NEF versus CSK families The talk will switch between two examples of kernel families K(µ) = {P θ (dx) : θ Θ}

NEF versus CSK families The talk will switch between two examples of kernel families K(µ) = {P θ (dx) : θ Θ} Natural exponential families (NEF) : P θ (dx) = 1 L(θ) eθx µ(dx) µ is a σ-finite measure, Θ = (θ, θ + ).

NEF versus CSK families The talk will switch between two examples of kernel families K(µ) = {P θ (dx) : θ Θ} Natural exponential families (NEF) : P θ (dx) = 1 L(θ) eθx µ(dx) µ is a σ-finite measure, Θ = (θ, θ + ). Cauchy-Stieltjes kernel families (CSK): P θ (dx) = 1 1 L(θ) 1 θx µ(dx) µ is a probability measure with support bounded from above. The generic choice for Θ is Θ = (0, θ + ).

A specific example of CSK Noncanonical parameterizations Let µ = 1 2 δ 0 + 1 2 δ 1 be the Bernoulli measure Noncanonical parametrization:

A specific example of CSK Noncanonical parameterizations Let µ = 1 2 δ 0 + 1 2 δ 1 be the Bernoulli measure Noncanonical parametrization: P θ = 1 θ 2 θ δ 0 + 1 2 θ δ 1, θ (, 1).

A specific example of CSK Noncanonical parameterizations Let µ = 1 2 δ 0 + 1 2 δ 1 be the Bernoulli measure Noncanonical parametrization: P θ = 1 θ 2 θ δ 0 + 1 2 θ δ 1, θ (, 1). Canonical parametrization: p = 1 2 θ Q p := P 2 1 = (1 p)δ 0 + pδ 1, p (0, 1) p

A specific example of CSK Noncanonical parameterizations Let µ = 1 2 δ 0 + 1 2 δ 1 be the Bernoulli measure Noncanonical parametrization: P θ = 1 θ 2 θ δ 0 + 1 2 θ δ 1, θ (, 1). Canonical parametrization: p = 1 2 θ Q p := P 2 1 = (1 p)δ 0 + pδ 1, p (0, 1) p Bernoulli family parameterized by probability of success p. p = xq p (dx) (parametrization by the mean)

Parametrization by the mean m(θ) = xp θ (dx) = L (θ) L(θ) L(θ) 1 θl(θ) NEF CSK For non-degenerate measure µ, function θ m(θ) is strictly increasing and has inverse θ = ψ(m).

Parametrization by the mean m(θ) = xp θ (dx) = L (θ) L(θ) L(θ) 1 θl(θ) NEF CSK For non-degenerate measure µ, function θ m(θ) is strictly increasing and has inverse θ = ψ(m). θ m(θ) maps (0, θ + ) onto (m 0, m + ), the domain of means.

Parametrization by the mean m(θ) = xp θ (dx) = L (θ) L(θ) L(θ) 1 θl(θ) NEF CSK For non-degenerate measure µ, function θ m(θ) is strictly increasing and has inverse θ = ψ(m). θ m(θ) maps (0, θ + ) onto (m 0, m + ), the domain of means. Parameterizations by the mean: K(µ) = {Q m (dx) : m (m 0, m + )} where Q m (dx) = P ψ(m) (dx)

Variance function V (m) = (x m) 2 Q m (dx) Variance function always exists for NEF.

Variance function V (m) = (x m) 2 Q m (dx) Variance function always exists for NEF. Variance function exists for CSK when µ(dx) has the first moment.

Variance function V (m) = (x m) 2 Q m (dx) Variance function always exists for NEF. Variance function exists for CSK when µ(dx) has the first moment. Variance function V (m) (together with the domain of means m (m, m + )) determines NEF uniquely (Morris (1982)).

Variance function V (m) = (x m) 2 Q m (dx) Variance function always exists for NEF. Variance function exists for CSK when µ(dx) has the first moment. Variance function V (m) (together with the domain of means m (m, m + )) determines NEF uniquely (Morris (1982)). Variance function V (m) (together with m 0 = m(0) R, the mean of µ) determines measure µ uniquely (hence determines CSK uniquely).

Example: a CSK with quadratic variance function Bernoulli measures Q m = (1 m)δ 0 + mδ 1 are parameterized by the mean, with the domain of means m (0, 1).

Example: a CSK with quadratic variance function Bernoulli measures Q m = (1 m)δ 0 + mδ 1 are parameterized by the mean, with the domain of means m (0, 1). The variance function is V (m) = m(1 m)

Example: a CSK with quadratic variance function Bernoulli measures Q m = (1 m)δ 0 + mδ 1 are parameterized by the mean, with the domain of means m (0, 1). The variance function is V (m) = m(1 m) The generating measure µ = 1 2 δ 0 + 1 2 δ 1 is determined uniquely once we specify its mean m 0 = 1/2. That is, there is no other µ that would have mean 1/2 and generate CSK with variance function V (m) that would equal to m(1 m) for all m (1/2 δ, 1/2 + δ)

All NEF with quadratic variance functions are known Morris class. Meixner laws The NEF with the variance function V (m) = 1 + am + bm 2 was described by Morris (1982), Ismail-May (1978)

All NEF with quadratic variance functions are known Morris class. Meixner laws The NEF with the variance function V (m) = 1 + am + bm 2 was described by Morris (1982), Ismail-May (1978) Letac-Mora (1990): cubic V (m)

All NEF with quadratic variance functions are known Morris class. Meixner laws The NEF with the variance function V (m) = 1 + am + bm 2 was described by Morris (1982), Ismail-May (1978) Letac-Mora (1990): cubic V (m) Various other classes Kokonendji, Letac,...

All CSK with quadratic variance functions are known Suppose m 0 = 0, V (0) = 1. Theorem (WB.-Ismail (2005)) End now

All CSK with quadratic variance functions are known Suppose m 0 = 0, V (0) = 1. Theorem (WB.-Ismail (2005)) 1. µ is the Wigner s semicircle (free Gaussian) law iff V (m) = 1 End now

All CSK with quadratic variance functions are known Suppose m 0 = 0, V (0) = 1. Theorem (WB.-Ismail (2005)) 1. µ is the Wigner s semicircle (free Gaussian) law iff V (m) = 1 K(µ) are the (atomless) Marchenko-Pastur (free Poisson) type laws End now

All CSK with quadratic variance functions are known Suppose m 0 = 0, V (0) = 1. Theorem (WB.-Ismail (2005)) 1. µ is the Wigner s semicircle (free Gaussian) law iff V (m) = 1 K(µ) are the (atomless) Marchenko-Pastur (free Poisson) type laws 2. µ is the Marchenko-Pastur (free Poisson) type law iff V (m) = 1 + am with a 0 3. µ is the free Gamma type law iff V (m) = (1 + bm) 2 with b > 0 End now

All CSK with quadratic variance functions are known Suppose m 0 = 0, V (0) = 1. Theorem (WB.-Ismail (2005)) 1. µ is the Wigner s semicircle (free Gaussian) law iff V (m) = 1 K(µ) are the (atomless) Marchenko-Pastur (free Poisson) type laws 2. µ is the Marchenko-Pastur (free Poisson) type law iff V (m) = 1 + am with a 0 3. µ is the free Gamma type law iff V (m) = (1 + bm) 2 with b > 0 4. µ is the free binomial type law (Kesten law, McKay law) iff V (m) = 1 + am + bm 2 with 1 b < 0 End now

Reproductive properties of NEF and CSK Theorem (NEF: Jörgensen (1997)) If µ is a probability measure in NEF with variance function V (m), then for r N the r-fold convolution µ r := µ r, is in NEF with variance function rv (m/r). Note End now

Reproductive properties of NEF and CSK Theorem (NEF: Jörgensen (1997)) If µ is a probability measure in NEF with variance function V (m), then for r N the r-fold convolution µ r := µ r, is in NEF with variance function rv (m/r). Theorem (CSK: WB-Ismail (2005), WB-Hassairi (2011)) If a probability measure µ generates CSK with variance function V µ (m), then the free additive convolution power µ r := µ r generates the CKS family with variance function rv µ (m/r). Note End now

Reproductive properties of NEF and CSK Theorem (NEF: Jörgensen (1997)) If µ is a probability measure in NEF with variance function V (m), then for r N the r-fold convolution µ r := µ r, is in NEF with variance function rv (m/r). Theorem (CSK: WB-Ismail (2005), WB-Hassairi (2011)) If a probability measure µ generates CSK with variance function V µ (m), then the free additive convolution power µ r := µ r generates the CKS family with variance function rv µ (m/r). Note If rv (m/r) is a variance function for all r (0, 1) then µ is infinitely divisible. End now

Reproductive properties of NEF and CSK Theorem (NEF: Jörgensen (1997)) If µ is a probability measure in NEF with variance function V (m), then for r N the r-fold convolution µ r := µ r, is in NEF with variance function rv (m/r). Theorem (CSK: WB-Ismail (2005), WB-Hassairi (2011)) If a probability measure µ generates CSK with variance function V µ (m), then the free additive convolution power µ r := µ r generates the CKS family with variance function rv µ (m/r). Note If rv (m/r) is a variance function for all r (0, 1) then µ is infinitely divisible. The domains of means behave differently. End now

Reproductive properties of NEF and CSK Theorem (NEF: Jörgensen (1997)) If µ is a probability measure in NEF with variance function V (m), then for r N the r-fold convolution µ r := µ r, is in NEF with variance function rv (m/r). Theorem (CSK: WB-Ismail (2005), WB-Hassairi (2011)) If a probability measure µ generates CSK with variance function V µ (m), then the free additive convolution power µ r := µ r generates the CKS family with variance function rv µ (m/r). Note If rv (m/r) is a variance function for all r (0, 1) then µ is infinitely divisible. The domains of means behave differently. The ranges of admissible r 1 are different. End now

Pseudo-Variance function for CSK The variance V (m) = 1 (x m) 2 L(ψ(m)) 1 ψ(m)x µ(dx) is undefined if m 0 = xµ(dx) =. (This issue does not arise for NEF)

Pseudo-Variance function for CSK The variance V (m) = 1 (x m) 2 L(ψ(m)) 1 ψ(m)x µ(dx) is undefined if m 0 = xµ(dx) =. (This issue does not arise for NEF) When V (m) exists, consider V(m) = m m m 0 V (m)

Pseudo-Variance function for CSK The variance V (m) = 1 (x m) 2 L(ψ(m)) 1 ψ(m)x µ(dx) is undefined if m 0 = xµ(dx) =. (This issue does not arise for NEF) When V (m) exists, consider It turns out that V(m) = m m m 0 V (m) ( ) 1 V(m) = m ψ(m) m (1) where ψ( ) is the inverse of θ m(θ) = xp θ (dx) on (0, θ + ).

Pseudo-Variance function for CSK The variance V (m) = 1 (x m) 2 L(ψ(m)) 1 ψ(m)x µ(dx) is undefined if m 0 = xµ(dx) =. (This issue does not arise for NEF) When V (m) exists, consider It turns out that V(m) = m m m 0 V (m) ( ) 1 V(m) = m ψ(m) m (1) where ψ( ) is the inverse of θ m(θ) = xp θ (dx) on (0, θ + ). Expression (1) defines a pseudo-variance function V(m) that is well defined for all non-degenerate probability measures µ with support bounded from above.

Properties of pseudo-variance function Uniqueness: measure µ(dx) is determined uniquely by V

Properties of pseudo-variance function Uniqueness: measure µ(dx) is determined uniquely by V Explicit formula for the CSK family: Q m (dx) = 1 L(ψ(m))(1 ψ(m)x) µ(dx) = V(m) V(m) + m(m x) µ(dx)

Properties of pseudo-variance function Uniqueness: measure µ(dx) is determined uniquely by V Explicit formula for the CSK family: Q m (dx) = 1 L(ψ(m))(1 ψ(m)x) µ(dx) Reproductive property still holds = V(m) V(m) + m(m x) µ(dx)

Properties of pseudo-variance function Uniqueness: measure µ(dx) is determined uniquely by V Explicit formula for the CSK family: Q m (dx) = 1 L(ψ(m))(1 ψ(m)x) µ(dx) Reproductive property still holds Theorem (WB-Hassairi (2011)) = V(m) V(m) + m(m x) µ(dx) Let V µ be a pseudo-variance function of the CSK family generated by a probability measure µ with support bounded from above and mean m 0 <. Then for m > rm 0 close enough to rm 0, V µ r (m) = rv µ (m/r). (2)

Example: CKS family with cubic pseudo-variance function Measure µ generating CSK with V(m) = m 3 has density f (x) = 1 4x 2πx 2 1 (, 1/4) (x) (3) From reproductive property it follows that µ is 1/2-stable with respect to, a fact already noted before: [Bercovici and Pata, 1999, page 1054], [Pérez-Abreu and Sakuma, 2008] { Q m (dx) = What is m +? 25 min? End now m2 } 1 4x 2π(m 2 + m x)x 2 1 (, 1/4)(x)dx : m (, m + )

Domain of means: {Q m : m (m 0, m + )} For V(m) = m 3 the domain of means is (, m + ), where: 1. θ m(θ) is increasing, so m + = lim θ θmax m(θ). This gives m + = 1 End now

Domain of means: {Q m : m (m 0, m + )} For V(m) = m 3 the domain of means is (, m + ), where: 1. θ m(θ) is increasing, so m + = lim θ θmax m(θ). This gives m + = 1 1 2. 1 θx 1 (, 1/4)(x) is positive for θ (0, ) (, 4). The domain of means can be extended to m + = lim θ 4 m(θ). This extends the domain of means up to m + = 1/2 End now

Domain of means: {Q m : m (m 0, m + )} For V(m) = m 3 the domain of means is (, m + ), where: 1. θ m(θ) is increasing, so m + = lim θ θmax m(θ). This gives m + = 1 1 2. 1 θx 1 (, 1/4)(x) is positive for θ (0, ) (, 4). The domain of means can be extended to m + = lim θ 4 m(θ). This extends the domain of means up to m + = 1/2 3. m 2 m 2 +m x 1 (, 1/4)(x) is positive for m 1/2. End now

Domain of means: {Q m : m (m 0, m + )} For V(m) = m 3 the domain of means is (, m + ), where: 1. θ m(θ) is increasing, so m + = lim θ θmax m(θ). This gives m + = 1 1 2. 1 θx 1 (, 1/4)(x) is positive for θ (0, ) (, 4). The domain of means can be extended to m + = lim θ 4 m(θ). This extends the domain of means up to m + = 1/2 3. m 2 m 2 +m x 1 (, 1/4)(x) is positive for m 1/2. But Q m (dx) < 1 for m > 1/2. End now

Domain of means: {Q m : m (m 0, m + )} For V(m) = m 3 the domain of means is (, m + ), where: 1. θ m(θ) is increasing, so m + = lim θ θmax m(θ). This gives m + = 1 1 2. 1 θx 1 (, 1/4)(x) is positive for θ (0, ) (, 4). The domain of means can be extended to m + = lim θ 4 m(θ). This extends the domain of means up to m + = 1/2 3. m 2 m 2 +m x 1 (, 1/4)(x) is positive for m 1/2. But Q m (dx) < 1 for m > 1/2. m Q m (dx) = 2 (1+2m)+ µ(dx) + δ (m 2 +m x) (m+1) 2 m+m 2 is well defined and parameterized by the mean for all m (, ). End now

Summary Kernels e θx and 1/(1 θx) generate NEF and CSK families Similarities

Summary Kernels e θx and 1/(1 θx) generate NEF and CSK families Similarities parameterizations by the mean

Summary Kernels e θx and 1/(1 θx) generate NEF and CSK families Similarities parameterizations by the mean Quadratic variance functions determine interesting laws

Summary Kernels e θx and 1/(1 θx) generate NEF and CSK families Similarities parameterizations by the mean Quadratic variance functions determine interesting laws Convolution affects variance function for NEF in a similar way as the additive free convolution affects the variance function for CSK

Summary Kernels e θx and 1/(1 θx) generate NEF and CSK families Similarities parameterizations by the mean Quadratic variance functions determine interesting laws Convolution affects variance function for NEF in a similar way as the additive free convolution affects the variance function for CSK Differences

Summary Kernels e θx and 1/(1 θx) generate NEF and CSK families Similarities parameterizations by the mean Quadratic variance functions determine interesting laws Convolution affects variance function for NEF in a similar way as the additive free convolution affects the variance function for CSK Differences The generating measure of a NEF is not unique.

Summary Kernels e θx and 1/(1 θx) generate NEF and CSK families Similarities parameterizations by the mean Quadratic variance functions determine interesting laws Convolution affects variance function for NEF in a similar way as the additive free convolution affects the variance function for CSK Differences The generating measure of a NEF is not unique. A CSK family in parameterizations by the mean may be well defined beyond the domain of means

Summary Kernels e θx and 1/(1 θx) generate NEF and CSK families Similarities parameterizations by the mean Quadratic variance functions determine interesting laws Convolution affects variance function for NEF in a similar way as the additive free convolution affects the variance function for CSK Differences The generating measure of a NEF is not unique. A CSK family in parameterizations by the mean may be well defined beyond the domain of means For CSK family, the variance function may be undefined. Instead of the variance function [Bryc and Hassairi, 2011] look at the pseudo-variance function m mv (m)/(m m 0 ) which is well defined for more measures µ.

Thank you

Thank you Back

References Bercovici, H. and Pata, V. (1999). Stable laws and domains of attraction in free probability theory. Ann. of Math. (2), 149(3):1023 1060. With an appendix by Philippe Biane.

References Bercovici, H. and Pata, V. (1999). Stable laws and domains of attraction in free probability theory. Ann. of Math. (2), 149(3):1023 1060. With an appendix by Philippe Biane. Bryc, W. (2009). Free exponential families as kernel families. Demonstr. Math., XLII(3):657 672. arxiv.org:math.pr:0601273.

References Bercovici, H. and Pata, V. (1999). Stable laws and domains of attraction in free probability theory. Ann. of Math. (2), 149(3):1023 1060. With an appendix by Philippe Biane. Bryc, W. (2009). Free exponential families as kernel families. Demonstr. Math., XLII(3):657 672. arxiv.org:math.pr:0601273. Bryc, W. and Hassairi, A. (2011). One-sided Cauchy-Stieltjes kernel families. Journ. Theoret. Probab., 24(2):577 594. arxiv.org/abs/0906.4073.

References Bercovici, H. and Pata, V. (1999). Stable laws and domains of attraction in free probability theory. Ann. of Math. (2), 149(3):1023 1060. With an appendix by Philippe Biane. Bryc, W. (2009). Free exponential families as kernel families. Demonstr. Math., XLII(3):657 672. arxiv.org:math.pr:0601273. Bryc, W. and Hassairi, A. (2011). One-sided Cauchy-Stieltjes kernel families. Journ. Theoret. Probab., 24(2):577 594. arxiv.org/abs/0906.4073. Bryc, W. and Ismail, M. (2005). Approximation operators, exponential, and q-exponential families. Preprint. arxiv.org/abs/math.st/0512224.

References Bercovici, H. and Pata, V. (1999). Stable laws and domains of attraction in free probability theory. Ann. of Math. (2), 149(3):1023 1060. With an appendix by Philippe Biane. Bryc, W. (2009). Free exponential families as kernel families. Demonstr. Math., XLII(3):657 672. arxiv.org:math.pr:0601273. Bryc, W. and Hassairi, A. (2011). One-sided Cauchy-Stieltjes kernel families. Journ. Theoret. Probab., 24(2):577 594. arxiv.org/abs/0906.4073. Bryc, W. and Ismail, M. (2005). Approximation operators, exponential, and q-exponential families. Preprint. arxiv.org/abs/math.st/0512224. Pérez-Abreu, V. and Sakuma, N. (2008). Free generalized gamma convolutions. Electron. Commun. Probab., 13:526 539.