Conditional independence and predictive copula

– In this paper, we address the concept of conditional independence between two random variables X and Y given the entity H . We identify the impact of conditional independence on the analytic form of the predictive 2-copula between X and Y. We obtain a representation of the predictive 2-copula between X and Y in terms of functions associated with the copulas between X and H and between Y and H . Through the concept of in ﬁ nite exchangeable sequences we amplify the validity of our results, obtaining the predictive 2-copula between two variables in terms of the copula between only one of these variables and the quantity H


Introduction
In this article, we address the notion of conditional independence and its impact on unconditional dependence. Here, unconditional dependence is being represented by a copula function and the impact of conditional independence on unconditional dependence is studied using the assumption of conditional independence in the copula function. Independence allows us to carry out the usual inferential process, for example, those obtained by calculating the likelihood function, and because of that is a relevant notion in statistical inference. Under the independence assumption is possible to carry out the classic techniques of point estimation, hypothesis testing, and all those procedures based on the likelihood function. As can then be verified, the concept of independence occupies a relevant space in statistics and for this reason, it is widely found in the literature. Several efforts have been done to promote its identification, see for instance García and González-López [1,2]. In such papers are proposed statistics for independence detection, see also Genest and Rémillard [3] and Hollander et al. [4]. Such statistics work under certain conditions imposed on the nature of the underlying variables. Other authors are concerned with measuring dependence. For example, García et al. [5] proposes an index with such purpose. Despite its enormous use, independence is a very restrictive case within the diversity of types of dependencies that could occur. For example, if we consider the random vector (X, Y) with bivariate Normal distribution, and density function f x; y ð Þ ¼ 1 2pð1 À q 2 Þ 1=2 exp À x 2 þ y 2 À 2qxy ð Þ 2 1 À q 2 ð Þ ; 8 x; y ð Þ 2 R; q 2 ½À1; 1; then, the independence only occurs when q = 0, which shows that independence is a restrictive assumption. The theory of copulas has been extensively used to investigate such diversity. Nelsen [6] and Joe [7] show the enormous diversity of possible dependencies, addressed from the notion of copulas. This article explores conditional independence, seeking to preserve some inferential procedures, and we also seek to characterize the impact of conditional independence on dependence. Conditional independence requires the identification of the entity H under which we can declare conditional independence (conditional to H), that requirement can be a constraint in practice since there is no clear identification procedure of H. The first question we seek to answer is how conditional independence impacts unconditional dependence. For the characterization of unconditional dependence, we will rely on the notion of copula. Secondly, we approach the impact of the notion of infinite exchangeability on conditional independence, based on the fact that such a notion guarantees conditional independence [8,9]. With such a notion in hand, we return to the investigation of the structure of the copula representing unconditional dependence. That is, in this paper, we investigate the impact of conditional independence on unconditional dependence and extend this question to a stronger notion as is the case of infinite exchangeability.
Independence is a very restrictive case within the diversity of types of dependencies, on the other hand, conditional independence (to H) is a much more flexible concept, and for this reason, it is the condition that we explore in this article. In our approach, we take the entity H as being a random variable since in this way we generalize the notion. In addition, we give space so that the reader can access the interpretation that de Finetti offers of such an entity.
Here we describe the content of this paper. In Section 2 we address the notion of conditional independence and we introduce a theorem that shows a representation of the dependence (copula). We also show corollaries addressing situations of interest that result from that theorem. In Section 3 we introduce the notion of infinite exchangeability, which allows us to identify the quantity H, guarantying conditional independence. We then show a result offering a representation of the dependence (copula density function), given the property of infinite exchangeability. The final considerations are given in Section 4.

Conditional independence
We begin this section by introducing the notion of conditional independence and then exemplifying it. We introduce in the sequence the notion of copula function. From these notions follows our main result, Theorem 2.1 which shows the impact of conditional independence on the dependence, that is, in this case, on the copula structure. We emphasize the consequences that result from Theorem 2.1 in two corollaries and we show also some examples.

Definition 2.1
Given two continuous random variables X and Y, X and Y are conditionally independent given with f X;Y jh ; f Xjh and f Y jh denoting the conditional density functions, given h, of (X,Y), X and Y respectively. The next example shows a case of conditional independence given H. The result that we show in Theorem 2.1 seeks to understand the meaning that H plays in the construction of the dependence between X and Y (not conditioned to H = h). For dependence between X and Y we understand the copula between them. In rough terms, the copula is a function that, combined with the marginal cumulative distributions, returns the joint cumulative distribution.
Next we formally introduce the notion of copula explaining how this notion serves to represent the dependence between X and Y.
According to Sklar's theorem [6], if H X,Y is the joint distribution function of (X, Y) with margins F X and F Y , there exists a copula C X,Y (called 2-copula of (X, Y)) such that 8x; y 2 R; If F X and F Y are continuous, then C X,Y is unique. Conversely, if C is a copula (following Definition 2.2), F and G are distribution functions, then the function H(x,y) = C(F(x),G(y)) is a joint distribution function with margins F and G.
From what is given in Sklar's theorem, the dependence between X and Y is represented by the 2-copula of (X,Y), C X,Y , since the margins F X and F Y depend on X and Y, respectively. Then, our purpose is to see the effect of Definition 2.1 in C X,Y .
The next theorem provides an analytical representation of the copula, knowing that the variables are conditionally independent given the entity H.

Theorem 2.1
Given X and Y under the assumptions of Definition 2.1, if C X,Y is the 2-copula of (X,Y), H the random variable of Definition 2.1 with density function p H , with C X,H and C Y,H denoting the 2-copulas of (X,H) and (Y,H) respectively, and F H denoting the cumulative distribution of H.
Proof. Since X, Y are continuous random variables we can write the cumulative distribution H X,Y between X and Y as follows, applying Definition 2.1, Considering the cumulative distribution between X and H, by Sklar's theorem, H X,H (s,h) = C X,H (F X (s), F H (h)) then, the joint density function between X and H is given by with c X,H denoting the density function related to the copula C X,H . Then, from equation (3), we obtain, We proceed to integrate the equation (4).
We have similar expressions for f Y,H (t,h) and f Y|h (t). Then, applying equation (5) in equation (2), we obtain Consider u and v in [0, 1], there exist values x and y, such that u = F X (x) and v = F Y (y), then, by Sklar's theorem and equation (6), We use in the following example a family of copulas that corresponds to weak dependence. It is the Farlie-Gumbel-Morgenstern family, and it encompasses the case of independence. See Nelsen [6].

Example 2.2
Consider (X,Y) with a 2-copula with parameters a 2 [À1, 1] and b 2 [À1, 1]. Then, (X,Y) has a Farlie-Gumbel-Morgenstern 2-copula with parameter Consider H as a continuous random variable, with Uniform distribution in (0,1). If we set the 2-copulas of (X, H) and (Y, H) as being Farlie-Gumbel-Morgenstern 2-copulas with parameters a 2 [ 1, 1] and b 2 [À1, 1], respectively. Since h~U(0,1), we have And, we obtain, As a consequence, One question that arises is whether every expression given by the right side of equation (1) is a copula, where C X,H and C Y,H are copulas. The Theorem 1, in González-López and Litvinoff Justus [10], states that the expression is always a copula. That is, postulating two copulas of (X, H) and (Y, H), we get a copula that could be used to deal with the dependence between X and Y. An example of what is stated in this paragraph appears in the previous example, since taking (X, H) and (Y, H) with Farlie-Gumbel-Morgenstern copulas, with parameters a and b, respectively, the resulting copula between X and Y (using the right side of equation (1)) is a Farlie-Gumbel-Morgenstern copula, with parameter d ¼ ab 3 : The following corollary shows the version of Theorem 2.1 in terms of density functions of copulas.

Corollary 2.1
Under the assumptions of Theorem 2.1, if c X,Y , c X,H and c Y,H are the density functions related to the 2-copulas C X,Y , C X,H and C Y,H , respectively, and, according to the equation (7), it is the copula density function of a FGM copula with parameter ab 3 : V.A. González-López and V. Litvinoff Justus: 4open 2022, 5, 20 The following corollary shows the version of Theorem 2.1 under the assumptions of identical conditional margins. The impact of the assumption appears in the integrating term of equation (1). This assumption about margins will become relevant later, in this paper.

Corollary 2.2
Under the assumptions of Theorem 2.1, if the variables X and Y are conditionally (to H) independent and identically distributed, Considering the Example 2.2, if a = b, the copula of (X,Y) is a Farlie-Gumbel-Morgenstern copula with c ¼ a 2 3 2 0; 1 3 ! : Then, the restriction (on the margins) imposes on the relationship between X and Y a positive Kendall's s coefficient . Theorem 2.1 shows that the conditional independence of X and Y given H defines the structure of the copula between X and Y. See equation (1) for the form of the copula and see Corollary 2.1 for the form of the density function of copula between X and Y. We exemplify both results in Examples 2.2 and 2.3, respectively. Corollary 2.2 shows the case covered by Corollary 2.1, under the assumption of identical margins. The latter is exemplified in Example 2.4. The Corollary 2.2 is especially interesting for the situation that we will address in the next section.The results of this section describe the analytical form of the copula between X and Y, no longer depending on H. This means we have offered a representation for the predictive copula between X and Y.
The key to determining conditional independence between X and Y is to know that there exists a variable H establishing this independence. This leads us to the following reflections, showing a context where such an entity can be identified. That is the goal of de Finetti's theorems of representation.

Exchangeability
We start the section with the definition of infinite exchangeability. This notion is weaker than independence and makes it possible to identify H. Then, we exemplify certain characteristics that result from this imposition. And, we present in Corollary 3.1, the impact of this notion on the dependence structure, that is, on the copula function.
The success in identifying the H entity stems from de Finetti's representation theorem, see de Finetti [8]. de Finetti [8] proves that, a binary sequence X 1 , X 2 ,.. such that for all n the joint probability mass function P ðX 1 ¼ x 1 ; . . . ; X n ¼ x n Þ ¼ F H is the cumulative distribution function of the entity H ¼ lim n!1 P n i¼1 X i n ; and x i 2 {0, 1}. As a consequence, the conditional mass probability is decomposed in a product of Bernoulli's being h a value of the variable H. Such a result is extended for all types of random variables in Hewitt and Savage [9]. For a sequence X 1 ,X 2 ,... of continuous random variables following Definition 3.1, there exists a variable H, such that the conditional density function f X 1 ;...;X n jh ðx 1 ; . . . ; x n Þ can be decomposed as f X 1 ;...;Xn jh ðx 1 ; . . . ; x n Þ ¼ Y n i¼1 f X i jh ðx i Þ: The previous relation happens for each value h of the variable H, then, the structure imposed by Definition 3.1 in the sequence X 1 ,X 2 ,Á Á Á guarantee the existence of the entity H.
We see in the remark to follow (Remark 3.1) that this notion leads us to conditional independence, and implies identical margins. But as we show in Example 3.1, exchangeability coexists with dependence.

Remark 3.1
Infinitely exchangeable variables (following Definition 3.1) are identically distributed conditionally to h (value of the variable H identified by de Finetti's theorem). If the infinitely exchangeable sequence is composed by absolutely continuous random variables and H has a density function p H (Á), for each i; j : The concept of infinite exchangeability could be considered unrealistic, but it can be seen applied in usual inferential strategies. If a sample is available, say X 1 ,. . .,X n , when constructing the likelihood function its elements are considered independent given a certain parameter. And how does this assumption is related to Definition 3.1? If the distribution of X 1 ,. . .,X n is invariant under permutations of the elements of X 1 ,. . .,X n , then we say that X 1 . . .,X n is exchangeable. If X 1 ,. . .,X n is exchangeable 8n 2 N, then, we say that X 1 , X 2 ,. . . is infinitely exchangeable. If X 1, . . .,X n can be embedded in an infinitely exchangeable sequence X 1 , X 2 ,. . ., then we say that X 1 ,. . .,X n is infinitely exchangeably extendable. Bayesian and frequentist approaches, in general, treat observable values as infinitely exchangeably extendable and, as a consequence of de Finetti's representations, independent and identically distributed conditional on some unknown entity H.
Although identically distributed independent random variables must be exchangeable, identically distributed exchangeable random variables need not be independent. Consider for instance the bivariate case of Gumbel distribution as shown by the next example.

Example 3.1
Let (X,Y) be a random vector, H its joint distribution function, given by H x; y ð Þ ¼ 1 À e Àx À e Ày þ e À xþyþaxy ð Þ ; x ! 0; y ! 0 and H(x,y) = 0, otherwise, with a a parameter in [0, 1]. We see that X and Y are exchangeable. Note also that X and Y have identical marginal distribution (exponential distribution). But, since the 2-copula of (X,Y) is C X ;Y ðu; vÞ ¼ u þ v À 1 þ ð1 À uÞð1 À vÞe Àa ln 1Àu ð Þln 1Àv ð Þ ; u; v 2 ½0; 1; X and Y are not independent. The previous example shows us a case where exchangeability occurs, the variables share the same marginal distribution, but X and Y are not independent. Thus we see that exchangeability is a distinctly different notion from the notion of independence. It also follows from Example 3.1 that identically distributed variables can be dependent since every copula is a joint distribution with uniform marginal distributions, in particular, the copula of Example 3.1.
The next example shows a case of dependent variables, with identical marginal distributions but not exchangeable.

Example 3.2
Let (X,Y) be a random vector, H its joint distribution function, given by H(x,y) = xy À x 3 y(1 À x)(1 À y), x,y 2 [0, 1]. Then, H is a copula, since fits the conditions of Definition 2.2. X and Y are identically distributed but, not exchangeable since H is not symmetrical.
Bearing in mind the previous example, it is now necessary to mention Theorem 2.7.4 [6] which shows how the margins should be and also the copula to guarantee the exchangeability between a pair of variables. Let X and Y be continuous random variables with joint distribution function H, margins F X and F Y , respectively, and copula C. Then X and Y are exchangeable if and only if F X = F Y and C is symmetric (C(u,v) = C(v,u), "(u,v) 2 [0, 1] 2 ).
In the treatment and modeling via copulas, it is very common to operate with the original variables transformed by their margins, so below we summarize the impact of Definition 3.1 on such variables.

Remark 3.2
A property that is straight from Definition 3.1, is the preservation of the infinite exchangeability from the original sequence to the standardized ones. Namely, if the sequence X 1 , X 2 , X 3 ,. . . of random variables is absolutely continuous, the sequence U 1 , U 2 , U 3 ,. . . is also infinitely exchangeable, where U i :¼ F X i X i ð Þ; i ! 1: Since, for any collection i 1 ,. . .i n , with arbitrary n, P ðU i 1 u 1 ; . . . ; U in u n Þ ¼ P ðU rði 1 Þ u 1 ; . . . ; U rðinÞ u n Þ; for any permutation r and arbitrary values u 1 ,. . ., u n 2 [0, 1].
where equation (10) results from Definition 3.1 and equation (11) follows from Remark 3.1 (F X i ðÁÞ ¼ F X 1 ðÁÞ; 8i ! 1). We note that if the variables U 1 , U 2 , U 3 ,. . . of Remark 3.2 are infinitely exchangeable, this does not imply that the variables X 1 , X 2 , X 3 ,. . . are infinitely exchangeable. It is enough that one of the marginal distributions F X i 0 is different from the rest fF X i g i6 ¼i 0 ; which would violate Remark 3.1.
Under the framework of Definition 3.1 and assuming the continuity of the variable identified by the de Finetti representation, say H, we will have the copula between pairs of variables of the infinitely exchangeable sequence given by the copula between a member of the sequence (the first one is enough) and the variable H, let's see.
for u ¼ F X i ðx i Þ and v ¼ F X j ðx j Þ: Using the previous proportionality we obtain c X i ;X j ðu; vÞ / ðx i þ kÞ 2 ðx j þ kÞ kðx i þ x j þ kÞ 3 : On the other hand, c X i ;H ðu; then, the right side of equation (12) is proportional to ðx i þ kÞ 2 ðx j þ kÞ 2 kðx i þ x j þ kÞ 3 : Verifying the Corollary 3.1.
Then, since we have F X i ðxÞ ¼ x ðk þ xÞ ; 8x ! 0; 8i; the copula density function c X i ;X j ðu; vÞ ¼ 2ð1 À uÞð1 À vÞ ð1 À uvÞ 3 ; u; v 2 ½0; 1; 8i; j ! 1; and the predictive 2-copula is We start the section with some notes and examples so that Corollary 3.1 is a natural consequence of the impact of conditional independence in the framework of infinite exchangeability. Remark 3.1 shows that infinite exchangeability implies conditionally independence and identical margins. While independence implies permutability, the opposite is not valid, see Example 3.1. Also, if the margins are all equal this does not imply permutability, see Example 3.2. Under Definition 3.1, we show in Corollary 3.1 that infinite exchangeability allows to obtain a representation of the copula between pair of elements of the infinite sequence in terms of the copula between each element of the infinite sequence and the entity H, quantity identified by de Finetti's theorem. Also, note that the right side of equation (12) reflects that is only necessary to know the copula between one element of the infinite sequence and H to obtain the predictive copula between pairs of elements of the infinite sequence. Example 3.3 exposes how works in practice the corollary.

Conclusion
In the framework of absolutely continuous random variables X and Y, we use the notion of conditional independence between X and Y (conditional on H), to obtain a representation of the predictive 2-copula between X and Y (Theorem 2.1, see also Corollary 2.1). Such a result tells us that the representation involves the 2-copula between X and H, the 2-copula between Y and H, and the distribution of H. Being the property of identically distributed useful for inferential procedures, in Corollary 2.2 we reveal the impact of such an assumption on the representation of the predictive 2-copula between X and Y. There are indications (see Example 2.4) that the restriction imposed by the condition of identical margins could lead to restrictions on the values of coefficients of dependence, such as Kendall's s coefficient, Spearman's q coefficient.
Since the notion of conditional independence requires the identification of the entity H making X and Y conditionally independent given H, we introduce the concept of infinite and exchangeable sequences, X 1 , X 2 , X 3 .... This condition appears as one of the inferential strategies applicable in the process of manipulating finite samples. It is reasonable to suppose that a finite sequence is part of an infinite sequence. Under Definition 3.1, by representation theorems [8,9] we can guarantee that H exists and that it is a random variable. To give the reader a more precise notion of the meaning of infinite exchangeability, we show that it imposes the same margins on X and Y (conditional or not, see Remark 3.1), but that it does not impose unconditional independence (see Example 3.1) and it does impose conditional independence on H. Then, under infinite exchangeability, we revisit the copula density function structure and prove Corollary 3.1, which shows the structure of the predictive 2-copula density function between X i and X j , i 6 ¼ j elements of the infinite sequence. We complete our findings with examples and notes for the reader. Corollaries 2.2 and 3.1, although they start from different assumptions, show that the predictive density function of the 2-copula between pairs of variables X and Y can be analytically expressed as a function of the density functions of the copula between only one variable (say X) and H. That is to say that the dependence, represented by the predictive 2-copula between X and Y, is only a consequence of the dependence between X and H. This fact shows the relevance of the identification of H, for the description of the predictive 2-copula.