Flavio Figueiredo1, José G. Fernandes1, Jackson N. Silva and Renato Assunção1,2 Universidade Federal de Minas Gerais ESRI Inc. Reproducibility: https://anonymous.4open.science/r/2cats-E765/ Contact Author: flavio@dcc.ufmg.br
Abstract
Copulas are powerful statistical tools for capturing dependencies across data dimensions. Applying Copulas involves estimating independent marginals, a straightforward task, followed by the much more challenging task of determining a single copulating function, , that links these marginals. For bivariate data, a copula takes the form of a two-increasing function , where . This paper proposes 2-Cats, a Neural Network (NN) model that learns two-dimensional Copulas without relying on specific Copula families (e.g., Archimedean). Furthermore, via both theoretical properties of the model and a Lagrangian training approach, we show that 2-Cats meets the desiderata of Copula properties. Moreover, inspired by the literature on Physics-Informed Neural Networks and Sobolev Training, we further extend our training strategy to learn not only the output of a Copula but also its derivatives. Our proposed method exhibits superior performance compared to the state-of-the-art across various datasets while respecting (provably for most and approximately for a single other) properties of C.
1 Introduction
Modeling univariate data is relatively straightforward for several reasons. Firstly, a wide range of probability distributions exist that are suitable for different data types, including Gaussian, Beta, Log-Normal, Gamma, and Poisson. This set is expanded by mixture models that use the elemental distributions as components and can model multimodality and other behavior. Secondly, visual tools such as histograms or Kernel Density Estimators (KDE) provide valuable assistance in selecting the appropriate distribution. Lastly, univariate models typically involve few parameters and may have exact solutions, simplifying the modeling process.
However, the process becomes more challenging when modeling multivariate data and obtaining a joint probability distribution. Firstly, only a few classes of multivariate probability distributions are available, with the primary exceptions being the multivariate Gaussian, Elliptical, and Dirichlet distributions. Secondly, simultaneously identifying the dependencies via conditional distributions based solely on empirical data is highly complex.
In 1959, Abe Sklar formalized the idea of Copulas[61, 60]. This statistical tool allows us to model a multivariate random variable of dimension by learning its cumulative multivariate distribution function (CDF) using independent marginal CDFs and a single additional Copulating function . A bivariate example may separately identify one marginal as a Log-Normal distribution and the other as a Beta distribution. Subsequently, the Copula model links these two marginal distributions.
However, the Copula approach to modeling multivariate data faced a significant limitation until recently. Our model choices were confined to only a handful of closed forms for the Copulas, such as the Gaussian, Frank, or Clayton Copulas[16, 25]. Unfortunately, this approach using closed-form Copula models proved inadequate for accurately capturing the complex dependencies in real data. The limited parameterization of these Copulas prevented them from fully representing the intricate relationships between variables.
This situation has changed recently.Neural Networks (NNs) are well-known for their universal ability to approximate any function. Consequently, researchers have started exploring NNs as alternatives to closed forms for [13, 27, 66, 31, 38, 51]. However, a limitation of NNs is that they neglect the importance of maintaining the fundamental mathematical properties of their domain of study.
For Copulas, only three properties define [50]:P1: , where ;P2: For any and , we have that the volume of is: ;and, P3: is grounded. That is, , . Moreover, and .
This paper proposes 2-Cats (2D Copula Approximating Transforms) as a NN approach that meets the above desiderata. How the model meets P1 to P3 is discussed in Section3. P1 and P2 are natural consequences of the model, whereas P3 is guaranteed via Lagragian Optimization.
To understand our approach, let represent our model (or Hypothesis). Let be any bivariate CDF with as support. We connect with as follows:
(1)
Here, is any NN that outputs a positive number and is the Logit function that is both monotonic and maps to , i.e., . Here,the function monotonically increases concerning because of the positive NN output; similarly, is monotonic on .
The monotonic properties of all transformations ensure that our output is monotonic for both and . Let and to simplify notation and a random vector with distribution function . Then, the probability transform below is valid[11, section 4.3]:
(2)
To ensure that our model derivatives also approximate the derivatives of , a second contribution of our work is that 2-Cats models are trained similarly to Physics Informed Neural Networks (PINNs)[32, 42] by employing Sobolev training[17] approximate Copula derivatives.
There have been several recent NN[13, 38, 51] and non-NN models[49, 5, 47] Copula-like models.We compare our 2-Cats model with these alternatives, and our results show that the 2-Cats performance (in negative log-likelihood) is better or statistically tied to baselines. Our contributions are:
•
We introduce the 2-Cats model, a novel NN Copula approximation. Unlike other NN approaches, we focus on satisfying (either provably or via constraints) the essential requirements of a Copula. We also demonstrate the empirical accuracy of 2-Cats;
•
We are the first to apply Sobolev training[17] in NN-based Copula methodologies. Our approach involves the introduction of a simple empirical estimator for the first derivative of a Copula, which is seamlessly integrated into our training framework.
2 Related Work
After being introduced by Sklar[61, 60], Copulas have emerged as valuable tools across various domains[25, 50], including engineering[56], geosciences[46, 72], and finance[12, 56]. Recently, Copulas have gained attention from the ML community, particularly with the advent of DL methods for Copula estimation and their utilization in generative models[13, 27, 66, 38, 31, 51].
Our proposed approach aligns closely with the emerging literature in this domain but makes significant advancements on two fronts. This paper primarily focuses on developing an NN-based Copula model that adheres to the fundamental properties of a Copula. Two methods sharing similar objectives are discussed in [38, 51] and [13]. However, our approach diverges from these prior works by not confining our method to Archimedean Copulas, as seen in[38] and [51].
As Chilinski and Silva[13], we begin by estimating densities through CDF estimations followed by automatic differentiation to acquire densities. However, our primary difference lies in our emphasis on relating 2-Cats to a Copula, an aspect not prioritized by the authors of that study.
To assess the effectiveness of our approach, we conduct comparative analyses against previous proposals, including the method presented by Janke et al.[31], proposed as a generative model. Furthermore, 2-Cats does not presume a fixed density function for copula estimation; instead, it is guided by the neural network (NN). Hence, our work shares similarities with non-parametric copula estimators used as baselines[49, 47, 5]. Ultimately, all these methods function as baseline approaches against which we evaluate and demonstrate the superiority of our proposed approach.
For training our models, we utilize Sobolev training as a regularizer. Sobolev training is tailored to estimate neural networks (NNs) by incorporating losses that not only address errors toward the primary objective but also consider errors toward their derivatives. This approach ensures a more comprehensive and accurate training of the NNs. A similar concept is explored in Physics Informed Neural Networks (PINNs)[32, 35], where NNs incorporate derivative information from physical models. This approach enables PINNs to model complex physics-related relationships.
In summary, our objective is to develop an NN-based Copula model that respects the mathematical properties of Copulas and benefits from considering derivative information. Consequently, although our focus is on 2D Copulas, the Pair Copula Decomposition (PCC)[1, 15] that requires derivative information and is used to model -dimensional data is valid for our models[16]. Due to space constraints, we leave the evaluation of such a decomposition for future work.
Our work also aligns with the literature on monotonic networks[58, 18, 71] and Integral Networks[36, 39]. The concept of using integral transforms to build monotonic models was introduced by[71].
3 On the Desiderata of a Copula and 2-Cats Properties
We now revisit the discussion on the 2-Cats model as introduced in Section1.
One of the building blocks of 2-Cats is the NN . In particular, we require that this NN outputs only positive numbers so that its integral captures a monotonic function. We achieve this by employing Exponential Linear Units (ELU) plus one activation (ELU+1) in every layer.
Let us consider the computation details of , with the understanding that the process for is analogous. In our implementation, we adopt Cumulative Trapezoid integration. Initially, we divide the range into equally spaced intervals. When computing , the value of is inserted into this interval while preserving the order. Subsequently, the NN is evaluated for the corresponding values of parameter . Trapezoidal integration is then applied to calculate the integral value at (for the numerator) and (for the denominator). Finally, when , .
Now, let us present our main properties. AppendixB and D complements these properties.
Theorem P1.
, where .
Proof.
Remind that and . Also, . These transforms are also able to cover the entire range of (when we have , when we have , the same goes for the transform). The function maps the domain into . Given that the bivariate CDFs work in this domain: , where .∎
Theorem P2.
The 2-Cats copula satisfies the non-negative volume property. That is,for any and , we have that .
Intuition..
This is a straightforward consequence of being a bivariate cumulative distribution function and the monotonicity of the , , and transforms.The same transform we explored above for the single variate case is valid for multiple variables. That is, fact that and are monotonic guarantees that defines a valid transform Eqn(2).
Considering a bivariate CDF, for any given point in its domain (e.g., ), the CDF represents the accumulated probability up to that point. Let the origin be . As either , , or both values increase, the CDF values can only increase or remain constant. Therefore, the volume under the CDF surface will always be positive. Now, consider the following Corollary of Eq2.∎
Corollary: The second derivative of 2-Cats is a pseudo-likelihood..
Let be the bivariate CDF associated with RVs and . With and being the marginal CDFs, we reach that: . By Eq(2), . 2-Cats is also capturing the same bivariate CDF of . The bivariate density, , thus is:
(3)
The above equation is solved via the Chain rule. A consequence of it, is that the likelihood , is proportional to . This term is known as a pseudo-likelihood[23, 22].∎
Proof P2..
Given that by definition, and also and : . The volume being the double integral of is always positive or zero.∎
, , , and . Notice that this property is a relaxation of the one presented in Section 1.
This last property of a Copula is the one that guarantees that the Copula will have uniform marginals. In particular, the terms: and are of utmost importance for sampling. To grasp this, consider some valid Copula . Here, , which is the CDF of the Uniform distribution. A similar argument exists for .
2-Cats does not provably meet P3 as it does P1 and P2. Nevertheless, we prove that: and , while approximating , and .
Lemma P3.1.
and
Proof..
Consider the valueof when . As , we have and hence. Given that is a bivariate CDF, for any . The same proof exists for .∎
Conjecture P3.2.
and .
Meeting P3.2 via Lagrangian..
To meet this relaxed property, we propose the following constrained optimization. Let the inputs of our model be comprised of the set , where . Moreover, let be some loss term used to optimize our model (see the next section). Our optimization for 2-Cats will focus on the following constrained problem:
(4)
Although it is trivial to understand why this optimization meets P3.2, performing such an optimization presents several challenges. The first challenge is how to model constraints. A natural first choice is to consider the square of constraints. However, squared constraint optimization is an ill-posed optimization problem[8, Chapter2.1], and [54, Section2.1]. Secondly, our model may correctly estimate that (or similarly that ), but this will not necessarily ensure that the partial derivatives of and are accurate (see AppendixA.3). Such derivatives must equal one for these inputs as the Copula marginals are distributed according to an . Thirdly, evaluating these constraints over the entire domain and is impossible. The final challenge is how to optimize constraints. Here, a natural approach is to consider Lagrangian multipliers[54, 8, 69]. However, in such cases, the solution to the optimization problem will lie on a saddle point of the loss surface, and gradient methods (used in NNs and 2-Cats) do not converge on saddle points (Section 3.1 of[54] provides a simple intuition on why this is so).
We tackle the first and second challenges by considering the derivative of the square of the above constraints, i.e: and , which via the chain rule will be defined as follows:
(5)
and stand for requirements on and , respectively. Both requirements are always positive or zero because . Consequently, both requirements will only equal zero when P3.2 is met (our goal). One advantage of this constraint is that it also considers our model’s derivative, which tackles our second challenge above.
We evaluated the constraints on our training set to tackle the third challenge. This leads to the optimization stated below. are the constraints on the training data. From the arguments above, we have that: and if and only if the constraints are met.
(6)
As in[7], to solve the above problem, we treat as a barrier and optimize as follows:
1.
Let and be hyper-parameters (e.g., we set and ). Also, let be the initial model parameters (initialized as in[34] – aka Lecun Normal).
2.
For every training iteration :
(a)
via ADAM[33]
(b)
3.
Set our model as
As , we have that , leading our optimization to focus on the constraints. Even though we may not start our optimization in a feasible solution to [7], the fact that our loss surface is fully differentiable allows the optimizer to reach such a solution if not stuck on some local minima/saddle. Nevertheless, we align with the empirical evidence that in overparametrized models, such as deep NNs, local minima are rare (see Chapter 6.7 – The Optimality of Backpropagation – of[6]) and that such stochastic optimizers, such as ADAM, are suitable for escaping saddle points[6]. Moreover, our initialization, , is suitable for satisfying regularizers[34].
For comparison, we shall present results with and without these Lagrangian terms. ∎
4 Sobolev Losses for 2-Cats Models
We have already discussed how our approach ensures the validity of Copula functions. To train our models, we designed our loss function as a weighted sum, i.e.:
(7)
Here, , , and are loss weights.
The first component, , stimulates the copula to closely resemble the empirical cumulative distribution function derived from the observed data. In this way, our model learns to capture the essential characteristics of the data.The second component, , imposes penalties on any disparities between the fitted copula’s first-order derivatives and the data-based estimates. A Copula’s first derivative is essential for sampling (see AppendixB) and Vine methods. The third component, , focuses on the copula’s second-order derivative, which is linked to the probability density function of the data, and evaluates its proximity to the empirical likelihood. Incorporating this aspect in our loss function enhances the model’s ability to capture the data’s distribution.
From the arguments above, all three terms play a role in Copula modeling. While obtaining the first component is relatively straightforward, the other two components necessitate some original work. We explain each of them in turn. Before continuing, we point out that AppendixA provides an example of how NNs behave when estimating derivatives and integrals of functions.
Let be the bivariate CDF for some random vector and be the empirical cumulative distribution function (ECDF) estimated from a sample of size .The Glivenko-Cantelli theorem states that: as [48].
We can explore these results to define the relationship between our model, the ECDF, and the true CDF. Being the estimated parameters and given that the CDF and the Copula evaluate to the same value (i.e., and are the inverse of the marginal CDFs of and respectively):
Considering these definitions, we can define the loss functions based on the ECDF and the output:
(8)
Now for our second term, , a natural step for the first derivatives would be to define a mean squared error (or similar) loss towards these derivatives. The issue is that we need to define empirical estimates from data for both derivatives: and . To estimate such derivatives, we could have employed conditional density estimators[53, 63, 26]. Nevertheless, methods as[53, 63] are also deep NNs that have the drawback of requiring extra parameters to be estimated. Even classical methods such as[26] suffer from this issue. As a minor contribution, we explore the underlying properties of a Copula to present an empirical approximation for such derivatives.
For a 2d Copula, the first derivative of has the form[50]:
The issue with estimating this function from data is that we cannot filter our dataset to condition on (we are working with data where , a real number). However, using Bayes rule we can rewrite this equation. First, let us rewrite the above equation in terms of density functions. That is, if is a cumulative function, and are the marginal copula densities (uniform by definition). Using Bayes rule, we have that:
(9)
Where the last identity above comes from the fact that for Copulas, the marginal distributions and are uniform, leading to . Also, we shall have that and [50]. Now, estimate .
To do so, let us define as an empirical estimate of such a density using data points. We employ the widely used Kernel Density Estimation (KDE) to estimate this function. A fast algorithm for our estimation will work as follows: we arrange our dataset as a table of two columns, where each row contains the pairs of (the columns). For efficiency, we create views of this table sorted on column or column . When we iterate this table sorted on , finding the points where is trivial, as these are simply the previously iterated rows. If we perform KDE on the column for these points, we shall have an empirical estimate of the density: . By simply multiplying this estimate with , we have our empirical estimation of the partial derivative with regards to , that is: . Similarly, we can estimate . Now, our loss term is:
(10)
Finally, we focus on the last part of our loss, the a pseudo-likelihood:
(11)
It is essential to understand why the three losses are needed. It might seem that simply minimizing this loss would be sufficient. However, we can show that integrals of NN are also approximations up to an asymptotical constant[39], but we still need to ensure that this constant is acceptable (see AppendixA.2). Moreover, AppendixC presents an ablation study on the impact of all three losses.
5 Experimental Results
We now present our experimental results on different fronts: (1) validating our Empirical estimates for the first derivative, a crucial step when training 2-Cats; (2) evaluating 2-Cats on synthetic and real datasets without the lagrangian term; and, (3) evaluating the impact of the lagrangian term.
Before presenting results, we note that our models were implemented using Jax333https://jax.readthedocs.io/en/latest//Flax444https://flax.readthedocs.io. Given that Jax does not implement CDF methods for Multivariate Normal Distributions, we ported a fast approximation of bivariate CDFs[67] and employed it for our models that rely on Gaussian CDFs/Copulas. Our code is available at https://anonymous.4open.science/r/2cats-E765/. Baseline methods were evaluated using the author-supplied code (details are on AppendixG).
5.1 Empirical First Derivative Estimator
First, we validate our empirical estimations for the first derivative of Copulas. We use the closed-form equations for the first derivatives of Gaussian, Clayton, and Frank Copulas described in [57].
These Copulas have a single parameter, which we varied. Gaussian Copulas () has a correlation parameter, and we set it to: 0.1, 0.5, and 0.9. Clayton and Frank copulas’ mixing parameter () was set to 1, 5, and 10. For each of these configurations, we measured the coefficient of determination () between empirical estimates and exact formulae. Overall, we found that R2 was above in every setting, validating that our estimates are pretty accurate in practive.
5.2 2-Cats on Datasets (Without Lagrangian Term)
We now turn to our main comparisons. Our primary focus, as is the case on baseline methods, will be on capturing the data likelihood (Eq(3)). In our experiments, the PDFs for and were estimated via KDE. The bandwidth for these methods is determined using Silverman’s rule[59].
As is commonly done, we evaluate the natural log of this quantity as the data log-likelihood and use its negative to compare 2-Cats with other methods. With this metric of interest in place, we considered two 2-Cats models: (1) 2-Cats-G (Final Layer Gaussian): 2-Cats with a Gaussian bivariate CDF as ; and, (2) 2-Cats-L (Final Layer Logistic): 2-Cats with a Logistic bivariate CDF as .
Thus, we considered two forms of . The first was the CDF of a bivariate Gaussian Distribution[67]. The second one is the CDF of the Flexible bivariate Logistic (see Section 11.8 of[4]):
, , , and are free parameters optimized by the model. The same goes for the free parameters of the bivariate Gaussian CDF (the means of each dimension, and , and the correlation parameter ). Moreover, for each model, we employed four-layer networks with hidden layers having sizes of 128, 64, 32, and 16. We now discuss loss weights.
Recall that the data density is proportional to the pseudo-likelihood. We primarily emphasized the likelihood loss (). This accentuates our pivotal component. Notably, the scale of its values differs from that of the MSE components, necessitating a proportional adjustment in weight selection. In light of these factors, we fixed , , and . We note that these hyper-parameters were sufficient to show how our approach is better than the literature. Our models were trained using early stopping. Here, for each training epoch, we evaluated the pseudo-log-likelihood on the training set and kept the weights of the epoch with the best likelihood on the training dataset.
In both our synthetic and real data experiments, we considered the following approaches as Deep Learning baselines: Deep Archimedian Copulas (ACNET) [38]; Generative Archimedian Copulas (GEN-AC) [51]; Neural Likelihoods (NL) [13]; and Implicit Generative Copulas (IGC) [31]. Here, we use the same hyperparameters employed by the authors. We also considered several Non-Deep Learning baselines. These were the Parametric and Non-Parametric Copula Models from[49]. They are referenced as: The best Parametric approach from several known parametric copulas (Par), non-penalized Bernstein estimator (Bern), penalized Bernstein estimator (PBern), penalized linear B-spline estimator (PSPL1), penalized quadratic B-spline estimator (PSPL2), transformation local likelihood kernel estimator of degree (TTL0), degree (TTL1), and (TTL2), and with nearest neughbors (TTL2nn). We also considered the recent non-parametric approach by Mukhopadhyay and Parzen[47] (MR), as well as the Probit transformation-based estimator from[21].
To assess the performance of our model on synthetic bivariate datasets, we generated Gaussian copulas with correlations , and Clayton and Frank copulas with parameters . The marginal distributions were chosen as uncorrelated Normal distributions with parameters (). We generated 1,500 training samples and 500 test samples. Results are for test samples. Before continuing, we point out that some deep methods require datasets to be scaled to [0, 1] intervals[38, 51]. Indeed, the author-supplied source code does not execute with data outside of this range (this choice was first used[38, Section 4.2] and followed by [51, Section 5.1.1]). For fair comparisons, in AppendixE, we present results for models with a min-max scaling of the data.
The results are presented in Table1, which contains the average negative log-likelihood on the test set for each copula. Results were estimated with a fixed seed of 30091985. The table also shows 95% Bootstrap Confidence Intervals[70]. Following suit with several guidelines, we choose to report CIs for better interpretability on the expected range of[20, 65, 3, 28, 19, 24, 30].
Initially, note that statistical ties (same average or overlapping intervals) are present throughout the table. The exception is for the NL[13] method, which severely underperforms in this setting, and the methods that did not execute (due to the required data scaling). When reading this table, it is imperative to consider that the simulated data comes from the models evaluated in the Par (first) row.Thus, parametric methods are naturally expected to overperform. Nevertheless, the statistical ties with such methods show how 2-Cats can capture this behavior even if it is a family-free approach.
Boston
INTC-MSFT
GOOG-FB
Non-Deep Learn
Par
Bern
PBern
PSPL1
PSPL2
TTL0
TTL1
TTL2
TLL2
MR
Probit
Deep Learn
ACNet
GEN-AC
NL
IGC
Our
2-Cats-L
2-Cats-G
We now turn to real datasets. As was done in previous work, we employ the Boston housing dataset, the Intel-Microsoft (INTC-MSFT), and Google-Facebook (GOOG-FB) stock ratios. These pairs are commonly used in Copula research; in particular, they are the same datasets used by[38, 51]. We employed the same train/test and pre-processing as previous work for fair comparisons. Thus, the data is scaled with the same code as the literature, and all deep methods are executed.
The results are presented in Table2. The winning 2-Cats models are highlighted in the table in Green. The best, on average, baseline methods are in Red. We consider a method better or worse than another when their confidence intervals do not overlap; thus, ties with the best model are shown in underscores. Overall, the table highlights the superior performance of the 2-Cats models across all datasets. Only in one setting, 2-Cats-G on INTC-MSOFT, did the model underperform.
5.3 On the Lagrangian Term for P3
So far, 2-Cats presents itself as a promising approach for Copula modeling. Nevertheless, to sample from Copulas, we require that the marginals of the model come from a uniform distribution. This is our primary for the Lagrangian terms used to meet P3 (see Section3).
We trained 2-Cats models on the three real-world datasets with and without the Lagrangian optimization of Eq(6). Due to space constraints, we present results for 2-Cats-L only. Here, models were trained for 1000 iterations. No early stopping was performed as our focus was on meeting P3.
We compare both the average absolute deviations (i.e., and ) as well as the average relative deviations (i.e., and ) for models trained with and without constraints.
For the GOOG-FB dataset, significant gains were achieved with the constraints, i.e.: went from to and improved from to . improved from to and went from to . The negative log-likelihood reduced slightly when Lagrangian was used (from -1.70 to -1.14). Nevertheless, the model is still better than the baselines (see Table2). Results for the negative log-likelihood do not exactly match the previous ones, as no early stopping was done here. Even though relative errors may appear large, note that such scores are severely sensitive to the tail of the distributions. Here, even a small deviation (e.g., 0.001 to 0.0015) incurs a large relative increase.
For INTC-MSOFT, results were similar across and dimensions. As in both versions, no significant gains were observed in nor ; these scores are pretty small (below ) in both cases. In relative terms, an increase was observed in from roughly 4% to 2%, with being 4% in both cases. No significant changes were observed in the Boston dataset where errors were already minor and below in both instances, with relative errors ranging from 2% to 4% regardless of the constraints being employed. This shows that 2-Cats may already approximate the Uniform marginals even without constraints.
These results show the efficacy of our Lagrangian penalty. In some settings, such as GOOG-GB, constraints may present significant improvements. Using the Lagrangian term will depend on whether simulation is essential to the end-user (see AppendixB).
6 Conclusions
In this paper, we presented Copula Approximating Transform models (2-Cats).Different from the literature when training our models; we focus not only on capturing the pseudo-likelihood of Copulas but also on meeting or approximating several other properties, such as the partial derivatives of the function. Moreover, a second major innovation is proposing Sobolev training for Copulas. Overall, our results show the superiority of 2-Cats on real datasets.
A natural follow-up for 2-Cats is on using the Pair Copula Construction (PCC)[1, 15] to port Vines[16, 41, 23, 22] to incorporate our method. PCC is a well-established algorithm to go from a 2D Copula to an ND One using Vines. We also believe that evaluating other non-negative NNs for is a promising direction[45, 55, 68, 62, 40, 36].
References
[1]Kjersti Aas, Claudia Czado, Arnoldo Frigessi, and Henrik Bakken.Pair-copula constructions of multiple dependence.Insurance: Mathematics and economics, 44(2), 2009.
[2]Mohomed Abraj, You-Gan Wang, and MHelen Thompson.A new mixture copula model for spatially correlated multiplevariables with an environmental application.Scientific Reports, 12(1), 2022.
[3]Douglas Altman, David Machin, Trevor Bryant, and Martin Gardner.Statistics with confidence: confidence intervals and statisticalguidelines.John Wiley & Sons, 2013.
[4]BarryC Arnold.Multivariate logistic distributions.Marcel Dekker New York, 1992.
[5]Yves INgounou Bakam and Denys Pommeret.Nonparametric estimation of copulas and copula densities byorthogonal projections.Econometrics and Statistics, 2023.
[6]Pierre Baldi.Deep learning in science.Cambridge University Press, 2021.
[7]Kevin Bello, Bryon Aragam, and Pradeep Ravikumar.Dagma: Learning dags via m-matrices and a log-determinant acyclicitycharacterization.In NeurIPS, 2022.
[8]DimitriP Bertsekas.Constrained optimization and Lagrange multiplier methods.Academic press, 2014.
[9]ChristopherM Bishop.Mixture density networks.Online Report., 1994.
[10]Yongqiang Cai.Achieve the minimum width of neural networks for universalapproximation.In ICLR, 2023.
[11]George Casella and RogerL Berger.Statistical inference.Cengage Learning, 2001.
[12]Umberto Cherubini, Elisa Luciano, and Walter Vecchiato.Copula methods in finance.John Wiley & Sons, 2004.
[13]Pawel Chilinski and Ricardo Silva.Neural likelihoods via cumulative distribution functions.In UAI, 2020.
[14]George Cybenko.Approximation by superpositions of a sigmoidal function.Mathematics of control, signals, and systems, 2(4), 1989.
[15]Claudia Czado.Pair-copula constructions of multivariate copulas.In Copula Theory and Its Applications. Springer, 2010.
[16]Claudia Czado and Thomas Nagler.Vine copula based modeling.Annual Review of Statistics and Its Application, 9, 2022.
[17]WojciechM Czarnecki, Simon Osindero, Max Jaderberg, Grzegorz Swirszcz, andRazvan Pascanu.Sobolev training for neural networks.In NeurIPS, 2017.
[18]Hennie Daniels and Marina Velikova.Monotone and partially monotone neural networks.IEEE Transactions on Neural Networks, 21(6), 2010.
[19]Pierre Dragicevic.Fair statistical communication in hci.Modern statistical methods for HCI, 2016.
[20]Martin Gardner and Douglas Altman.Confidence intervals rather than p values: estimation rather thanhypothesis testing.Br Med J (Clin Res Ed), 292(6522), 1986.
[21]Gery Geenens, Arthur Charpentier, and Davy Paindaveine.Probit transformation for nonparametric kernel estimation of thecopula density.Bernoulli, 2017.
[22]Christian Genest and Anne-Catherine Favre.Everything you always wanted to know about copula modeling but wereafraid to ask.Journal of hydrologic engineering, 12(4), 2007.
[23]Christian Genest, Kilani Ghoudi, and L-P Rivest.A semiparametric estimation procedure of dependence parameters inmultivariate families of distributions.Biometrika, 82(3), 1995.
[24]Sander Greenland, StephenJ Senn, KennethJ Rothman, JohnB Carlin, CharlesPoole, StevenN Goodman, and DouglasG Altman.Statistical tests, p values, confidence intervals, and power: a guideto misinterpretations.European journal of epidemiology, 31, 2016.
[25]Joshua Größer and Ostap Okhrin.Copulae: An overview and recent developments.Wiley Interdisciplinary Reviews: Computational Statistics,14(3), 2022.
[26]Peter Hall, RodneyCL Wolff, and Qiwei Yao.Methods for estimating a conditional distribution function.Journal of the American Statistical association, 94(445), 1999.
[28]Joses Ho, Tayfun Tumkaya, Sameer Aryal, Hyungwon Choi, and Adam Claridge-Chang.Moving beyond p values: data analysis with estimation graphics.Nature methods, 16(7), 2019.
[29]Kurt Hornik, Maxwell Stinchcombe, and Halbert White.Multilayer feedforward networks are universal approximators.Neural networks, 2(5), 1989.
[30]GuidoW Imbens.Statistical significance, p-values, and the reporting of uncertainty.Journal of Economic Perspectives, 35(3), 2021.
[32]GeorgeEm Karniadakis, IoannisG Kevrekidis, LuLu, Paris Perdikaris, SifanWang, and Liu Yang.Physics-informed machine learning.Nature Reviews Physics, 3(6), 2021.
[33]DiederikP Kingma and Jimmy Ba.Adam: A method for stochastic optimization.In ICLR, 2015.
[34]Günter Klambauer, Thomas Unterthiner, Andreas Mayr, and Sepp Hochreiter.Self-normalizing neural networks.In NeurIPS, 2017.
[35]Ivan Kobyzev, SimonJD Prince, and MarcusA Brubaker.Normalizing flows: An introduction and review of current methods.IEEE transactions on pattern analysis and machine intelligence,43(11), 2020.
[36]Ryan Kortvelesy.Fixed integral neural networks.arXiv preprint arXiv:2307.14439, 2023.
[37]Georg Lindgren, Holger Rootzén, and Maria Sandsten.Stationary stochastic processes for scientists and engineers.CRC press, 2013.
[39]Yucong Liu.Neural networks are integrable.arXiv preprint arXiv:2310.14394, 2023.
[40]Lorenzo Loconte, Stefan Mengel, Nicolas Gillis, and Antonio Vergari.Negative mixture models via squaring: Representation and learning.In The 6th Workshop on Tractable Probabilistic Modeling, 2023.
[41]Rand KwongYew Low, Jamie Alco*ck, Robert Faff, and Timothy Brailsford.Canonical vine copulas in the context of modern portfolio management:Are they worth it?Asymmetric Dependence in Finance: Diversification, Correlationand Portfolio Management in Market Downturns, 2018.
[42]LuLu, Raphael Pestourie, Wenjie Yao, Zhicheng Wang, Francesc Verdugo, andStevenG Johnson.Physics-informed neural networks with hard constraints for inversedesign.SIAM Journal on Scientific Computing, 43(6), 2021.
[43]Yulong Lu and Jianfeng Lu.A universal approximation theorem of deep neural networks forexpressing probability distributions.In NeurIPS, 2020.
[44]Zhou Lu, Hongming Pu, Feicheng Wang, Zhiqiang Hu, and Liwei Wang.The expressive power of neural networks: A view from the width.In NeurIPS, 2017.
[45]Ulysse Marteau-Ferey, Francis Bach, and Alessandro Rudi.Non-parametric models for non-negative functions.In NeurIPS, 2020.
[46]Sadegh Modiri, Santiago Belda, Mostafa Hoseini, Robert Heinkelmann, JoséMFerrándiz, and Harald Schuh.A new hybrid method to improve the ultra-short-term prediction oflod.Journal of geodesy, 94, 2020.
[47]Subhadeep Mukhopadhyay and Emanuel Parzen.Nonparametric universal copula modeling.Applied Stochastic Models in Business and Industry, 36(1),2020.
[48]Michael Naaman.On the tight constant in the multivariatedvoretzky–kiefer–wolfowitz inequality.Statistics & Probability Letters, 173, 2021.
[49]Thomas Nagler, Christian Schellhase, and Claudia Czado.Nonparametric estimation of simplified vine copula models: comparisonof methods.Dependence Modeling, 5(1), 2017.
[50]RogerB Nelsen.An introduction to copulas.Springer, 2006.
[51]Yuting Ng, Ali Hasan, Khalil Elkhalil, and Vahid Tarokh.Generative archimedean copulas.In UAI, 2021.
[52]TTin Nguyen, HienD Nguyen, Faicel Chamroukhi, and GeoffreyJ McLachlan.Approximation by finite mixtures of continuous density functions thatvanish at infinity.Cogent Mathematics & Statistics, 7(1), 2020.
[53]George Papamakarios, Theo Pavlakou, and Iain Murray.Masked autoregressive flow for density estimation.In NeurIPS, 2017.
[54]John Platt and Alan Barr.Constrained differential optimization.In NeurIPS, 1987.
[55]Alessandro Rudi and Carlo Ciliberto.Psd representations for effective probability models.In NeurIPS, 2021.
[56]GSalvadori and Carlo DeMichele.Frequency analysis via copulas: Theoretical aspects and applicationsto hydrological events.Water resources research, 40(12), 2004.
[57]Ulf Schepsmeier and Jakob Stöber.Derivatives and fisher information of bivariate copulas.Statistical papers, 55(2), 2014.
[59]BernardW Silverman.Density estimation for statistics and data analysis, volume26.CRC press, 1986.
[60]Abe Sklar.Random variables, distribution functions, and copulas: a personallook backward and forward.Lecture notes-monograph series, 1996.
[61]MSklar.Fonctions de répartition à n dimensions et leurs marges.Annales de l’ISUP, 8(3), 1959.
[62]AleksanteriMikulus Sladek, Martin Trapp, and Arno Solin.Encoding negative dependencies in probabilistic circuits.In The 6th Workshop on Tractable Probabilistic Modeling, 2023.
[63]Kihyuk Sohn, Honglak Lee, and Xinchen Yan.Learning structured output representation using deep conditionalgenerative models.In NeurIPS, 2015.
[64]BSohrabian.Geostatistical prediction through convex combination of archimedeancopulas.Spatial Statistics, 41, 2021.
[65]JonathanAC Sterne and GeorgeDavey Smith.Sifting the evidence—what’s wrong with significance tests?Physical therapy, 81(8), 2001.
[66]Natasa Tagasovska, Firat Ozdemir, and Axel Brando.Retrospective uncertainties for deep models using vine copulas.In AISTATS, 2023.
[67]Wen-Jen Tsay and Peng-Hsuan Ke.A simple approximation for the bivariate normal integral.Communications in Statistics-Simulation and Computation, 52(4),2023.
[68]Russell Tsuchida, ChengSoon Ong, and Dino Sejdinovic.Squared neural families: A new class of tractable density models.arXiv preprint arXiv:2305.13552, 2023.
[69]GRWalsh.Saddle-point property of lagrangian function.Methods of Optimization. New York: John Wiley & Sons, 1975.
[70]Larry Wasserman.All of statistics: a concise course in statistical inference,volume26.Springer, 2004.
[71]Antoine Wehenkel and Gilles Louppe.Unconstrained monotonic neural networks.In NeurIPS, 2019.
[72]Liuyang Yang, Jinghuai Gao, Naihao Liu, Tao Yang, and Xiudi Jiang.A coherence algorithm for 3-d seismic data analysis based on themutual information.IEEE Geoscience and Remote Sensing Letters, 16(6), 2019.
[73]Qingyang Zhang and Xuan Shi.A mixture copula bayesian network model for multimodal genomic data.Cancer Informatics, 16, 2017.
Appendix A On Derivatives and Integrals of NNs
In this appendix, we present an example of the issue of approximating derivatives/integrals of NNs.
A.1 An Example on why approximating derivatives and integrals fail
We begin with a pictorial example of the issue. Before doing so, we take some time to revise the universal approximation properties of NNs[14, 29, 43]. We also shift our focus from Copulas for this argument. It is expected to state that the universal approximation theorem (UAT) guarantees that functions are approximated with NNs. However, given that when learning models from training data, it would be more correct to say that function evaluations at training points are approximated.
Consider a simple 4-Layer Relu NN with layers of 128 neurons. This will be our model. To avoid overfitting, this model will be optimized with a validation set. The problem with relying only on the UAT is shown in Figure1. The figure shows a ground truth data-generating process of , where is some noise term. On the first column, when , the model will approximate the data generating function (middle row), its derivative (bottom row), and integral (top row). When , the model will approximate the data points, but not necessarily the derivatives.
This example may seem to contradict[17, 29] others that looked into the universal approximation properties of NN regarding derivative. We point out that this is not the case. In AppendixA.3, we show how the variance will increase when estimating the derivative, leading to the figure’s oscillating behavior. From the figure, also note that integrals are approximated up to an asymptotic constant (see AppendixA.2, and [39]); however, we still need to control for this constant (note the growing deviation on the top plot of the middle row).
We now present some results on the Integral (exemplified on the top row of Figure1) and Derivative (bottom row) that tie in with our example. Furthermore, we briefly discuss the last column of the figure. Notice that on the extreme left and extreme right, the derivative of the network begins to diverge from the underlying process. The NN was not trained with points sampled after these boundaries. Under data shifts (e.g., data from a similar process not seen on the training set), the issues we point out throughout this Appendix should increase.
A.2 On the Integral of Neural Networks
To present a view of the theoretical bounds on the error of approximation of integrals, let be some NN parametrized by . Also, let represent our training dataset. We also make the reasonable assumption our dataset comes from some real function, , sampled under a symmetric additive noise, i.e., , with an expected value equal to zero: . Gaussian noise is such a case.
If a NN is a universal approximator[14, 29, 42, 10, 44] and it is trained on the dataset , we reach:
Most UAT proofs are valid for any norm ( norms bound one another), and assuming the 1 norm, we reach: for some constant . As a consequence, we reach:
Given that integration is a linear operator, we can show that:
To do so, we shall use the law of the unconscious statistician as follows:
where is a constant as is also a constant. However, this constant will grow depending on the integrating interval. This growing error is visible on the plot on the center top of Figure1.Controlling for will depend on the dataset’s quality and the training procedure.
A similar result is discussed in[39].
A.3 On the Derivative of Underlying Process and its impact on Neural Networks
To understand the oscillations on the derivatives of the example, let us rewrite the underlying process that generated the data as a Gaussian process. That is: . is the stochastic process and is some phase component necessary for stationarity. The variance of this process will be given by the kernel function , which is now controlled by parameter . Given that this process has the same variance over time, , it is stationary, and a stationary Gaussian process has [37]. Before continuing, we note that in our example, we sampled from a Gaussian with a standard deviation of and variance of . This leads to a value of that is of since .
Now, let be the derivative of this process. This may be defined as:
if:
when: .
Given that our process is stationary and has the expected value of the noise term as zero, we can estimate the value of the derivative process as zero, i.e., [37]. We also have that the variance is equal to the second derivative of : [37].
From here, we can derive twice to show that: . Recall that and . Now, by solving a simple inequality, we can show that the variance of the derivative process, , is greater than the variance of the original process , i.e., when (as is our example where ).
This result explains the oscillation in the bottom middle plot of Figure1. The derivative process has a higher variation; thus, we expect a higher variance in estimations of the derivative.
Overall, even if an NN is a universal approximation of derivatives[17, 29], the variance of the noise term will likely increase (here we began with a variance of 0.25 and it is already sufficient to show such an increase) even for the derivative of the NN.
Appendix B Sampling from 2-Cats
We now detail how to sample from 2-Cats. Let the first derivatives of 2-Cats be: , and . Now, let be the first derivative of , and similarly . These derivatives are readily available in the Hessian matrix that we estimate symbolically for 2-Cats (see Section3).
Now, let us determine the inverse of , that is: (the subscript stands for inverse). The same arguments are valid for and , and thus we omit them. Notice that with this inverse, we can sample from the CDF defined by using the well-known Inverse transform sampling. Now, notice that by definition, is already the derivative of with regards to .
Where, again, the derivative inside the parenthesis is readily available to us via symbolic computation.
With such results in hand, the algorithm for sampling is:
1.
Generate two independent values and .
2.
Set . This is an Inverse transform sampling for
3.
Now we have the pair. We can again use Inverse transform sampling to determine:
(a)
(b)
Appendix C Ablation Study
We now provide an ablation study to understand the impact of our three losses. In this study, our model was trained without the Lagrangian terms of Property P3.
BOSTON
Only for training
0.115
0.033
-0.448
and for training
0.107
0.035
-0.454
and for training
0.103
0.023
-0.236
All Three Losses (paper)
0.107
0.020
-0.630
INCT-MSOFT
Only for training
0.131
0.007
-0.327
and for training
0.141
0.004
-0.314
and for training
0.137
0.007
-0.320
All Three Losses (paper)
0.141
0.018
-0.402
GOOG-FB
Only for training
0.166
6.88
-3.324
and for training
0.133
0.066
-2.163
and for training
0.195
2.50
-3.201
All Three Losses (paper)
0.136
0.087
-2.881
The three tables (see Table3) of this section below present the values for (the squared error of the cumulative function C), (the squared error for the first derivatives of C), and (the copula likelihood). We present one table for each real-world dataset.
is the score of interest when comparing models and the one we report in our paper. However, as discussed in AppendixA of our manuscript, when an NN approximates one aspect of a function (here being the Copula density ), the NN may miss other elements, such as the integral and derivatives of the function. This is the reason why we now present results for all three metrics. The metrics are in the columns of the tables. Each row presents a different 2-Cats approach using: (1) only for training (the metric of most interest); (2) and ; and ; and, (3) all three losses (as in the main text).
From the tables, we can see that every loss plays a role when training the model. When using all three losses, gains/ties are achieved in 5 out of 9 cases (bold). When ignoring such terms, large losses (italic) will also arise
Appendix D Conjecture: A Mixture of 2-Cats is a Universal Copula Approximator
With sufficient components, any other CDF (or density function) may be approximated by a mixture of other CDFs[52]. As a consequence, a mixture of 2-Cats of the form:
where is the -th 2-Cats model. is the parameter vector comprised of concatenating the individual parameters of each mixture component: , and are real numbers related to the mixture weights.
This model may be trained as follows:
1.
Map mixture parameters to the Simplex: . if the -th position of this vector.
2.
When training, backpropagate to learn and .
This is similar to the Mixture Density NN[9]. Over the last few years, mixture of Copulas approaches has been gaining traction in several fields[2, 64, 73]; here, we are proposing an NN variation. By definition, this is a valid Copula.
Table4 presents results without scaling the input data as is done in Deep Learning methods. The table presents the average negative log-likelihood and the 95% confidence interval.
The colors and highlights on this table match the main text. The winning 2-Cats models are highlighted in the table in Green. The best, on average, baseline methods are in Red. The table shows that 2-Cats is better than baseline methods when the dependency ( or ) increases. For small dependencies, methods not based on Deep Learning outperform Deep ones (including 2-Cats). Nevertheless, 2-Cats is the winning method in 6 out of 9 datasets and is tied with the best in one case.
Appendix F A Flexible Variation of the Model
A Flexible 2-Cats model works similarly to our 2-Cats. However, the transforms are different:
1.
Let be an MLP outputs positive numbers. We achieve this by employing Elu plus one activation in every layer as in[71].
2.
Define the transforms:
3.
The 2-Cats-FLEX hypothesis is now defined as the function: , where is any bivariate CDF on the domain (e.g., the Bivariate Logistic or Bivariate Gaussian).
This model meets P2, but not P1 nor P3. As already stated in the introduction, the fact that and are monotonic, one-to-one, guarantees that defines a valid probability transform to a new CDF. A major issue with this approach is that we have no guarantee that the model’s derivatives are conditional cumulative functions (first derivatives) or density functions (second derivatives). That’s why we call it Flexible (FLEX).
F.1 Full Results
Boston
INTC-MSFT
GOOG-FB
Par
Bern
PBern
PSPL1
PSPL2
TTL0
TTL1
TTL2
TLL2
MR
Probit
ACNet
GEN-AC
NL
IGC
2-Cats-FLEX-L
2-Cats-FLEX-G
2-Cats-G
2-Cats-L
In Table5, we show the results for these models on real datasets. The definitions of the models are as follows: (2-Cats-P Gaus.) A parametric mixture of 10 Gaussian Copula densities.; (2-Cats-P Frank) Similar to the above, but is a mixture of 10 Frank Copula densities; (2-Cats-FLEX-G) A Flexible model version where the final activation layer is Bivariate Gaussian CDF; (2-Cats-FLEX-L) A Flexible version of the model where the final activation layer is Bivariate Logistic CDF.
Models were trained with the same hyperparameters and network definition as the ones in our main text.
Introduction: My name is Geoffrey Lueilwitz, I am a zealous, encouraging, sparkling, enchanting, graceful, faithful, nice person who loves writing and wants to share my knowledge and understanding with you.
We notice you're using an ad blocker
Without advertising income, we can't keep making this site awesome for you.