The search session has expired. Please query the service again.
The search session has expired. Please query the service again.
The search session has expired. Please query the service again.
The search session has expired. Please query the service again.
The search session has expired. Please query the service again.
The search session has expired. Please query the service again.
The search session has expired. Please query the service again.
Gaussian process modeling is one of the most popular approaches for building a metamodel in the case of expensive numerical simulators. Frequently, the code outputs correspond to physical quantities with a behavior which is known a priori: Chemical concentrations lie between 0 and 1, the output is increasing with respect to some parameter, etc. Several approaches have been proposed to deal with such information. In this paper, we introduce a new framework for incorporating constraints in Gaussian...
The purpose of feature selection in machine learning is at least two-fold - saving measurement acquisition costs and reducing the negative effects of the curse of dimensionality with the aim to improve the accuracy of the models and the classification rate of classifiers with respect to previously unknown data. Yet it has been shown recently that the process of feature selection itself can be negatively affected by the very same curse of dimensionality - feature selection methods may easily over-fit...
High-dimensional data models abound in genomics studies, where often inadequately small sample sizes create impasses for incorporation of standard statistical tools. Conventional assumptions of linearity of regression, homoscedasticity and (multi-) normality of errors may not be tenable in many such interdisciplinary setups. In this study, Kendall's tau-type rank statistics are employed for statistical inference, avoiding most of parametric assumptions to a greater extent. The proposed procedures...
In this paper the Kagan divergence measure is extended in order to establish a measure of the information that a random sample gives about a Dirichlet process as a whole. After studying some of its properties, the expression obtained in sampling from the step n to the step n+1 is given, and its Bayesian properties are studied. We finish proving the good behaviour of a stopping rule defined on the basis of the information obtained in sampling when passing from a step to the following.
In this work we introduce a nonparametric recursive aggregation process called Multilayer Aggregation (MLA). The name refers to the fact that at each step the results from the previous one are aggregated and thus, before the final result is derived, the initial values are subjected to several layers of aggregation. Most of the conventional aggregation operators, as for instance weighted mean, combine numerical values according to a vector of weights (parameters). Alternatively, the MLA operators...
We study sample-based estimates of the expectation of the function
produced by the empirical minimization algorithm. We investigate the
extent to which one can estimate the rate of convergence of the
empirical minimizer in a data dependent manner. We establish three
main results. First, we provide an algorithm that upper bounds the
expectation of the empirical minimizer in a completely
data-dependent manner. This bound is based on a structural result
due to Bartlett and Mendelson, which relates...
We study the performance of empirical risk minimization (ERM), with respect to the quadratic risk, in the context of convex aggregation, in which one wants to construct a procedure whose risk is as close as possible to the best function in the convex hull of an arbitrary finite class . We show that ERM performed in the convex hull of is an optimal aggregation procedure for the convex aggregation problem. We also show that if this procedure is used for the problem of model selection aggregation,...
We consider the problem of providing optimal uncertainty quantification (UQ) – and hence rigorous certification – for partially-observed functions. We present a UQ framework within which the observations may be small or large in number, and need not carry information about the probability distribution of the system in operation. The UQ objectives are posed as optimization problems, the solutions of which are optimal bounds on the quantities of interest; we consider two typical settings, namely parameter...
Let X be a one dimensional positive recurrent diffusion continuously observed on [0,t] . We consider a non parametric estimator of the drift function on a given interval. Our estimator, obtained using a penalized least square approach, belongs to a finite dimensional functional space, whose dimension is selected according to the data. The non-asymptotic risk-bound reaches the minimax optimal rate of convergence when t → ∞. The main point of our work is that we do not suppose the process to be in...
Let X be a one dimensional positive recurrent diffusion continuously observed on [0,t] . We consider a non parametric estimator of the drift function on a given interval. Our estimator, obtained using a penalized least square approach, belongs to a finite dimensional functional space, whose dimension is selected according to the data. The non-asymptotic risk-bound reaches the minimax optimal rate of convergence when t → ∞. The main point of our work is that we do not suppose the process to be in...
Let (X, Y) be a random couple in S×T with unknown distribution P. Let (X1, Y1), …, (Xn, Yn) be i.i.d. copies of (X, Y), Pn being their empirical distribution. Let h1, …, hN:S↦[−1, 1] be a dictionary consisting of N functions. For λ∈ℝN, denote fλ:=∑j=1Nλjhj. Let ℓ:T×ℝ↦ℝ be a given loss function, which is convex with respect to the second variable. Denote (ℓ•f)(x, y):=ℓ(y; f(x)). We study the following penalized empirical risk minimization problem
which is an empirical version of the problem
(hereɛ≥0...
Currently displaying 21 –
40 of
46