SOTAVerified

Gaussian Processes

Gaussian Processes is a powerful framework for several machine learning tasks such as regression, classification and inference. Given a finite set of input output training data that is generated out of a fixed (but possibly unknown) function, the framework models the unknown function as a stochastic process such that the training outputs are a finite number of jointly Gaussian random variables, whose properties can then be used to infer the statistics (the mean and variance) of the function at test values of input.

Source: Sequential Randomized Matrix Factorization for Gaussian Processes: Efficient Predictions and Hyper-parameter Optimization

Papers

Showing 76100 of 1963 papers

TitleStatusHype
The Neural Process Family: Survey, Applications and PerspectivesCode1
Light curve completion and forecasting using fast and scalable Gaussian processes (MuyGPs)Code1
Stationary Kernels and Gaussian Processes on Lie Groups and their Homogeneous Spaces I: the compact caseCode1
Low-Precision Arithmetic for Fast Gaussian ProcessesCode1
Volatility Based Kernels and Moving Average Means for Accurate Forecasting with Gaussian ProcessesCode1
Transformer Neural Processes: Uncertainty-Aware Meta Learning Via Sequence ModelingCode1
Supernova Light Curves Approximation based on Neural Network ModelsCode1
LIMO: Latent Inceptionism for Targeted Molecule GenerationCode1
Neural Diffusion ProcessesCode1
Active Bayesian Causal InferenceCode1
Posterior and Computational Uncertainty in Gaussian ProcessesCode1
Bayesian Active Learning with Fully Bayesian Gaussian ProcessesCode1
High-dimensional additive Gaussian processes under monotonicity constraintsCode1
Probabilistic Estimation of Instantaneous Frequencies of Chirp SignalsCode1
Meta-learning Adaptive Deep Kernel Gaussian Processes for Molecular Property PredictionCode1
Gaussian Processes for Missing Value ImputationCode1
GP-BART: a novel Bayesian additive regression trees approach using Gaussian processesCode1
Diverse Text Generation via Variational Encoder-Decoder Models with Gaussian Process PriorsCode1
AutoIP: A United Framework to Integrate Physics into Gaussian ProcessesCode1
Invariance Learning in Deep Neural Networks with Differentiable Laplace ApproximationsCode1
Supervising the Multi-Fidelity Race of Hyperparameter ConfigurationsCode1
Bayesian Optimization of Function NetworksCode1
Transformers Can Do Bayesian InferenceCode1
Gaussian Process Regression With Interpretable Sample-Wise Feature WeightsCode1
State-space deep Gaussian processes with applicationsCode1
Show:102550
← PrevPage 4 of 79Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ICKy, periodicRoot mean square error (RMSE)0.03Unverified