The distributional learning theory or learning of probability distribution is a framework in computational learning theory. It has been proposed from Michael Kearns, Yishay Mansour, Dana Ron, Ronitt Rubinfeld, Robert Schapire and Linda Sellie in 1994 [1] and it was inspired from the PAC-framework introduced by Leslie Valiant.[2]
In this framework the input is a number of samples drawn from a distribution that belongs to a specific class of distributions. The goal is to find an efficient algorithm that, based on these samples, determines with high probability the distribution from which the samples have been drawn. Because of its generality, this framework has been used in a large variety of different fields like machine learning, approximation algorithms, applied probability and statistics.
This article explains the basic definitions, tools and results in this framework from the theory of computation point of view.
^M. Kearns, Y. Mansour, D. Ron, R. Rubinfeld, R. Schapire, L. Sellie On the Learnability of Discrete Distributions. ACM Symposium on Theory of Computing, 1994 [1]
^L. Valiant A theory of the learnable. Communications of ACM, 1984
and 23 Related for: Distribution learning theory information
The distributionallearningtheory or learning of probability distribution is a framework in computational learningtheory. It has been proposed from Michael...
Statistical learningtheory is a framework for machine learning drawing from the fields of statistics and functional analysis. Statistical learningtheory deals...
recognition and computational learningtheory in artificial intelligence. In 1959, Arthur Samuel defined machine learning as a "field of study that gives...
computational learningtheory, probably approximately correct (PAC) learning is a framework for mathematical analysis of machine learning. It was proposed...
reinforcement learning is studied in many disciplines, such as game theory, control theory, operations research, information theory, simulation-based...
probability theory and statistics, the Gumbel distribution (also known as the type-I generalized extreme value distribution) is used to model the distribution of...
from statistics, fuzzy logic, and probability theory. There is a close connection between machine learning and compression. A system that predicts the posterior...
In probability theory and statistics, the beta distribution is a family of continuous probability distributions defined on the interval [0, 1] or (0, 1)...
In probability theory and statistics, a probability distribution is the mathematical function that gives the probabilities of occurrence of different...
In probability theory and statistics, the binomial distribution with parameters n and p is the discrete probability distribution of the number of successes...
Unsupervised learning is a method in machine learning where, in contrast to supervised learning, algorithms learn patterns exclusively from unlabeled data...
mechanics and mathematics, a Boltzmann distribution (also called Gibbs distribution) is a probability distribution or probability measure that gives the...
combined use of computer hardware, software, and educational theory and practice to facilitate learning. When referred to with its abbreviation, "EdTech," it...
In statistics and machine learning, ensemble methods use multiple learning algorithms to obtain better predictive performance than could be obtained from...
In probability theory and statistics, the Weibull distribution /ˈwaɪbʊl/ is a continuous probability distribution. It models a broad range of random variables...
In machine learning, the kernel embedding of distributions (also called the kernel mean or mean map) comprises a class of nonparametric methods in which...
Empirical risk minimization is a principle in statistical learningtheory which defines a family of learning algorithms based on evaluating performance over a...
to as "quantum learningtheory". Quantum-enhanced machine learning refers to quantum algorithms that solve tasks in machine learning, thereby improving...
In computational learningtheory (machine learning and theory of computation), Rademacher complexity, named after Hans Rademacher, measures richness of...
In computational learningtheory, Occam learning is a model of algorithmic learning where the objective of the learner is to output a succinct representation...
Supervised learning (SL) is a paradigm in machine learning where input objects (for example, a vector of predictor variables) and a desired output value...
probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization...