[go: nahoru, domu]

Jump to content

User:Scott.linderman/sandbox

From Wikipedia, the free encyclopedia

In computational learning theory, probably approximately correct learning (PAC learning) is a framework for mathematical analysis of machine learning. It was proposed in 1984 by Leslie Valiant.[1]

In this framework, the learner receives samples and must select a generalization function (called the hypothesis) from a certain class of possible functions. The goal is that, with high probability (the "probably" part), the selected function will have low generalization error (the "approximately correct" part). The learner must be able to learn the concept given any arbitrary approximation ratio, probability of success, or distribution of the samples.

The model was later extended to treat noise (misclassified samples).

An important innovation of the PAC framework is the introduction of computational complexity theory concepts to machine learning. In particular, the learner is expected to find efficient functions (time and space requirements bounded to a polynomial of the example size), and the learner itself must implement an efficient procedure (requiring an example count bounded to a polynomial of the concept size, modified by the approximation and likelihood bounds).

Definitions and terminology[edit]

In order to give the definition for something that is PAC-learnable, we first have to introduce some terminology.[2] [3]

For the following definitions, two examples will be used. The first is the problem of character recognition given an array of bits encoding a binary-valued image. The other example is the problem of finding an interval that will correctly classify points within the interval as positive and the points outside of the range as negative.

The instance space, , is a set called of all possible inputs to the learning algorithm. In the character recognition problem, the instance space is , where is the length of the inputs. In the interval problem the instance space is , where denotes the set of all real numbers.

A concept is a subset . One concept is the set of all patterns of bits in that encode a picture of the letter "P". In the interval learning problem, an example of a concept is the interval . A concept class is a set of concepts over . This could be the set of intervals of length less than one, for example. The goal of the learning problem will be to identify a hypothesis concept that is close. in a sense that will be made precise, to the true concept given a set of labeled samples .

Paralleling the concept class is the hypothesis class. , from which the hypothesis is drawn. In many cases the hypothesis class is equal to the concept class, though this need not be the case. The hypothesis class may be restricted to a smaller set as a form of regularization. Alternatively, the hypothesis class may be infinitely large in order to capture as many potential concepts as possible.

Let be a procedure that draws an example, , using a probability distribution and gives the correct label , that is 1 if and 0 otherwise.

Say that there is an algorithm that given access to and inputs and that, with probability at least , outputs a hypothesis that has error less than or equal to with examples drawn from according to . If there is such an algorithm for every concept , for every distribution over , and for all and then is PAC learnable (or distribution-free PAC learnable) by . We can also say that is a PAC learning algorithm for .

An algorithm runs in time if it draws at most examples and requires at most time steps. A concept class is efficiently PAC learnable if it is PAC learnable by an algorithm that runs in time polynomial in , and instance length.

Example: Learning intervals[edit]

Consider the aforementioned problem of learning intervals on the real line. You are given a set of labeled points , where and if and otherwise. The locations of these points are drawn from an unknown distribution . Our goal is to provide an algorithm that, given such a set of points, outputs an interval that, with high probability, will be "close" to the true interval in that the expected classification error on new points drawn from will be small. We will show that the following algorithm satisfies this property.

Algorithm: Given a set of labeled points , output the interval where and . To evaluate the performance of this simple algorithm, we prove the following theorem,
Theorem: For all , if examples are given, then with probability the interval returned by the above algorithm will be incorrect on less than an fraction of examples drawn from . That is, the probability of an error will be less than .
Proof: First, notice that the algorithm can only underestimate the true interval since it always returns the tightest interval containing the positively labeled points. We call an interval -good if and , where the probability is taken with respect to distribution . Such a hypothesis will have error less than since
.
Assuming that the total probability of the interval is greater than , let and be the maximum and minimum boundaries, respectively, of an -good interval.
Now we analyze the probability of returning an interval that is not -good. This can happen if none of the points in are less than or greater than . Since the points are independently sampled, the probability of each event is less than , and, by a union bound, the probability of either event occurring is less than the sum of their probabilities, .
To achieve our confidence guarantee, we solve for the number of examples necessary to guarantee that the interval will be -good with probability . We want , or . Using the inequality , this is satisfied if or . This is the lower bound on the required by the theorem, and thus concludes the proof.

Occam's Razor[edit]

(This page is under development by user AlexE.)

How does one go about finding an efficient learning algorithm for a given concept class? It turns out that a very simple principle known as Occam's Razor, first articulated by the thirteenth century philosopher William of Occam, provides a guide. Simply finding a succinct hypothesis that is consistent with the given examples is enough to guarantee that the hypothesis will generalize to new examples as well. This statement can be formalized in terms of the cardinality of the hypothesis class (i.e. the number of potential hypotheses), yielding a general upper bound on the sample complexity of the form,

.

This can also be seen from an information theoretic perspective: if the examples can be compressed and represented by a succinct hypothesis, then in the PAC framework it can be shown that this hypothesis will, with high probability, generalize to new examples with low error. More details can be found in Occam Learning.

Reconsider the Learning Intervals above. The algorithm we gave shares the spirit of Occam learning and has a very similar sample complexity. We simply returned the smallest interval consistent with the given examples. However, the Occam learning theory does not seem to apply since there are an infinite number of hypotheses (intervals) to choose from. That is, the hypothesis class has infinite cardinality. Still, the concept of closed intervals on the real line is really quite simple. To extend these ideas to real-valued concepts, we leverage tools of Vapnik–Chervonenkis theory.

Relationship to Vapnik Chervonenkis theory[edit]

Vapnik-Chervonenkis (VC) theory introduces the notion of the VC dimension of a concept class. The VC dimension is the maximum number such that there exist points and concepts in that classify the points in all possible ways. This can substitute for , the log cardinality of the concept class, in the Occam learning bounds discussed above. A similar upper bound of

can be shown for the sample complexity of PAC-learning a concept class with VC-dimension . This upper bound can be shown to be tight to within a factor of , though the proof is nontrivial. For more details, see [4].

A similar relationship can be shown (under some regularity conditions) when is a Glivenko-Cantelli class.

Boosting[edit]

Boosting addresses the question of whether many weak learners can be combined to produce a strong learner. Concretely, if we have a learning algorithm that is weak in the sense that it returns a concept with error (i.e. only slightly better than a random guess), then boosting seeks to use this weak learning algorithm to produce a learning algorithm that returns a concept with arbitrarily low error. It turns out that this is indeed possible. The most common means of boosting accuracy is by maintaining a set of candidate learning algorithms and weighting them according to how well they perform on a set of training examples. We can iteratively reweight these candidates such that weighted average response of the candidates achieves the desired accuracy, even though each individual has poor overall accuracy.

Learning in the presence of noise[edit]

The PAC framework described above assumes that examples are labeled according to the true underlying concept. This is, arguably, the least realistic assumption since, in reality, our labels are imperfect. When these labels are corrupted by noise, either random or malicious, then we place even greater demands upon our learning algorithms. The simplest noise model is the classification error model in which labels are independently flipped with probability . To formalize the concepts that can be learned in the presence of such noise, Kearns[5] introduced the statistical query model, a strictly weaker model than the PAC-learning model. Any concept that can be efficiently learned using statistical queries is efficiently PAC learnable in the presence of classification noise.

There are other, more restrictive noise models as well, such as the malicious noise model in which an adversary is allowed to corrupt each label with probability in order to make the concept more challenging to learn. Here, the limits of what may be efficiently PAC learned are more severe.

Finally, though not necessarily noise, it is also likely the case that the true process that generates our labels does not belong to our hypothesis class. In this case, the disagreement in labels is due to model misspecification. This problem is studied under the heading of agnostic learning. In this scenario, we attempt to make as few assumptions as possible about the underlying set of functions or distributions that give rise to our examples, and seek to find the concept in our hypothesis class that is closest (according to some loss function) to the underlying concept. This is a particularly challenging learning model. In general, those concepts which can be agnostically learned are closely related to those which have the property of uniform convergence.

(Add page from AlexE.)

Biological applications of PAC learning[edit]

PAC theory unites the computational and statistical requirements of learning algorithms. Though these are most evident in machine learning, they are equally applicable to biological learning problems. Leslie Valiant has extended the PAC framework to the problems of evolution [6] and neural computation [7]. Though not bound by the same constraints as computers, there is no reason to believe these systems can circumvent the pragmatic constraints of polynomial time computability. In his book[8], Valiant argues that the PAC framework provides important constraints for guiding our understanding of these biological processes.

References[edit]

  1. ^ L. Valiant. A theory of the learnable. Communications of the ACM, 27, 1984.
  2. ^ Kearns and Vazirani, pg. 1-12,
  3. ^ Balas Kausik Natarajan, Machine Learning , A Theoretical Approach, Morgan Kaufmann Publishers, 1991
  4. ^ Kearns and Vazirani, Chapter 3
  5. ^ Kearns, Michael. "Efficient noise-tolerant learning from statistical queries." Journal of the ACM (JACM) 45.6 (1998): 983-1006.
  6. ^ Valiant, Leslie G. "Evolvability." Journal of the ACM (JACM) 56.1 (2009): 3.
  7. ^ Leslie G. Valiant. Circuits of the Mind. Oxford University Press, 1994.
  8. ^ Valiant, L. Probably Approximately Correct. Basic Books, 2013.

Further reading[edit]

Category:Computational learning theory