NIPS Workshop on Learning when Test and Training Inputs Have Different Distributions, Whistler 2006

NIPS Workshop on Learning when Test and Training Inputs Have Different Distributions, Whistler 2006

2 Lectures · Dec 9, 2006

About

Many machine learning algorithms assume that the training and the test data are drawn from the same distribution. Indeed many of the proofs of statistical consistency, etc., rely on this assumption. However, in practice we are very often faced with the situation where the training and the test data both follow the same conditional distribution, p(y|x), but the input distributions, p(x), differ. For example, principles of experimental design dictate that training data is acquired in a specific manner that bears little resemblance to the way the test inputs may later be generated.

The aim of this workshop is to try and shed light on the kind of situations where explicitly addressing the difference in the input distributions is beneficial, and on what the most sensible ways of doing this are.

Related categories

Uploaded videos:

Invited Talks

video-img
30:28

Projection and Projectability

David Corfield

Feb 25, 2007

 · 

3507 Views

Invited Talk

Lectures

video-img
21:53

Using features of probability distributions to achieve covariate shift

Alex Smola

Feb 25, 2007

 · 

6644 Views

Lecture