Information Theoretic Model Validation by Approximate Optimization
published: Jan. 5, 2011, recorded: December 2010, views: 4760
Report a problem or upload filesIf you have found a problem with this lecture or would like to send us extra material, articles, exercises, etc., please use our ticket system to describe your request and upload the data.
Enter your e-mail into the 'Cc' field, and we will keep you updated with your request's status.
Model selection in pattern recognition requires (i) to specify a suitable cost function for the data interpretation and (ii) to control the degrees of freedom depending on the noise level in the data. We advocate an information theoretic perspective where the uncertainty in the measurements quantizes the solution space of the underlying optimization problem, thereby adaptively regularizing the cost function. A pattern recognition model, which can tolerate a higher level of fluctuations in the measurements than alternative models, is considered to be superior provided that the solution is equally informative. The optimal tradeoff between ‘‘informativeness’’ and ‘‘robustness’’ is quantified by the approximation capacity of the selected cost function. Empirical evidence for this model selection concept is provided by cluster validation in computer security, i.e., multilabel clustering of Boolean data for role based access control, but also in high dimensional Gaussian mixture models and the analysis of microarray data. Furthermore, the approximation capacity of the SVD cost function suggests a cutoff value for the SVD spectrum.
Link this pageWould you like to put a link to this lecture on your homepage?
Go ahead! Copy the HTML snippet !