published: Oct. 9, 2014, recorded: August 2014, views: 74765
Report a problem or upload filesIf you have found a problem with this lecture or would like to send us extra material, articles, exercises, etc., please use our ticket system to describe your request and upload the data.
Enter your e-mail into the 'Cc' field, and we will keep you updated with your request's status.
Building intelligent systems that are capable of extracting high-level representations from high-dimensional sensory data lies at the core of solving many AI related tasks, including visual object or pattern recognition, speech perception, and language understanding. Theoretical and biological arguments strongly suggest that building such systems requires deep architectures that involve many layers of nonlinear processing. Many existing learning algorithms use shallow architectures, including neural networks with only one hidden layer, support vector machines, kernel logistic regression, and many others. The internal representations learned by such systems are necessarily simple and are incapable of extracting some types of complex structure from high-dimensional input. In the past few years, researchers across many different communities, from applied statistics to engineering, computer science and neuroscience, have proposed several deep (hierarchical) models that are capable of extracting useful, high-level structured representations. An important property of these models is that they can extract complex statistical dependencies from high-dimensional sensory input and efficiently learn high-level representations by re-using and combining intermediate concepts, allowing these models to generalize well across a wide variety of tasks. The learned high-level representations have been shown to give state-of-the-art results in many challenging learning problems, where data patterns often exhibit a high degree of variations, and have been successfully applied in a wide variety of application domains, including visual object recognition, information retrieval, natural language processing, and speech perception. A few notable examples of such models include Deep Belief Networks, Deep Boltzmann Machines, Deep Autoencoders, and sparse coding-based methods.
Download slides: kdd2014_salakhutdinov_deep_learning.pdf (53.7 MB)
Link this pageWould you like to put a link to this lecture on your homepage?
Go ahead! Copy the HTML snippet !
Reviews and comments:
Power systems are complex structure that require deep (hierarchical) models that are capable of extracting useful, high-level structured representations, for many applications such DSS, diagnosis and prediction systems. I am interested in using Deep Learning for model representation on processes real database. I would like to know your opinion. Intelligent systems are fundamental about the implementation of smart grid.
check this blog out! <a href="http://baahkast.com/">baahkast.com</a>
This lecture is very interesting, it helped me a lot with my research here http://dazzle.ru/proizvodstvo/7036-ot...
The intelligence of computer science and technologies is increasing day by day. Machine learning education evolution is included in the university and academics. Learners also get the advantage of Machine learning and I also learned the SPSS Statistics at the stage of @https://sixdollaressay.com/spss-assignment-help/ and this is capable of providing the high-level of SPSS writing.
Great content by the way very informative I gained a lot of knowledge about the science topics from this page as well as from https://freeroku.wordpress.com/ this page so keeps up the informative work.
Write your own review or comment: