Material Detail

Model Compression: Bagging your Cake and Eating it too (part 2)

Model Compression: Bagging your Cake and Eating it too (part 2)

This video was recorded at NIPS Workshop on Efficient Machine Learning, Whistler 2007. The ever increasing size of available data to be processed by machine learning algorithms has yielded several approaches, from online algorithms to parallel and distributed computing on multi-node clusters. Nevertheless, it is not clear how modern machine learning approaches can either cope with such parallel machineries or take into account strong constraints regarding the available time to handle training and/or test examples. This workshop explores two alternatives: 1. modern machine learning approaches that can handle real time processing at train and/or at test time, under strict computational constraints (when the flow of incoming data is continuous and needs to be handled), and 2. modern machine learning approaches that can take advantage of new commodity hardware such as multicore, GPUs, and fast networks. This two-day workshop aims to set the agenda for future advancements by fostering a discussion of new ideas and methods and by demonstrating the potential uses of readily-available solutions. It brings together both researchers and practitioners to offer their views and experience in applying machine learning to large scale learning. Find out more at the Workshop website.


  • User Rating
  • Comments
  • Learning Exercises
  • Bookmark Collections
  • Course ePortfolios
  • Accessibility Info

More about this material


Log in to participate in the discussions or sign up if you are not already a MERLOT member.