Learning Machines 101 - podcast cover

Learning Machines 101

Richard M. Golden, Ph.D., M.S.E.E., B.S.E.E.www.learningmachines101.com
Smart machines based upon the principles of artificial intelligence and machine learning are now prevalent in our everyday life. For example, artificially intelligent systems recognize our voices, sort our pictures, make purchasing suggestions, and can automatically fly planes and drive cars. In this podcast series, we examine such questions such as: How do these devices work? Where do they come from? And how can we make them even smarter and more human-like? These are the questions that will be addressed in this podcast series!

Episodes

LM101-056: How to Build Generative Latent Probabilistic Topic Models for Search Engine and Recommender System Applications

In this NEW episode we discuss Latent Semantic Indexing type machine learning algorithms which have a PROBABILISTIC interpretation. We explain why such a probabilistic interpretation is important and discuss how such algorithms can be used in the design of document retrieval systems, search engines, and recommender systems. Check us out at: www.learningmachines101.com and follow us on twitter at: @lm101talk

Sep 20, 201628 minSeason 1Ep. 56

LM101-055: How to Learn Statistical Regularities using MAP and Maximum Likelihood Estimation (Rerun)

In this rerun of Episode 10, we discuss fundamental principles of learning in statistical environments including the design of learning machines that can use prior knowledge to facilitate and guide the learning of statistical regularities. In particular, the episode introduces fundamental machine learning concepts such as: probability models, model misspecification, maximum likelihood estimation, and MAP estimation. Check us out at: www.learningmachines101.com...

Aug 16, 201635 minSeason 1Ep. 55

LM101-054: How to Build Search Engine and Recommender Systems using Latent Semantic Analysis (RERUN)

Welcome to the 54th Episode of Learning Machines 101 titled "How to Build a Search Engine, Automatically Grade Essays, and Identify Synonyms using Latent Semantic Analysis" (rerun of Episode 40). The principles in this episode are also applicable to the problem of "Market Basket Analysis" and the design of Recommender Systems. Check it out at: www.learningmachines101.com and follow us on twitter: @lm101talk

Jul 25, 201630 minSeason 1Ep. 54

LM101-053: How to Enhance Learning Machines with Swarm Intelligence (Particle Swarm Optimization)

In this 53rd episode of Learning Machines 101, we introduce the concept of a Swarm Intelligence with respect to Particle Swarm Optimization Algorithms. The essential idea of “Swarm Intelligence” is that you have a group of individual entities which behave in a coordinated manner yet there is no master control center providing directions to all of the individuals in the group. The global group behavior is an “emergent property” of local interactions among individuals in the group! We will analyze...

Jul 11, 201627 minSeason 1Ep. 53

LM101-052: How to Use the Kernel Trick to Make Hidden Units Disappear

Today, we discuss a simple yet powerful idea which began popular in the machine learning literature in the 1990s which is called “The Kernel Trick”. The basic idea of the “Kernel Trick” is that you specify similarity relationships among input patterns rather than a recoding transformation to solve a nonlinear problem with a linear learning machine. It's a great magic trick...check it out at: www.learningmachines101.com where you can obtain transcripts of this episode and download free machine le...

Jun 13, 201629 minSeason 1Ep. 52

LM101-051: How to Use Radial Basis Function Perceptron Software for Supervised Learning[Rerun]

This particular podcast is a RERUN of Episode 20 and describes step by step how to download free software which can be used to make predictions using a feedforward artificial neural network whose hidden units are radial basis functions. This is essentially a nonlinear regression modeling problem. We show the performance of this nonlinear learning machine is substantially better on test data set than the linear learning machine software presented in Episode 13. Basically performance for the linea...

May 24, 201629 minSeason 1Ep. 51

LM101-050: How to Use Linear Machine Learning Software to Make Predictions (Linear Regression Software)[RERUN]

In this episode we will explain how to download and use free machine learning software from the website: www.learningmachines101.com . This podcast is concerned with the very practical issues associated with downloading and installing machine learning software on your computer. If you follow these instructions, by the end of this episode you will have installed one of the simplest (yet most widely used) machine learning algorithms on your computer. You can then use the software to make virtually...

May 04, 201631 minSeason 1Ep. 50

LM101-049: How to Experiment with Lunar Lander Software

In this episode we continue the discussion of learning when the actions of the learning machine can alter the characteristics of the learning machine’s statistical environment. We describe how to download free lunar lander software so you can experiment with an autopilot for a lunar lander module that learns from its experiences and describe the results of some simulation studies. To learn more, visit: www.learningmachines101.com to download the free lunar lander software which illustrates princ...

Apr 22, 201635 minSeason 1Ep. 49

LM101-048: How to Build a Lunar Lander Autopilot Learning Machine (Rerun)

In this episode we consider the problem of learning when the actions of the learning machine can alter the characteristics of the learning machine’s statistical environment. We illustrate the solution to this problem by designing an autopilot for a lunar lander module that learns from its experiences. For more information, check out: www.learningmachines101.com and visit us a twitter: @lm101talk #machinelearning #statistics #artificialintelligence...

Mar 29, 201631 minSeason 1Ep. 48

LM101-047: How Build a Support Vector Machine to Classify Patterns (Rerun)

We explain how to estimate the parameters of such machines to classify a pattern vector as a member of one of two categories as well as identify special pattern vectors called “support vectors” which are important for characterizing the Support Vector Machine decision boundary. The relationship of Support Vector Machine parameter estimation and logistic regression parameter estimation is also discussed. For more information..check us out at: www.learningmachines101.com also check us out on twitt...

Mar 14, 201635 minSeason 1Ep. 47

LM101-046: How to Optimize Student Learning using Recurrent Neural Networks (Educational Technology)

In this episode, we briefly review Item Response Theory and Bayesian Network Theory methods for the assessment and optimization of student learning and then describe a poster presented on the first day of the Neural Information Processing Systems conference in December 2015 in Montreal which describes a Recurrent Neural Network approach for the assessment and optimization of student learning called “Deep Knowledge Tracing”. For more details check out: www.learningmachines101.com and follow us on...

Feb 23, 201623 minSeason 1Ep. 46

LM101-045: How to Build a Deep Learning Machine for Answering Questions about Images

In this episode we discuss just one out of the 102 different posters which was presented on the first night of the 2015 Neural Information Processing Systems Conference. This presentation describes a system which can answer simple questions about images. Check out: www.learningmachines101.com f or additional details!!

Feb 08, 201622 minSeason 1Ep. 45

LM101-044: What happened at the Deep Reinforcement Learning Tutorial at the 2015 Neural Information Processing Systems Conference?

This is the third of a short subsequence of podcasts providing a summary of events associated with Dr. Golden’s recent visit to the 2015 Neural Information Processing Systems Conference. This is one of the top conferences in the field of Machine Learning. This episode reviews and discusses topics associated with the Introduction to Reinforcement Learning with Function Approximation Tutorial presented by Professor Richard Sutton on the first day of the conference. Check out: www.learningmachines1...

Jan 26, 201632 minSeason 1Ep. 44

LM101-043: How to Learn a Monte Carlo Markov Chain to Solve Constraint Satisfaction Problems (Rerun of Episode 22)

Welcome to the 43rd Episode of Learning Machines 101! We are currently presenting a subsequence of episodes covering the events of the recent Neural Information Processing Systems Conference. However, this week will digress with a rerun of Episode 22 which nicely complements our previous discussion of the Monte Carlo Markov Chain Algorithm Tutorial. Specifically, today we discuss the problem of approaches for learning or equivalently parameter estimation in Monte Carlo Markov Chain algorithms. T...

Jan 12, 201628 minSeason 1Ep. 43

LM101-042: What happened at the Monte Carlo Markov Chain (MCMC) Inference Methods Tutorial at the 2015 Neural Information Processing Systems Conference?

This is the second of a short subsequence of podcasts providing a summary of events associated with Dr. Golden’s recent visit to the 2015 Neural Information Processing Systems Conference. This is one of the top conferences in the field of Machine Learning. This episode reviews and discusses topics associated with the Monte Carlo Markov Chain (MCMC) Inference Methods Tutorial held on the first day of the conference. Check out: www.learningmachines101.com to listen or download this podcast episode...

Dec 29, 201526 minSeason 1Ep. 42

LM101-041: What happened at the 2015 Neural Information Processing Systems Deep Learning Tutorial?

This is the first of a short subsequence of podcasts which provides a summary of events associated with Dr. Golden’s recent visit to the 2015 Neural Information Processing Systems Conference. This is one of the top conferences in the field of Machine Learning. This episode introduces the Neural Information Processing Systems Conference and reviews the content of the Morning Deep Learning Tutorial which took place on the first day of the conference. Check out: www.learningmachines101.com f or add...

Dec 16, 201530 minSeason 1Ep. 41

LM101-040: How to Build a Search Engine, Automatically Grade Essays, and Identify Synonyms using Latent Semantic Analysis

In this episode we introduce a very powerful approach for computing semantic similarity between documents. Here, the terminology “document” could refer to a web-page, a word document, a paragraph of text, an essay, a sentence, or even just a single word. Two semantically similar documents, therefore, will discuss many of the same topics while two semantically different documents will not have many topics in common. Machine learning methods are described which can take as input large collections ...

Nov 24, 201528 minSeason 1Ep. 40

LM101-039: How to Solve Large Complex Constraint Satisfaction Problems (Monte Carlo Markov Chain and Markov Fields)[Rerun]

In this episode we discuss how to solve constraint satisfaction inference problems where knowledge is represented as a large unordered collection of complicated probabilistic constraints among a collection of variables. The goal of the inference process is to infer the most probable values of the unobservable variables given the observable variables. Concepts of Markov Random Fields and Monte Carlo Markov Chain methods are discussed. For additional details and technical notes, please visit the w...

Nov 09, 201535 minSeason 1Ep. 39

LM101-038: How to Model Knowledge Skill Growth Over Time using Bayesian Nets

In this episode, we examine the problem of developing an advanced artificially intelligent technology which is capable of tracking knowledge growth in students in real-time, representing the knowledge state of a student a skill profile, and automatically defining the concept of a skill without human intervention! The approach can be viewed as a sophisticated state-of-the-art extension of the Item Response Theory approach to Computerized Adaptive Testing Educational Technology described in Episod...

Oct 27, 201524 minSeason 1Ep. 38

LM101-037: How to Build a Smart Computerized Adaptive Testing Machine using Item Response Theory

In this episode, we discuss the problem of how to build a smart computerized adaptive testing machine using Item Response Theory (IRT). Suppose that you are teaching a student a particular target set of knowledge. Examples of such situations obviously occur in nursery school, elementary school, junior high school, high school, and college. However, such situations also occur in industry when top professionals in a particular field attend an advanced training seminar. All of these situations woul...

Oct 12, 201535 minSeason 1Ep. 37

LM101-036: How to Predict the Future from the Distant Past using Recurrent Neural Networks

In this episode, we discuss the problem of predicting the future from not only recent events but also from the distant past using Recurrent Neural Networks (RNNs). A example RNN is described which learns to label images with simple sentences. A learning machine capable of generating even simple descriptions of images such as these could be used to help the blind interpret images, provide assistance to children and adults in language acquisition, support internet search of content in images, and ...

Sep 28, 201525 minSeason 1Ep. 36

LM101-035: What is a Neural Network and What is a Hot Dog?

In this episode, we address the important questions of “What is a neural network?” and “What is a hot dog?” by discussing human brains, neural networks that learn to play Atari video games, and rat brain neural networks. Check out: www.learningmachines101.com for videos of a neural network that learns to play ATARI video games and transcripts of this podcast!!! Also follow us on twitter at: @lm101talk See you soon!!

Sep 15, 201529 minSeason 1Ep. 35

LM101-034: How to Use Nonlinear Machine Learning Software to Make Predictions (Feedforward Perceptrons with Radial Basis Functions)[Rerun]

Welcome to the 34th podcast in the podcast series Learning Machines 101 titled "How to Use Nonlinear Machine Learning Software to Make Predictions". This particular podcast is a RERUN of Episode 20 and describes step by step how to download free software which can be used to make predictions using a feedforward artificial neural network whose hidden units are radial basis functions. This is essentially a nonlinear regression modeling problem. Check out: www.learningmachines101.com and follow us ...

Aug 25, 201529 minSeason 1Ep. 34

LM101-033: How to Use Linear Machine Learning Software to Make Predictions (Linear Regression Software)[RERUN]

In this episode will explain how to download and use free machine learning software which can be downloaded from the website: www.learningmachines101.com . The software can be used to make predictions using your own data sets. Although we will continue to focus on critical theoretical concepts in machine learning in future episodes, it is always useful to actually experience how these concepts work in practice.This is a rerun of Episode 13....

Aug 10, 201531 minSeason 1Ep. 33

LM101-032: How To Build a Support Vector Machine to Classify Patterns

In this 32nd episode of Learning Machines 101, we introduce the concept of a Support Vector Machine. We explain how to estimate the parameters of such machines to classify a pattern vector as a member of one of two categories as well as identify special pattern vectors called “support vectors” which are important for characterizing the Support Vector Machine decision boundary. The relationship of Support Vector Machine parameter estimation and logistic regression parameter estimation is also dis...

Jul 13, 201535 minSeason 1Ep. 32

LM101-030: How to Improve Deep Learning Performance with Artificial Brain Damage (Dropout and Model Averaging)

Deep learning machine technology has rapidly developed over the past five years due in part to a variety of actors such as: better technology, convolutional net algorithms, rectified linear units, and a relatively new learning strategy called "dropout" in which hidden unit feature detectors are temporarily deleted during the learning process. This article introduces and discusses the concept of "dropout" to support deep learning performance and makes connections of the "dropout" concept to conce...

Jun 08, 201532 minSeason 1Ep. 30

LM101-029: How to Modernize Deep Learning with Rectilinear units, Convolutional Nets, and Max-Pooling

This podcast discusses talks, papers, and ideas presented at the recent International Conference on Learning Representations 2015 which was followed by the Artificial Intelligence in Statistics 2015 Conference in San Diego. Specifically, commonly used techniques shared by many successful deep learning algorithms such as: rectilinear units, convolutional filters, and max-pooling are discussed. For more details please visit our website at: www.learningmachines101.com !...

May 25, 201536 minSeason 1Ep. 29

LM101-028: How to Evaluate the Ability to Generalize from Experience (Cross-Validation Methods)[RERUN]

This rerun of an earlier episode of Learning Machines 101 discusses the problem of how to evaluate the ability of a learning machine to make generalizations and construct abstractions given the learning machine is provided a finite limited collection of experiences. Check out: www.learningmachines101.com to obtain transcripts of this podcast and download free machine learning software!

May 11, 201535 minSeason 1Ep. 28
For the best experience, listen in Metacast app for iOS or Android
Open in Metacast