Chapter 1 - Introduction
2015-04-12 19:27
253 查看
1. Applications and problems
Applications
Text or document classification, e.g., spam detection;
Natural language processing, e.g., morphological analysis, part-of-speech tagging, statistical parsing, named-entity recognition;
Speechrecognition, speech synthesis, speaker verification;
Optical character recognition (OCR);
Computational biology applications, e.g., protein function or structured prediction;
Computer vision tasks, e.g., image recognition, face detection;
Faud detection (credit card, telephone) and network intrusion;
Games, e.g., chess, backgammon;
Medical diagnosis;
Recommendation systems, search enginesm information extraction systems.
Problems
Classification
Regression
Ranking
Clustering
Dimensionality reduction or manifold learning
1.2 Definitions and terminology
Examples
Features
Labels
Training sample
Validation sample
Test sample
Loss function
Hypothesis set
1.3 Cross-validation
In practice, the amount of labeled data available is often too small to set aside a validation sample since that would leave an insufficient amount of training data. Instead, a widely adopted method known as n-fold cross-validation is used to exploit the labeled data both for model selection (selection of the free parameters of the algorithm) and for training.
1.4 Learning scenarios
Supervised learning
The learner receives a set of labeled examples as training data and makes predictions for all unseen points.
Unsupervised learning
The learner exclusively receives unlabeled training data,and makes predictions for all unseen points.
Semi-unsupervised learning
The learner receives a training sample consisting of both labeled and unlabeled data, and makes predictions for all unseen points.
Transductive inference
As in the semi-supervised scenario, the learner receives a labeled training sample along with a set of unlabeled test points. However, the objective of transductive inference is to predict labels only for these particular test points.
On-line learning
In contrast with the previous scenarios, the online scenario
involves multiple rounds and training and testing phases are intermixed. At each
round, the learner receives an unlabeled training point, makes a prediction, receives
the true label, and incurs a loss
Reinforcement learning
The training and testing phases are also intermixed in reinforcement learning. To collect information, the learner actively interacts with the environment and in some cases affects the environment, and receives an immediate reward for each action. The object of the learner is to maximize his reward over a course of actions and iterations with the environment.
Active learning
The learner adaptively or interactively collects training examples,
typically by querying an oracle to request labels for new points.
Applications
Text or document classification, e.g., spam detection;
Natural language processing, e.g., morphological analysis, part-of-speech tagging, statistical parsing, named-entity recognition;
Speechrecognition, speech synthesis, speaker verification;
Optical character recognition (OCR);
Computational biology applications, e.g., protein function or structured prediction;
Computer vision tasks, e.g., image recognition, face detection;
Faud detection (credit card, telephone) and network intrusion;
Games, e.g., chess, backgammon;
Medical diagnosis;
Recommendation systems, search enginesm information extraction systems.
Problems
Classification
Regression
Ranking
Clustering
Dimensionality reduction or manifold learning
1.2 Definitions and terminology
Examples
Features
Labels
Training sample
Validation sample
Test sample
Loss function
Hypothesis set
1.3 Cross-validation
In practice, the amount of labeled data available is often too small to set aside a validation sample since that would leave an insufficient amount of training data. Instead, a widely adopted method known as n-fold cross-validation is used to exploit the labeled data both for model selection (selection of the free parameters of the algorithm) and for training.
1.4 Learning scenarios
Supervised learning
The learner receives a set of labeled examples as training data and makes predictions for all unseen points.
Unsupervised learning
The learner exclusively receives unlabeled training data,and makes predictions for all unseen points.
Semi-unsupervised learning
The learner receives a training sample consisting of both labeled and unlabeled data, and makes predictions for all unseen points.
Transductive inference
As in the semi-supervised scenario, the learner receives a labeled training sample along with a set of unlabeled test points. However, the objective of transductive inference is to predict labels only for these particular test points.
On-line learning
In contrast with the previous scenarios, the online scenario
involves multiple rounds and training and testing phases are intermixed. At each
round, the learner receives an unlabeled training point, makes a prediction, receives
the true label, and incurs a loss
Reinforcement learning
The training and testing phases are also intermixed in reinforcement learning. To collect information, the learner actively interacts with the environment and in some cases affects the environment, and receives an immediate reward for each action. The object of the learner is to maximize his reward over a course of actions and iterations with the environment.
Active learning
The learner adaptively or interactively collects training examples,
typically by querying an oracle to request labels for new points.
相关文章推荐
- 读书笔记之 Advanced Bash-Scripting Guide Chapter 4 Introduction to Variables and Parameters
- Chapter 3. Introduction to Visual Studio .NET
- chapter 1 Introduction to objects
- Chapter 1. Introduction to Web Automation
- 深入理解Linux内核 Chapter1 introduction
- Introduction to Algorithm(chapter 7)
- Learning CCNA - Chapter 2: Introduction to TCP/IP
- Chapter 1:Introduction
- PRML Notes-Chapter1 Introduction(1.2 Probability Theory)
- 《Linux Kernel Development》chapter 1 Introduction to Linux Kernel
- Chapter1 Introduction to Databases
- Introduction to Algorithm - Summary of Chapter 7 - Quicksort
- SymmetricDS文档翻译--【Chapter 1. 简介(Introduction)】
- 【PRML读书笔记-Chapter1-Introduction】1.4 The Curse of Dimensionality
- 移动端web开发 chapter 1 – introduction
- PRML学习心得:Chapter1—Introduction
- Introduction to Algorithm(chapter 14)
- OpenGL ES2.0 Programming Guide - chapter 3:An introduction to EGL
- Chapter 1 Introduction
- Introduction to Algorithms Chapter 12: Binary Search Tree