Identifying Feature Relevance using a Random Forest thumbnail
Pause
Mute
Subtitles
Playback speed
0.25
0.5
0.75
1
1.25
1.5
1.75
2
Full screen

Identifying Feature Relevance using a Random Forest

Published on Feb 25, 200712613 Views

Many feature selection algorithms are limited in that they attempt to identify relevant feature subsets by examining the features individually. This paper introduces a technique for determining featur

Related categories

Chapter list

Identifying Feature Relevance Using a Random Forest00:00
Overview00:19
Random Forest01:17
Random Forest (cont...)01:46
Feature Relevance: Ranking03:05
Feature Relevance: Subset Methods04:19
Relevance Identification using Average Information Gain05:56
Node Complexity Compensation07:18
Unique & Non-Unique Arrangements08:01
Node Complexity Compensation (cont…)08:49
Information Gain Density Functions09:43
Information Gain Density Functions10:34
Employing Feature Relevance11:25
Parallel12:51
Convergence Rates14:08
Results15:03
Irrelevant Features16:19
Expected Information Gain16:56
Expected Information Gain17:54
Bounds on Expected Information Gain18:56
Irrelevant Features: Bounds19:39
Friedman20:13
Simple21:42
Results22:08
Summary23:49