![]() |
|
![]() |
![]() |
|
![]() |
![]() |
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[Subscribe]
Re: DM: CHAID vs. CARTŪFrom: Ronny Kohavi Date: Fri, 26 Sep 1997 09:58:51 -0400 (EDT)
Mike> Ronny Kohavi wrote:
>> Since the theory says that it's impossible for one algorithm to
>> uniformly beat any other on generalization accuracy in
>> classification tasks (where uniformly is for all possible target
>> concepts), the question is *when* (under what conditions) one
>> algorithm is better than another, not *whether*.
>>
Mike> Do you have a reference to the paper or book that describes such
Mike> a theory? Thanks.
Nicholas mentioned Schaffer's paper which I won't repeat. The theorem
is actually trivial and has been mentioned through the years in
various paper (Tom Mitchell's version spaces and some older refs in
the late 60's).
David Wolpert makes things very formal in his paper, including
interesting repercussions for meta-level learning:
author = "David H. Wolpert",
title = "The Relationship between {PAC}, the statistical physics
framework, the {Bayesian} framework, and the {VC}
framework",
year = 1994,
booktitle = {The Mathematics of Generalization},
publisher = {Addison Wesley},
editor = {David H. Wolpert}
The proof is based on the following simple idea for Boolean labels:
Without any assumptions, if algorithm A beats algorithm B
on a dataset, then there exists a dataset with the opposite labels
and
it is just as likely, so B will beat A on that.
In real life, there are smoothness assumptions, especially on
real-valued attributes, which is why the theorem isn't too interesting
and why work in the field is still going on. Identifying those
smoothness assumptions (or other types of assumptions) is the big
problem. There are very few rules of thumb as to when one
should use one algorithm over another.
--
Ronny Kohavi (ronnyk@sgi.com, http://robotics.stanford.edu/~ronnyk)
Engineering Manager, Analytical Data Mining.
Silicon Graphics, Inc.
|
MHonArc 2.2.0