Machine Learning

Friday, February 25, 2005

Knowledge Intensive Machine Learning, Part 1

Machine learning is all about learning from extensional information (datapoints). The incorporation of intensional information (domain knowledge) is generally done by hand via data encoding and feature and model selection. Doing this well is arguably the most difficult part of a successful machine learning system. It would sure be nice if our learning algorithm could make direct use of qualitative domain knowledge, say for model or feature selection, or reducing the parameter space? This is the idea behind "knowledge intensive machine learning".

Probably the best (though almost trivial) example of this is the simple Bayesian network. A Bayes net is a probabilistic model which can be learned from data, but for which the user can very easily specify high-level qualitative domain knowledge (causality or conditional independence) which significantly constrains the parameter space. What else do we have? Not much.

Here's a motivating example: suppose you are building a simple anti-smoking propaganda Bayes net. You have domain knowledge that tells you what nodes should be connected (for example, smoking and lung cancer). But we know more than that structure: the more a person smokes the more at risk they are for lung cancer. The qualitative reasoning folks would write this as "SmokingFrequency Q+ LungCancer", indicating a qualitative (perhaps stochastic) monotonic influence. Now, instead of using that statement for qualitative simulation, we use it for putting a prior on our CPTs that enforces stochastic monotonicity, resulting in better classifiers in low-data situations.

Wednesday, February 23, 2005

"Hello World"

My name is Eric Altendorf. I'm currently studying with Tom Dietterich at Oregon State University (I apologize if that link takes you to a large photograph of cows with glasses; I am absolutely not responsible for the department's graphic design choices). I've been convinced to share some thoughts on machine learning by my friend Yaroslav, though I am quite sure I will have not so much to share as he.