Professional Documents
Culture Documents
Pat Langley
langley@rtna.daimlerchrysler.com
Adaptive Systems Group, DaimlerChrysler Research and Technology Center, 1510 Page Mill Road, Palo Alto,
California 94304 USA
Abstract
1. Introduction
Although machine learning has become a scientic discipline, the eective communication of its ideas remains an art. Nevertheless, there exist rules of thumb
even for practicing art, and in this essay we present
some heuristics that we maintain can help machine
learning authors improve their papers. Much of this
advice applies equally well to other branches of articial intelligence and even to scientic elds in general,
but we will cast it in terms specic to our discipline.
Each section addresses a dierent facet of publications
on machine learning. We rst address the content appropriate for papers, considering brie
y the topics that
should appear in any scholarly work. After this, we
discuss issues of evaluation at greater length, as they
have come to play a central role in papers on machine
learning. In closing, we give advice about matters of
communication, ranging from high-level organization
to the individual words used in the text. We hope
that, taken together, these suggestions help authors
to convey their ideas eectively to their audience.
1
0.8
0.6
0.4
Naive Bayes
Nearest neighbor
Budds classifier
0.2
0
0
0.2
0.4
0.6
0.8
1
False positive rate
ROC curves for building detection in nadir images when trained and tested on dierent locations.
Figure 1.
1 of 1 concept
3 irrelevants
3 of 3 concept
5 irrelevants
5 of 5 concept
0.5
0.5
0.6
1 irrelevant
(b)
0.6
(a)
10
15
20
25
30
35
40
Number of training instances
10
15
20
25
30
35
40
Number of training instances
Theoretical and experimental learning curves, with 95% condence intervals, for naive Bayes when (a) the
domain involves a `2 of 2' target concept and varying numbers of irrelevant attributes, and (b) for a domain with one
irrelevant attribute and a conjunctive concept with varying numbers of relevant features.
Figure 2.
4. Issues of Communication
The organization of a paper into sections and paragraphs helps readers place a structure on the material
and understand its contributions. Thus, you should
put some eort into designing a good organization.
For instance, your paper will benet from informative
section and subsection headings, rather than generic
ones like `Representation' or `Evaluation', as they let
the reader browse through the paper and still have
some idea what it is about. For the same reason, never
use pronouns (such as `it') in headings, and do not
treat a section heading as if it were part of the text.
Include introductory sentences at the beginnings of
sections and subsections to help readers make the transition. Make your sections roughly the same length,
except possibly for the introduction and conclusion.
Be consistent about whether you include an introductory paragraph before the rst subsection. Also, never
include only one subsection in a section, since subsections are designed to divide a section into components.
For the same reasons, avoid subsections that contain
only one paragraph; if you have only one paragraph's
worth of material, embed it in another subsection.
Within each section or subsection, you should further
partition the paper into paragraphs, each of which
should discuss a distinct idea and
ow naturally from
its predecessor. The ideal paragraph will run no more
than six sentences and no fewer than three sentences.
Neither should the sentences themselves say too much
or too little; rather, they should convey ideas in bites
the reader can digest.
On occasion, you may want to use footnotes1 to provide readers with additional information about a topic
without interrupting the
ow of the paper. For the
sake of readability, footnotes should take the form of
complete sentences.
Successful communication begins at the level of individual words, so the choices you make here also in
uence the readability of your paper. Remember that
precision is not enough; your audience must be able to
recall what they have read.
One important step is avoid abbreviations, especially
if you invoke them only a few times. Even if you use an
5. Concluding Remarks
Acknowledgements
References
Dietterich, T. G., (1990). Exploratory research in machine learning. Machine Learning , 5 , 5{10.
John, G. H., & Langley, P. (1995). Estimating continuous distributions in Bayesian networks. Proceedings of the Eleventh Conference on Uncertainty in
Articial Intelligence (pp. 338{345). San Francisco:
Morgan Kaufmann.
Maloof, M. A., Langley, P., Binford, T. O., & Nevatia,
R. (1998). Generalizing over aspect and location for
rooftop detection. Proceedings of the Fourth IEEE
Workshop on Applications of Computer Vision (pp.
194{199). Princeton, NJ: IEEE Press.
Nordhausen, B., & Langley, P. (1993). An integrated
framework for empirical discovery. Machine Learning, 12, 17-47.
Langley, P. (1986). Human and machine learning. Machine Learning , 1 , 243{248.
Langley, P. (1990). Advice to authors of machine
learning papers. Machine Learning , 5 , 233{237.
Langley, P. (October, 1996). Relevance and insight in
experimental studies. IEEE Expert , 11{12.
Langley, P., & Sage, S. (1999). Tractable average-case
analysis of naive Bayesian classiers. Proceedings
of the Sixteenth International Conference on Machine Learning (pp. 220{228). San Francisco: Morgan Kaufmann.
Provost, F., Fawcett, T., & Kohavi, R. (1998). The
case against accuracy estimation for comparing induction algorithms. Proceedings of the Fifteenth International Conference on Machine Learning (pp.
445{453). San Francisco: Morgan Kaufmann.
Provost, F., & Kohavi, R. (1998). On applied research
in machine learning. Machine Learning , 30, 127{
132.