Statistical Science

A Conversation with Jerry Friedman

N. I. Fisher

Full-text: Open access

Abstract

Jerome H. Friedman was born in Yreka, California, USA, on December 29, 1939. He received his high school education at Yreka High School, then spent two years at Chico State College before transferring to the University of California at Berkeley in 1959. He completed an undergraduate degree in physics in 1962 and a Ph.D. in high-energy particle physics in 1968 and was a post-doctoral research physicist at the Lawrence Berkeley Laboratory during 1968–1972. In 1972, he moved to Stanford Linear Accelerator Center (SLAC) as head of the Computation Research Group, retaining this position until 2006. In 1981, he was appointed half time as Professor in the Department of Statistics, Stanford University, remaining half time with his SLAC appointment. He has held visiting appointments at CSIRO in Sydney, CERN and the Department of Statistics at Berkeley, and has had a very active career as a commercial consultant. Jerry became Professor Emeritus in the Department of Statistics in 2007. Apart from some 30 publications in high-energy physics early in his career, Jerry has published over 70 research articles and books in statistics and computer science, including co-authoring the pioneering books Classification and Regression Trees and The Elements of Statistical Learning. Many of his publications have hundreds if not thousands of citations (e.g., the CART book has over 21,000). Much of his software is incorporated in commercial products, including at least one popular search engine. Many of his methods and algorithms are essential inclusions in modern statistical and data mining packages. Honors include the following: the Rietz Lecture (1999) and the Wald Lectures (2009); election to the American Academy of Arts and Sciences (2005) and the US National Academy of Sciences (2010); a Fellow of the American Statistical Association; Paper of the Year ( JASA 1980, 1985; Technometrics 1998, 1992); Statistician of the Year (ASA, Chicago Chapter, 1999); ACM Data Mining Lifetime Innovation Award (2002), Emanuel & Carol Parzen Award for Statistical Innovation (2004); Noether Senior Lecturer (American Statistical Association, 2010); and the IEEE Computer Society Data Mining Research Contribution Award (2012).

The interview was recorded at his home in Palo Alto, California during 3–4 August 2012.

Article information

Source
Statist. Sci. Volume 30, Number 2 (2015), 268-295.

Dates
First available in Project Euclid: 3 June 2015

Permanent link to this document
https://projecteuclid.org/euclid.ss/1433341483

Digital Object Identifier
doi:10.1214/14-STS509

Mathematical Reviews number (MathSciNet)
MR3353108

Zentralblatt MATH identifier
1332.01045

Keywords
ACE boosting CART machine learning MARS MART projection pursuit RuleFit statistical computing statistical graphics statistical learning

Citation

Fisher, N. I. A Conversation with Jerry Friedman. Statist. Sci. 30 (2015), no. 2, 268--295. doi:10.1214/14-STS509. https://projecteuclid.org/euclid.ss/1433341483


Export citation

References

  • Breiman, L. (1996). Arcing classifiers. Technical Report 460, Univ. California, Berkeley.
  • Breiman, L. and Friedman, J. H. (1985). Estimating optimal transformations for multiple regression and correlation. J. Amer. Statist. Assoc. 80 580–619.
  • Breiman, L. and Friedman, J. H. (1997). Predicting multivariate responses in multiple linear regression. J. R. Stat. Soc. Ser. B Stat. Methodol. 59 3–54.
  • Breiman, L., Friedman, J. H., Olshen, R. A. and Stone, C. J. (1984). Classification and Regression Trees. Wadsworth, Belmont, CA.
  • Brillinger, D. R. (2002). John W. Tukey: His life and professional contributions. Ann. Statist. 30 1535–1575. In memory of John W. Tukey.
  • Cover, T. M. and Hart, P. E. (1967). Nearest neighbor pattern classification. IEEE Trans. Inform. Theory IT-13 2–27.
  • de Boor, C. (2001). A Practical Guide to Splines, Revised ed. Applied Mathematical Sciences 27. Springer, New York.
  • Efron, B., Hastie, T., Johnstone, I. and Tibshirani, R. (2004). Least angle regression. Ann. Statist. 32 407–499.
  • Fisher, N. I. (2015). Supplement to “A conversation with Jerry Friedman.” DOI:10.1214/14-STS509SUPP.
  • Fithian, W. and Hastie, T. (2013). Finite-sample equivalence in statistical models for presence-only data. Ann. Appl. Stat. 7 1917–1939.
  • Frank, I. E. and Friedman, J. H. (1993). A statistical view of some chemometrics regression tools. Technometrics 35 109–148.
  • Freund, Y. and Shapire, R. E. (1996). Experiments with a new boosting algorithm. In Machine Learning: Proceedings of the Thirteenth International Conference 148–156. Morgan Kaufmann, San Francisco, CA.
  • Friedman, J. H. (1984). A variable span smoother. Technical Report 5, Laboratory for Computational Statistics, Stanford Univ., Stanford, CA.
  • Friedman, J. H. (1987). Exploratory projection pursuit. J. Amer. Statist. Assoc. 82 249–266.
  • Friedman, J. H. (1989a). Regularized discriminant analysis. J. Amer. Statist. Assoc. 84 165–175.
  • Friedman, J. H. (1989b). Modern statistics and the computer revolution. In Symposium on Statistics in Science, Industry, and Public Policy, Part 3 14–29. National Academies Press, Washington, DC.
  • Friedman, J. H. (1991). Multivariate adaptive regression splines. Ann. Statist. 19 1–141.
  • Friedman, J. H. (1997). On bias, variance, 0/1-loss, and the curse-of-dimensionality. Data Min. Knowl. Discov. 1 55–77.
  • Friedman, J. H. (2001a). Greedy function approximation: A gradient boosting machine. Ann. Statist. 29 1189–1232.
  • Friedman, J. H. (2001b). The role of statistics in the data revolution? Int. Stat. Rev. 69 5–10.
  • Friedman, J. H., Bentley, J. L. and Finkel, R. A. (1977). An algorithm for finding best matches in logarithmic time. ACM Trans. Math. Software 3 209–226.
  • Friedman, J. H. and Fisher, N. I. (1999). Bump hunting in high-dimensional data. Stat. Comput. 9 123–162.
  • Friedman, J. H., Grosse, E. and Stuetzle, W. (1983). Multidimensional additive spline approximation. SIAM J. Sci. Statist. Comput. 4 291–301.
  • Friedman, J., Hastie, T. and Tibshirani, R. (2000). Additive logistic regression: A statistical view of boosting. Ann. Statist. 28 337–407.
  • Friedman, J. H. and Rafsky, L. C. (1979). Multivariate generalizations of the Wald–Wolfowitz and Smirnov two-sample tests. Ann. Statist. 7 697–717.
  • Friedman, J. H. and Rafsky, L. C. (1983). Graph-theoretic measures of multivariate association and prediction. Ann. Statist. 11 377–391.
  • Friedman, J. H. and Stuetzle, W. (1981). Projection pursuit regression. J. Amer. Statist. Assoc. 76 817–823.
  • Friedman, J. H. and Stuetzle, W. (2002). John W. Tukey’s work on interactive graphics. Ann. Statist. 30 1629–1639. In memory of John W. Tukey.
  • Friedman, J. H., Stuetzle, W. and Schroeder, A. (1984). Projection pursuit density estimation. J. Amer. Statist. Assoc. 79 599–608.
  • Friedman, J. H. and Tukey, J. W. (1974). A projection pursuit algorithm for exploratory data analysis. IEEE Trans. Comput. C-23 881–889.
  • Hastie, T., Tibshirani, R. and Friedman, J. (2001). The Elements of Statistical Learning. Data Mining, Inference, and Prediction. Springer, New York.
  • Kipling, R. (1886). Part of the second last couplet of “The Betrothed.” First published in Departmental Ditties. Available at http://en.wikipedia.org/wiki/The_Betrothed_%28Kipling_poem%29.
  • Morgan, J. N. and Sonquist, J. A. (1963). Problems in the analysis of survey data, and a proposal. J. Amer. Statist. Assoc. 58 415–435.
  • Orear, J. (1982). Notes on statistics for physicists, revised. Available at http://ned.ipac.caltech.edu/level5/Sept01/Orear/frames.html.
  • Quinlan, J. R. (1986). Induction of decision trees. Machine Learning 1 81–106. Reprinted in Readings in Machine Learning (J. W. Shavlik and T. G. Dietterich, eds.). Morgan Kaufmann, San Francisco, 1990, and also, in Readings in Knowledge Acquisition and Learning (B. G. Buchanan and D. Wilkins, eds.). Morgan Kaufmann, San Francisco, 1993.
  • Tukey, J. W. (1962). The future of data analysis. Ann. Math. Statist. 33 1–67.
  • Varian, H. (2009). Hal Varian on how the Web challenges managers. Available at http://www.mckinsey.com/insights/innovation/hal_varian_on_how_the_web_challenges_managers.
  • Wold, S. and Sjostrom, M. (1977). SIMCA: A method for analyzing chemical data in terms of similarity and analogy. In Chemometrics Theory and Application (B. R. Kowalski, ed.). American Chemical Society Symposium Series 52 243–282. American Chemical Society, Washington, D.C.
  • Zou, H. and Hastie, T. (2005). Regularization and variable selection via the elastic net. J. R. Stat. Soc. Ser. B Stat. Methodol. 67 301–320.

Supplemental materials

  • Supplement to “A conversation with Jerry Friedman”. The supplementary materials associated with this article comprise a number of anecdotes, plus an example of one way in which John Tukey communicated his research ideas to Jerry in the course of their collaboration. They are available from Fisher (2015).