I'm not sure how best to get the message out to non-statisticians--or to many statisticians for that matter--but regarding AI, we should essentially claim credit. Perhaps it's worth recounting the broad arc of AI history. There were computers, and then computer scientists developed ways to easily encode complex deterministic logic algorithms and there was great hope for AI, and then it fizzled. Research funding dried up in the so-called "AI winter". In recent years we've had a remarkable thaw. Why? The application of models that comprehend that observed data comes from noisy processes. That's statistical in nature, even if the people developing the models don't call themselves statisticians.
To your original question, i.e., statistical challenges, one general topic could be "What types of AI problems cannot be addressed by probabilistic modeling methods, if any?"
Another statistical challenge is assessing the performance of AI methods on a theoretical basis, and if possible explaining performance, is of particular concern to statisticians. What practices tend to work better than others for certain classes of problems? What are those classes of problems? I can think of two examples where statistical methodological research has yielded benefits. First, it's my understanding that Dr. Jerome Friedman took it upon himself to figure out why "boosting" works. He found that it optimizes a criterion similar to likelihood. Why not use likelihood itself? Thus was gradient boosting born.
Second, I recall seeing an article (I don't have references in front of me at the moment) proving that for every neural net that is fitted without weight decay, there exists a neural net fitted with non-zero weight decay that dominates it. Further, weight decay can be seen as a roughness penalty, and also as approximately Bayesian. In these examples, statistical (decision-theoretic) research applied to computer science ideas yielded both improved performance and increased knowledge about why or how these models work.
------------------------------
Jim Garrett, PhD
Sr. Assoc. Dir. of Biostatistics
Novartis
Original Message:
Sent: 07-11-2016 11:01
From: Amy Nussbaum
Subject: Statistical Challenges of AI
Hi all!
Some of you may be aware of new White House efforts to reach out to the scientific community regarding Artificial Intelligence. The Office of Science and Technology Policy has formed a series of workshops exploring opportunities brought forth by AI , and the National Science and Technology Council has formed a Subcommittee on Machine Learning and Artificial Intelligence, which will continue to monitor new developments and advances in the field.
The ASA wants to make sure the important role of statistics in AI is recognized. Specifically, we would like to know what important statistical challenges need to be addressed. We would also appreciate suggestions for framing such a broad question. Please respond below, or send me your ideas at nussbaum@amstat.org. Thanks in advance!
------------------------------
Amy Nussbaum
Science Policy Fellow
American Statistical Association
------------------------------