ASA Connect

 View Only
  • 1.  What does AI mean for statisticians?

    Posted 07-20-2016 17:40

    Recently, ASA Science Policy Fellow Amy Nussbaum asked about the statistical challenges posed by AI.

    In the public's mind, 'AI' often has the flash and shallowness of movie robots. But what, in your opinion (and experience!), does AI mean for statisticians and statistical work?

    ------------------------------
    Lara Harmon
    Marketing and Online Community Coordinator
    American Statistical Association
    ------------------------------


  • 2.  RE: What does AI mean for statisticians?

    Posted 07-21-2016 18:04

    Artificial Intelligence, or AI, has potentially broad-based implications for the statisticians, from both practical and theoretical perspectives. Consider some of the intrinsically statistical algorithmic applications e.g. boosting and the like; which provide 'highly accurate' predictors in machine learning applications.  The aforementioned boosting techniques - and many other algorithmic applications - having foundations in statistics and subsequently subject to emerging statistical processes, may provide otherwise abstract approaches for statisticians to explore in advancing applications in AI.  Just a thought.

    ------------------------------
    AJ Perry
    Graduate Candidate



  • 3.  RE: What does AI mean for statisticians?

    Posted 07-21-2016 19:48

    The most obvious impact AI can have for statistics is to push up the starting salaries of statisticians. I think optimistically that computing related developments help us if we are proactive in embracing such advancements in our program designs at the statistics departments. By creating statisticians who also are "mini-AI scientists" and "mini-Data Scientists" at the time of graduation, statistics will remain at the Top of Best Graduate Programs in the USA.  Think positive.

    ------------------------------
    Raid Amin
    Professor
    University of West Florida



  • 4.  RE: What does AI mean for statisticians?

    Posted 07-22-2016 04:11

    Good morning,

    I posted my 2cts in the original track from Amy Nussbaum.

    Best regards,
    Christian

    ------------------------------
    Christian Graf
    Dipl.-Math.
    Qualitaetssicherung & Statistik

    "To call in the statistician after the experiment is done may be no more than asking him to perform a post-mortem examination: he may be able to say what the experiment died of."

    Ronald Fisher in 'Presidential Address by Professor R. A. Fisher, Sc.D., F.R.S. Sankhyā: The Indian Journal of Statistics (1933-1960), Vol. 4, No. 1 (1938), pp. 14-17'



  • 5.  RE: What does AI mean for statisticians?

    Posted 07-25-2016 14:11

    As a chemist, I think AI means the same thing automation means for chemists and other professions: competition and new possibilities. With automated chemical analyses, chemists can perform more tests, perhaps making some chemists redundant. However, at the same time, because of the abundance of knowledge, chemists can push the frontiers of scientific knowledge.

    Machine learning can take a load off of statisticians, but the entrepreneurial statisticians will use this as an opportunity to make discoveries.

    ------------------------------
    Herman Autore
    Chemist, stats grad student



  • 6.  RE: What does AI mean for statisticians?

    Posted 07-26-2016 06:47
    AI opens the door wide to multiple testing false positives.




  • 7.  RE: What does AI mean for statisticians?

    Posted 07-27-2016 06:18
    This is one reasonable view, but there are several others:
    1. Starry eyed Computer Scientists view the holy grail of AI to be the
    construction of computers that "work like the human mind". While they
    may not be particularly close yet, they have their share of rather
    amazing successes (a poster child being self-driving cars). It seems
    likely this will continue over the long term in the future, and there
    are folks who definitely have their eye on that goal.
    2. If you look at the actual contributions over the years, there is a
    rather long history of big ballyhoos (often very successful at
    attracting gobs of research funding):
    a) AI itself was just such a buzzword at one point. Some of you may
    remember the gross over-advertisement that was done (I think around the
    1970s?) of "computers will put doctors out of business at the task of
    medical diagnosis". After it didn't quite work out as advertised, the
    community learned from the backlash by physicians that threatening
    people is not a great way to get one's methods adopted.
    b) Another major event in that spirit was "neural networks", which
    really was about rather directly attempting to mimic how the brain
    works. There is an important set of problems that have been solved by
    this approach, including the voice recognition software we deal with
    when we phone any major corporation today. Once again this came with
    serious over-advertising (and a massive infusion of research funding).
    Folks again claimed to have methods that could solve all problems, and
    of course that later turned out to be too ambitious, with many attendant
    failures.
    c) Another round of this type of history is "machine learning". This is
    mostly about trying to do statistical tasks, but with much less emphasis
    on probability distributions, and much more on optimization. Again gobs
    of research funding followed the hot new ideas, and again there is a set
    of analytical problems where that set of ideas has proved to be very
    effective.
    d) One more round relevant to this AI discussion is "Deep Learning".
    This is really just a repackaging of the Neural Nets in (b) above, but
    there are two major differences: [1] typical data sets are now far
    larger, and it is appearing that many of the earlier failures came from
    data sets too small to effectively train the methods. [2] Vastly better
    computational resources are now available, to the point that much better
    training of these things is giving much better results. It is worth
    noting that Deep Learning methods have almost completely taken over very
    active research areas such as Computer Vision (an important aspect of
    e.g. self-driving cars).
    3. Now let's consider the issue of multiple false positives. First, from
    a classical statistical viewpoint, I completely agree. However, it could
    be useful to also consider a bigger picture. One relevant dichotomy of
    data analytic tasks is "Causal" versus "Correlational". Here Causal
    refers to the goal of really understanding the driving factors in an
    analysis, and the desire to quantify the strength of the evidence for
    that. The poster child for that is the "Scientific Method" (developed
    over quite some period more than 100 years ago), and the basis of most
    modern statistical inference. From the causal perspective, many of the
    approaches mentioned above have precisely this weakness. Correlational
    is just about getting answers to very specific problems (voice
    recognition, or keeping a self-driving car on task). Note that these
    explicitly do NOT attempt to say anything about underlying driving
    factors, but in fact just give particular (but often very useful)
    practical results.
    4. Where should we go from here? For me both approaches have their
    place. I agree with the concerns expressed, but I also believe we
    should teach more correlation methodologies (e.g. machine learning or
    deep learning related) in our courses, as they are important tools in
    the future of data analysis.
    Best,
    Steve




  • 8.  RE: What does AI mean for statisticians?

    Posted 07-27-2016 14:57

    One of the big 'Dises" against AI and Machine Learning is that it doesn't involve the human element when it makes models, correlations, etc. There seems to be an assumption that humans are flawless and typical statistical methods are great always. Unfortunately, this is not the case. Each method has it's strengths and it's weaknesses.  

    Another issue that comes up is that algorithms like ANN, Random Forests, etc, are really complicated models. Most of us, (statisticians), learned about parsimony. This insane assumption that we can model complex systems with simple methods. These models fall victim to Simpson's Paradox. Sometimes, we (statisticians) are afraid to find things out too. Currently I am working with college data sets. The department heads want simple analysis of the data, like, "How many students passed the class?" or "What is the pass rate of this class?" What good does such simple data tell us? Basically, that on average, XX.xx% of students taking a class pass. If we ask better questions like, which professors pass the highest rate of students and how well do professors prepare students for future classes, we get strikingly different answers. If we add in data about the start time of the classes, you can get even different results form the first 2 models. I can use this type of data to tell a student, "Take Prof A for class101 at time 'T2' or later. If you get a B- or better, take Prof B for class102 between times 'T2' and 'T3'. Otherwise, take Prof C for class102 between times 'T4' and 'T5'. This path gets you the best grades and the most knowledge." I'd like traditional stats methods provide a student with that level of knowledge.

    As someone that has run and tuned logistic regressions and ensembles of Random Forests on small data sets, I'll never use Logistic Regression again.... Especially if I need a good prediction. LR can't compete...unless you start adding interaction terms like those you get in RF. Beyond that, the Odds Ratios are nice for scaring people.... "There is a 20% greater chance of disease when you do...." However, that 20% great chance is meaningless if the original probability of getting the disease is 0.000001%. The RF model will tell you if you do (activity) you probably still won't get the disease. What sounds better, more accurate and less scary if the info comes from your doctor? Would you rather hear, "You have a greater chance of getting a genetically linked disease by doing Z." or " You have a low chance of getting the disease, whether or not you do Z." One sounds like I might need a new medication to calm my nerves. The other sounds like I can rest easy, medication free. I'll take useful and actionable over scaremongering.        

    ------------------------------
    Andrew Ekstrom

    Statistician, Chemist, HPC Abuser;-)



  • 9.  RE: What does AI mean for statisticians?

    Posted 07-28-2016 17:59

    As one of the earlier posts above pointed out, artificial intelligence is an attempt to build computers that work like the human mind.  (A more general definition might be "to work like the human mind, only better").  There are essentially two components to this, deductive reasoning and inductive reasoning.  Statistical inference is the formal arm of inductive reasoning which tries to draw general conclusions from specific sets of examples.  The conclusions drawn typically are then said to be true with a certain measure of probability. Deductive reasoning on the other hand starts with certain propositions which are assumed to be true, then, by a process of logical steps reaches conclusions that are (at least formally) true as well. 

     

    To put this into terms that are a bit less philosophically oriented, inductive reasoning is that portion of the process involved with finding and verifying models, while deductive reasoning involves that part that uses the models to draw further conclusions.  

     

    The early attempts to come up with artificial intelligence systems focused on the deductive reasoning side, modelling human decision making through sets of rules drawn from observed human decision makers that could then be combined to draw inferences.  This lead to the (largely failed) attempt to build "expert systems" that modelled the reasoning of human experts in a domain. A much more successful example of deductive inference AI systems (although almost never characterized as such) are things such as autopilots and anti-skid brakes which take well established models (mostly from physics) and use these to replace or augment humans in the execution of some task.

     

    One of the earliest examples of inductive inference in artificial intelligence would be in the area of hand written character recognition initially developed to read zip codes on envelopes and packages. Although this also was seldom characterized as an example of artificial intelligence, it really does provide an example of teaching machines to do something that previously only a human could accomplish. (That this is still difficult for machines can be seen in the widespread use of CAPTCHA (the weird distorted characters you have to match) to login to certain websites).

     

    Other areas of artificial intelligence in which inductive inference (e.g. statistics) played a key role are automated translation, speech recognition, financial trading systems and automated game playing systems (e.g. TD-Gammon and AlphaGo) in all of which machines are reaching or exceeding human capabilities based on statistical models. 

     

    The next question is: What is the difference between traditional statistics and artificial neural networks, data science, machine learning, big data or whichever current buzzword the computer science crowd has come up with for their brand of inductive inference that would be more closely associated in the public mind with artificial intelligence? In terms of the objectives of the disciplines and basic methodology, essentially nothing.  In terms of research style and emphasis on particular areas of statistical practice though, there are significant differences.

     

    As noted previously, the inductive inference problem can be divided into two components:

    -Model selection/hypothesis formation.

    -Model/hypothesis testing or verification.

     

    In traditional statistics, the first of these steps is something of a red headed step child, while in machine learning (or whichever name is currently popular) exploratory data analysis and non-parametric modelling methods are at the center of the enterprise.   Traditional statistics tends to favor model/hypothesis testing based on in-sample statistics using asymptotic values (normal, F, chi-squared, etc.) where in machine learning, tests are usually based on out-of-sample tests, using hold out sets or various types of data reuse methods (k-fold cross validation or bootstrapping).  The machine learning people are generally more willing to just "try something out" even if there is no theoretical underpinning to the method than the traditional statistics community and to do bake-offs comparing methods on simulated data or large data sets comparing results on out-of-sample data.  Finally, the machine learning crowd has a much higher density of Bayesians than is found in the traditional statistics community.

     

    It should be noted though, that almost every method used in the machine learning world can be traced back to methods originally developed by someone in the traditional statistics community. (Particularly influential examples are Bradley Efron, Jerome Friedman, Trevor Hastie, Robert Tibshirani and Vladimir Vapnik). 

     

    So where does all this leave us? Artificial intelligence attempts to emulate or improve on various types of human reasoning capabilities, one of which is to learn from experience (inductive inference).  To get machines to have these capabilities, one must give them a formalized system to accomplish this, which statistical methodology provides (under whichever name it is hiding). To reach true artificial intelligence though, one must take external human guidance and decision making out of all the steps of a statistical modelling process, from which we are still well removed.

     






  • 10.  RE: What does AI mean for statisticians?

    Posted 07-29-2016 10:54
    The statement "There seems to be an assumption that humans are flawless and typical statistical methods are great always." Seems to imply that modeling is a black and white situation. Humans, of course, are flawed; but, so are automated modeling methods. In fact, one of the human flaws, is blind acceptance of automated models without understanding the underlying structure of the model. Too many bad models have come out of automated systems to use them to make crucial business decisions without understanding the drivers. Blind acceptance of a complex model, just because it has a miniscule improvement over a much simpler model could be disastrous. A complex model could have a trivial flaw (due to some complex interaction) that results in a multimillion dollar decision which would overwhelm the small profits made on other decisions.

    In my experience, mostly financial risk and marketing models, complex algorithms like ANN, RF, etc. are an improvement over simpler logistic or other regression models; however, the vast majority of the time the difference is trivial. And, in most of these modeling situations, there are needs to see what is driving the model. For example, regulatory requirements such as reason codes or other compliance issues (to avoid discrimination or other legal constraints). In addition, business decision makers should want to understand what is driving the model and whether these drivers make sense or are they just some temporary correlate that is a function of the time frame (or other design flaw) from which the data was pulled.

    A good example of this is house pricing models based only on the late 90's to 2008 where the price of housing never had a significant decrease. These models would indicate that prices will always increase. A complex blind algorithm would have time as a positive driver of price. This is an example where there was a need to understand what was driving the model, and whether or not it made logical sense.

    If there were a way for the modeling algorithm to understand and identify the logical external relationships between the dependent and independent variables, including directional information,  above and beyond the correlations, then  maybe, just maybe, those algorithms could be trusted to create models without logical flaws. But, it is hard for me to imagine an algorithm that, just out of the blue, would say: 

    "Wait a minute, housing prices can't always increase over time. I mean that's what the data tells me, but that just doesn't make sense."

    So, until we can create an AI that includes Subject Matter Expertise it will always need some level of intervention.


    ​​
    Michael L. Mout, MS, Cstat, Csci
    MIKS & Assoc. - Senior Consultant/Owner
    4957 Gray Goose Ln, Ladson, SC 29456
    804-314-5147(Mbl), 843-871-3039 (Home)