Artificial intelligence has emerged as the defining technology of our era, as transformative over time as the steam engine, electricity, computers, and the Internet. AI technologies are approaching or surpassing human levels of performance in vision, speech recognition, language translation, and other human domains. Machine learning (ML) advances, like deep learning, have played a central role in AI’s recent achievements, giving computers the ability to be trained by ingesting and analyzing large amounts of data instead of being explicitly programmed.
Deep learning is a powerful statistical technique for classifying patterns using large training data sets and multi-layer AI neural networks. Each artificial neural unit is connected to many other such units, and the links can be statistically strengthened or decreased based on the data used to train the system. But such statistical methods are not equally suitable for all tasks. Tasks that are particularly suitable for machine learning, exhibit several key criteria, such as the availability of large data sets of well-defined input-output pairs for training ML classifiers, - e.g., carefully labeled cat, not-cat pictures for cat recognition classifiers, and english-french document pairs for machine translation algorithms.
The methods behind a machine learning prediction, - subtle adjustments to the numerical weights that interconnect its huge number of artificial neurons, - are very difficult to explain because they’re so different from the methods used by humans. The bigger the training data set, the more accurate the prediction, but the more difficult it will be to provide a detailed, understandable explanation to a human of how the prediction was made.
A few weeks ago I attended an online seminar, - How Can You Trust Machine Learning? by Stanford professor Carlos Guestrin, - on the difficulty of understanding machine learning predictions. Guestrin’s seminar was based on a 2016 article he co-authored with Marco Tulio Ribeiro and Sameer Singh , “Why Should I Trust You?” Explaining the Predictions of Any Classifier.
Continue reading "How Can You Trust the Predictions of a Large Machine Learning Model?" »