Limitations of ML
Limitations of ML
Limitations of ML
Machine learning is now seen as a silver bullet for solving all problems, but sometimes it is not
the answer.
Limitation 1 — Ethics
Machine learning, a subset of artificial intelligence, has evolutionalized the world as we know it
in the past decade. The information explosion has resulted in the collection of massive amounts of
data, especially by large companies such as Facebook and Google. This amount of data, coupled
with the rapid development of processor power and computer parallelization, has now made it
possible to obtain and study huge amounts of data with relative ease.
It is easy to understand why machine learning has had such a profound impact on the world, what
is less clear is exactly what its capabilities are, and perhaps more importantly, what its limitations
are. Yuval Noah Harari famously coined the term ‘dataism’, which refers to a putative new stage
of civilization we are entering in which we trust algorithms and data more than our own judgment
and logic.
Whilst you may find this idea laughable, remember the last time you went on vacation and
followed the instructions of a GPS rather than your own judgment on a map — do you question
the judgment of the GPS? People have literally driven into lakes because they blindly followed
The idea of trusting data and algorithms more than our own judgment has its pros and cons.
Obviously, we benefit from these algorithms, otherwise, we wouldn’t be using them in the first
place. These algorithms allow us to automate processes by making informed judgments using
available data. Sometimes, however, this means replacing someone’s job with an algorithm,
which comes with ethical ramifications. Additionally, who do we blame if something goes
wrong?
The most commonly discussed case currently is self-driving cars — how do we choose how the
vehicle should react in the event of a fatal collision? In the future will we have to select which
ethical framework we want our self-driving car to follow when we are purchasing the vehicle?
Whilst these are all fascinating questions, they are not the main purpose of this article. Clearly,
however, machine learning cannot tell us anything about what normative values we should accept,
i.e. how we should act in the world in a given situation. As David Hume famously said, one
This is a limitation I personally have had to deal with. My field of expertise is environmental
science, which relies heavily on computational modeling and using sensors/IoT devices.
Machine learning is incredibly powerful for sensors and can be used to help calibrate and correct
sensors when connected to other sensors measuring environmental variables such as temperature,
pressure, and humidity. The correlations between the signals from these sensors can be used to
develop self-calibration procedures and this is a hot research topic in my research field of
atmospheric chemistry.
However, things get a bit more interesting when it comes to computational modeling.
Running computer models that simulate global weather, emissions from the planet, and transport
that a research-level simulation can take weeks even when running on a supercomputer.
Good examples of this are MM5 and WRF, which are numerical weather prediction models that
are used for climate research and for giving you weather forecasts on the morning news. Wonder
what weather forecasters do all day? Run and study these models.
Running weather models is fine, but now that we have machine learning, can we just use this
instead to obtain our weather forecasts? Can we leverage data from satellites, weather stations,
and use an elementary predictive algorithm to discern whether it is going to rain tomorrow?
The answer is, surprisingly, yes. If we have knowledge of the air pressures around a certain
region, the levels of moisture in the air, wind speeds, and information about neighboring points
and their own variables, it becomes possible to train, for example, a neural network. But at what
cost?
Using a neural network with a thousand inputs to determine whether it will rain tomorrow in
Boston is possible. However, utilizing a neural network misses the entire physics of the weather
system.
A neural network does not understand Newton’s second law, or that density cannot be
physical constraints to neural networks and other algorithms so that they can be used for purposes
such as this.
Limitation 3 — Data
This is the most obvious limitation. If you feed a model poorly, then it will only give you poor
results. This can manifest itself in two ways: lack of data, and lack of good data.
Lack of Data
Many machine learning algorithms require large amounts of data before they begin to give useful
results. A good example of this is a neural network. Neural networks are data-eating machines
that require copious amounts of training data. The larger the architecture, the more data is needed
to produce viable results. Reusing data is a bad idea, and data augmentation is useful to some
Despite the appearance, this is not the same as the above comment. Let’s imagine you think you
can cheat by generating ten thousand fake data points to put in your neural network. What
It will train itself, and then when you come to test it on an unseen data set, it will not perform
well. You had the data but the quality of the data was not up to scratch.
In the same way that having a lack of good features can cause your algorithm to perform poorly,
having a lack of good ground truth data can also limit the capabilities of your model. No company
is going to implement a machine learning model that performs worse than human-level error.
Similarly, applying a model that was trained on a set of data in one situation may not necessarily
apply as well to a second situation. The best example of this I have found so far is in breast cancer
prediction.
Mammography databases have a lot of images in them, but they suffer from one problem that has
caused significant issues in recent years — almost all of the x-rays are from white women. This
may not sound like a big deal, but actually, black women have been shown to be 42 percent more
likely to die from breast cancer due to a wide range of factors that may include differences in
detection and access to health care. Thus, training an algorithm primarily on white women
What is needed in this specific case is a larger number of x-rays of black patients in the training
database, more features relevant to the cause of this 42 percent increased likelihood, and for the
algorithm to be more equitable by stratifying the dataset along the relevant axes.
If you are skeptical of this or would like to know more, I recommend you look at this article.
Limitation 4 — Misapplication
Related to the second limitation discussed previously, there is purported to be a “crisis of machine
learning in academic research” whereby people blindly use machine learning to try and analyze
succeed, but the algorithm which not be learning the relationship between the two variables, and
will not know when it is violating physical laws. We simply gave some inputs and outputs to the
system and told it to learn the relationship — like someone translating word for word out of a
dictionary, the algorithm will only appear to have a facile grasp of the underlying physics.
For stochastic (random) systems, things are a little less obvious. The crisis of machine learning for
P-hacking
P-hacking
When one has access to large data, which may have hundreds, thousands, or even millions of
variables, it is not too difficult to find a statistically significant result (given that the level of
statistical significance needed for most scientific research is p < 0.05). This often leads to
spurious correlations being found that are usually obtained by p-hacking (looking through
mountains of data until a correlation showing statistically significant results is found). These are
not true correlations and are just responding to the noise in the measurements.
This has resulted in individuals ‘fishing’ for statistically significant correlations through large data
sets, and masquerading these as true correlations. Sometimes, this is an innocent mistake (in
which case the scientist should be better trained), but other times, it is done to increase the number
of papers a researcher has published — even in the world of academia, competition is strong and
There are inherent differences in the scope of the analysis for machine learning as compared with
inherently exploratory.
We can consider confirmatory analysis and models to be the kind of thing that someone does in a
Ph.D. program or in a research field. Imagine you are working with an advisor and trying to
develop a theoretical framework to study some real-world system. This system has a set of pre-
defined features that it is influenced by, and, after carefully designing experiments and developing
hypotheses you are able to run tests to determine the validity of your hypotheses.
Exploratory, on the other hand, lacks a number of qualities associated with the confirmatory
analysis. In fact, in the case of truly massive amounts of data and information, the confirmatory
approaches completely break down due to the sheer volume of data. In other words, it simply is
not possible to carefully lay out a finite set of testable hypotheses in the presence of hundreds,
Therefore and, again, broadly speaking, machine learning algorithms and approaches are best
suited for exploratory predictive modeling and classification with massive amounts of data and
computationally complex features. Some will contend that they can be used on “small” data but
why would one do so when classic, multivariate statistical methods are so much more
informative?
ML is a field which, in large part, addresses issues derived from information technology,
computer science, and so on, these can be both theoretical and applied problems. As such, it is
related to fields such as physics, mathematics, probability, and statistics but ML is really a field
unto itself, a field which is unencumbered by the concerns raised in the other disciplines. Many of
the solutions ML experts and practitioners come up with are painfully mistaken…but they get the
job done.
Limitation 5 — Interpretability
Interpretability is one of the primary problems with machine learning. An AI consultancy firm
trying to pitch to a firm that only uses traditional statistical methods can be stopped dead if they
do not see the model as interpretable. If you cannot convince your client that you understand how
the algorithm came to the decision it did, how likely are they to trust you and your expertise?
“A business manager is more likely to accept the [machine learning method] recommendations if
These models as such can be rendered powerless unless they can be interpreted, and the process of
human interpretation follows rules that go well beyond technical prowess. For this reason,
interpretability is a paramount quality that machine learning methods should aim to achieve if
they are to be applied in practice.
The blossoming -omics sciences (genomics, proteomics, metabolomics and the like), in particular,
have become the main target for machine learning researchers precisely because of their
dependence on large and non-trivial databases. However, they suffer from the lack of
the emergence of a mindset that can be best described as “AI solutionism”. This is the philosophy
that, given enough data, machine learning algorithms can solve all of humanity’s problems.
As I hope I have made clear in this article, there are limitations that, at least for the time being,
prevent that from being the case. A neural network can never tell us how to be a good person, and,
at least for now, do not understand Newton’s laws of motion or Einstein’s theory of relativity.
There are also fundamental limitations grounded in the underlying theory of machine learning,
called computational learning theory, which are primarily statistical limitations. We have also
discussed issues associated with the scope of the analysis and the dangers of p-hacking, which can
lead to spurious conclusions. There are also issues with the interpretability of results, which can
negatively impact businesses that are unable to convince clients and investors that their methods
Whilst in this article I have covered very broadly some of the most important limitations of AI, to
finish, I will outline a list published in an article by Peter Voss in October 2016, outlining a more
comprehensive list on the limitations of AI. Whilst current mainstream techniques can be very
powerful in narrow domains, they will typically have some or all of a list of constraints that he
All that being said, machine learning and artificial intelligence will continue to revolutionize
industry and will only become more prevalent in the coming years. Whilst I recommend you
utilize machine learning and AI to their fullest extent, I also recommend that you remember the
Source : https://towardsdatascience.com/the-limitations-of-machine-learning-a00e0c3040c6