Just as science-fiction authors have always predicted, Artificial Intelligence (AI) is increasingly becoming an everyday part of our lives. From personal assistants such as Siri or Cortana to cutting-edge applications in healthcare and across industry, self-learning machines have arrived and are busy helping us find new ways to solve problems.
As the predictions and dreams of yesterday’s futurologists have solidified into today’s tools and technology, a sometimes confusing lexicon has sprung up around the subject. As these new words and phrases are often attempts to outline the fundamental thinking behind this ongoing robotic revolution, understanding them is key to getting to grips with what AIs actually are, where they’ve come from, what they want, and most importantly, how you can put them to work yourself!
So, here’s my brief guide to some of the most common and some of the latest terminology being used when discussing cutting-edge AI, in alphabetical order.
AlphaGo is an AI which became the first computer program to beat a professional player at the board game Go. Game playing has often been a field in which computer scientists have sought to prove that machines can outperform humans. However earlier applications such as chess computers are not considered “true” AI today because they don’t really learn – they simply rely on brute force to consider every permutation of a structured dataset (all of the moves possible in a game of chess.) AlphaGo uses deep learning to refine its algorithms based on the results of historical games, and from running simulated games against itself. This means it can be considered to be learning and comes closer to what we consider “true” (human-like) intelligence.
This is the original, catch-all term for “machines which can think”, first conceived by philosophers and storytellers in ancient times. Technological advancement has brought them closer to reality and also redefined what we consider “intelligence” when it comes to machines. Rather than walking, talking automatons, today’s AI’s are more likely to take the form of discreet computer code dedicated to handing a particular task in an intelligent way.
The “fuel” of AI. Knowledge unlocks understanding and wisdom. AI platforms leverage the huge volume, variety and velocity of information available in today’s digitized world to learn faster and make increasingly well-informed recommendations and decisions.
Cognitive computing is the process by which computers think and learn, as well as the development of these processes. These give rise to artificial intelligence, machine learning, deep learning and all technologies which involve simulating human thought and decision-making. In practice, the term is often used synonymously with modern, application-focused AI.
This is a subfield of machine learning (see below) which uses many layers of artificial neural networks to handle processing of data in increasingly complex ways. This means that classification (sorting into sets) can be done more precisely and pattern recognition is more sophisticated. These are two of the most useful fundamental tasks that AI carries out today, meaning Deep Learning is a cutting-edge and very active field of research. Layers of neural networks stacked on top of each other to be used in deep learning are known as deep neural networks.
Generalized AI is a concept – widely thought to still be some way off – of a machine which can carry out any job it is told to do. An android such as those seen in Star Trek or Blade Runner – who could be given a mop and told to clean a floor, or given a weapon and told to defend against attacking Klingons, would be an archetypal example. While advances such as machine learning and deep neural networks point towards it being something that we will achieve in the future, currently the majority of AI research focuses on creating applied or specialized AIs (see below).
Teaching machines to recognise and classify objects visually – by inputting visual data – is an important foundation of AI because visual information is so valuable to humans, and AI seeks to emulate human thought processes. Either using cameras or raw image data such as picture or video files, computers are being taught to classify images according to what they depict, using pattern recognition to identify key features. Advances in machine learning have greatly improved the ability of computers to do this, as they have become able to teach themselves from vast image databases, increasing their probability of outputting accurate results.
Often used synonymously with AI these days, but there is an important distinction. While AI applies to the entire concept of “thinking” machines from sci-fi robots to self-learning computer code being developed by business and academia today, Machine Learning (ML) is the practical implementation that is generating the biggest breakthroughs in the real world. At its most basic it is technology designed around the principle that rather than have to teach machines to carry out every task, we should just be able to feed them data and allow them to work out the rules by themselves. This is done through a process of simulated trial-and-error where machines crunch datasets through algorithms which are capable of adapting, based on what they learn from the data, in order to more efficiently process subsequent data.
Natural Language Processing
Natural Language Processing (NLP) technology is concerned with building machines which can understand human speech patterns. Because spoken communication comes far more naturally to us than writing computer code, it makes sense that machines, with their superior processing powers, learn to adapt to us by understanding and speaking our language, rather than us adapt to them! Due to the huge variance in human languages and the way they are used, machine learning is employed to pick out patterns, tonal variances and colloquial or non-literal use of language and interpret what we are trying to express. ML-derived NLP can be seen or heard in action in virtual assistants such as Apple’s Siri, Microsoft’s Cortana and Amazon’s Alexa.
Algorithmic models structured as hierarchical networks of nodes which all pass information (data) between themselves, extrapolating more and more precise meaning and value from it as it passes along the chain. Their complex, interconnected nature allows data to be processed far more comprehensively than traditional, linear algorithms allow, enabling them more insightful output from big, messy and unstructured datasets.
The more precise and correct term, artificial neural networks (ANNs), is often simply shortened to “neural network”, the term for the system of biological neurons in the animal brain which machine learning attempts to emulate.
The form of AI becoming commonplace in business, scientific research and our everyday lives – usually in the form of applications designed to carry out one specific task in an increasingly efficient way. This could be anything from giving you tips on improving your fitness by monitoring exercise patterns, to predicting when machinery will break down on a production line, to spotting genetic indicators of illness in a human gene sequence.
Supervised learning is a term used for machine learning processes where the output of the algorithm is checked, and the results fed back to the computer to enable it to know how accurate they are. It can then use this knowledge to increase the probability that it will return with an acceptably accurate result next time around. As a simple example imagine an AI fraud detection algorithm designed to flag up suspicious transactions by a bank. In unsupervised learning, data is matched against previous outcomes to look for patterns in financial transactions, such as their point of origin, size or time of day they take place, which may indicate they are suspicious. As new suspicious transactions are identified, the algorithm adapts to “learn” that other features of the newly-identified suspicious transactions may also be an indicator of fraud. In this way, a supervised learning system can learn to identify fraud from characteristics that were not highlighted in its initial training data as indicators of fraud.
Unsupervised learning is the flip-side of the coin from supervised learning and involves giving computers the ability to increasingly accurately recognise and classify data without needing a human, or initial training data, to check it if it is right or wrong. In unsupervised learning the algorithm only ever sees the input data, and it classifies it according to patterns that it recognises from other input data that it has previously processed. This is generally done through a statistical process known as clustering, where objects (financial transactions to carry on my example from above) are grouped together according to qualities and attributes that they share. This approach to the problem of data classification has tremendous potential for developing machines which more closely emulate our own thought and decision-making processes, but also requires huge amounts of processing power compared to supervised learning.
Bernard Marr is an internationally bestselling author, futurist, keynote speaker, and strategic advisor to companies and governments. He advises and coaches many of the world’s best-known organisations on strategy, digital transformation and business performance. LinkedIn has recently ranked Bernard as one of the top 5 business influencers in the world and the No 1 influencer in the UK. He has authored 16 best-selling books, is a frequent contributor to the World Economic Forum and writes a regular column for Forbes. Every day Bernard actively engages his almost 2 million social media followers and shares content that reaches millions of readers.