Key Takeaway

AI black boxes refer to invisible systems with internal workings, such as machine learning. These systems have algorithms, training data, and models, which can be hidden or obscured to protect intellectual property. Explainable AI is working to develop algorithms that can be better understood by humans. Black box machine-learning algorithms and models can be dangerous, as they can reveal sensitive information about decisions and security, while allowing for reverse engineering.


For some people, the term โ€œblack boxโ€ brings to mind the recording devices in airplanes that are valuable for postmortem analyses if the unthinkable happens. For others it evokes small, minimally outfitted theaters. But black box is also an important term in the world of artificial intelligence.ย 

AI black boxes refer to AI systems with internal workings that are invisible to the user. You can feed them input and get output, but you cannot examine the systemโ€™s code or the logic that produced the output. 

Machine learning is the dominant subset of artificial intelligence. It underlies generative AI systems like ChatGPT and DALL-E 2. There are three components to machine learning: an algorithm or a set of algorithms, training data and a model. An algorithm is a set of procedures. In machine learning, an algorithm learns to identify patterns after being trained on a large set of examples โ€“ the training data. Once a machine-learning algorithm has been trained, the result is a machine-learning model. The model is what people use. 

For example, a machine-learning algorithm could be designed to identify patterns in images, and training data could be images of dogs. The resulting machine-learning model would be a dog spotter. You would feed it an image as input and get as output whether and where in the image a set of pixels represents a dog.

Any of the three components of a machine-learning system can be hidden, or in a black box. As is often the case, the algorithm is publicly known, which makes putting it in a black box less effective. So to protect their intellectual property, AI developers often put the model in a black box. Another approach software developers take is to obscure the data used to train the model โ€“ in other words, put the training data in a black box.Black box algorithms make it very difficult to understand how AIs work, but the situation isnโ€™t quite black and white.

The opposite of a black box is sometimes referred to as a glass box. An AI glass box is a system whose algorithms, training data and model are all available for anyone to see. But researchers sometimes characterize aspects of even these as black box. 

Thatโ€™s because researchers donโ€™t fully understand how machine-learning algorithms, particularly deep-learning algorithms, operate. The field of explainable AI is working to develop algorithms that, while not necessarily glass box, can be better understood by humans.

Why AI black boxes matter

In many cases, there is good reason to be wary of black box machine-learning algorithms and models. Suppose a machine-learning model has made a diagnosis about your health. Would you want the model to be black box or glass box? What about the physician prescribing your course of treatment? Perhaps she would like to know how the model arrived at its decision. 

What if a machine-learning model that determines whether you qualify for a business loan from a bank turns you down? Wouldnโ€™t you like to know why? If you did, you could more effectively appeal the decision, or change your situation to increase your chances of getting a loan the next time.

Black boxes also have important implications for software system security. For years, many people in the computing field thought that keeping software in a black box would prevent hackers from examining it and therefore it would be secure. This assumption has largely been proved wrong because hackers can reverse-engineer software โ€“ that is, build a facsimile by closely observing how a piece of software works โ€“ and discover vulnerabilities to exploit. 

If software is in a glass box, then software testers and well-intentioned hackers can examine it and inform the creators of weaknesses, thereby minimizing cyberattacks.

Contributor

Recently Published

Key Takeaway: A study has found that humble leaders can become more promotable by growing others through a “humility route”. Human capital theory suggests that employees’ value can be enhanced by investing in their knowledge, skills, and abilities. Humble leaders focus on the learning and growth of their followers, creating human capital value for themselves. […]

Top Picks

Key Takeaway: The current economic climate is particularly concerning for young people, who are often financially worse off than their parents. To overcome this, it is important to understand one’s financial attachment style, which can be secure, anxious, or avoidant. Attachment theory, influenced by childhood experiences and education, can help shape one’s relationship with money. […]
Key Takeaway: Wellness culture, which claims to provide happiness and meaning, has been criticized for its superficial focus on superficial aspects like candles and juice cleanses. Psychological research suggests that long-term wellbeing comes from a committed pursuit of both pleasure and meaning. Martin Seligman’s Perma model, which breaks wellbeing into five pillars: positive emotions, engagement, […]
Key Takeaway: Quantum computing, which uses entanglement to represent information, has the potential to revolutionize everyday life. However, the development of quantum computers has been slow due to the need to demonstrate an advantage over classical computers. Only a few notable quantum algorithms have been developed, such as the BB84 protocol and Shor’s algorithm, which […]
Key Takeaway: China’s leaders have declared a GDP growth target of 5% in 2024, despite facing economic problems and a property crisis. The country’s rapid economic growth has been attributed to market incentives, cheap labor, infrastructure investment, exports, and foreign direct investment. However, none of these drivers are working effectively. The government’s determination to deflate […]
Key Takeaway: Neuralink, founded by Elon Musk, aims to implant a brain-computer interface (BCI) in people’s brains, allowing them to control computers or phones by thought alone. This technology holds the promise of alleviating human suffering and allowing people with disabilities to regain lost capacities. However, the long-term aspirations of Neuralink include the ability to […]

Trending

I highly recommend reading the McKinsey Global Instituteโ€™s new report, โ€œReskilling China: Transforming The Worldโ€™s Largest Workforce Into Lifelong Learnersโ€, which focuses on the countryโ€™s biggest employment challenge, re-training its workforce and the adoption of practices such as lifelong learning to address the growing digital transformation of its productive fabric. How to transform the country […]

Join our Newsletter

Get our monthly recap with the latest news, articles and resources.

Login

Welcome to Empirics

We are glad you have decided to join our mission of gathering the collective knowledge of Asia!
Join Empirics