Connect with us

Hi, what are you looking for?


Artificial Intelligence

Are Artificial Intelligence and Machine Learning Just a Temporary Advantage to Defenders?

Attackers Explore How to Defeat Machine Learning (ML)-Based Defenses and Use ML in Their Own Attacks

Attackers Explore How to Defeat Machine Learning (ML)-Based Defenses and Use ML in Their Own Attacks

Artificial intelligence (AI), or more usually machine learning (ML), is the new kid on the block. It has become de rigueur for any new product or new version of an existing product to tout its AI/ML credentials. But the technology that was originally sold as the answer to cybercrime is now being questioned: is it a silver bullet or just a temporary advantage to the defenders?

Adam Kujawa, director of Malwarebytes Labs, has been considering the potential for bad actors to usurp machine learning for their own purposes. His report looks at some of the methods by which cybercriminals could use ML as offensive weapons against industry, focusing on three areas: poisoning defenders’ ML algorithms, DeepFakes, and artificially intelligent malware.

There are two fundamentals to machine learning: the algorithms that teach the machine how and what to learn, and a large amount of data (big data) to learn from. And there are two fundamental methodologies: unsupervised learning (which is effectively pure AI, where the machine teaches itself without reference to direct human intervention), and supervised learning (where the learning process is guided by human experts).

Algorithms are improving all the time, but suffer from one major weakness: they are subject to the conscious or subconscious bias of the designer. This is a bigger problem outside of cyber — where ML decisions can have life or death, freedom or prison implications — but will nevertheless be apparent in cybersecurity applications. Unsupervised machine learning will more rapidly evolve along the designers’ prejudices.

Poisoned data is a bigger problem; and again, unsupervised ML will respond to the poison faster. The danger has already been shown in the ML-based Twitter chatbot (Tay) developed and rapidly withdrawn by Microsoft. “A Twitter bot based on unsupervised machine learning,” says Kujawa, “had to be taken offline rather quickly when it started imitating unbecoming behavior that it ‘learned’ from other Twitter users. This was almost a perfect showcase of how easily machine learning can be corrupted when left without human supervision.”

The problem with all AI is that it cannot understand social context. Twitter is replete with bad language, extreme views, hate and false news that to humans (unless we already share those views) are easily recognizable. This is a data pool, already poisoned by its very nature, from which Microsoft’s bot learned bad behavior as normal. “‘Tay’ went from ‘humans are super cool’ to full nazi in <24 hrs,” commented @geraldmellor.

Advertisement. Scroll to continue reading.

The same principal can be used by cybercriminals to subvert the data pool used by products to learn the patterns of suspect behavior. “Threat actors,” warns Kujawa, “could also dirty the sample for machine learning, flagging legitimate packages as malware, and training the platform to churn out false positives.” The higher the concentration of false positives, the greater the likelihood for the security team to ignore alerts in their triaging process.

A second concern for Kujawa is the use of ML in social engineering. Deep, automated AI-enhanced social media scanning could rapidly build profiles of targets and their employees for compelling spear-phishing campaigns. But perhaps the most newsworthy current development is the evolution of what is called the DeepFake video

DeepFake videos can be generated by using AI to match a target’s facial imagery to words spoken by a voice imitator. In the future, the voice itself might also be generated by AI. A recent example appeared to show Mark Zuckerberg delivering a deeply cynical message. “”Imagine this for a second,” he appears to say: “One man, with total control of billions of people’s stolen data, all their secrets, their lives, their futures. I owe it all to Spectre. Spectre showed me that whoever controls the data, controls the future.”

One danger is that this technology could be married to BEC attacks — already a phenomenally successful and attractive attack for criminals (the latest FBI report says that $1.3 billion was lost through BEC and EAC attacks during 2018). “Now imagine getting a video call from your boss telling you she needs you to wire cash to an account for a business trip that the company will later reimburse,” says Kujawa. 

The latest Verizon DBIR strongly suggests that criminals are moving to the well-trusted and easier methods of earning their income. The marriage of ML and social engineering is less a possibility than an inevitability.

The evolution of ML-enhanced malware is also inevitable. For example, he writes, “Imagine worms that are capable of avoiding detection by learning from each detection event. If such a family of worms is able to figure out what got them detected, they will avoid that behavior or characteristic in the next infection attempt.” 

IBM has already shown the potential with its DeepLocker project. DeepLocker was a research project to examine what could be done with AI-enhanced malware. “DeepLocker,” IBM told SecurityWeek, “uses AI to hide any malicious payload invisibly within a benign, popular application — for example, any popular web conferencing application. With DeepLocker we can embed a malicious payload and hide it within the videoconferencing application. Using AI,” it added, “the conditions to unlock the malicious behavior will be almost impossible to reverse engineer.” The result is malware that is completely invisible until it detects its precise target, at which point it detonates.

Machine learning is out of the bag. It currently gives the advantage to the defenders — but this is primarily because they are the latest users. This may not always be the case. There is an old maxim: developing better security creates better attackers. Attackers are already exploring how to defeat ML-based defenses, and use ML in their own attacks. Defenders must recognize that ML was never a silver bullet, but merely the temporary advantage in the never-ending game of leapfrog between defense and attack.

Despite this, Kujawa is not ultimately pessimistic. The nature of ML used by attackers is that it must necessarily be largely unsupervised in operation. This is its weakness. “Our advantage over AI continues to be the sophistication of human thought patterns and creativity,” he states; “therefore, human-powered intelligence paired with AI and other technologies will still win out over systems or attacks that rely on AI alone.”

Related: It’s Time For Machine Learning to Prove Its Own Hype 

Related: Things to Consider Before Incorporating Machine Learning into Your Security

Related: Hunting the Snark with Machine Learning, AI, and Cognitive Computing 

Related: Demystifying Machine Learning: Turning the Buzzword Into Benefits

Written By

Kevin Townsend is a Senior Contributor at SecurityWeek. He has been writing about high tech issues since before the birth of Microsoft. For the last 15 years he has specialized in information security; and has had many thousands of articles published in dozens of different magazines – from The Times and the Financial Times to current and long-gone computer magazines.

Click to comment

Daily Briefing Newsletter

Subscribe to the SecurityWeek Email Briefing to stay informed on the latest threats, trends, and technology, along with insightful columns from industry experts.

Join security experts as they discuss ZTNA’s untapped potential to both reduce cyber risk and empower the business.


Join Microsoft and Finite State for a webinar that will introduce a new strategy for securing the software supply chain.


Expert Insights

Related Content

Artificial Intelligence

The CRYSTALS-Kyber public-key encryption and key encapsulation mechanism recommended by NIST for post-quantum cryptography has been broken using AI combined with side channel attacks.

Artificial Intelligence

The release of OpenAI’s ChatGPT in late 2022 has demonstrated the potential of AI for both good and bad.

Artificial Intelligence

ChatGPT is increasingly integrated into cybersecurity products and services as the industry is testing its capabilities and limitations.

Artificial Intelligence

The degree of danger that may be introduced when adversaries start to use AI as an effective weapon of attack rather than a tool...

Endpoint Security

Today, on January 10, 2023, Windows 7 Extended Security Updates (ESU) and Windows 8.1 have reached their end of support dates.

Artificial Intelligence

Two of humanity’s greatest drivers, greed and curiosity, will push AI development forward. Our only hope is that we can control it.

Application Security

Virtualization technology giant VMware on Tuesday shipped urgent updates to fix a trio of security problems in multiple software products, including a virtual machine...

Artificial Intelligence

Microsoft and Mitre release Arsenal plugin to help cybersecurity professionals emulate attacks on machine learning (ML) systems.