Connect with us

Hi, what are you looking for?

SecurityWeekSecurityWeek

Artificial Intelligence

Insider Q&A: Artificial Intelligence and Cybersecurity In Military Tech

Shift5 founder Josh Lospinoso discusses AI and how software vulnerabilities in weapons systems are a major threat to the U.S. military.

AI

Josh Lospinoso’s first cybersecurity startup was acquired in 2017 by Raytheon/Forcepoint.. His second, Shift5, works with the U.S. military, rail operators and airlines including JetBlue. A 2009 West Point grad and Rhodes Scholar, the 36-year-old former Army captain spent more than a decade authoring hacking tools for the National Security Agency and U.S. Cyber Command.

Lospinoso recently told a Senate Armed Services subcommittee how artificial intelligence can help protect military operations. The CEO/programmer discussed the subject with The Associated Press as well how software vulnerabilities in weapons systems are a major threat to the U.S. military. The interview has been edited for clarity and length.

Q: In your testimony, you described two principal threats to AI-enabled technologies: One is theft. That’s self-explanatory. The other is data poisoning. Can you explain that?

A: One way to think about data poisoning is as digital disinformation. If adversaries are able to craft the data that AI-enabled technologies see, they can profoundly impact how that technology operates.

Q: Is data poisoning happening?

A: We are not seeing it broadly. But it has occurred. One of the best-known cases happened in 2016. Microsoft released a Twitter chatbot it named Tay that learned from conversations it had online. Malicious users conspired to tweet abusive, offensive language at it. Tay began to generate inflammatory content. Microsoft took it offline.

Q: AI isn’t just chatbots. It has long been integral to cybersecurity, right?

A: AI is used in email filters to try to flag and segregate junk mail and phishing lures. Another example is endpoints, like the antivirus program on your laptop – or malware detection software that runs on networks. Of course, offensive hackers also use AI to try defeat those classification systems. That’s called adversarial AI.

Advertisement. Scroll to continue reading.

Q: Let’s talk about military software systems. An alarming 2018 Government Accountability Office report said nearly all newly developed weapons systems had mission critical vulnerabilities. And the Pentagon is thinking about putting AI into such systems?

A: There are two issues here. First, we need to adequately secure existing weapons systems. This is a technical debt we have that is going to take a very long time to pay. Then there is a new frontier of securing AI algorithms – novel things that we would install. The GAO report didn’t really talk about AI. So forget AI for a second. If these systems just stayed the way that they are, they’re still profoundly vulnerable.

We are discussing pushing the envelope and adding AI-enabled capabilities for things like improved maintenance and operational intelligence. All great. But we’re building on top of a house of cards. Many systems are decades old, retrofitted with digital technologies. Aircraft, ground vehicles, space assets, submarines. They’re now interconnected. We’re swapping data in and out. The systems are porous, hard to upgrade, and could be attacked. Once an attacker gains access, it’s game over.

Sometimes it’s easier to build a new platform than to redesign existing systems’ digital components. But there is a role for AI in securing these systems. AI can be used to defend if someone tries to compromise them.

Q: You testified that pausing AI research, as some have urged, would be a bad idea because it would favor China and other competitors. But you also have concerns about the headlong rush to AI products. Why?

A: I hate to sound fatalistic, but the so-called “burning-use” case seems to apply. A product rushed to market often catches fire (gets hacked, fails, does unintended damage). And we say, ‘Boy, we should have built in security.’ I expect the pace of AI development to accelerate, and we might not pause enough to do this in a secure and responsible way. At least the White House and Congress are discussing these issues.

Q: It seems like a bunch of companies – including in the defense sector — are rushing to announce half-baked AI products.

A: Every tech company and many non-tech companies have made almost a jarring pivot toward AI. Economic dislocations are coming. Business models are fundamentally going to change. Dislocations are already happening or are on the horizon — and business leaders are trying to not get caught flat-footed.

Q: What about the use of AI in military decision-making such as targeting?

A: I do not, categorically do not, think that artificial intelligence algorithms — the data that we’re collecting — are ready for prime time for a lethal weapon system to be making decisions. We are just so far from that.

Read: OT Security Firm Shift5 Raises $50M to Protect Planes, Trains, and Tanks From Cyberattacks

Written By

Click to comment

Trending

Daily Briefing Newsletter

Subscribe to the SecurityWeek Email Briefing to stay informed on the latest threats, trends, and technology, along with insightful columns from industry experts.

Understand how to go beyond effectively communicating new security strategies and recommendations.

Register

Join us for an in depth exploration of the critical nature of software and vendor supply chain security issues with a focus on understanding how attacks against identity infrastructure come with major cascading effects.

Register

Expert Insights

Related Content

Artificial Intelligence

The CRYSTALS-Kyber public-key encryption and key encapsulation mechanism recommended by NIST for post-quantum cryptography has been broken using AI combined with side channel attacks.

Artificial Intelligence

The release of OpenAI’s ChatGPT in late 2022 has demonstrated the potential of AI for both good and bad.

Artificial Intelligence

ChatGPT is increasingly integrated into cybersecurity products and services as the industry is testing its capabilities and limitations.

Artificial Intelligence

The degree of danger that may be introduced when adversaries start to use AI as an effective weapon of attack rather than a tool...

Artificial Intelligence

Two of humanity’s greatest drivers, greed and curiosity, will push AI development forward. Our only hope is that we can control it.

Artificial Intelligence

Microsoft and Mitre release Arsenal plugin to help cybersecurity professionals emulate attacks on machine learning (ML) systems.

Application Security

Thinking through the good, the bad, and the ugly now is a process that affords us “the negative focus to survive, but a positive...

Artificial Intelligence

Two new surveys stress the need for automation and AI – but one survey raises the additional specter of the growing use of bring...