Microsoft launches open source tool to prevent AI hacking

By Sam Steers
Share
Microsoft has launched a counterfeit open source tool in an attempt to prevent AI hacking and allow businesses to evaluate machine learning security...

Microsoft has announced that it has launched a counterfeit open source tool to try and prevent AI systems being hacked.

The Counterfeit project, released on GitHub, allows business developers to evaluate the severity of a cyber attack by simulating a threat against an AI system. 

In a statement, Microsoft said: “This tool is part of broader efforts at Microsoft to empower engineers to securely develop and deploy AI systems.”

Security professionals are able to set up the project in three specific ways: scanning AI systems for vulnerabilities, logging attacks against AI models, and by pen testing and red teaming AI systems. 

Scanning AI systems

Scanning AI systems regularly for vulnerabilities allows businesses to gain an understanding of potential weaknesses in their system’s environment. It also helps in preventing cyber attacks that could severely damage valuable software. 

What is pen testing?

Also known as penetration testing, unlike manual testing, pen testing allows for increased system and software security by ensuring that there are no loopholes in it once the testing is complete. The results of tests are also more accurate than those of a manual test, meaning they would be more reliable. 

Benefits of using artificial intelligence to prevent cyber attacks

There are several benefits to using artificial intelligence to help stop cyber threats. Firstly, AI can process much larger volumes of data than a human can, meaning that they can pick up any threats earlier and faster. Another advantage is it reduces the likelihood of any errors in a company's cybersecurity software, allowing for security that is more trustworthy. 

AI also increases the response and detection times when searching for threats. This also allows threats to be spotted and thwarted more quickly and efficiently than a cybersecurity system not enhanced by AI. Artificial intelligence has the ability to spot multiple threats at once, which increases the wall of security around the software and information that needs to be protected. While securing data is not easy, artificial intelligence can make it easier by taking care of threat prevention as it works with the cybersecurity programme.

Microsoft says that the tool comes with attack algorithms preloaded with developers and security experts being able to use the cmd2 scripting engine built into the tool to carry out the tests. 

The company also claims that companies can alternatively create baselines by scanning AI systems using the attack simulations which aims to help measure the company’s progress. 

According to Microsoft, several of its partners and government agencies have collaborated with the company to test the tool in their own environments. 

Share

Featured Articles

Why Dynatrace Signs Analytics Deal With F1 Team VCARB

Dynatrace partners with Formula One team VCARB to leverage AI observability and performance analytics tech for driver performance and race optimisation

NTT: How Global CEOs Are Planning For an AI Investment Surge

NTT reports that global CEOs prioritise AI investments for future profitability, with 89% ranking AI as critical tech & 77% planning budget increases

Why Kaspersky Joined The EU AI Pact Ahead of Regulation

Kaspersky joins the European Commission's AI Pact, committing to responsible AI governance, risk mapping and literacy promotion

JLR Harnesses AI Power of Tata Communications's MOVE

Technology

Why Persona and Audience Segmentation is a Marketer's Edge

AI Applications

WEF Report: The Impact of AI Driving 170m New Jobs by 2030

AI Applications