Analytics

5/30/2018
02:30 PM
Craig Hinkley
Craig Hinkley
Commentary
Connect Directly
Twitter
LinkedIn
RSS
E-Mail vvv
50%
50%

Machine Learning, Artificial Intelligence & the Future of Cybersecurity

The ability to learn gives security-focused AI and ML apps unrivaled speed and accuracy over their more basic, automated predecessors. But they are not a silver bullet. Yet.

Machine learning (ML) and artificial intelligence (AI) are not what most people imagine them to be. Far removed from R2-D2 or WALL-E, today's bots, sophisticated algorithms, and hyperscale computing can "learn" from past experiences to influence future outcomes.

This ability to learn gives cybersecurity-focused Al and ML applications unrivaled speed and accuracy over their more basic, automated predecessors. This might sound like the long-awaited silver bullet, but AI and ML are unlikely, at least in the near future, to deliver the much-heralded "self-healing network." The technology does, however, bring to the table a previously unavailable smart layer that forms a critical first-response defense from hackers.

The Double-Edged Sword
AI and ML would be complete game changers for cybersecurity teams if not for the fact that hackers have also embraced the technologies. This means that, although AI and ML form an increasing part of the cybersecurity solution, they more frequently contribute to the cybersecurity problem.

So, when thinking about AI and ML, it's important not to take an insular approach. Don't just focus on what your company needs in isolation. Consider what your competitors might be adopting in regard to scanning technology for locating security defects in code or vulnerabilities in production — and how you can best keep up. Think about what hackers could be deploying — and how you can counter it. Working in this way will help identify the new policies, procedures, processes, and countermeasures that must be put in place to keep your organization safe and to get the full benefit from any investment in AI and ML.

Cybersecurity Job Prospects
When the IT world first started talking about AI and ML, there was a deep-rooted concern that "the robots" would take over human jobs. In the cybersecurity sector, nothing could be further from the truth. No enterprise actually wants to give up human control of their security systems and, in fact, most organizations will need more security experts and data scientists to operate or "teach" the software.

Let's take a minute to understand why. Without human monitoring and continuous input, the current generation of AI and ML software cannot reliably learn and adapt; neither can it highlight when the data sets it relies on are becoming corrupted, question whether its conclusions are correct, or guarantee compliance. Indeed, most AI and ML projects fail when either the software hasn't been programmed to ask the right questions in order to learn, or, when trying to learn, the software is presented with flawed data. More will fail in the future if they cannot demonstrate compliance with global legislation and industry-specific regulations. 

Longer term, use of AI and ML to combat cybersecurity threats might bring about closer coordination between cybersecurity professionals and data scientists. It's not unfeasible that cybersecurity teams might recruit data scientists or that companies will begin to look for cybersecurity experts with specific data science expertise. Eventually both roles and disciplines could even merge.

So, far from discouraging graduates to study cybersecurity, AI, and data science, the growth in both technologies should encourage students to take these courses and acquire some specialization in the field. Looking broadly across the IT security sector, the current skills and knowledge gap is unlikely to go away — and, in fact, as companies struggle to understand AI on a practical level, the number of open job vacancies could increase.

Who Is in Charge?
It's important that we as humans don't lose the capacity to oversee and manage AI and ML technology — in particular, that we don't abdicate responsibility for the outcomes produced by AI and ML software. The law has some catching up to do in this regard, but we already are seeing a lot more written about AI and ML transparency, trustworthiness, and interoperability — particularly for those using AI or ML within regulated markets such as banking and insurance.

It is a brave new world out there. So, stay abreast of new AI- and ML-based cybersecurity technologies, products, and services. Some of these are going to be real industry turning points, and you don't want to be the last person finding out about them. As AI and ML begin to play even more direct and obvious roles in IT infrastructures, it's vital for cybersecurity folks to keep their knowledge current and relevant. Try to get to at least one conference a year on the topic, jump on a webinar once a quarter, read some quality independent research each month so you have a real feel for what's happening out there.

This is the next frontier, and it's time to boldly go.

Related Content:

Craig Hinkley joined WhiteHat Security as CEO in early 2015, bringing more than 20 years of executive leadership in the technology sector to this role. Craig is driving a customer-centric focus throughout the company and has broadened WhiteHat's global brand and visibility ... View Full Bio
Comment  | 
Print  | 
More Insights
Comments
Newest First  |  Oldest First  |  Threaded View
Meet 'Bro': The Best-Kept Secret of Network Security
Greg Bell, CEO, Corelight,  6/14/2018
Containerized Apps: An 8-Point Security Checklist
Jai Vijayan, Freelance writer,  6/14/2018
Four Faces of Fraud: Identity, 'Fake' Identity, Ransomware & Digital
David Shefter, Chief Technology Officer at Ziften Technologies,  6/14/2018
Register for Dark Reading Newsletters
White Papers
Video
Cartoon
Current Issue
Flash Poll
Twitter Feed
Dark Reading - Bug Report
Bug Report
Enterprise Vulnerabilities
From DHS/US-CERT's National Vulnerability Database
CVE-2018-0291
PUBLISHED: 2018-06-20
A vulnerability in the Simple Network Management Protocol (SNMP) input packet processor of Cisco NX-OS Software could allow an authenticated, remote attacker to cause the SNMP application on an affected device to restart unexpectedly. The vulnerability is due to improper validation of SNMP protocol ...
CVE-2018-0292
PUBLISHED: 2018-06-20
A vulnerability in the Internet Group Management Protocol (IGMP) Snooping feature of Cisco NX-OS Software could allow an unauthenticated, adjacent attacker to execute arbitrary code and gain full control of an affected system. The attacker could also cause an affected system to reload, resulting in ...
CVE-2018-0293
PUBLISHED: 2018-06-20
A vulnerability in role-based access control (RBAC) for Cisco NX-OS Software could allow an authenticated, remote attacker to execute CLI commands that should be restricted for a nonadministrative user. The attacker would have to possess valid user credentials for the device. The vulnerability is du...
CVE-2018-0294
PUBLISHED: 2018-06-20
A vulnerability in the write-erase feature of Cisco FXOS Software and Cisco NX-OS Software could allow an authenticated, local attacker to configure an unauthorized administrator account for an affected device. The vulnerability exists because the affected software does not properly delete sensitive...
CVE-2018-0295
PUBLISHED: 2018-06-20
A vulnerability in the Border Gateway Protocol (BGP) implementation of Cisco NX-OS Software could allow an unauthenticated, remote attacker to cause a denial of service (DoS) condition due to the device unexpectedly reloading. The vulnerability is due to incomplete input validation of the BGP update...