The Race for AI-Powered Security Platforms Heats Up | #hacking | #cybersecurity | #infosec | #comptia | #pentest | #ransomware

When a major vulnerability shakes up the cybersecurity world — such as the recent XZ backdoor or the Log4J2 flaws of 2021 — the first question that most companies ask is, “Are we affected?” In the absence of well-written playbooks, the simple question can require a great deal of effort to answer.

Microsoft and Google are investing heavily in generative AI systems that can turn large security questions into concrete actions, assist security operations. and, increasingly, taking automated actions. Microsoft offers overworked security operations centers with Security Copilot, a generative-AI-based service that can identify breaches, connect threat signals, and analyze data. And Google’s Gemini in Security is a collection of security capabilities powered by the company’s Gemini generative AI.

Startup Simbian is joining the race with its new generative-AI-based platform for helping companies tackle their security operations. Simbian’s system combines large language models for summarizing data and understanding native language, other machine learning models to connect disparate data points, and a software based expert system based on security information culled from the Internet.

Where configuring a security information and event management system (SIEM) or a security orchestration, automation, and response (SOAR) system could take weeks or months, using AI cuts the time to — in some cases — seconds, says Ambuj Kumar, co-founder and CEO of Simbian.

“With Symbian, literally, these things are done in seconds,” he says. “You ask a question, you express your goal in natural language, we break into steps code execution, and this is all done, automatically, it’s self sufficient.”

Helping overworked security analysts and incident responders streamline their jobs is a perfect application for the more powerful capabilities of generative AI, says Eric Doerr, vice president of engineering at Google Cloud.

“The opportunity in security is particularly acute given the elevated threat landscape, the well publicized talent gap in cybersecurity professionals, and the toil that is the status quo in most security teams,” Doerr says. “Accelerating productivity and driving down mean time to detect, respond, and contain [or] mitigate threats through the use of GenAI will enable security teams to catch up and defend their organizations more successfully.”

Different Starting Points, Different ‘Advantages’

Google’s advantages in the market are evident. The information-technology and internet giant has the budget to stay the course, has the technical expertise in machine learning and AI from its DeepMind projects to innovate, and has access to a lot of training data — a critical consideration for creating large language models (LLMs).

“We have a tremendous amount of proprietary data that we’ve used to train a custom security LLM — SecLM — which is part of Gemini for Security,” Doerr says. “This is the superset of 20 years of Mandiant intelligence, VirusTotal, and more, and we’re the only platform that has an open API — part of Gemini for Security — that allows partners and enterprise customers to extend our security solutions and have a single AI that can operate with all the context of the enterprise.”

Like Simbian’s guidance, Gemini in Security Operations — one capability under the Gemini in Security umbrella — will assist in investigations starting at the end of April, guiding the security analyst and recommending actions from within Chronicle Enterprise.

Simbian uses natural language queries to generate results, so asking, “Are we affected by the XZ vulnerability?” will produce a table of IP addresses of vulnerably applications. Depending on the systems to which the Simbian platform connects, the systems also uses curated security knowledge gathered from internet, to create guidebooks for security analysts that shows them a script of prompts to give to the system to accomplish a specific task.

“The guidebook is a way of personalizing or creating a trusted content,” says Simbian’s Kumar. “Right now, we are creating the guidebooks, but once … people just start to use it, then they can create their own.”

Strong ROI Claims for LLMs

The returns on investment will grow as companies move from a manual process to an assisted process to autonomous activity. Most generative AI-based systems have only advanced to the stage of an assistant or copilot, when it suggests actions or takes only a limited series of actions, after gaining the users permissions.

The real return on investment will come later, Kumar says.

“What we are excited about building is autonomous — autonomous is making decisions on your behalf that are within the scope of guidance you have given it,” he says.

Google’s Gemini also seems to straddle the gap between an AI assistant and an automated engine. Financial services firm Fiserv is using Gemini in Security Operations for creating detections and playbooks faster and with less effort, and for helping security analysts to quickly find answers using natural language search, boosting the productivity of security teams, Doerr says.

Yet, trust is still an issue and a hurdle for increased automation, he says. To bolster trust in the system and solutions, Google remains focused on creating explainable AI systems that are transparent in how they come to a decision.

“When you use a natural language input to create a new detection, we show you the detection language syntax and you choose to run that,” he says. “This is part of the process of building confidence and context with Gemini for Security.”


Click Here For The Original Source.

National Cyber Security