menu
techminis

A naukri.com initiative

google-web-stories
Home

>

ML News

>

OMNIGUARD:...
source image

Arxiv

6d

read

234

img
dot

Image Credit: Arxiv

OMNIGUARD: An Efficient Approach for AI Safety Moderation Across Modalities

  • OMNIGUARD is an approach for detecting harmful prompts across languages and modalities in large language models (LLMs).
  • The approach identifies internal representations of LLMs/MLLMs that are aligned across languages or modalities to build a language-agnostic or modality-agnostic classifier for detecting harmful prompts.
  • OMNIGUARD improves harmful prompt classification accuracy significantly in multilingual setting, image-based prompts, and audio-based prompts.
  • By repurposing embeddings computed during generation, OMNIGUARD is very efficient and sets a new state-of-the-art for audio-based prompts.

Read Full Article

like

14 Likes

For uninterrupted reading, download the app