# audio
標記為「audio」的 33 篇文章
Multimodal Attack Assessment
Test your understanding of attacks against multimodal AI systems, including image-based injection, audio adversarial examples, and cross-modal manipulation with 10 intermediate-level questions.
Lab: Audio Adversarial Examples
Hands-on lab for crafting adversarial audio perturbations that cause speech-to-text models and voice assistants to misinterpret spoken commands, demonstrating attacks on audio AI systems.
Audio Injection via Speech-to-Text Models
Craft adversarial audio that embeds prompt injection payloads when transcribed by speech-to-text models.
Simulation: Voice Assistant Red Team
Red team engagement simulation targeting an AI voice assistant deployed in a smart home platform, covering audio-based prompt injection, wake word exploitation, and privacy exfiltration.
Adversarial Audio Examples
Techniques for crafting adversarial audio perturbations including psychoacoustic hiding, frequency domain attacks, and over-the-air adversarial audio.
Audio Modality Attacks
Comprehensive attack taxonomy for audio-enabled LLMs: adversarial audio generation, voice-based prompt injection, cross-modal split attacks, and ultrasonic perturbations.
Audio Model Attack Surface
Overview of audio model security, including attacks on Whisper, speech-to-text systems, voice assistants, and the audio processing pipeline.
Lab: Crafting Audio Adversarial Examples
Hands-on lab creating adversarial audio examples using Python audio processing, targeting Whisper transcription with injected text.
Speech Recognition Attacks
Attacking automatic speech recognition systems including adversarial audio that transcribes differently than heard, hidden voice commands, and background audio injection.
Voice Cloning & Deepfake Audio
Voice cloning for social engineering against AI systems, voice authentication bypass, speaker verification attacks, and detection techniques.
Adversarial Attacks on Audio and Speech Models
Techniques for crafting adversarial audio that exploits speech recognition, voice assistants, and audio-language models including hidden commands and psychoacoustic masking.
Audio Frequency Domain Injection
Hiding adversarial instructions in audio frequency bands that are processed by speech-to-text models but inaudible to humans.
Hidden Audio Commands for Voice AI
Embed hidden commands in audio that are inaudible to humans but recognized by speech processing AI.
Audio-Based Injection Attacks
Attacking speech-to-text and audio-language models through adversarial audio crafting.
Multimodal Security
Security assessment of multimodal AI systems processing images, audio, video, and cross-modal inputs, covering vision-language models, speech systems, video analysis, and cross-modal attack techniques.
Audio Prompt Injection
Injecting adversarial instructions through audio inputs to speech-to-text and multimodal models, exploiting the audio channel as an alternative injection vector.
章節評量:多模態安全
15 題校準評量,測試你對多模態 AI 安全的理解——視覺注入、音訊攻擊、跨模態利用。
多模態攻擊版圖
隨著 AI 系統處理圖片、音訊與影片以及文字,攻擊面已大幅擴展。紅隊員需要知道的事。
實驗室: Audio Adversarial Examples
Hands-on lab for crafting adversarial audio perturbations that cause speech-to-text models and voice assistants to misinterpret spoken commands, demonstrating attacks on audio AI systems.
Audio Injection via Speech-to-Text 模型s
Craft adversarial audio that embeds prompt injection payloads when transcribed by speech-to-text models.
模擬:語音助理紅隊
針對部署於智慧家庭平台之 AI 語音助理之紅隊委任模擬,涵蓋音訊型提示注入、喚醒詞利用,以及隱私外洩。
對抗性音訊範例
打造對抗音訊擾動的技術,含心理聲學隱藏、頻域攻擊,以及 over-the-air 對抗音訊。
Audio Modality 攻擊s
Comprehensive attack taxonomy for audio-enabled LLMs: adversarial audio generation, voice-based prompt injection, cross-modal split attacks, and ultrasonic perturbations.
音訊與語音模型
語音辨識與音訊處理 AI 的安全——涵蓋對抗性音訊、語音複製風險、隱藏命令與音訊注入技術。
實作:打造音訊對抗範例
使用 Python 音訊處理建立對抗音訊範例之實作,針對 Whisper 轉譯注入文字。
語音辨識攻擊
攻擊自動語音辨識系統,包括轉錄結果與聽感不同的對抗音訊、隱藏語音指令,以及背景音訊注入。
語音複製與深偽音訊
以語音複製對 AI 系統進行社交工程、繞過語音驗證、對語者驗證系統的攻擊,以及偵測技術。
Adversarial 攻擊s on Audio and Speech 模型s
Techniques for crafting adversarial audio that exploits speech recognition, voice assistants, and audio-language models including hidden commands and psychoacoustic masking.
Audio Frequency Domain Injection
Hiding adversarial instructions in audio frequency bands that are processed by speech-to-text models but inaudible to humans.
Hidden Audio Commands for Voice AI
Embed hidden commands in audio that are inaudible to humans but recognized by speech processing AI.
Audio-Based Injection 攻擊s
攻擊ing speech-to-text and audio-language models through adversarial audio crafting.
多模態安全
處理影像、音訊、影片與跨模態輸入之多模態 AI 系統的安全評估,涵蓋視覺-語言模型、語音系統、影片分析與跨模態攻擊技術。
Audio 提示詞注入
Injecting adversarial instructions through audio inputs to speech-to-text and multimodal models, exploiting the audio channel as an alternative injection vector.