July 18, 2025

Adobe Firefly introduces feature to generate custom sound effects from user audio cues

Adobe Firefly introduces feature to generate custom sound effects from user audio cues

Adobe Firefly Learns to Listen: AI Now Generates Sound Effects From Your Voice

Adobe's Firefly, the generative AI engine integrated across the Creative Cloud suite, has already transformed visual workflows for millions. Initially celebrated for its ability to generate commercially-safe images, vectors, and text effects from simple text prompts, Firefly established itself as a co-pilot for designers and artists. It's the AI assistant living right inside Photoshop and Illustrator, trained on Adobe's vast library of stock imagery to ensure creators can use its output without copyright headaches. Until now, its senses have been purely visual, interpreting words to create pictures.

That all changes with the groundbreaking introduction of AI-powered sound effect generation. But this isn't just another text-to-audio model. Adobe's new feature allows users to provide their own audio cues as a prompt. Imagine you need a sci-fi "whoosh" for a video transition. Instead of typing "fast futuristic whoosh sound," you can now simply record yourself making a whoosh sound with your mouth. Firefly's AI analyzes the sonic characteristics of your input—the pitch, tempo, and texture—and generates a polished, high-fidelity, and completely unique sound effect based on it. This new capability fundamentally alters the creative process, making it more intuitive, personal, and immediate.

The implications for creators are immense. Video editors, podcasters, game developers, and social media managers can now craft bespoke audio without needing foley artists or spending hours sifting through stock sound libraries. This feature democratizes professional-grade sound design, placing a powerful tool directly into the hands of a broader creative community. It drastically accelerates workflows, allowing creators to maintain their creative momentum by generating the exact sound they envision in seconds. By teaching its AI to listen, not just read, Adobe isn't just adding a feature; it's pushing the entire industry toward a more intuitive, multimodal future for generative AI, where our natural expressions become the prompts for digital creation.

Saiki Sarkar

Saiki Sarkar

ATM @ Together CFO