Since rolling out the redesign of its Firefly app in April, Adobe has been releasing major updates for the generative AI hub at a near monthly clip. Today, the company is introducing a handful of new features to assist those who use Firefly’s video capabilities.
To start, Adobe is making it easier to add sound effects to AI-generated clips. Right now, the majority of video models create footage without any accompanying audio. Adobe is addressing this with a nifty little feature that allows users to first describe the sound effect they want to generate and then record themselves making it. The second part isn’t so Adobe’s model can mimic the sound. Rather, it’s so the system can get a better idea of the intensity and timing the user wants from the effect.
In the demo Adobe showed me, one of the company’s employees used the feature to add the sound of a zipper being unzipped. They made a “zzzztttt” sound, which Adobe’s model faithfully used to reproduce the effect at the intended volume. The translation was less convincing when the employee used the tool to add the sound of footsteps on concrete, though if you’re using the feature for ideation as Adobe intended, that may not matter. When adding sound effects, there’s a timeline editor along the bottom of the interface to make it easy to time the audio properly.
The other new features Adobe is adding today are called Composition Reference, Keyframe Cropping and Video Presets. The first of those allows you to upload a video or image you captured to guide the generation process. In combination with Video Presets, you can define the style of the final output. Some of the options Adobe is offering at launch allow you to create clips with anime, black and white or vector art styles. Lastly, with Keyframe Cropping you can upload the first and final frame of a video and select an aspect ratio. Firefly will then generate a video that stays within your desired format.
In June, Adobe added , and this month it’s doing the same. Most notable is the inclusion of , which Google premiered at its in May. At the moment, Veo 3 is one of the only AI models that can generate video with sound. Like with all the other partner models Adobe offers in Firefly, Google has agreed not to use data from Adobe users for training future models. Every image and video people create through Firefly is digitally signed with the model that was used to create it. That is one of the safeguards Adobe includes so that Firefly customers don’t accidentally ship an asset that infringes on copyrighted material.
According to Zeke Koch, vice president of product management for Adobe Firefly, users can expect the fast pace of updates to continue. “We’re relentlessly shipping stuff almost as quickly as we can,” he said. Koch adds Adobe will continue to integrate more third-party models, as long as their providers agree to the company’s data privacy terms.
Credit: Source link