• VP Land
  • Posts
  • AI Tool Updates That Should Be On Your Radar: Pikadditions, Flora, and More

AI Tool Updates That Should Be On Your Radar: Pikadditions, Flora, and More

In the rapidly evolving landscape of AI-driven creative tools, staying informed about the latest developments can be challenging for media professionals. The newest episode of Denoised cuts through the noise to highlight several significant AI tool updates that are expanding what's possible for content creators, filmmakers, and visual effects artists.

From production-ready video generation tools to integrated workflow platforms, these tools are worth knowing about as they continue to mature and find their place in professional pipelines.

Pika Labs Introduces Swap and Additions Features

Pika Labs has released two notable features that demonstrate impressive capabilities for video manipulation and creation. While Pika initially positioned itself as a consumer-focused platform rather than a professional production tool, recent updates show how quickly these technologies are advancing toward production-quality applications.

The first feature, Pikaswaps, allows users to upload a video and replace elements within it using either text prompts or image references. This inpainting capability enables creative object replacement while maintaining surprisingly good spatial awareness.

Joey Daoud tested this by replacing his co-host with a robot in a multi-camera podcast setup: ā€œIt was freakishly accurate in some ways. In the clip I gave it, it had some angle cutting... and it honored the 3D geography of the space. From the wide shot, the robot looked a little distorted, but then it cut to a different camera and had the robot arm partially in frame, which would be what you expect. And then it cut to the other camera, and it was a full shot of the robot in perfect perspective.ā€

The second feature, Pikadditions, lets users add new objects into existing video while honoring the geography of the 3D space. For example, users can upload a video and an image of another object, then tell the system what they want the object to do within the scene.

Joey demonstrated this by adding a computer-generated octopus crawling across a table in a video. The AI maintained proper perspective even with camera angle changes.

Key takeaways about Pika's updates:

  • The system shows surprising spatial awareness across multiple camera angles

  • While not production-quality yet, it represents a significant step forward for accessible effects creation

  • The speed difference is dramatic - what might take skilled VFX artists days or weeks can be approximated in seconds

For comparison, Addy Ghani explained the traditional VFX workflow for a similar octopus effect would involve:

  • Animating eight legs convincingly (potentially days of work)

  • Creating an artificial 3D table for floor registration

  • Perspective matching a 3D camera to the live-action camera

  • Rendering, compositing in software like Nuke, color matching, and ensuring consistent image-based lighting

While professional VFX still offers significantly more control and refinement, these tools show how the technical barrier is lowering for visual storytelling.

Flora: An Intelligent Canvas for AI Workflows

Another noteworthy addition to the AI toolset is Flora, described as an "intelligent canvas" that serves as a middle ground between simple one-off AI tools and more complex technical platforms like ComfyUI.

Flora provides a node-based workflow platform that connects various AI models through API integrations, making it easier for users to build custom AI pipelines without requiring deep technical knowledge.

ā€œIt's a bit of a middle ground where you want to do some more creative stuff with AI that you can't really do with a single tool. You want to connect a variety of tools, but you're not as technically inclined to use something complex like ComfyUI,ā€ explained Joey.

The hosts compared Flora to the relationship between professional cinema cameras and more accessible options: "If ComfyUI is your pro tool like the Venice 2, then Flora is like your Blackmagic. It still retains functionality but is looking at a much bigger demographic with a nice user-friendly, clean interface."

Some notable capabilities of Flora include:

  • Connecting multiple AI tools through a visual node-based interface

  • Creating workflows that analyze hand-drawn storyboards and convert them into more detailed visualizations

  • Providing a central hub for different AI models without constant switching between platforms

  • Offering a more accessible entry point for creative professionals who need more than basic tools but aren't ready for complex programming

As Addy noted: ā€œThe more you could centralize and not have to bounce around between different apps, the better and faster it's going to be to create.ā€

Alibaba's Wan 2.1: Free Open-Source Video Generation

Alibaba has released an updated video model called Wan2.1 that is drawing comparisons to more prominent models like OpenAI's Sora. Unlike many high-end models, Wan2.1 is free and open-source, allowing users to run it on their local machines through platforms like ComfyUI.

The model supports text-to-video, image-to-video, and potentially video-to-video generation. While the hosts didn't consider it equal to Google's Veo 2 in quality (particularly with physics simulations), they noted it offers impressive capabilities for a free, open-source option.

"The physics were interesting. It's open source, it's free. Another kind of Chinese model just thrown out there, making things faster, cheaper, free," Addy commented.

For creators interested in experimenting with video generation without subscription costs, this model provides a valuable opportunity to test concepts locally.

For those seeking access to Google's Veo 2 model (which many consider currently superior), the hosts mentioned that Freepik offers access through their premium plan, though at approximately $2 per generation—making experimentation potentially costly.

ElevenLabs' Scribe: Advanced Speech-to-Text

ElevenLabs, known primarily for their text-to-speech technology, has now entered the reverse domain with a new speech-to-text model called Scribe. According to the company, Scribe outperforms previous speech recognition models including OpenAI's Whisper and Google's Gemini 2.0 on accuracy benchmarks.

This development continues the trend of AI companies expanding across the entire audio processing pipeline, from transcription to voice synthesis.

The discussion of speech models led to an interesting sidebar about the use of AI voice technology in film production, specifically referencing recent controversy around AI-generated voices in The Brutalist. Addy noted that while there might be issues with using such technology for high-profile artistic performances, "for the other 99.999% of use cases, I'm sure it's fine... If you have to dub a telenovela in five other languages in the course of one day, there's no other way to do it."

Midjourney Style References and Moodboards

While not a new tool, the hosts highlighted a valuable technique for achieving consistent results in Midjourney: using style reference codes and moodboards.

Rather than crafting elaborate text prompts, experienced Midjourney users often rely on style reference codes—a more reliable way to achieve specific visual styles. These codes can be stacked together to create distinctive looks.

The "moodboard" approach takes this further by allowing users to train the system on their own collection of reference images:

"Moodboards is you can make a moodboard and then you can just give it all of the images that you like of your reference images... you give it the images and then it creates a new style reference code based on your images. You're basically training a style reference."

This technique proves especially valuable for pre-visualization, storyboarding, and concept development, where maintaining visual consistency is crucial.

James Cameron's "No AI" Statement

The episode concluded with a discussion of James Cameron's recent statement that the next Avatar film might open with a title card stating "no generative AI was used in the making of this movie."

The hosts viewed this as potentially hypocritical given the extensive use of advanced computer graphics techniques in the Avatar franchise, and noted that Cameron himself now sits on the board of Stability AI.

They suggested this stance might be related to new Academy Award qualification rules regarding AI disclosure, or as a response to recent controversies surrounding AI use in film.

The hosts emphasized the need for more nuanced language around AI technologies: "We just need more language around AI and more terminology because AI is such a broad term... are you going to tell me that no machine learning was used in how you animate the waves or how you figure out the movement of the waves?"

Conclusion

The rapid advancement of AI tools for media creation continues to blur the line between consumer-level applications and professional production workflows. While these tools don't yet match the control and refinement offered by traditional production pipelines, they're advancing at a remarkable pace.

As Addy noted during the discussion, when evaluating these technologies, "It's not about what we see right now. It's about where we'll be two iterations from now, and how fast things are changing."

For media professionals, staying informed about these developments isn't just about keeping up with trends—it's about understanding how these tools might integrate into existing workflows to enable new creative possibilities.

For the complete discussion of these tools and more insights from Addy and Joey, be sure to listen to the full episode of Denoised and subscribe for twice-weekly updates on the most important developments in creative technology.

Reply

or to participate.