Hi Creatives! 👋

This week’s highlights point to a more settled phase in AI tools for creative work. Across voice, video, image workflows, and production systems, the shift is less about novelty and more about how these tools perform in actual use.

Some of these updates make creative processes easier to manage. Others are a reminder that speed, access, and convenience do not always mean long-term reliability. For creatives, the question is becoming less about what a tool can do once and more about whether it fits into a workflow that needs consistency over time.

This Week’s Highlights:

  • Freepik’s new 3D Scenes brings camera control into AI image workflows

  • Gemini 3.1 Flash Live makes voice AI more practical

  • Sora’s shutdown is a reminder that AI video platforms can shift fast

  • ElevenLabs Flows brings node-based creative pipelines into one canvas

  • Midjourney’s Version system is becoming a workflow choice, not just a settings toggle

  • A Sheep’s Big Ambition

  • From Prompting to Production

Freepik’s new 3D Scenes brings camera control into AI image workflows

Freepik’s new 3D Scenes feels less like another image generator and more like a virtual set builder. You can start from a preset scene, your own image, a Street View panorama, a 3D model, or even a blank canvas, then place objects, move the camera, and capture multiple shots from different angles without rebuilding the environment each time. Freepik’s official docs focus on that workflow, while Freepik’s own social post describes the feature as “Powered by World Labs spatial AI.”

What it actually does

  • Turns an uploaded image into an explorable 3D environment.

  • Lets you add subjects from a built-in 3D library, upload your own GLB model, or generate a 3D object from an image on the fly.

  • Gives you Compose mode for building the scene and Shoot mode for framing shots with field of view, zoom, aspect ratio, and cinematic settings.

  • Keeps the environment persistent between captures, so lighting, objects, and scene structure stay consistent across angles.

  • Offers a free Standard preview mode and a credit-based High Quality render mode for final images.

Insight

What makes this interesting is that it pushes AI image-making a little closer to directing instead of just generating. A lot of image tools are still built around starting over with every new angle or composition change. This feels different because the environment holds together, which makes the process more useful for campaigns, product storytelling, and any project where consistency matters across multiple shots.

It also points to a bigger shift. AI tools are starting to move from single outputs toward systems you can actually work inside. That tends to be where things get more useful for real production. Not because the tool does everything for you, but because it reduces how often you have to restart the whole process just to change the camera.

Read full details here.

Gemini 3.1 Flash Live makes voice AI more practical

Gemini 3.1 Flash Live is built for real-time voice interaction, with the focus on faster responses, more natural rhythm, and stronger performance in longer or messier conversations. It is available in preview through the Live API and is also being rolled into voice-facing products and enterprise use cases.

What to watch for

  • It is designed for live dialogue, not just speech output
    The model is optimized for low-latency, audio-to-audio interaction and voice-first applications.

  • The bigger pitch is reliability in real use
    It is positioned for multi-step task execution, stronger tonal understanding, and better handling of noisy environments and interruptions.

  • It supports more than simple chat
    The preview model supports function calling, Live API access, and Google Search grounding, which makes it more useful for assistants and interactive tools.

  • This is meant to show up across the stack
    The model is being pushed into developer tooling, enterprise customer experience, and consumer-facing voice products.

Insight

This feels like a more practical step for voice AI. A lot of systems can already talk, but that does not always mean they are good at staying on track, reacting naturally, or handling the kind of interruptions people actually bring into a conversation. That is where this update gets more interesting.

If the model really holds up in longer, more dynamic exchanges, then the opportunity is bigger than better voice chat. It points to more useful hands-free tools, stronger live assistants, and voice interfaces that feel less like demos and more like something people would actually keep using.

Read more here.

Sora’s shutdown is a reminder that AI video platforms can shift fast

OpenAI’s reported decision to discontinue the Sora video platform points to a bigger issue in AI video right now: platform stability. What makes this notable is not only the product decision itself, but how quickly the direction appears to have changed. Just days earlier, OpenAI’s Sora release notes were still rolling out new editing features, while its developer docs now show a formal deprecation path for the Videos API and Sora 2 model aliases, with removal scheduled for September 24, 2026.

For anyone following AI video, this looks less like a small product cleanup and more like a reminder that even high-profile tools can shift course fast. Sora helped set the tone for the current wave of cinematic text-to-video, but the latest reporting suggests OpenAI is redirecting focus toward areas it sees as more commercially important, including coding and enterprise products.

What to watch for

  • This looks bigger than just an app change
    The move appears tied to a broader pullback from Sora as a standalone video product, not only its front-end experience.

  • The direction changed quickly
    Sora had still been receiving product updates, which makes the reported shutdown feel abrupt.

  • Workflow flexibility matters more than ever
    If access, APIs, or core tools can change this fast, teams need systems that can adapt without breaking production.

  • Exports and asset management matter
    When platforms change direction, having organized files, references, and backups becomes just as important as the tool itself.

Insight

What stands out to me is that this is less about one company making one product decision and more about the current reality of AI video. The tools are improving fast, but the platform layer is still unstable. A product can look impressive, generate a lot of attention, and still not become a reliable long-term foundation for a workflow.

That is probably the more useful lesson here. The creative opportunity is still real, but it makes more sense to treat these platforms as flexible production tools rather than permanent home bases. The work should be portable. The process should be able to survive a product shift.

Read the full article here.

ElevenLabs Flows brings node-based creative pipelines into one canvas

ElevenLabs has introduced Flows inside ElevenCreative, a node-based creative canvas designed to connect image generation, video, text to speech, lip-sync, sound effects, and music in one workspace. The pitch is straightforward: instead of bouncing between separate tools, exporting files, and rebuilding the same sequence over and over, teams can map a pipeline once and reuse it with different inputs. ElevenLabs says Flows is available now in ElevenCreative, with API access coming soon.

For creatives, the more interesting part is not just that another platform added more models. It is that ElevenLabs is pushing a workflow-first idea. A Flow can be reused, rerun, and adjusted at the step level, which makes it easier to test variations without regenerating an entire project from scratch

What to watch for

  • Flows is built for end-to-end pipelines
    You can chain visuals, voice, lip-sync, and music in one canvas, then rework a single step without regenerating everything.

  • Reuse and batch testing are a big part of the value
    ElevenLabs says one Flow can be reused with different products, hooks, avatars, voices, or soundtracks, which makes it relevant for ad testing and repeatable content production.

  • It sits inside a larger multimodal setup
    ElevenLabs recently expanded its Image & Video offering with models such as Veo, Sora, Kling, Wan, Seedance, Nanobanana, Flux Kontext, GPT Image, and Seedream, plus its own audio tools.

  • API access is not here yet
    Flows is available now in ElevenCreative, while API access is listed as coming soon.

Insight

For the creative community, the value here is not simply “more AI in one place.” Plenty of platforms are moving toward all-in-one positioning. What stands out is the shift toward repeatable creative infrastructure.

That matters because a lot of production time is lost in the handoff layer:

  • exporting and re-uploading assets

  • rebuilding the same prompt chain

  • changing one step and having to redo the rest

  • managing version drift across tools

Flows is trying to reduce that friction by turning a creative process into something more modular and reusable. The upside is speed, testing, and more consistency across outputs. The trade-off is that a more automated pipeline can also make it easier to produce a large volume of average work. For creatives, the difference will still come from direction, taste, and knowing where a workflow should stay flexible instead of becoming too templated.

Read it full here.

Midjourney’s Version system is becoming a workflow choice, not just a settings toggle

Midjourney’s Version control lets users switch between model generations with --version or --v, and the current default is V7. The official docs now position versioning less as a technical detail and more as a practical creative choice, since each model handles prompts, style, speed, and feature compatibility differently.

The bigger shift is that Midjourney now has a clearer split between stable production use and early testing. V7 remains the default model, while V8 Alpha launched on March 17, 2026 on the alpha site only, not on the main site or in Discord. Midjourney says V8 Alpha is its fastest model so far, with standard jobs rendering about 4 to 5 times faster than earlier versions, and adds optional 2K HD generation through --hd.

What to watch for

  • V7 is still the safer default
    It remains the most practical choice for stable workflows and broader feature support.

  • V8 Alpha is faster, but still early
    It is meant more for testing than for fully established production use right now.

  • Feature support varies by version
    Some tools and prompt controls do not work the same way across models, so version choice can affect your workflow as much as the output.

  • Speed and consistency tools matter
    Features like Draft Mode and Omni Reference make Midjourney more useful for ideation and repeatable visual systems.

💡 Insight

What stands out to me is that Midjourney’s Version system is becoming less of a background setting and more of a real production choice. It is no longer just about chasing the newest model. It is about understanding what each version is actually good at and where it fits in the process.

V7 feels like the steadier option when consistency and compatibility matter more. V8 Alpha is interesting because of the speed, but it still feels more like a testing ground than something to fully rely on for established workflows. The bigger point is that version choice now affects how smoothly the work moves, not just how the final image looks.

If you want, I can also revise the entire short version so the whole piece flows with this more personal insight style.

Read the full details here.

A Sheep’s Big Ambition

Edmond Yang and the Nerve Lab team just released a new promotional film for Rednote Redgalas, inspired by the story of a sheep determined to become a sheepdog. It is a strong example of how AI can support a larger production pipeline when paired with clear creative direction, collaboration, and craft across story, art direction, sound, and post. The project was built entirely in Freepik, with Spaces used for collaboration, iteration, and finalization.

Read more here

From Prompting to Production

Henry Duabrez shared JUNKYARD KING — Episode 0 is an early look at how AI video tools are moving from short-form generation into more structured storytelling. Built over roughly four weeks, the project explores how a small creative setup can develop an animation prototype with more of a studio mindset, using prepared assets, consistent visual references, and a more directed workflow. What stands out is not just the final piece, but the process behind it: using tools like Seedance 2.0, Google Flow, and Gemini to build a more intentional pipeline for series development.

Check full details here.

💡 Insight

Taken together, this week’s updates suggest that creative work is becoming more shaped by workflow decisions, not just model quality. The more useful question now is not only what a tool can generate, but whether it supports the way a team actually works, adapts when platforms change, and helps maintain quality from draft to delivery.

For the creative community, that makes judgment, flexibility, and process design just as important as the tools themselves. The technology keeps moving, but the value still comes from knowing what to use, when to use it, and where to stay cautious.

That’s it for this week, folks.

🔔 Stay in the Loop!

Did you know that you can now read all our newsletters about The AI Stuff Creators and Artist Should Know that you might have missed.

Don't forget to follow us on Instagram 📸: @clairexue and @moodelier. Stay connected with us for all the latest and greatest! 🎨👩‍🎨

Stay in the creative mood and harness the power of AI,

Moodelier and Claire 🌈

Keep Reading