• Home
  • Cinema
  • Studio
  • Digital
  • Royalty Free Music
  • web agency
  • blog
  • Home
  • Cinema
  • Studio
  • Digital
  • Royalty Free Music
  • web agency
  • blog

BLOG! BLOG! BLOG! BLOG! BLOG! BLOG! BLOG! BLOG! BLOG! BLOG! BLOG! BLOG!

blog

articoli, news, update

IAMCCS NEWSLETTER — MAY 10, 2026

bla bla bla

  • Aprile 10, 2026

Hi folks, this is CCS.

First of all — thank you. And I don’t mean that in the usual sign-off way. I mean it in a very specific, almost technical sense: seeing IAMCCS_nodes, workflows, and experimental pipelines actually living inside real production chains — not just test renders, not just proof-of-concepts — is the only metric I actually care about. It means we’re not tinkering with curiosities. We’re building something that works, and slowly shaping a new language for moving images.

And then there are your messages. I read them all, and I want you to know they mean something real to me — not as engagement metrics, not as validation, but as the voice of people who are doing the same work I’m doing, from different corners of the world, with the same obsession. We are colleagues in this. We are assembling, together, the camera of our ideas — and that’s not a metaphor I use lightly. Cinema has always been built by people who refused to wait for the industry to catch up with their vision. That’s exactly what we’re doing here, and I’m genuinely grateful to be in this with you.


PART 1 — AI RIGHT NOW

Let me cut through the noise and focus on what actually matters for filmmaking.


GEMS — Spatial Intelligence is Finally Getting Serious

GEMS introduces a closed-loop system for spatial logic and text rendering inside image generation — and yes, it outperforms Nano Banana 2 on GenEval2. But forget the benchmark for a second. What this really signals is that models are starting to understand composition as a structural problem, not just an aesthetic one. For anyone thinking in terms of frames and staging, that’s a meaningful shift — because the moment a model genuinely understands spatial relationships, you stop wrestling with the output and start directing it.

GitHub: https://github.com/lcqysl/GEMS
Paper: https://arxiv.org/abs/2603.28088

ComfyUI Post-Processing Suite — Photorealism Done Properly

This is one of those tools that will get underestimated on first glance, and then quietly become indispensable. Thezveroboy’s suite simulates sensor noise, analog artifacts, and even embeds camera metadata through base64 EXIF transfer and calibrated DNG writing. What that means in practical terms is that you stop generating images and start generating something that reads like footage — and that gap between AI output and cinematic credibility has been, for a long time, the hardest thing to close.

https://github.com/thezveroboy/ComfyUI-zveroboy-photo/blob/main/img1.jpg

CutClaw — Editing is Becoming Autonomous

CutClaw is an open multi-agent system that takes hours of raw footage and shapes it into narrative shorts. I’ll be honest about this one: it’s both genuinely powerful and worth approaching with some critical distance, because editing is not just structure — it’s intention, and that distinction matters. That said, used thoughtfully, this becomes something like a pre-edit intelligence layer — a way to explore narrative possibilities faster before you commit to anything final.

GitHub: https://github.com/GVCLab/CutClaw
Paper: https://huggingface.co/papers/2603.29664

Gemma 4

What makes Gemma 4 interesting is the combination: 256K token context, true multimodality across text, image, and audio, and open licensing that actually lets you build on top of it without depending on closed APIs. In practical terms, it’s part of a broader shift toward AI that runs locally, on your own hardware, accessible to independent creators rather than just large infrastructure. That democratization is slow and imperfect, but it’s real.

https://huggingface.co/google/gemma-4-E4B-it

Flux FaceIR — Face Restoration Without Compromise

This one is straightforward and useful. A Flux-2-klein LoRA built for blind or reference-guided face restoration — which means fixing broken generations, stabilizing identity across shots, and improving continuity in narrative work. If you’re doing any kind of character-driven AI filmmaking, this isn’t an optional extra. It’s pipeline infrastructure.

https://github.com/cosmicrealm/ComfyUI-Flux-FaceIR

LTX 2.3 Cameraman LoRA — Camera Motion as a Transferable Language

This one deserves attention. A LoRA for LTX 2.3 that extracts camera motion from reference videos and transfers it directly into new generated scenes — no trigger words, no workarounds. Which means you’re no longer describing movement through prompts. You’re capturing it from something real and reapplying it with intent. If this stabilizes in practice, it changes something fundamental: camera movement becomes an asset you can design, store, and reuse — motion as material.

https://huggingface.co/Cseti/LTX2.3-22B_IC-LoRA-Cameraman_v1

daVinci MagiHuman — Multimodal Video Generation

This is where things get genuinely interesting. A new multimodal system designed for video generation with stronger semantic coherence — and what I find most compelling about its direction is that it’s moving toward models that don’t just generate frames, but start to understand human presence as something continuous across time. That’s the real frontier, and we’re only beginning to touch it.

https://huggingface.co/GAIR/daVinci-MagiHuman

Hunyuan OmniWeaving — Multimodal Video is Evolving Fast

Tencent drops OmniWeaving — a new multimodal system for video generation where the relevant part isn’t just quality, but integration: image, motion, structure, and semantics starting to converge inside a single model logic. We’re moving toward systems that don’t just produce isolated shots but begin to understand sequences as connected things. Still early, but the direction is becoming clear.

https://github.com/Tencent-Hunyuan/OmniWeaving

I’m currently studying both models in depth. The goal is always the same: understand what can actually be integrated into a real filmmaking pipeline, and what genuinely adds control, consistency, and speed to our work.


My position on all of this (unchanged, and worth repeating):

We are inside incremental chaos right now. New models every hour, benchmarks everywhere, and very little actual direction. If you follow everything, you build nothing. The most promising tools — the ones that actually push filmmaking forward — will get dedicated deep-dives from me. Because the only thing that matters is keeping the line straight: understanding what works, understanding what serves your vision, and having the discipline to ignore the rest.


PART 2 — GOYAICANVAS

This week is significant. I’m fully inside beta testing now, and goyAIcanvas is already solving problems I couldn’t solve with workflows alone — which is precisely the kind of shift I was waiting for before talking about it publicly.

Inside it, there are multiple internal modes guiding you through videoclip generation, short-form content, post-production, editing, writing, animation, and even 3D logic integration. This is becoming a creative operating system in the real sense — not a tool that sits at the edge of your process, but something that runs through the middle of it.

Previews are coming soon. And I want to be transparent about how this is structured: Premium and especially VIP supporters will get access to the full system as it develops, because what I’m building inside goyAIcanvas isn’t a cosmetic upgrade — it’s a set of production layers that took months to develop and that I’m actively testing in my own filmmaking pipeline. That depth has to be sustainable, and your support is what makes it possible. Every tier of this community matters to me — but if you’re serious about compressing the learning curve and working with the most advanced configurations as they come out, the Premium and VIP path is where that happens. I’ll be here to guide all of it, filmmaker to filmmaker, as best I can.


PART 3 — WHAT WE BUILT THIS MONTH

This month started with a clear direction: long-form video generation. We pushed hard on temporal coherence, audio-driven structures, and narrative continuity — and here’s what went live:

  • [VID] WAN 2.2 SVI PRO v4 – First & Last Frame Video Generation — update 01.03.26 | Patreon

  • [VID] WAN 2.2 SVI PRO v4 – First & Last Frame Video Generation – update 01.03.26 (Patreon supporters)

  • [VID] LTX-2.3: The New King of AI Video? 🚀 Full Workflow & Test | Patreon

  • [VID] LTX-2.3 PROMPTING MASTERCLASS — DIRECTING THE LTX-2.3 MODEL (Patreon supporters)

  • [VID+AUD] 🎬 Directing with Sound: LTX-2.3 Audio-Guided Performance & Lipsync 👄 | Patreon

  • [VID+AUD] Directing LTX-2.3: From Audio-Guided Lipsync to Full Video Pipeline (Patreon supporters)

  • [VID+AUD]🎧LTX-2.3: From Audio + Image to Long-Form, Lip-Synced full video🎬 | Patreon

  • [VID+AUD] 🎧 LTX-2.3: Audio + Image Lip-Synced Long Video advanced workflows🎬 Smart Settings for Video & Monologue (Patreon supporters)

  • [VID+AUD] 🎭 LTX-2.3 ID-LoRA — Consistent Identity Across Audio & Video | Patreon

  • [VID] 🎬 LTX-2.3 VIDEO GENERATION Workflow Breakthrough — Advanced Setup & Optimization (Patreon supporters)

PART 4 — WHAT’S NEXT

The next posts will follow the same direction they always have: rooted in a filmmaker’s perspective on AI cinema, with a constant eye toward long-form generation, stability, continuity, and a few surprises still in the works — not least goyAIcanvas itself, which I’m hoping to release as soon as I get through the last remaining bugs.

Some early-access workflows are already out – IMAGE+AUDIO TO VIDEO INFINITE-LENGHT -and the LTX-2.3 VIDEO TO VIDEO GENERATION EXTENDED, thanx to our 1.4.0 version IAMCCS-nodes.
Coming soon: enhancer and upscale-to-4K workflows, a long-generation-length wrapper for better control over extended sequences, continued work on long-length WAN 2.2, and many other surprises — all of it in constant update, because quality and usability aren’t a destination, they’re an ongoing commitment.

PART 5 — CLOSING

When I was shooting with a RED or an Alexa — and I still do — the thing that drove me was never the gear. It was the moment when something that existed only inside my head became real light hitting a sensor. That gap between imagination and materialization, between the idea and the image: that’s the whole game. That’s always been the whole game. And there’s something that still catches me off guard sometimes, even after all this time: the image that comes back is occasionally more beautiful than the one I had visualized. More precise. More alive. The material world has a way of exceeding the mental draft of it.

Working with AI video generation gives me the same feeling. The tools are different, the pipeline is radically different, but the obsession is identical — to make the stuff of thought crystallize in front of your eyes, to take what lives in your mind and give it form, movement, presence. And sometimes, just like on a film set, what renders back is beyond what you imagined. That moment of surprise — that’s what I’m chasing here, in both worlds, with both cameras.

AI is not the filmmaker. It’s a tool — one of the most interesting tools I’ve worked with, but still a tool. Experimentation is the path. It always has been. Real creativity happens outside presets, outside trend cycles, outside the comfort of what already works.

One more thing worth saying: stay close, because as this gets more serious — and it is getting more serious — the posts won’t be static. They’ll evolve: updates, reconfigurations, simplifications. Some of you want full raw workflows with everything exposed. Others want speed and clarity. Both of those things matter, and I’m moving toward multiple versions of the same systems — raw pipelines, simplified workflows, and built-app versions inside ComfyUI. Less spaghetti. Same depth of control.

I’m the first one using these in real filmmaking pipelines. This is not theory.

More soon.

— CCS

Share:

Categorie

  • Cinema

  • Digital

  • Musica

  • Web design

Altri post

[EARLY ACCESS] LTX-2.3 Audio + Image to Video — Infinite Loop Workflow

Read More »

[VID] LTX-2.3 V2V Extended — Something New is Coming

Read More »

[VID+AUD] 🎭 LTX-2.3 ID-LoRA — Consistent Identity Across Audio & Video

Read More »

[VID+AUD]🎧LTX-2.3: From Audio + Image to Long-Form, Lip-Synced full video🎬

Read More »

FAIDENBLASS
studio

Hai bisogno di video-editing, post-produzione per il tuo progetto audio-video? Contattaci!
CONTATTACI

FAIDENBLASS
digital

Utilizziamo gli ultimi sistemi di Stable diffusion in locale di generazione di immagini + video. Contattaci per ogni info o richiesta
CONTATTACI

FAIDENBLASS
agency

Costruiamo siti web dinamici e responsivi su wordpress
CONTATTACI

Send Us A Message

PrevPrevious[EARLY ACCESS] LTX-2.3 Audio + Image to Video — Infinite Loop Workflow

FAIDENBLASS

Faidenblass: il punto di incontro tra arte analogica e digitale.

links

  • Carmine Cristallo Scalzi
  • Mitologia Elfica
  • faidenblass web agency

pagine sito

  • Home
  • Cinema
  • Studio
  • Digital
  • Royalty Free Music
  • web agency
  • blog
  • Home
  • Cinema
  • Studio
  • Digital
  • Royalty Free Music
  • web agency
  • blog