FILM AND VFX TESTS POWERED BY AI.

*For AI & VFX consultancy services, please fill out the contact page form.:

Contact — Freddy Chavez Olmos

Combinando y testeando algunas técnicas de VFX y AI para recrear al personaje de Don Ramon y El Chavo del Ocho en los años 70s.

I had the chance to early test Beeble Switchlight's Plugin for Foundry Nuke. Pretty incredible! It's one of the many new AI-powered innovative tools coming out soon in their Switchlight Studio Release. Oh, and now you can also extract ambient occlusion in addition to normals, albedo, specular, roughness, depth, and LDR-to-HDR light maps.

Plate Footage by ActionVFX

Music by Mubert

A new tool worth sharing. In this test, DUST3R estimated camera positions and generated a point cloud using single frames of multiple shots in a scene with minimum or no overlap in just a few seconds using a RTX4090. The tool currently allows to only export a GLB file but it looks quite promising where this is going once it's able to generate moving cameras.

We have partnered up with Beeble Switchlight to to find solutions for the inconsistent lighting on greenscreen driving plates. It's been almost a year since Switchlight first caught my attention. Initially developed as an app for relighting single frame portraits, I had the pleasure of meeting Hoon in person during last year’s Siggraph event in LA. We quickly became good friends and realized the tool's versatility in addressing common issues in VFX. Despite still being in beta and developing to a professional level, Switchlight has evolved into a more robust tool, capable of extracting Normals, Albedo, Specular, Roughness, and Depth maps from a plate. Moreover, it can generate an HDR from an LDR background plate and a Neural Render for use in compositing along with other maps. I hope that tools like these can help overcome the negative stigma around AI/ML and showcase their potential for real-world applications.

Bye-Bye was shot in only 2 hours in a virtual production stage. These are the VFX Breakdowns featuring GenAI and Machine Learning tools such as Nuke’s Copycat, Leonardo and Photoshop’s Generative to assist and speed up the post-production process.

Real-Time GenAI is progressing rapidly, now processing at 24fps and 30fps with a webcam. Although this tech still needs to achieve greater consistency and resolution, it's very exciting to test ideas and see instant results thanks to what @burkaygur and @fal_ai_data are building.

Taking stills from the uncanny valley for a makeover with Magnific.ai

Music by Mubert

[EN] Excited about the future of this tech once it runs 24fps, especially when combined with practical effects. AI real time rendering will open new doors for filmmakers. In this example, I used an animatronic that was built for one of my previous horror short films. Next time I will need to record myself wearing a greenscreen spandex ;)

Warning, some of these images might cause nightmares.

[ES] Emocionado por el futuro de esta tecnología una vez que corra a 24 fps, especialmente cuando se combina con efectos prácticos. Real-time rendering con AI abrirá nuevas puertas para los cineastas. En este ejemplo, utilicé un animatrónico que se hizo para uno de mis cortometrajes de terror anteriores. La próxima vez necesitaré grabarme usando spandex verde ;)

Advertencia, algunas de estas imágenes podrían causar pesadillas.

[EN] Testing the new lip-sync AI tool for language translation by HeyGen (currently in Beta). It not only generated Spanish, French, Hindi, Italian and Portuguese translations but also maintained a similar tone to the original. Additionally, I ran the tool a few times using the same footage, and it provided a few variations of each translation. The integration of this tech into real-time video conferencing may be on the horizon.

*Footage Property of WingNut Films/TriStar Pictures. For Educational Purposes Only.

[ES] Test de la nueva herramienta de AI para lip-sync de traducción de HeyGen (actualmente en Beta). No solo generó traducciones en Español, Francés, Hindi, Italiano y Portugués, sino que también mantuvo un tono similar al original. Además, al usar la herramienta varias veces utilizando la misma grabación, generó variaciones de cada traducción. La integración de esta tecnología en videoconferencias en tiempo real podría no estar muy lejos.

[EN] Test generating stock elements for vfx with Gen-2 from Runway and using the interpolation and upres features to improve the quality. This is still in its infancy stage of course but shows the potential where this generative tech is going with potential applications in vfx. Improvements in resolution, dynamic range, frame rate (among other things), is something we will probably see in the next Runway Gen versions.

Music created with Mubert.

[ES] TEST GENERANDO ELEMENTOS STOCK DE VFX CON I.A.

Test generando elementos de stock para vfx con Gen-2 de Runway usando las funciones de interpolación y upres para mejorar la calidad. Por supuesto, esto todavía está en su etapa inicial, pero muestra el potencial hacia dónde va esta tecnología generativa con aplicaciones a vfx. Mejoras en resolución, rango dinámico y frame rate (entre otras cosas), es algo que probablemente veremos en las próximas versiones de Runway Gen.

Música creada con Mubert.

Short test using a customized model in Leonardo Ai, ControlNet and Foundry Nuke’s CopyCat.

[EN] TEST GENERATING CLEAN PLATE FRAMES FOR VFX WITH A.I.

The new generative fill tool in Adobe Photoshop is simply amazing. Soon, having no access to clean plates in some situations won’t be the end of the world and paint tasks will be less time consuming.

Although the tool is not tech perfect by any means, that can change rapidly when this feature moves out of beta. A real clean plate will always be ideal but for those situations when there is none, this tool can make a big difference and assist artists in a meaningful way.

Music created with Mubert.

*Still Images property of VLAD CIOPLEA/NETFLIX. For Educational Purposes Only*

[ES] TEST GENERANDO FRAMES DE CLEAN PLATES PARA VFX CON I.A.

La nueva herramienta de Photoshop esta increíble. Pronto, el no tener acceso a clean plates en algunas situaciones no será el fin del mundo y los tasks de paint consumirán menos tiempo.

Música creada con Mubert.

*Imágenes propiedad de VLAD CIOPLEA/NETFLIX. Solo con fines educativos*

[EN] CREATING DIGITAL MAKEUP FX CONCEPTS WITH A.I.

Test using Leonardo AI, Foundry Nuke’s CopyCat and EbSynth. Leonardo’s ability to use customized models and controlNet features allows you to do some pretty fun stuff for horror projects.

Music created with Mubert.

*Footage property of Paramount Pictures. For Educational Purposes Only*

[ES] CREANDO EFECTOS DE MAQUILLAJE DIGITAL CONCEPTUAL CON A.I.

Test utilizando Leonardo AI, CopyCat de Nuke y EbSynth. La capacidad de usar Leonardo para utilizar modelos personalizados y funciones de controlNet te permite hacer cosas bastante divertidas para proyectos de terror.

Música creada con Mubert.

*Material propiedad de Paramount Pictures. Solo con fines educativos*

CONVIRTIENDO A “EL CHAVO DEL OCHO” EN “EL CHABOT DEL OCHO” CON A.I.

[ES] Mi primer test usando Wonder Studio (Beta) reemplazando a el Chavo del Ocho con un Bot. Por el momento, Wonder Studio es limitado en los tools que genera para obtener control del CG y el Clean Plate dentro de su plataforma. Incluso el Alpha que te genera, no es del CG sino del Clean Plate por lo que es difícil hacer cualquier modificación en el personaje 3D a menos que hagas las modificaciones en el archivo de Blender. Estoy seguro que en futuras versiones esto se irá implementando. Por ahora, mi workaround fue hacer un difference key en Nuke y de esa manera obtuve a el alpha del personaje para añadirle compresión y artefactos para que se integrara mejor en el plate. Wonder Studio acaba de implementar una herramienta para añadir tu propio personaje 3D por lo que vendrán más updates y tests de este app. A pesar de esto, es increíble ver lo que Wonder Dynamics ha logrado usando A.I. en tan poco tiempo. Recomiendo esta entrevista de Allan McKay con Nikola Todorovic (Co-Fundador y CEO de Wonder Dynamics). Es muy inspirador escuchar su visión:

https://www.youtube.com/watch?v=qRZXWtACIdc&t=1465s

[EN] My first test using Wonder Studio (Beta) replacing El Chavo del Ocho with a Bot. At the moment, Wonder Studio tools are limited when it comes down to controls for the CG and Clean Plate when working within its platform. The Alpha it generates is not for the 3D character but rather for the Clean Plate, so it is difficult to make any further tweaks to the CG in the renders unless you make those modifications using the Blender file. I have no doubt that in future versions this will be implemented. For now, my workaround was to do a difference key in Nuke to obtain the alpha for the character to add compression and artifacts so it feels more integrated to the plate. Wonder Studio has already implemented a feature to add your own character, so more updates and tests for this app are in the works. Despite all of this, I'm amazed to see what the team at Wonder Dynamics have achieved with A.I. in such a short time. If you haven't already, I would recommend to listen to Allan McKay 's interview with Nikola Todorovic (Co-Founder and CEO at Wonder Dynamics). It's very inspiring to hear his vision:

https://www.youtube.com/watch?v=qRZXWtACIdc&t=1465s

[EN] DEAGING MICHAEL KEATON AS "BATMAN" (1989) IN "THE FLASH" TRAILER (2023) USING A.I.

New test using DeepFaceLab, Photoshop Neural and Blackmagic Design DaVinci Resolve’s Magic Mask.

Music created with Mubert: music powered by A.I.

*Trailer Footage Property of Warner Bros / DC. For Educational Purposes Only*.

[ES] REJUVENECIENDO A MICHAEL KEATON COMO "BATMAN" (1989) EN EL TRÁILER DE "THE FLASH" (2023) USANDO A.I.

Nuevo test usando DeepFaceLab, Photoshop Neural y Magic Mask de Blackmagic Design DaVinci Resolve.

Música creada con Mubert.

*Footage del Tráiler es Propiedad de Warner Bros / DC. Solo para fines educativos.*

The new Switchlight app by Beeble allows you to key, add a background and relight a single frame. The Beta Tester version has very limited features at the moment but it has lots of potential for Compositing vfx. In this test, I used a jpg image from FxGuide featuring The Mandalorian over a green screen. I ended up denoising/upresing to 6K using Topaz Gigapixel and then added the grain back in Photoshop. Once the app developers add features such as HDR inputs/outputs, different aspect ratios, per light/color/edge controls, customized backgrounds, this tool will be something to watch out for. Nuke plugin anyone?

Music created with MUBERT: music powered by AI.

*Green screen footage Property of Disney. For Educational Purposes Only.

Personal test using AI to age Ewan McGregor as an older Obi-Wan Kenobi. Something that could look cool in a flash forward scene. Lots of trial/error on this attempt learning tools like ebsynth, stable diffusion, photoshop neural, etc…

Music created with AI using AIVA.

*Footage Property of Disney. For Educational Purposes Only.

Testing some aging techniques with AI as a proof of concept for potential proposals and uses in music videos.

Test creating parallax in Midjourney's v5 environment images using AI generated depth maps and Pixop Deep Restoration for Upres.

Music created with AI using AIVA

Generating photorealistic CG on a smartphone using 3D assets and AI with the Simulon App (currently in Beta). At the moment the beta version of the iOS app only allows you to generate full-res renders on a still frame using HDRIs but I went a bit further using EbSynth to reproject the render on the AR geo already available in the app to add motion. The app developers are working on some new great features like point light controls, better tracking and other cool stuff coming up soon.

Music created using Mubert: music powered by AI