AI Art Weekly #110

Hello, my fellow dreamers, and welcome to issue #110 of AI Art Weekly! πŸ‘‹

It has been unusually quiet this week, almost like we’re about to be hit by a big storm. Nonetheless, there were some cool updates, like Luma launching a new video model with incredible motion capabilities, as well as some other interesting papers that I cross my fingers will get a code release (I’ve already bookmarked them with Premium to get notified when they do).

Enjoy your weekend and until next week!


News & Papers

Highlights

Luma Ray2

Luma has launched Ray2, a powerful AI video generation model that represents a 10x computational increase over its predecessor. It currently supports text-to-video with advanced motion handling, ultra-realistic details and physics-simulation capabilities. Image-to-video and video-to-video transformation features are on their way as well.

Ship captain smokes a pipe, turns and looks at a looming storm in the distance

3D

BioPose: Biomechanically-accurate 3D Pose Estimation from Monocular Videos

BioPose can predict accurate 3D human poses from single videos.

BioPose example

Image

SynthLight: Portrait Relighting with Diffusion Model by Learning to Re-render Synthetic Faces

SynthLight can relight human portraits using environment map lighting to create realistic effects like highlights and shadows.

SynthLight example

FramePainter: Endowing Interactive Image Editing with Video Diffusion Priors

FramePainter can edit images using sketches, allowing for highly coherent modifications. It excels in generalizing to out-of-domain scenarios, such as transforming a clownfish into a shark-like shape, while maintaining temporal consistency.

FramePainter example

1-2-1: Renaissance of Single-Network Paradigm for Virtual Try-On

1-2-1 can achieve high-quality virtual try-on for images and videos. It uses a Modality-specific Normalization strategy to process different inputs and reduces the need for complex systems, making it easier to use for high-resolution applications.

1-2-1 example

Video

Go-with-the-Flow: Motion-Controllable Video Diffusion Models Using Real-Time Warped Noise

Go-with-the-Flow can control motion patterns in video diffusion models using real-time warped noise from optical flow fields. It allows users to manipulate object movements and camera motions while keeping high image quality and not needing changes to existing models.

Go-with-the-Flow example

FlexiClip: Locality-Preserving Free-Form Character Animation

FlexiClip can generate smooth animations from clipart images while keeping key points in the right place.

FlexiClip examples

RepVideo: Rethinking Cross-Layer Representation for Video Generation

RepVideo can improve video generation by making visuals look better and ensuring smooth transitions.

with/without RepVideo comparison

LayerAnimate: Layer-specific Control for Animation

LayerAnimate can animate single anime frames from text prompts or interpolate between two frames with or without sketch-guidance. It allows users to adjust foreground and background elements separately.

LayerAnimate example

Multi-subject Open-set Personalization in Video Generation

Video Alchemist can generate personalized videos using text prompts and reference images. It supports multiple subjects and backgrounds without long setup times, achieving high-quality results with better subject fidelity and text alignment.

Multi-subject Open-set Personalization in Video Generation example

Perception-as-Control: Fine-grained Controllable Image Animation with 3D-aware Motion Representation

Perception-as-Control can achieve fine-grained motion control for image animation by creating a 3D motion representation from a reference image.

Perception-as-Control example

Audio

AnCoGen: Analysis, Control and Generation of Speech with a Masked Autoencoder

AnCoGen can analyze and generate speech by estimating key attributes like speaker identity, pitch, and loudness. It can also perform tasks such as speech denoising, pitch shifting, and voice conversion using a unified masked autoencoder model.

AnCoGen architecture

Enjoy your weekend dreamhead!

And that my fellow dreamers, concludes yet another AI Art weekly issue. If you like what I do, you can support me by:

  • Sharing it πŸ™β€οΈ
  • Following me on Twitter: @dreamingtulpa
  • Buying me a coffee (I could seriously use it, putting these issues together takes me 8-12 hours every Friday πŸ˜…)
  • Buying my Midjourney prompt collection on PROMPTCACHE πŸš€
  • Buying access to AI Art Weekly Premium πŸ‘‘

Reply to this email if you have any feedback or ideas for this newsletter.

Thanks for reading and talk to you next week!

– dreamingtulpa

by @dreamingtulpa