Short-Form Video App Storiaverse Touts ‘Read-Write’ Format

Mobile entertainment platform Storiaverse is connecting writers and animators around the world to create content for what it claims is a unique “read-watch” format. Available on iOS and Android, Storiaverse combines animated video, audio and text into a narrative that “enhances the reading experience for digital native adults.” Created by Agnes Kozera and David Kierzkowski, co-founders of the Podcorn podcast sponsorship marketplace, Storiaverse caters to graphic novel fans interested in discovering original, short-form animated stories that range from 5-10 minutes in length. At launch there will be 25 original titles. Continue reading Short-Form Video App Storiaverse Touts ‘Read-Write’ Format

Stable Video 3D Generates Orbital Animation from One Image

Stability AI has released Stable Video 3D, a generative video model based on the company’s foundation model Stable Video Diffusion. SV3D, as it’s called,  comes in two versions. Both can generate and animate multi-view 3D meshes from a single image. The more advanced version also let users set “specified camera paths” for a “filmed” look to the video generation. “By adapting our Stable Video Diffusion image-to-video diffusion model with the addition of camera path conditioning, Stable Video 3D is able to generate multi-view videos of an object,” the company explains. Continue reading Stable Video 3D Generates Orbital Animation from One Image

Midjourney Creates a Feature to Advance Image Consistency

Artificial intelligence imaging service Midjourney has been embraced by storytellers who have also been clamoring for a feature that enables characters to regenerate consistently across new requests. Now Midjourney is delivering that functionality with the addition of the new “–cref” tag (short for Character Reference), available for those who are using Midjourney v6 on the Discord server. Users can achieve the effect by adding the tag to the end of text prompts, followed by a URL that contains the master image subsequent generations should match. Midjourney will then attempt to repeat the particulars of a character’s face, body and clothing characteristics. Continue reading Midjourney Creates a Feature to Advance Image Consistency

Alibaba’s EMO Can Generate Performance Video from Images

Alibaba is touting a new artificial intelligence system that can animate portraits, making people sing and talk in realistic fashion. Researchers at the Alibaba Group’s Institute for Intelligent Computing developed the generative video framework, calling it EMO, short for Emote Portrait Alive. Input a single reference image along with “vocal audio,” as in talking or singing, and “our method can generate vocal avatar videos with expressive facial expressions and various head poses,” the researchers say, adding that EMO can generate videos of any duration, “depending on the length of video input.” Continue reading Alibaba’s EMO Can Generate Performance Video from Images

Generative Tech Enables Multiple Versions of the Same Movie

Filmmaker Gary Hustwit and artist Brendan Dawes aspire to change the way audiences experience film. Their startup, Anamorph, has launched with an app that can reassemble different versions of the same film. The app debuted with “Eno,” a Hustwit-directed documentary about the music iconoclast Brian Eno that premiered in January at the Sundance Film Festival, where every “Eno” showing presented the audience with a unique viewing experience. Drawing scenes from a repository of over 500 hours of “Eno” material, the Anamorph app would potentially be able to generate what the company says is billions of different configurations. Continue reading Generative Tech Enables Multiple Versions of the Same Movie

Apple’s Keyframer AI Tool Uses LLMs to Prototype Animation

Apple has taken a novel approach to animation with Keyframer, using large language models to add motion to static images through natural language prompts. “The application of LLMs to animation is underexplored,” Apple researchers say in a paper that describes Keyframer as an “animation prototyping tool.” Based on input from animators and engineers, Keyframer lets users refine their work through “a combination of prompting and direct editing,” the paper explains. The LLM can generate CSS animation code. Users can also use natural language to request design variations. Continue reading Apple’s Keyframer AI Tool Uses LLMs to Prototype Animation

CES: Digital Hollywood Session Explores AI at Inflection Point

Impact and opportunity surfaced as the dominant theme of a full day of Digital Hollywood sessions devoted to artificial intelligence at CES 2024. We are in a period of disruption similar to the early 90s when the Internet went mainstream, said Forbes columnist Charlie Fink, moderating a panel of industry leaders from CAA, Paramount, HTC, Nvidia and Google. Yet despite the transformation already underway, panelists agreed that this is neither the first nor last technology to shift the status quo, more the latest example of inevitable change and adjustment. The current conversations around AI at CES are a refreshing departure after a few years of evolutionary, not revolutionary tech confabs. Continue reading CES: Digital Hollywood Session Explores AI at Inflection Point

VideoPoet: Google Launches a Multimodal AI Video Generator

Google has unveiled a new large language model designed to advance video generation. VideoPoet is capable of text-to-video, image-to-video, video stylization, video inpainting and outpainting, and video-to-audio. “The leading video generation models are almost exclusively diffusion-based,” Google says, citing Imagen Video as an example. Google finds this counter intuitive, since “LLMs are widely recognized as the de facto standard due to their exceptional learning capabilities across various modalities.” VideoPoet eschews the diffusion approach of relying on separately trained tasks in favor of integrating many video generation capabilities in a single LLM. Continue reading VideoPoet: Google Launches a Multimodal AI Video Generator

Stability Introduces GenAI Video Model: Stable Video Diffusion

Stability AI has opened research preview on its first foundation model for generative video, Stable Video Diffusion, offering text-to-video and image-to-video. Based on the company’s Stable Diffusion text-to-image model, the new open-source model generates video by animating existing still frames, including “multi-view synthesis.” While the company plans to enhance and extend the model’s capabilities, it currently comes in two versions: SVD, which transforms stills into 576×1024 videos of 14 frames, and SVD-XT that generates up to 24 frames — each at between three and 30 frames per second. Continue reading Stability Introduces GenAI Video Model: Stable Video Diffusion

Meta Touts Its Emu Foundational Model for Video and Editing

Having made the leap from image generation to video generation over the course of a few months in 2022, Meta Platforms introduces Emu, its first visual foundational model, along with Emu Video and Emu Edit, positioned as milestones in the trek to AI moviemaking. Emu uses just two diffusion models to generate 512×512 four-second long videos at 16 frames per second, Meta said, comparing that to 2022’s Make-A-Video, which requires a “cascade” of five models. Internal research found Emu video generations were “strongly preferred” over the Make-A-Video model based on quality (96 percent) and prompt fidelity (85 percent). Continue reading Meta Touts Its Emu Foundational Model for Video and Editing

Unity Opens Beta for Muse AI, Sets General Release for 2024

Unity has officially released its Muse AI platform for general use in early access. Muse is a suite of AI-powered tools that streamline game development. The Muse package includes Muse Chat to source answers and generate code, Muse Sprite for 2D sprites generation, and Muse Texture, providing 2D and 3D ready textures. Originally announced in July, Muse is now offered at a $30 per month subscription. Also announced at the firm’s annual Unite conference was the next major software update, Unity 6, for 2024, and the deployment of Unity Cloud to connect development tools across projects and pipelines. Continue reading Unity Opens Beta for Muse AI, Sets General Release for 2024

Social Startup Plai Labs Debuts Free Text-to-Video Generator

The entrepreneurs behind the Myspace social network and gaming company Jam City have shifted their focus to generative AI and web3 with a new venture, Plai Labs, a social platform that provides AI tools for collaboration and connectivity. Plai Labs has released a free text-to-video generator, PlaiDay, which will compete with other GenAI video tools from the likes of OpenAI (DALL-E 2), Google (Imagen), Meta Platforms (Make-A-Video) and Stable Diffusion. But PlaiDay hopes to set itself apart by offering the ability to personalize videos with selfie likenesses. Continue reading Social Startup Plai Labs Debuts Free Text-to-Video Generator

Startup Kaiber Launches Mobile GenAI App for Music Videos

Kaiber, the AI-powered creative studio whose credits include music video collaborations with artists such as Kid Cudi and Linkin Park, has launched a mobile version of its creator tools designed to give musicians and graphic artists on-the-go access to its suite of GenAI tools offering text-to-video, image-to-video and video-to-video, “now with curated music to reimagine the music video creation process.” Users can select artist tracks to accompany visuals to build a music video “with as much or little AI collaboration as they wish.” Users can also upload their own music or audio and tap Kaiber for visuals. Continue reading Startup Kaiber Launches Mobile GenAI App for Music Videos

YouTube Adds Dozens of Mobile Playback, Creator Controls

YouTube has introduced three dozen new features that do everything from improve mobile playback and search to expand creator tools. When the words “like and subscribe” are uttered by creators, the buttons will respond with animations. Bigger preview thumbnails and a “lock-screen” feature designed to prevent mishaps are also in the new toolset. The improvements build on the “new look and feel” YouTube debuted last year, adding “more modern design elements and features that helped our viewers feel more immersed,” the company says. Some of the improvements target smartphones and connected TVs, while others are for the web app. Continue reading YouTube Adds Dozens of Mobile Playback, Creator Controls

ETC Will Host Sessions at SIGGRAPH Conference This Week

ETC@USC will host its 8th vETC virtual conference at SIGGRAPH 2023 in Los Angeles, August 8-10. The event – which highlights significant presentations of emerging technologies and their impact on the M&E industry – will explore how generative AI, machine learning, and other compelling new tools help simplify building 3D worlds and tackle today’s computer vision challenges. Three days of sessions will be recorded and posted on ETC’s YouTube channel. For those attending SIGGRAPH who may be interested in attending the sessions (located at Z by HP Booth 215), visit the program guide, which includes a full schedule and speaker bios. Continue reading ETC Will Host Sessions at SIGGRAPH Conference This Week