The model is designed to accurately replicate human speech, movement, and gestures. According to the company’s website, whether it’s a portrait, half-body shot, or full-body image, OmniHuman can generate lifelike movements, natural gestures, and stunning attention to detail. At its core, OmniHuman is a multimodality-conditioned human video generation model. This means it integrates various input types, such as images and audio clips, to create highly realistic videos. OmniHuman is currently in the research phase and is not yet accessible to the public. The developers have shared demos and hinted at a possible code release in the future. The release of this model marks another Chinese breakthrough in the AI industry following DeepSeek’s large language model (LLM) DeepSeek-v3. ByteDance’s OmniHuman-1 is a direct competitor to OpenAI’s video-generating model, Sora, which was released in December 2024, as well as other video-generating models like Runway’s Gen-3 Alpha and Luma AI’s Dream Machine.
Comments
No comments yet. Be the first to comment!
Leave a Comment