Yall dont understand: Sora is going to be a GAME CHANGER for fan fiction.

bnew

Veteran
Joined
Nov 1, 2015
Messages
57,329
Reputation
8,496
Daps
159,976












1/27
@EccentrismArt
🏰🏰 It’s here, “Castle Mates - Part 2”!!

Part 2 of my comedy mockumentary. If you haven’t seen Part 1, see it 👇

Please like and share 🙏🏻

🔊 On



https://video.twimg.com/amplify_video/1861791580732145664/vid/avc1/1280x720/QEwKrnDyVLbeiaZb.mp4

2/27
@EccentrismArt
Part 1 here:

[Quoted tweet]
🏰 “Castle Mates - Part 1” 🏰

A comedy mockumentary and spinoff to my “Monster Flatmates”

Part 2 coming soon. Please like and share 🙏🏻

🔊 On
#lotr #skyrim #Witcher


https://video.twimg.com/amplify_video/1859272340473085952/vid/avc1/1280x720/6tF1jQBzfzM64bFY.mp4

3/27
@EccentrismArt
Workflow:
🏞️Midjourney and @freepik
🎥@freepik & @KaiberAI and @runwayml (Act-One)
🫦Lip sync by Runway Act-One and @hedra_labs with my facial capture as well.
🎶Suno V4
📜 (Titles): @ideogram_ai



4/27
@EccentrismArt
My comedy friends! Castle Mates Part 2 is here!! Let me all know if you want a Part 3 🤪
@steviemac03
@koba_1975
@iamtomblake
@madpencil
@ikewagh
@pressmanc
@alexgnewmedia
@blvcklightai
@captainhahaa
@Stonekaiju
@empath_angry
@_lev3lup
@jeepersmedia
@Uncanny_Harry
@TheReelRobot
@NeuralViz



5/27
@dlennard
😄 awesome job!



6/27
@EccentrismArt
Thank you!



7/27
@TashaCaufield
Lol, interviewing next to each other is a legit complaint 😅😅😅... Very nice... 👏🏽👏🏽👏🏽



8/27
@EccentrismArt
Haha right!? Thanks Tasha



9/27
@dankchungusYT
youtube?



10/27
@EccentrismArt
Working on it. I will probably combine part 1&2 and post it on YouTube. Also do an upscale with topaz



11/27
@Iamtomblake
Another one well done! 😆👏🏼



12/27
@EccentrismArt
Thanks Tom! Appreciate it. Hope someday Runway has Eleven Labs speech change built right into Act-One



13/27
@sergiosuave23
This is awesome and hilarious! 💯🤣



14/27
@EccentrismArt
Haha thanks! Glad you like it. Did you watch Part 1?



15/27
@HBCoop_
I like part 2 :smile:



16/27
@EccentrismArt
Thanks again and thank you for the repost 🙏🏻 Working on a Christmas one now stay tuned 😆



17/27
@madpencil_
Well done my friend 👍😄 👏



18/27
@EccentrismArt
Thought I replied to you! Thank you appreciate it my friend



19/27
@MrDavids1
RIP Stabby The Ass Kicker, hahaha had a good laugh, nice one Jer!



20/27
@EccentrismArt
Thanks Travis!



21/27
@JyeBeats
That's glorious bro, well done.



22/27
@DreamStarter_1
Haha,good one!



23/27
@empath_angry
This is just so fabulous.

I don't laugh this much at proper TV sitcoms. 🤣🤣🤣



24/27
@levaiteart
What ai you used



25/27
@BLVCKLIGHTai
Lfg!!!



26/27
@Alterverse_AI
Yep, that's a series alright!🔥🔥



27/27
@stonekaiju
So good!!!! You kept to your word, the King got his revenge 😉 😂 🧡 🔥




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
57,329
Reputation
8,496
Daps
159,976







1/12
@guicastellanos1
I had the chance to play with the new @LumaLabsAI /search?q=#DreamMachine a bit in advance, and I have to tell you that is a very powerful tool to let your imagination have a wild ride!
very intuitive and easy to use, here it is my first piece made it with it, Hope you will like it 😶‍🌫️😉



https://video.twimg.com/ext_tw_video/1861307253002760192/pu/vid/avc1/1300x720/GyENXdVUJOv0-seV.mp4

2/12
@guicastellanos1
Thanks Rob, that’s a good idea I will try as well. Still exploring the possibilities!



3/12
@techhalla
Epic



4/12
@guicastellanos1
Graaaacias Luis, tu feedbak es muy importante para mi, aprendo muchisimo de tus post!!



5/12
@ernst320
I like how the voice matches the story and the objects morph into different things as the camera perspective shifts. Very cool effects.



6/12
@guicastellanos1
Thanks Ernst. I tried a couple of voices with elevenlabs and cutting the tale by sentences worked better with the pace. And for the effects it was with keyframes in luma playing with the camera controls to make it more fluid. 🙏



7/12
@HendoAi
Stylish camera pulls dude

Looks awesome



8/12
@guicastellanos1
Thaanks Hendo I’m loving this new dream machine!



9/12
@amazonjeff67
Dream !



10/12
@guicastellanos1
Yeees it has that feeling right?



11/12
@LudovicCreator
Nice work Guillermo



12/12
@guicastellanos1
Thaanks Ludovic. How do you like the new Dream Machine? Me I love 2 generations at the same time and the board to review it, is very helpful!




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
57,329
Reputation
8,496
Daps
159,976






1/11
@Alterverse_AI
I had the privilege of exploring the all-new Dream Machine update from @LumaLabsAI, and I can’t overstate how impressive it is!
Think of it as your personal AI creative partner for image and video generation on the platform.

The best part? It’s available now for everyone!



https://video.twimg.com/ext_tw_video/1861085219412942850/pu/vid/avc1/1920x1080/NLKmGxhjxrllAhM7.mp4

2/11
@Kyrannio
Just gorgeous!! I love the jellyfish :smile: motion is stunning!!!



3/11
@Alterverse_AI
Hey Kiri! Thanks, glad you liked it! I love abyssal creatures! The motion on the new model is breathtaking honestly! Super honored to be part of @LumaLabsAI creative partners!



4/11
@_lev3lup
Wow, that looks awesome!



5/11
@Alterverse_AI
Thanks! I really love the amount of motion the new model can output!



6/11
@JeepersMedia
It’s like being in a Dream, and you are!



7/11
@Alterverse_AI
Thank you! It is like you are conjuring a dream!



8/11
@majdoubTn
I agree it's really wonderful



9/11
@Alterverse_AI
🙏 thank you for the kind words!



10/11
@guicastellanos1
Wooow Epic and the music is perfect



11/11
@Alterverse_AI
❤️ thank you, Guillermo!!! Glad you liked it!




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
57,329
Reputation
8,496
Daps
159,976



1/11
@venturetwins
I don't think most people understand how far AI video has come this year.

12 months ago, we couldn't get a few seconds of coherent motion.

Today, we have consistent characters with controllable movement and realistic facial expressions.

(from u/Kind_Distance9504 on Hailuo)



https://video.twimg.com/ext_tw_video/1863664582906220548/pu/vid/avc1/720x720/EoOSQXY5i0udVDxc.mp4

2/11
@umesh_ai
I am the original creator of this video.

[Quoted tweet]
How AI sees X payout reactions!


https://video.twimg.com/ext_tw_video/1855079130599182336/pu/vid/avc1/720x720/RFYnS2_lHzA50DUH.mp4

3/11
@venturetwins
amazing!!



4/11
@UnityEagle
Why didn’t you mention the original creator ? Or repost his original such a shame big account keep doing this



5/11
@venturetwins
Hey! I saw it on Reddit and credited the creator there in my post.

Unfortunately there’s no way for me to know if that creator took it from someone else who originally posted it on another platform 😔



6/11
@Scobleizer
You are right.

Most of my friends and family have zero clues about what is going down.

You all do because you are following my AI Artists and AI Filmmaker lists.

The next 12 months are gonna be insane in music, and AI-driven creative tools. (Watch my Holodeck list to see it all go down).



7/11
@123wimi
like 😍



8/11
@koltregaskes
And we get 3D world generators from images today. Crazy times. :-)



9/11
@TheErikLens
And in one year?



10/11
@AngieCordeiro57
In a manner is speaking AI video animation has become the next level, or should I say, the next dimension of animation.



11/11
@SadiaUme1
Agreed—seeing this evolution is mind-blowing.

Can’t wait to explore more of these capabilities! 🙌




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
57,329
Reputation
8,496
Daps
159,976

1/70
@RuudNL
Sora v2 release is impending:

* 1-minute video outputs
* text-to-video
* text+image-to-video
* text+video-to-video

OpenAI's Chad Nelson showed this at the C21Media Keynote in London. And he said we will see it very very soon, as @sama has foreshadowed.



https://video.twimg.com/ext_tw_video/1865424798098063360/pu/vid/avc1/1280x720/o2CgvaQxW-BwzuLv.mp4

2/70
@crownut984
Bro no one has access to to v1 why should I care about v2



3/70
@AI_Car_Pool
Monday - noon EST - 200$ per plan



4/70
@MicahJanke
Here’s to hoping the price isn’t insane



5/70
@caojiaming1
Looking forward to it!



6/70
@Goran_Majic
Put all the OpenAI employees in a room with all the computing power in the world, but no access to the internet and anyone else's content. Let them rebuild from scratch. It might be interesting to see what that would look like.



7/70
@clay_ripma
The convergence of this + VR is going to open up an entirely new category of personalized entertainment.

Most big releases are already 2 hours of CGI with a thin plot. Yawn.

But put me in there in VR where I can change around the story? It’ll be like lucid dreaming. I’m in.



8/70
@Nexus_AiAcademy
this is wild



9/70
@majdoubTn
Definitely need this 😁



10/70
@TB_Motion
Holy s... I need more storage space 🤪😱



11/70
@pratjoey
Text and image to video will be a rich area.



12/70
@LittleTinRobot
I remain optimistic as have been waiting around this tech for a while now to test it out.
All sounds very intriguing.



13/70
@creativstrm
so there was a Sora V1? probably in another dimension... and even if a sora v2 will exist, it will be for professional users. So, whatever.



14/70
@boneGPT
big



15/70
@MaxZiebell
This is impressive—when is this from? Either it’s highly cherry-picked, or they’ve been hard at work. I had written off Sora as being overtaken, but could this mark a major comeback?



16/70
@TheAI_Frontier
It will be on day 3, I can feel it.



17/70
@koltregaskes
Looks great but I'm wondering release to who? Hollywood only?



18/70
@ZMadoc
Can’t wait!



19/70
@StevieMac03
V2? What happened to v1?



20/70
@jphilipp
Here's hoping for them to eventually release a lipsync tool, too... we need higher-quality and more face-detecting ones.



21/70
@MotionMark1111
Whoa



22/70
@rchitectopteryx
Where is v1?



23/70
@BenPielstick
If we can get 1 minute continuous shots and consistent characters between shots we are going to have a very interesting year.



24/70
@U79916881747113
Thats a big screen and the quality looks highly impressive so it must be 1080p or may be 2K and also the frame rate seems to be high so may be 24fps👍



25/70
@PlutusCosmos
Looking forward to it!



26/70
@iamdeepaklenka
Wow,it’s just incredible



27/70
@thegenioo
stunning and amazing



28/70
@imagineadil
wow



29/70
@PJaccetturo
Now I’m hyped.



30/70
@Emily_Escapor
I am ready 😊 the prompt is ready 🙏



31/70
@azed_ai
Wild!



32/70
@edalgomezn
@nicdunz



33/70
@bilawalsidhu
Okay this has me hyped — the military characters and spaceship generations are right up my alley



34/70
@NeuralHours
Incredible



35/70
@AliceAI314
I genuinely can't wait to get my hands on that. Look at the face consistency where the blood is splattered on her nose. It's identical to the other shot. Sora is looking stunning. And the biggest thing for me is little to no morphing and consistent characters.



36/70
@Yang_ML_Estate
Currently I have been using Kling. Will try Sora v2 after its release



37/70
@nebsh83
Bet they will drop it for xmas



38/70
@ChukwumaNwaugha
Wow, this is super impressive.



39/70
@mark_k
Very cool, thanks for posting!



40/70
@techguyver
What about API availability?



41/70
@Marconium2
What do you mean v2? We never got to see v1 lol



42/70
@DreschHorbach
Will it be pro only?



43/70
@Eric520CC
Amazing, very much looking forward to SORA V2, will it be released tomorrow?



44/70
@rajeshkannan_mj
I need to check it live before trusting them. Last time their demo was very cherry pick though. Only time will tell about the quality.



45/70
@AntiZionist1917
I think Sora is also image generator. Maybe it's just frame of videos that are acting as image generation. But it has some form of image generation.



46/70
@yc3t_
holy



47/70
@D41Dev
I've been waiting to see text+video-to-video models, this is where significant leaps can be made.

- Full length movies can be made by inserting script snippets and a 1min clip of the previous scene. This can be done in a loop.

- Enhanced and simplified video editing



48/70
@seeupsidedown
LET. US. COOK.



49/70
@ShivOperator
Gradually, then suddenly



50/70
@GinChou
@HuoQubot



51/70
@AlexanderNL
haha hoefde je geen NDA te tekenen?



52/70
@emon_whatever
holy! another big bomb is about to drop



53/70
@AI_Car_Pool


[Quoted tweet]
Is it finally coming? @OpenAI


GePR23WWQAAdIY6.jpg


54/70
@bowtiedwhitebat
we bout to end the woke in movies



55/70
@Emily_Escapor
I am ready 😁



56/70
@evren_mercan
Unlimited 2000 $/month 😁



57/70
@SomBh1
Crazy. Imagine on a million GPU cluster. Can make a 15 min short movie.



58/70
@crypto_sam_974
who cares if it is not released...



59/70
@BreezeChai
Crazy



60/70
@_NamelessBrain_
Sora Turbo, not v2. Very different beasts.



61/70
@api_prlabs
Unlock the power of advanced AI with PR Labs' ChatGPT4 API! 🚀 Starting at just $5/month with up to 2.4M requests,it's perfect for developers and businesses aiming to scale effortlessly.
Check now:ChatGPT 4
website: PR LABS API
/search?q=#ChatGPT4 /search?q=#OpenAI /search?q=#NewsAPI



62/70
@nextdao




GeUpFJfacAIoBrC.jpg

GeUpFJkaUAA8K3t.jpg

GeUpFJlasAAW7iL.jpg


63/70
@PrescienceDS
Exciting to see the Sora v2 release on the horizon! The addition of text-to-video and text+image-to-video features will definitely open up new creative possibilities. Can't wait to see how this will impact content creation and storytelling!



64/70
@YolaoDude
Looks great. But if the price tag is high, $200 or so, then i will jump in IF...
-It can generate FAR more than just 5 to 10 seconds videos!
-you can have full control of the movement of the characters, video to video.
-the character are consistent.



65/70
@aidesignss
sora v1 has not been released yet



66/70
@JogishE45866
Wonderful



67/70
@nextdao
http://OpenAI.chathttp://OpenAI.com

http://OpenAI.techhttp://OpenAI.com

http://OpenAI.toolshttp://OpenAI.com

Now only two domain names for sale.
@sama @OpenAI

🔶openai.im is available for purchase - Sedo.com

🔷openai.gg is available for purchase - Sedo.com

/search?q=#OpenAI 🎉



68/70
@Opic3D
Cool! 😎



69/70
@farzadhss
Is it open source?



70/70
@irrealer
wild.




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
57,329
Reputation
8,496
Daps
159,976






1/11
@_philschmid
WTF?! New open-source physics AI engine absolutely insane! 🤯 Genesis is a new physics engine that combines ultra-fast simulation with generative capabilities to create dynamic 4D worlds for robotics and physics.

TL;DR:
🚀 430,000x faster than real-time physics simulation, processes 43M FPS on a single RTX 4090
🐍 Built in pure Python, 10-80x faster than existing GPU solutions like Isaac Gym
🌐 Cross-platform support: Linux, MacOS, Windows, with CPU, NVIDIA, AMD, and Apple Metal backends
🧪 Unified framework combining multiple physics solvers: Rigid body, MPM, SPH, FEM, PBD, Stable Fluid
🤖 Extensive robot support: arms, legged robots, drones, soft robots; supports MJCF, URDF, obj, glb files
🎨 Built-in photorealistic ray-tracing rendering
⚡ Takes only 26 seconds to train real-world transferrable robot locomotion policies
💻 Simple installation via pip: pip install genesis-world
🤝 Physics engine and simulation platform are fully open-sourced
🔜 ”.generate” method/generative framework coming soon.



https://video.twimg.com/ext_tw_video/1869639134156840960/pu/vid/avc1/1280x720/rffXgweJaKVMxGFQ.mp4

2/11
@_philschmid
Code: GitHub - Genesis-Embodied-AI/Genesis: A generative world for general-purpose robotics & embodied AI learning.
Documentation: Genesis — Genesis 0.2.0 documentation

GitHub - Genesis-Embodied-AI/Genesis: A generative world for general-purpose robotics & embodied AI learning.



3/11
@ma5701458
@pikaso_me screenshot this



4/11
@pikaso_me
Enjoy your screenshot and...

Get 10 stickers for $1: $1 for 10 | Sticker Mule



GfKZj-vWEAAOwQr.jpg


5/11
@PamirSevincel
this is insane. you probably won't need that much real world data to train physical systems e.g. humanoid robots.

@adcock_brett what do you think?



6/11
@Mooie_89
@khouuba 👀



7/11
@BalajiAkiri
A new era of Robotics begins



8/11
@360
Reality breaks more every day



9/11
@AIBuzzNews
@desy @desynews das könnte spannend und auch nützlich sein.



10/11
@glaksmono
This is super cool 🔥



11/11
@EducatingwithAI
Generative AI has hit a wall 😂

In before: but this is not pure GenAI.

No, but it uses it with other tools - exactly what @ilyasut mentioned.











1/11
@EHuanglu
This New AI Just Nuked the Competition

Genesis AI is here to wipe out every outdated simulation
and totally dominate AI world.

- Open sourced
- Create data from prompt
- 430,000x faster than real-time
- Create 4D physical worlds in seconds with text prompt

Let's dive in:



https://video.twimg.com/ext_tw_video/1869694710941618176/pu/vid/avc1/1280x720/dZ67k0pVaJpn20Ny.mp4

2/11
@EHuanglu
1. Generating 4D dynamical & physical world

• VLM-based generative agent
• Worlds provide diverse data modalities.
• Generates physically accurate, view consistent videos with camera and object motion.



https://video.twimg.com/ext_tw_video/1869709610762055682/pu/vid/avc1/1280x720/vznCJHqYHQh66u3C.mp4

3/11
@EHuanglu
2. Precise control

• Provides complete control over physics
• Handles camera motion and parameters with precision



https://video.twimg.com/ext_tw_video/1869709664302313472/pu/vid/avc1/1280x720/AnS2VyDilYPHlN2a.mp4

4/11
@EHuanglu
3. Character Motion Generation

• A Chinese soldier performs the Gangnam Style dance
• A Japanese samurai performs boxing
• A Roman soldier walks forward like a zombie



https://video.twimg.com/ext_tw_video/1869709732405182464/pu/vid/avc1/1280x720/tSN-btsf_6i4GXb_.mp4

5/11
@EHuanglu
4. Robotic Policy Generation

• Generative robotic agent + advanced physics engine
• Generates demonstration data for diverse skills
• Enables robots to learn and adapt autonomously.
• Powered by AI models for next-level automation.



https://video.twimg.com/ext_tw_video/1869709846528110592/pu/vid/avc1/1280x720/3fM-cF-Dmh5Ta9As.mp4

6/11
@EHuanglu
5. 3D & Fully Interactive Scene Generation

"A home interior scene with a living room (including a dinning space), a restroom, a study and a bedroom."



https://video.twimg.com/ext_tw_video/1869709919198515200/pu/vid/avc1/1280x720/frU1kO_8cGhMx4Fz.mp4

7/11
@EHuanglu
6. Speech Audio, Facial Animation & Emotion Generation



https://video.twimg.com/ext_tw_video/1869709968406142976/pu/vid/avc1/1280x720/LOXAphf2Ddtic8v1.mp4

8/11
@EHuanglu
Open sourced code:
GitHub - Genesis-Embodied-AI/Genesis: A generative world for general-purpose robotics & embodied AI learning.

Project Page:
https://genesis-embodied-ai.github.io



9/11
@EHuanglu
If you enjoyed reading this post,

Follow @EHuanglu for more great stuff!

And support with like/repost the post below 👇

[Quoted tweet]
This New AI Just Nuked the Competition

Genesis AI is here to wipe out every outdated simulation
and totally dominate AI world.

- Open sourced
- Create data from prompt
- 430,000x faster than real-time
- Create 4D physical worlds in seconds with text prompt

Let's dive in:
[media=twitter]1869710532934254954[/media]

https://video.twimg.com/ext_tw_video/1869694710941618176/pu/vid/avc1/1280x720/dZ67k0pVaJpn20Ny.mp4

10/11
@zeng_wt
This is just 2024. Not even 2025. And we have this! wow



11/11
@EHuanglu
I always thought 2025 is the year of 3D, it seems I was short sighted 😀














1/11
@zhou_xian_
Everything you love about generative models — now powered by real physics!

Announcing the Genesis project — after a 24-month large-scale research collaboration involving over 20 research labs — a generative physics engine able to generate 4D dynamical worlds powered by a physics simulation platform designed for general-purpose robotics and physical AI applications.

Genesis's physics engine is developed in pure Python, while being 10-80x faster than existing GPU-accelerated stacks like Isaac Gym and MJX. It delivers a simulation speed ~430,000 faster than in real-time, and takes only 26 seconds to train a robotic locomotion policy transferrable to the real world on a single RTX4090 (see tutorial: https://genesis-world.readthedocs.io/en/latest/user_guide/getting_started/locomotion.html).

The Genesis physics engine and simulation platform is fully open source at GitHub - Genesis-Embodied-AI/Genesis: A generative world for general-purpose robotics & embodied AI learning.. We'll gradually roll out access to our generative framework in the near future.

Genesis implements a unified simulation framework all from scratch, integrating a wide spectrum of state-of-the-art physics solvers, allowing simulation of the whole physical world in a virtual realm with the highest realism.

We aim to build a universal data engine that leverages an upper-level generative framework to autonomously create physical worlds, together with various modes of data, including environments, camera motions, robotic task proposals, reward functions, robot policies, character motions, fully interactive 3D scenes, open-world articulated assets, and more, aiming towards fully automated data generation for robotics, physical AI and other applications.

Open Source Code: GitHub - Genesis-Embodied-AI/Genesis: A generative world for general-purpose robotics & embodied AI learning.
Project webpage: https://genesis-embodied-ai.github.io/

Documentation: https://genesis-world.readthedocs.io/

1/n



https://video.twimg.com/ext_tw_video/1869510892880097280/pu/vid/avc1/1280x720/_dc5zlB1WmdE01Jh.mp4

2/11
@zhou_xian_
Nvidia brought GPU acceleration to robotic simulation, speeding up simulation speed by more than one order of magnitude compared to CPU-based simulation. This brought numerous amazing robotic skills to life by leveraging large-scale GPU-parallelized simulation. Genesis pushes up this speed by another order of magnitude. Note that the speed improvement is achieved with no compromise in simulation accuracy.
2/n



https://video.twimg.com/ext_tw_video/1869506856051261441/pu/vid/avc1/1280x720/F5sz3RnivNvuO6Dj.mp4

3/11
@zhou_xian_
Genesis supports simulating various types of physical phenomena. We developed from scratch a unified physics engine that integrates various SOTA physics solvers (MPM, SPH, FEM, Rigid Body, PBD, etc.), supporting simulation of a wide range of materials: rigid body, articulated body, Cloth, Liquid, Smoke, Deformables, Thin-shell materials, Elastic/Plastic Body, Robot Muscles, etc.

3/n



https://video.twimg.com/ext_tw_video/1869507363788541952/pu/vid/avc1/1280x720/NEU8FPHKy4Z_bAk6.mp4

4/11
@zhou_xian_
Genesis is the first-ever platform providing comprehensive support for soft muscles and soft robot and their interaction with rigid robots. Genesis also ships with a URDF-like soft-robot configuration system.

4/n



https://video.twimg.com/ext_tw_video/1869512601123594241/pu/vid/avc1/1280x720/UUrRbSSpOXEOIhY_.mp4

5/11
@zhou_xian_
Genesis's generative framework supports generating 3D and fully interactive scenes for training robotic skills
5/n



https://video.twimg.com/ext_tw_video/1869512781008973825/pu/vid/avc1/1280x720/62RDqV516piUnQM5.mp4

6/11
@zhou_xian_
Our generative agent autonomously proposes robotic tasks, design environments, write reward functions, and ultimately leading to automated generation of robotic policies.

6/n



https://video.twimg.com/ext_tw_video/1869512976299864064/pu/vid/avc1/1280x720/0YflDqvt8WnNtXNj.mp4

7/11
@zhou_xian_
Genesis's generative framework supports data generation beyond robotics, such as character motion:

7/n



https://video.twimg.com/ext_tw_video/1869513206672117760/pu/vid/avc1/1280x720/-sNINHa-_cgzgdiI.mp4

8/11
@zhou_xian_
Genesis's GPU parallellized IK solver is able to solve IK for 10,000 Franka arms simultaneously, under 2ms:

8/n



https://video.twimg.com/ext_tw_video/1869513330584338432/pu/vid/avc1/1280x720/JzGz03g2Il-_wzLD.mp4

9/11
@zhou_xian_
We support native non-convex collision handling:

9/n



https://video.twimg.com/ext_tw_video/1869513464055492608/pu/vid/avc1/1280x720/B6sCV4PMyOtFHggC.mp4

10/11
@zhou_xian_
Genesis supports a physically accurate tactile sensing simulation module: https://github.com/Genesis-Embodied-AI/DiffTactile. (Will be integrated into the main branch in a future release soon)

10/n



11/11
@zhou_xian_
Finally, a cute interactive physical Tetris game made with Genesis :smile:

Thanks to all the amazing collaborators who together made everything possible over the last two years! There's no space here to @ every single one, but a huge kudos to the whole Genesis team!

We welcome everyone from the open-source community to come join us and build Genesis with us together!

11/11



https://video.twimg.com/ext_tw_video/1869514325922058240/pu/vid/avc1/1280x720/kOCJbJ4KQnUJw1ht.mp4
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
57,329
Reputation
8,496
Daps
159,976

A new, uncensored AI video model may spark a new AI hobbyist movement​


Will Tencent's "open source" HunyuanVideo launch an at-home "Stable Diffusion" moment for uncensored AI video?


Benj Edwards – Dec 19, 2024 10:50 AM |
112

Still images from three videos generated with Tencent's HunyuanVideo.



Still images from three videos generated with Tencent's HunyuanVideo. Credit: Tencent


The AI-generated video scene has been hopping this year (or twirling wildly, as the case may be). This past week alone we've seen releases or announcements of OpenAI's Sora, Pika AI's Pika 2, Google's Veo 2, and Minimax's video-01-live. It's frankly hard to keep up, and even tougher to test them all. But recently, we put a new open-weights AI video synthesis model, Tencent's HunyuanVideo, to the test—and it's surprisingly capable for being a "free" model.

Unlike the aforementioned models, HunyuanVideo's neural network weights are openly distributed, which means they can be run locally under the right circumstances (people have already demonstrated it on a consumer 24 GB VRAM GPU) and it can be fine-tuned or used with LoRAs to teach it new concepts.

Notably, a few Chinese companies have been at the forefront of AI video for most of this year, and some experts speculate that the reason is less reticence to train on copyrighted materials, use images and names of famous celebrities, and incorporate some uncensored video sources. As we saw with Stable Diffusion 3's mangled release, including nudity or pornography in training data may allow these models achieve better results by providing more information about human bodies. HunyuanVideo notably allows uncensored outputs, so unlike the commercial video models out there, it can generate videos of anatomically realistic, nude humans.

Putting HunyuanVideo to the test​

To evaluate HunyuanVideo, we provided it with an array of prompts that we used on Runway's Gen-3 Alpha and Minimax's video-01 earlier this year. That way, it's easy to revisit those earlier articles and compare the results.

We generated each of the five-second-long 864 × 480 videos seen below using a commercial cloud AI provider. Each video generation took about seven to nine minutes to complete. Since the generations weren't free (each cost about $0.70 to make), we went with the first result for each prompt, so there's no cherry-picking below. Everything you see was the first generation for the prompt listed above it.



"A highly intelligent person reading 'Ars Technica' on their computer when the screen explodes"


"commercial for a new flaming cheeseburger from McDonald's"


"A cat in a car drinking a can of beer, beer commercial"


"Will Smith eating spaghetti"


"Robotic humanoid animals with vaudeville costumes roam the streets collecting protection money in tokens"


"A basketball player in a haunted passenger train car with a basketball court, and he is playing against a team of ghosts"


"A beautiful queen of the universe in a radiant dress smiling as a star field swirls around her"


"A herd of one million cats running on a hillside, aerial view"


"Video game footage of a dynamic 1990s third-person 3D platform game starring an anthropomorphic shark boy"




"A muscular barbarian breaking a CRT television set with a weapon, cinematic, 8K, studio lighting"


"A scared woman in a Victorian outfit running through a forest, dolly shot"


"Low angle static shot: A teddy bear sitting on a picnic blanket in a park, eating a slice of pizza. The teddy bear is brown and fluffy, with a red bowtie, and the pizza slice is gooey with cheese and pepperoni. The sun is setting, casting a golden glow over the scene"


"Aerial shot of a small American town getting deluged with liquid cheese after a massive cheese rainstorm where liquid cheese rained down and dripped all over the buildings"


Also, we added a new one: "A young woman doing a complex floor gymnastics routine at the Olympics, featuring running and flips."



Weighing the results​

Overall, the results shown above seem fairly comparable to Gen-3 Alpha and Minimax video-01, and that's notable because HunyuanVideo can be downloaded for free, fine-tuned, and run locally in an uncensored way (given the appropriate hardware).

There are some flaws, of course. The vaudeville robots are not animals, the cat is drinking from a weird transparent beer can, and the man eating spaghetti is obviously not Will Smith. There appears to be some celebrity censorship in the metadata/labeling of the training data, which differs from Kling and Minimax's AI video offerings. And yes, the gymnast has some anatomical issues.



Right now, HunyuanVideo's results are fairly rough, especially compared to the state-of-the-art video synthesis model to beat at the moment, the newly-unveiled Google Veo 2. We ran a few of these prompts through Sora as well (more on that later in a future article), and Sora created more coherent results than HunyuanVideo but didn't deliver on the prompts with much fidelity. We are still early days of AI, but quality is rapidly improving while models are getting smaller and more efficient.

Even with these limitations, judging from the history of Stable Diffusion and its offshoots, HunyuanVideo may still have significant impact: It could be fine-tuned at higher resolutions over time to eventually create higher-quality results for free that may be used in video productions, or it could lead to people making bespoke video pornography, which is already beginning to appear in trickles on Reddit.

As we've mentioned before in previous AI video overviews, text-to-video models work by combining concepts from their training data—existing video clips used to create the model. Every AI model on the market has some degree of trouble with new scenarios not found in their training data, and that limitation persists with HunyuanVideo.

Future versions of HunyuanVideo could improve with better prompt interpretation, different training data sets, increased computing power during training, or changes in the model design. Like all AI video synthesis models today, users still need to run multiple generations to get desired results. But it looks like the “open weights” AI video models are already here to stay.[]
 
Top