Welp, I spent all day yesterday messing with this, here are some collected thoughts.
One, it's amazing that I experienced two firsts on this site: Aurora Borealis AND the first fat-compatible Hunyuan video lora were both posted here, amazing.
As for the rest. Well, wow. It's incredible that this tech has filtered down so quickly from giant super computer and subscription required to kling or pixverse or whatever to consumer machines. Does Huyuan completely blow other, older methods of animation like Pyramidflow, CogVideo, and AnimatedDiff out of the water? Absolutely. Is it as good as Kling/Pixverse/etc? Ehh, on mid-grade consumer hardware, I'd say no. Maybe it can compete on 24gb VRAM and above. Maybe.
And that brings me to the real kind of "downer" here: me spending 8 hours running this stuff (with tons of fails and tests) probably cost me around $8 in electricity total where I live, and I have to make a lot of concessions: small initial res, conservative upscale, no interpolation, long (5+ minute) generation times for 3 seconds clips. The reality is, it would have been cheaper, and faster, and would have had better looking results, if I had rented a runpod to do this experimenting than to run it at home on what was, four years ago, the second best GPU on the market (3080 10gb). So that's a stark look at how quickly this tech has evolved and how fast the tech a lot of people have sitting around is beginning to get outclassed. Not just in video but most actually good LLMs as well don't really fit in 10GB. Thankfully it's so cold right now that any heat I make from generating stuff is probably just money saved by causing the heat pump to not run as much (both are 600w when running with load), so it *probably* evens out, but still.
As for the examples, I am thrilled that already, even in these infancy stages, it's possible to combine loras in Hunyuan video and still get a decent result. That means that eventually, just like every other model, character loras will be coming to Hunyuan at an accelerated rate, and eventually some amazing things will come of it. As of RIGHT THIS SECOND though, unless you're renting a runpod or have a 3090 or above, I would say this is *mostly* just a fun toy, and nobody should go out of their way to make this work on their home machines unless they have cheap electricity and a beefy (nvidia) GPU. If anyone wanted to donate some money to the site specifically for a BBWChan runpod for Hunyuan, I'd certainly set it up and take requests, but as of now these little Jinx clips will probably be the only "I was here when it happened" examples from me. Fascinating tech. Thanks for cluing us in, and great work on that first lora drewski. Way to take the initiative.
For anyone curious, I just ran a slightly modified version of this workflow: https://civitai.com/models/1134115/hunyuan-yawyet-another-workflow-fastt2vv2vmulti-resaudiopreview-pauseupscalemulti-lora-aio-interpolateteacacheoptions
To make it work on 10GB below, just make sure to us the lowest initial res it comes with (might even have to reduce that a bit on 8GB cards), disable interpolation and audio, only use the intermediate V2V upscale and not the better upscaler. Oh and change both areas with the the TileSize parameter from 256 to 128, and the overlap from 64 to 32.
Edited last time by admin on 01/17/2025 (Fri) 13:31:10.