Yall dont understand: Sora is going to be a GAME CHANGER for fan fiction.

bnew

Veteran
Joined
Nov 1, 2015
Messages
63,932
Reputation
9,808
Daps
174,252
Extensive list of generative tools curated by Eyal Gruss

 

bnew

Veteran
Joined
Nov 1, 2015
Messages
63,932
Reputation
9,808
Daps
174,252
Introducing Runway Gen-4 | Runway



Channel Info Runway Subscribers: 98.2K subscribers

Description
Introducing Runway Gen-4: Our next-generation series of state of the art AI models for media generation and world consistency.

A new generation of consistent and controllable media is here.

With Gen-4, you are now able to precisely generate consistent characters, locations and objects across scenes. Simply set your look and feel and the model will maintain coherent world environments while preserving the distinctive style, mood and cinematographic elements of each frame. Allowing you to regenerate those elements from multiple perspectives and positions within your scenes.

Get started at runwayml.com
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
63,932
Reputation
9,808
Daps
174,252
gpt-4o mini

script to calculate cost of generating a video on kling based on 4/19/2025 info.


Python:
import readline

def calculate_movie_cost(movie_length_seconds, credits_per_5_seconds, credits_per_month, additional_credit_cost):
    # Calculate total credits needed for the movie
    total_credits = (movie_length_seconds / 5) * credits_per_5_seconds
    
    # Calculate additional credits needed if total credits exceed monthly credits
    additional_credits_needed = max(0, total_credits - credits_per_month)
    
    # Calculate total cost in dollars
    total_cost = (additional_credits_needed / 100) * additional_credit_cost
    
    return total_credits, total_cost

def convert_to_seconds(length, unit):
    if unit == 'hours':
        return length * 3600
    elif unit == 'minutes':
        return length * 60
    elif unit == 'seconds':
        return length
    else:
        raise ValueError("Invalid time unit. Please use 'hours', 'minutes', or 'seconds'.")

def main():
    # Constants for each plan with explanations
    plans = {
        "Basic": {
            "credits_per_month": 166,
            "additional_credit_cost": 0,  # Free plan does not allow additional credits
            "explanation": "Best for casual users who want to explore Kling AI without committing to a paid plan."
        },
        "Standard": {
            "credits_per_month": 660,
            "additional_credit_cost": 1.06,
            "explanation": "Ideal for content creators and digital artists who need better AI-generated visuals with faster processing."
        },
        "Pro": {
            "credits_per_month": 3000,
            "additional_credit_cost": 0.87,
            "explanation": "Perfect for professional designers and video creators who need faster processing and more credits."
        },
        "Premier": {
            "credits_per_month": 8000,
            "additional_credit_cost": 0.81,
            "explanation": "Designed for high-level professionals and studios needing large amounts of high-quality AI content with exclusive perks."
        }
    }

    credits_per_5_seconds = 100  # Credits for every 5 seconds

    # User input for movie length
    print("Enter the length of the movie:")
    length = float(input("Length (e.g., 2.5 for 2 hours 30 minutes, 120 for 120 minutes): "))

    # Define possible units and set up readline for autocomplete
    units = ['hours', 'minutes', 'seconds']
    readline.set_startup_hook(lambda: readline.insert_text('hours'))  # Default suggestion
    readline.parse_and_bind("tab: complete")
    readline.set_completer(lambda text, state: [unit for unit in units if unit.startswith(text)][state])

    unit = input("Unit (hours/minutes/seconds): ").strip().lower()
    readline.set_startup_hook()  # Clear the startup hook

    # Convert the input length to seconds
    try:
        movie_length_seconds = convert_to_seconds(length, unit)
    except ValueError as e:
        print(e)
        return

    # Calculate and display results for each plan
    for plan_name, plan_details in plans.items():
        print(f"\n--- {plan_name} Plan ---")
        print(f"Explanation: {plan_details['explanation']}")
        total_credits, total_cost = calculate_movie_cost(
            movie_length_seconds,
            credits_per_5_seconds,
            plan_details["credits_per_month"],
            plan_details["additional_credit_cost"]
        )
        
        print(f"Movie Length: {length} {unit}")
        print(f"Total Credits Required: {total_credits:.2f} credits")
        print(f"Total Cost: ${total_cost:.2f}")

if __name__ == "__main__":
    main()
 
Last edited:

bnew

Veteran
Joined
Nov 1, 2015
Messages
63,932
Reputation
9,808
Daps
174,252
New layer addition to Transformers radically improves long-term video generation



Posted on Tue Apr 8 15:30:23 2025 UTC


Fascinating work coming from a team from Berkeley, Nvidia and Stanford.

They added a new Test-Time Training (TTT) layer to pre-trained transformers. This TTT layer can itself be a neural network.

The result? Much more coherent long-term video generation! Results aren't conclusive as they limited themselves to a one minute limit. But the approach can potentially be easily extended.

Maybe the beginning of AI shows?

Link to repo:
One-Minute Video Generation with Test-Time Training

One-Minute Video Generation with Test-Time Training


Abstract​


Transformers today still struggle to generate one-minute videos because self-attention layers are inefficient for long context. Alternatives such as Mamba layers struggle with complex multi-scene stories because their hidden states are less expressive. We experiment with Test-Time Training (TTT) layers, whose hidden states themselves can be neural networks, therefore more expressive. Adding TTT layers into a pre-trained Transformer enables it to generate one-minute videos from text storyboards. For proof of concept, we curate a dataset based on Tom and Jerry cartoons. Compared to baselines such as Mamba 2, Gated DeltaNet, and sliding-window attention layers, TTT layers generate much more coherent videos that tell complex stories, leading by 34 Elo points in a human evaluation of 100 videos per method. Although promising, results still contain artifacts, likely due to the limited capability of the pre-trained 5B model. The efficiency of our implementation can also be improved. We have only experimented with one-minute videos due to resource constraints, but the approach can be extended to longer videos and more complex stories.

Paper

Code

Adding TTT Layers to a Pre-Trained Transformer​


Adding TTT layers into a pre-trained Transformer enables it to generate one-minute videos with strong temporal consistency and motion smoothness.









1/12
@hyperbolic_labs
We’re proud to have supported the team behind One-Minute Video Generation with Test-Time Training with compute infrastructure.

Incredible to see our platform enabling breakthroughs in long-form video generation. Congrats to the authors!

@danielkoceja @GashonHussein @Jerry_XU_Jiarui @__yuezhao__ @jankautz @guestrin @tatsu_hashimoto @sanmikoyejo @YejinChoinka @xiaolonw @karansdalal

[Quoted tweet]
Today, we're releasing a new paper – One-Minute Video Generation with Test-Time Training.

We add TTT layers to a pre-trained Transformer and fine-tune it to generate one-minute Tom and Jerry cartoons with strong temporal consistency.

Every video below is produced directly by the model in a single shot, without editing, stitching, or post-processing. Every story is newly created.

Demos: test-time-training.github.io…
Paper: test-time-training.github.io…


GolAoksW8AA1YCm.png

GolApZAWIAA8wJC.jpg


https://video.twimg.com/ext_tw_video/1909310443530944513/pu/vid/avc1/720x480/S8MsN5qN0o9f_Lnx.mp4

2/12
@hyperbolic_labs
Read the full paper: https://test-time-training.github.io/video-dit/assets/ttt_cvpr_2025.pdf



3/12
@Quangduycbq
so cool i will make meaningful video🥰🥰🥰



4/12
@hyperbolic_labs
love it



5/12
@ChetaOfAllTrade
Incredible, Hyperbolic built to see developers actually reach their potential and not getting stucked by computing resource.

Congrats to the team



6/12
@hyperbolic_labs
🥂



7/12
@ericspo29
So now I can make my own cartoons, this is awesome!



8/12
@hyperbolic_labs
Pretty wild tech



9/12
@Just_marhk
Great 👍👏



10/12
@hyperbolic_labs
💯💯



11/12
@Bruhbears985
That's so great 🤘🏻



12/12
@hyperbolic_labs
amazing what AI can do now




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
















1/22
@karansdalal
Today, we're releasing a new paper – One-Minute Video Generation with Test-Time Training.

We add TTT layers to a pre-trained Transformer and fine-tune it to generate one-minute Tom and Jerry cartoons with strong temporal consistency.

Every video below is produced directly by the model in a single shot, without editing, stitching, or post-processing. Every story is newly created.

Demos: One-Minute Video Generation with Test-Time Training
Paper: http://test-time-training.github.io/video-dit/assets/ttt_cvpr_2025.pdf



https://video.twimg.com/ext_tw_video/1909310443530944513/pu/vid/avc1/720x480/S8MsN5qN0o9f_Lnx.mp4

2/22
@karansdalal
Test-time training (TTT) layers are RNN layers where the hidden state is a machine learning model and the update rule is a step of gradient descent. See this thread for previous work.

[Quoted tweet]
I’m excited to share a project I’ve been working on for over a year, which I believe will fundamentally change our approach to language models.

We’ve designed a new architecture, which replaces the hidden state of an RNN with a machine learning model. This model compresses context through actual gradient descent on input tokens. We call our method “Test-Time-Training layers.”

TTT layers directly replace attention, and unlock linear complexity architectures with expressive memory, allowing us to train LLMs with millions (someday billions) of tokens in context.

Our instantiations, TTT-Linear and TTT-MLP, both match or beat the strongest Transformers and Mamba. Arxiv: arxiv.org/abs/2407.04620


GR-cpVpawAABD38.png


3/22
@karansdalal
Our approach simply adds TTT layers to a pre-trained Diffusion Transformer and fine-tunes it on long videos with text annotations. To keep costs manageable, we limit self-attention to local segments and let TTT (linear complexity) operate globally.



Gn88CAKbwAMyW2Q.jpg


4/22
@karansdalal
We create an “On-Chip Tensor Parallel” algorithm to implement an efficient TTT-MLP kernel. Specifically, we shard the weights of the “hidden state model” across Streaming Multiprocessors, and use the DSMEM feature Hopper GPUs implement AllReduce among SMs.

This avoids costly transfers between global memory (HBM) and shared memory (SMEM), while still fitting the large hidden state into the small amount of fast SMEM.

More details in the paper. Kernel code: GitHub - test-time-training/ttt-tk



Gn88LCwbwAMkVY_.jpg


5/22
@karansdalal
Grateful for wonderful collaborators. This work will be presented at CVPR 2025.

@danielkoceja @GashonHussein @Jerry_XU_Jiarui @__yuezhao__ @jankautz @guestrin @tatsu_hashimoto @sanmikoyejo @YejinChoinka @xiaolonw



Gn89FL3bwAM8v5j.jpg


6/22
@karansdalal
+ our wonderful collaborators without Twitter – Shihao Han, Ka Chun Cheung, Youjin Song, and Yu Sun.



7/22
@menhguin
what the fukk (complimentary)

ok for like a solid 30 seconds I thought this was the Test-Time Training used for the ARC AGI MIT submission and I was rly confused



8/22
@karansdalal
Same thing, different application! Best characterization would be "End to End" vs "Non E2E" test-time training.

Test-Time Training Project Website



9/22
@ruslanjabari
damn and this is only ~50 hours of training runs



10/22
@karansdalal
With a 5B model 🫣



11/22
@reborn_agi
This is incredible work — generating coherent, one-minute-long animated stories with zero post-processing is a huge leap in video generation. The TTT approach looks super promising for maintaining temporal consistency. Huge respect to you and the team.



12/22
@karansdalal
Thank you



13/22
@willdepue
very cool work karan! do you have any baselines of what it looks like without test time training?



14/22
@karansdalal
Thank you Will, sorry to miss this! Here's the World Trade Center video with the local attention baseline* We have some examples of comparing TTT to other RNNs on the project page.

* Disclaimer – this model does have less parameters than the one with added TTT layers.



https://video.twimg.com/ext_tw_video/1909798570049650689/pu/vid/avc1/720x480/0agZ6XihQUKUJ9iC.mp4

15/22
@TheGrizztronic
Pretty cool. TTT should get more love. Hope this helps!



16/22
@karansdalal
🙏



17/22
@jc_stack
Really interested in your pre-training approaches. Have you seen much impact on compute/memory overhead with the TTT layers? Thinking about startup resource constraints here.



18/22
@karansdalal
TTT layers have linear complexity, so long context inference is far better than self-attention. But we still have some way to go on kernel optimization when compared to other modern RNN layers.

Figure 6 from our paper:



Gn9WhFzbwAE84O3.jpg


19/22
@john7rho
Amazing work Karan



20/22
@karansdalal
Thank you John!



21/22
@jam3scampbell
🤔

[Quoted tweet]
in b4 ttt is the new q*


22/22
@nearcyan
hmmmm




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
63,932
Reputation
9,808
Daps
174,252

1/1
@DoctorGoldOval
FantasyTalking: Realistic Talking Portrait Generation via Coherent Motion Synthesis



https://video.twimg.com/amplify_video/1913085247140642816/vid/avc1/540x360/CVrR8DPm6YFRcDqD.mp4


To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196













1/7
@Gradio
FantasyTalking -- Realistic talking portrait generations have never been this good! Checkout the links in the thread for learning more 👇



https://video.twimg.com/ext_tw_video/1917550851037577217/pu/vid/avc1/960x720/jjANKUFGkaZ3QpbT.mp4

2/7
@Gradio
Build Fantasy Talking on your own machine: GitHub - Fantasy-AMAP/fantasy-talking: FantasyTalking: Realistic Talking Portrait Generation via Coherent Motion Synthesis



3/7
@Gradio
App is live on @huggingface: FantasyTalking - a Hugging Face Space by acvlab



4/7
@ericreator
this one is not ready



5/7
@FearCryptoGreed
Looks solid



6/7
@NeuralKnight_
These realistic portraits sound intriguing. Curious to learn about the techniques involved.



7/7
@WendyCarlosa
you do python go away




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196









1/7
@Tech_Transforms
China just dropped a new open-source AI model — FantasyTalking and this could totally reshape how we create videos 👀

See it's amazing work and how it's different from other models here:



2/7
@Tech_Transforms
Realistic lips synchronization

Generate realistic lip synchronization ensuring characters lip movements match the audio provided



https://video.twimg.com/amplify_video/1912112753851949056/vid/avc1/1056x720/psGZlBAYNErtb2j4.mp4

3/7
@Tech_Transforms
Realistic talking videos

Generation of realistic talking videos with varied body types and angles — from close-ups to full-body, front-facing to side views.



https://video.twimg.com/amplify_video/1912112957669916672/vid/avc1/576x720/QvzcEjcxp0x8V8p8.mp4

4/7
@Tech_Transforms
Various Avatars

From humans to cartoons to animated characters, FantasyTalking can create it all seamlessly.



https://video.twimg.com/amplify_video/1912113080890204160/vid/avc1/720x480/k0-uXQg2R42V2-PR.mp4

5/7
@Tech_Transforms
Comparing with other Models

FantasyTalking’s output when compared to OmniHuman-1 with the same source video.



https://video.twimg.com/amplify_video/1912114521356783618/vid/avc1/980x646/U-N9udMUbZruhf2h.mp4

6/7
@Tech_Transforms
Architecture overview

Built on the Wan2.1 video diffusion model, FantasyTalking creates ultra-realistic talking portraits with precise audio-visual alignment.
It ensures identity consistency and natural motion using face-focused modeling and a motion control network.



GokxF69XEAEZpm8.jpg


7/7
@Tech_Transforms
Is FastasyTalking better than other Video Generation models?

Follow @Tech_Transforms for more such updates!




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196















1/13
@ItsKevinNexus
China's Open Source AI is Next-Level 🤯

Alibaba just dropped FantasyTalking — an insane AI that lip-syncs characters with realistic facial and full-body motion.

It outperforms current SOTA methods like OmniHuman-1, Sonic, and Hallo 3.

🔥 Here are 10 jaw-dropping examples you’ve gotta see:



https://video.twimg.com/amplify_video/1911729207111016448/vid/avc1/1066x720/-kPWm_FtHWGiulJX.mp4

2/13
@ItsKevinNexus
1. 🧠 Generated Videos with FantasyTalking

Delivers highly realistic lip-syncing, perfectly matching mouth movements to audio.
Supports a wide range of avatar styles — from realistic to cartoon.
Generates high-quality conversational videos with full facial and body motion.



https://video.twimg.com/amplify_video/1911729287842988032/vid/avc1/480x480/TZhDRuBJjjpSYVzr.mp4

3/13
@ItsKevinNexus
2. 🎥 Realistic Talking Videos

Supports generating lifelike talking videos across multiple body ranges: close-up portraits, half-body, and full-body.
Handles various orientations, including front-facing and side-facing poses — all with natural motion and detail.



https://video.twimg.com/amplify_video/1911729352720461824/vid/avc1/480x600/NvmmcA9L1SJuTjld.mp4

4/13
@ItsKevinNexus
3. 🎭 Diverse Character Styles

Animate both humans and animals in a wide range of styles — from realistic to highly stylized.
Produces dynamic, expressive, and naturally realistic animations that bring any character to life.



https://video.twimg.com/amplify_video/1911729417186885632/vid/avc1/720x480/zdczvitwUOj-qmCJ.mp4

5/13
@ItsKevinNexus
4. 📊 Comparison with Closed-Source Methods

FantasyTalking outperforms current state-of-the-art (SOTA) approaches in multimodality-conditioned human video generation, setting a new benchmark for realism and control.



https://video.twimg.com/amplify_video/1911729480709574656/vid/avc1/1296x720/h2qDwFheHwaucntG.mp4

6/13
@ItsKevinNexus
5. 🗣️ Lip Sync with Half-Body Motion

Achieves precise lip-syncing synchronized with natural half-body movements, creating more immersive and lifelike character animations.



https://video.twimg.com/amplify_video/1911729737988292608/vid/avc1/720x576/rXRGyVlJRbAZi6rx.mp4

7/13
@ItsKevinNexus
6. 🗣️ Lip Sync with Half-Body Motion

Achieves precise lip-syncing synchronized with natural half-body movements, creating more immersive and lifelike character animations.



https://video.twimg.com/amplify_video/1911729783928414208/vid/avc1/900x720/rQrB5YDkU1HnkV8x.mp4

8/13
@ItsKevinNexus
7. 🎨 Diverse Character Styles

Supports a wide range of character types — from realistic humans to stylized avatars and animals.
Generates expressive, dynamic animations tailored to each style, making every character feel alive.



https://video.twimg.com/amplify_video/1911729859144876032/vid/avc1/720x432/7i9XdUF7AIrhw60M.mp4

9/13
@ItsKevinNexus
8. 🧍‍♂️ Diverse Characters with Full-Body Motion

Animate a variety of characters — from realistic to stylized, including animals — with natural full-body movement.
Delivers smooth, expressive motion across different poses, styles, and perspectives.



https://video.twimg.com/amplify_video/1911729974064635904/vid/avc1/720x720/Ausu2YSjnZnywTwg.mp4

10/13
@ItsKevinNexus
9. 🗣️ Lip Sync with Half-Body Motion

Achieves precise lip-syncing synchronized with natural half-body movements, creating more immersive and lifelike character animations.



https://video.twimg.com/amplify_video/1911730045078347776/vid/avc1/720x480/vaGkfumjQT-H47Y5.mp4

11/13
@ItsKevinNexus
10. 🗣️ Lip Sync with Half-Body Motion

Achieves precise lip-syncing synchronized with natural half-body movements, creating more immersive and lifelike character animations.



https://video.twimg.com/amplify_video/1911730093610696704/vid/avc1/480x800/uvI9f6T0OC62VFAh.mp4

12/13
@ItsKevinNexus
Paper page of huggingface:
Paper page - FantasyTalking: Realistic Talking Portrait Generation via Coherent Motion Synthesis



13/13
@ItsKevinNexus
Thanks for reading

If you enjoyed this post, please support it with like / repost the post below

[Quoted tweet]
China's Open Source AI is Next-Level 🤯

Alibaba just dropped FantasyTalking — an insane AI that lip-syncs characters with realistic facial and full-body motion.

It outperforms current SOTA methods like OmniHuman-1, Sonic, and Hallo 3.

🔥 Here are 10 jaw-dropping examples you’ve gotta see:
[media=twitter]1911730347290550747[/media]

https://video.twimg.com/amplify_video/1911729207111016448/vid/avc1/1066x720/-kPWm_FtHWGiulJX.mp4








1/4
@susumuota
[29/30] 58 Likes, 7 Comments, 1 Posts
FantasyTalking: Realistic Talking Portrait Generation via Coherent Motion Synthesis cs․CV, 07 Apr 2025

🆕FantasyTalking: Realistic Talking Portrait Generation via Coherent Motion Synthesis

Mengchao Wang, Qiang Wang, Fan Jiang, Yaqi Fan, Yunpeng Zhang, Yonggang Qi, Kun Zhao, Mu Xu



GpGRYwDXsAAYmm5.png


2/4
@susumuota
Twitter: https://twitter.com/search?q=arxiv.org/abs/2504.04842 OR arxiv.org/pdf/2504.04842.pdf
Reddit: https://www.reddit.com/search/?q="2504.04842"&sort=top



3/4
@susumuota
(1/1) 58 Likes, 7 Comments, 17 Apr 2025, Reddit
https://redd.it/1k16klz



4/4
@susumuota
FantasyTalking: Realistic Talking Portrait Generation via Coherent Motion Synthesis
一枚の静止画からリアルなアニメーション可能なアバターを作るのは、依然として難しい。既存のアプローチでは、微妙な表情やそれに伴う全体的な体の動き、ダイナミックな背景を捉えるのに苦労することが多い。これらの限界に対処するために、我々は、制御可能なモ...



GpGRaBOWcAAl85X.jpg



To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196



1/2
@wildmindai
FantasyTalking: Realistic Talking Portrait Generation via Coherent Motion Synthesis
Released inference code and model weights for audio conditions.
- Wan2.1-I2V-14B-720P (Base model)
- Wav2Vec (Audio encoder)
- FantasyTalking model (condition weights)
FantasyTalking: Realistic Talking Portrait Generation via Coherent Motion Synthesis



https://video.twimg.com/amplify_video/1916797115893698561/vid/avc1/1280x720/qirUBXJbzz86yxOi.mp4

2/2
@wildmindai
Code: GitHub - Fantasy-AMAP/fantasy-talking: FantasyTalking: Realistic Talking Portrait Generation via Coherent Motion Synthesis
Model: acvlab/FantasyTalking · Hugging Face




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196




 

bnew

Veteran
Joined
Nov 1, 2015
Messages
63,932
Reputation
9,808
Daps
174,252



1/11
@victormustar
Great gameplay, amazing graphics: 10/10

...being able to generate videos of this quality in 20 seconds is game changer!

⬇️ Sharing the (free) Hugging Face App



https://video.twimg.com/amplify_video/1923304231563341824/vid/avc1/1152x768/QMYnCA3wGNaCHUbZ.mp4

2/11
@victormustar
app: LTX Video Fast - a Hugging Face Space by Lightricks



3/11
@victormustar
Credits to @Lightricks - follow them on HF for more!
Lightricks (Lightricks)



4/11
@aykutkardas
this is awesome 🫠



5/11
@CaptainHaHaa
Very nice! Cheers mate!



6/11
@samuelwoods_
20 seconds? That's insane speed



7/11
@crislenta
Did you text or img to vid?



8/11
@freddyunivers3
It’s crazy how much AI has improved



9/11
@soushi888
3D Princess Mononoke... Awesome !!!



10/11
@bytesray
Wow, AI has improved so much. The graphics resemble Genshin Impact.



11/11
@anhcuongrau
Make a scam game with text and promote app ?!
Shame as you !




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
63,932
Reputation
9,808
Daps
174,252



1/10
@victormustar
🤯 It’s here: sub-10 second video generation is now real with LTX-Video-13B-distilled!

⬇️ Try it now on Hugging Face



https://video.twimg.com/amplify_video/1922926265511604224/vid/avc1/1352x1080/BY4lfJy5In-8VFlN.mp4

2/10
@victormustar
LTX Video Fast - a Hugging Face Space by Lightricks



3/10
@kingnish24
What's the prompt ??



4/10
@victormustar
something like "fpv gameplay" (image-to-video)



5/10
@Hathibel
Just tried LTX-Video-13B-distilled out. Took about 30 seconds to generate this.



https://video.twimg.com/amplify_video/1923267445260943363/vid/avc1/768x768/JDNsJk948jm-p9od.mp4

6/10
@Ren_Simmons
It’s incredible



7/10
@kasznare
Is this open source?



8/10
@bradsmithcoach
Sub-10 second video generation is a game changer!



9/10
@turbotardo
How soon? 🐋®️2️⃣



10/10
@picatrix_picori
prompt share, bro 😆




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
63,932
Reputation
9,808
Daps
174,252





1/5
@victormustar
💩 Your AI landing page sucks but it doesn't have to...

⬇️ Sharing my secret to go from (1) to (2) in one shot



Gq74JHlWcAAoOVV.jpg


2/5
@victormustar
First, we asked the Gemini 2.5 Pro to create a landing page for a luxury house rental service using HTML and Tailwind CSS. The result is this:



Gq74ZMRWcAAqKeT.jpg


3/5
@victormustar
Now here is the secret: DeepSeek-V3-0324 is the best model for UI design (by far).

Use your free Hugging Face Inference Credits to copy the result from Gemini to the HF Playground, select DeepSeek-V3-0324 and set the temperature to 1 (this is key) and ask for a redesign 🤯



Gq75ygBWEAEZzoU.jpg


4/5
@victormustar
Initial Gemini result:
Tailwind Play
Redesign by DeepSeek-V3-0324 result (first try):
Tailwind Play



https://video.twimg.com/amplify_video/1922748191046275072/vid/avc1/1368x1080/bBvFag0lEki9aIP2.mp4

5/5
@victormustar
Get started with the Playground:
Playground - Hugging Face




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 
Top