<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
    <channel>
        <title>AI Video on KnightLi Blog</title>
        <link>https://www.knightli.com/en/tags/ai-video/</link>
        <description>Recent content in AI Video on KnightLi Blog</description>
        <generator>Hugo -- gohugo.io</generator>
        <language>en</language>
        <lastBuildDate>Fri, 08 May 2026 22:27:10 +0800</lastBuildDate><atom:link href="https://www.knightli.com/en/tags/ai-video/index.xml" rel="self" type="application/rss+xml" /><item>
        <title>miHoYo LPM 1.0 Explained: How an AI Video Model Could Reshape Game NPCs</title>
        <link>https://www.knightli.com/en/2026/05/08/lpm-1-0-ai-video-character-performance/</link>
        <pubDate>Fri, 08 May 2026 22:27:10 +0800</pubDate>
        
        <guid>https://www.knightli.com/en/2026/05/08/lpm-1-0-ai-video-character-performance/</guid>
        <description>&lt;p&gt;LPM 1.0 is easy to mistake for another AI video generation model. Judging only by demos, it may not look as visually explosive as some text-to-video systems. But viewed through the paper&amp;rsquo;s goal, it is not mainly trying to generate a good-looking clip. It is trying to make a digital character feel present during interaction.&lt;/p&gt;
&lt;p&gt;That is the biggest difference between LPM 1.0 and ordinary video models. A typical video model focuses on image quality, camera continuity, and prompt following. LPM 1.0 focuses on character performance: lip sync, rhythm, and expression while speaking; nods, gaze, pauses, and micro-expressions while listening; and stable identity across long interactions.&lt;/p&gt;
&lt;h2 id=&#34;from-generating-video-to-generating-performance&#34;&gt;From generating video to generating performance
&lt;/h2&gt;&lt;p&gt;LPM stands for Large Performance Model. The name matters because it shifts the task boundary from &amp;ldquo;video&amp;rdquo; to &amp;ldquo;performance&amp;rdquo;.&lt;/p&gt;
&lt;p&gt;In real conversation, whether someone feels natural is not only about what they say. Listening is part of communication: the timing of nods, the direction of gaze, and subtle emotional changes all affect whether we believe a character is alive.&lt;/p&gt;
&lt;p&gt;Many digital human systems still attach text, speech, and lip motion to a character. The character can talk, but may not truly listen. It can output lines, but may not react continuously to the previous second of input. LPM 1.0 aims to turn passive playback into real-time interaction.&lt;/p&gt;
&lt;h2 id=&#34;the-three-hard-problems&#34;&gt;The three hard problems
&lt;/h2&gt;&lt;p&gt;The LPM 1.0 paper describes a trilemma in AI character performance: expressiveness, real-time inference, and long-horizon identity stability. A system may look detailed but be slow, respond quickly but feel rigid, or stay stable briefly but drift over time. Achieving all three is much harder.&lt;/p&gt;
&lt;p&gt;To address this, LPM 1.0 uses richer character conditioning. Instead of giving the model only one reference image, it introduces multi-granularity identity references, including global appearance, multi-view body images, and facial expression examples. The goal is to reduce hallucinated details such as profile shape, teeth, expression texture, and body proportions.&lt;/p&gt;
&lt;p&gt;The paper also separates speaking and listening behavior. Speaking audio mainly drives lip sync, speech rhythm, head motion, and body rhythm. Listening audio triggers gaze, nodding, posture changes, and micro-expressions. If both signals are mixed into one control stream, the model can easily learn the wrong behavior. LPM 1.0 models speaking and listening separately, then connects them in one online interaction system.&lt;/p&gt;
&lt;h2 id=&#34;base-lpm-and-online-lpm&#34;&gt;Base LPM and Online LPM
&lt;/h2&gt;&lt;p&gt;According to the public paper, LPM 1.0 is built on a 17B-parameter Diffusion Transformer. Base LPM learns high-quality, controllable, identity-consistent character performance video. Online LPM is a distilled streaming generator designed for low-latency, long-running interaction.&lt;/p&gt;
&lt;p&gt;This split is important. Offline models can focus on quality, but interactive systems cannot make users wait. When a user starts speaking, the character should begin listening immediately. When the character starts speaking, lip sync, expression, and body motion must follow at once. Online LPM is valuable because it compresses complex video generation into something closer to real-time interaction.&lt;/p&gt;
&lt;p&gt;So LPM 1.0 is not just a short-video asset tool for creators. It is closer to a visual engine for conversational agents, virtual streamers, and game NPCs: the language model understands and generates content, the speech model provides the voice, and LPM makes the on-screen character perform credibly.&lt;/p&gt;
&lt;h2 id=&#34;what-it-means-for-games&#34;&gt;What it means for games
&lt;/h2&gt;&lt;p&gt;In games, LPM 1.0 points less toward prettier cutscenes and more toward the next generation of interactive characters.&lt;/p&gt;
&lt;p&gt;Traditional NPCs rely on prewritten scripts, fixed animations, and limited branches. Players can talk to them, but their responses are usually predesigned. In the AI era, the target goes further: different players may experience different story paths in the same world, and the same character may respond with actions, emotions, and dialogue that fit each player&amp;rsquo;s context.&lt;/p&gt;
&lt;p&gt;That is what a truly personalized game experience needs underneath. Language models can generate lines, and behavior systems can choose goals, but if the character on screen still looks stiff, players will struggle to believe it understands them. LPM 1.0 tries to fill that visual and performance layer.&lt;/p&gt;
&lt;h2 id=&#34;not-a-finished-magic-product&#34;&gt;Not a finished magic product
&lt;/h2&gt;&lt;p&gt;LPM 1.0 should still be understood as a technical direction, not an immediately scalable commercial product. The paper and demos show a possibility: real-time, full-duplex, identity-stable character video generation is getting closer to usable. But before it can enter games broadly, there are still problems around cost, latency, edge deployment, content safety, character rights, multiplayer scenes, and engine integration.&lt;/p&gt;
&lt;p&gt;A more realistic path may start with virtual streamers, AI companions, story interaction, character support agents, and educational coaching. As model cost falls and latency improves, the technology can move into more complex game systems.&lt;/p&gt;
&lt;h2 id=&#34;summary&#34;&gt;Summary
&lt;/h2&gt;&lt;p&gt;The value of LPM 1.0 is not whether it can generate the most spectacular video clip. It is that it pushes AI video from &amp;ldquo;image generation&amp;rdquo; toward &amp;ldquo;character presence&amp;rdquo;.&lt;/p&gt;
&lt;p&gt;If future games become more personalized, more dynamic, and more dependent on AI characters, language, speech, motion, expression, and identity consistency must be designed together. LPM 1.0 offers one possible path: digital characters that do not just talk, but listen, react, and remain recognizably themselves over long interactions.&lt;/p&gt;
&lt;p&gt;References:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://arxiv.org/abs/2604.07823&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;arXiv: LPM 1.0: Video-based Character Performance Model&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a class=&#34;link&#34; href=&#34;https://large-performance-model.github.io/&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;LPM 1.0 project page&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
</description>
        </item>
        <item>
        <title>Pixelle-Video: An Open-Source AI Engine for Generating Short Videos From One Topic</title>
        <link>https://www.knightli.com/en/2026/05/07/pixelle-video-ai-short-video-engine/</link>
        <pubDate>Thu, 07 May 2026 20:25:17 +0800</pubDate>
        
        <guid>https://www.knightli.com/en/2026/05/07/pixelle-video-ai-short-video-engine/</guid>
        <description>&lt;p&gt;&lt;a class=&#34;link&#34; href=&#34;https://github.com/AIDC-AI/Pixelle-Video&#34;  target=&#34;_blank&#34; rel=&#34;noopener&#34;
    &gt;Pixelle-Video&lt;/a&gt; is an open-source fully automated short-video generation engine from AIDC-AI. Its goal is direct: the user enters a topic, and the system automatically writes the script, generates AI images or videos, creates voice narration, adds background music, and renders the final video.&lt;/p&gt;
&lt;p&gt;This kind of tool is useful for batch short-video creation, knowledge explainers, talking-head content, novel recaps, history and culture videos, and self-media experiments. It is not a single text-to-video model. It is a production pipeline that connects several AI capabilities.&lt;/p&gt;
&lt;h2 id=&#34;what-it-automates&#34;&gt;What It Automates
&lt;/h2&gt;&lt;p&gt;Pixelle-Video&amp;rsquo;s default flow can be summarized as:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;enter a topic or fixed script;&lt;/li&gt;
&lt;li&gt;use an LLM to generate narration;&lt;/li&gt;
&lt;li&gt;plan scenes and generate images or video clips;&lt;/li&gt;
&lt;li&gt;use TTS to create voice narration;&lt;/li&gt;
&lt;li&gt;add background music;&lt;/li&gt;
&lt;li&gt;apply a video template and render the final result.&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;The README describes the flow as &amp;ldquo;script generation → image planning → frame-by-frame processing → video composition.&amp;rdquo; The modular design is clear: each step can be replaced, tuned, or connected to a custom workflow.&lt;/p&gt;
&lt;h2 id=&#34;key-features&#34;&gt;Key Features
&lt;/h2&gt;&lt;p&gt;The project covers a fairly complete set of capabilities:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;AI script writing: automatically generate narration from a topic;&lt;/li&gt;
&lt;li&gt;AI image generation: create illustrations for each line or scene;&lt;/li&gt;
&lt;li&gt;AI video generation: connect to video generation models such as WAN 2.1;&lt;/li&gt;
&lt;li&gt;TTS voice: support Edge-TTS, Index-TTS, and other options;&lt;/li&gt;
&lt;li&gt;background music: use built-in BGM or custom music;&lt;/li&gt;
&lt;li&gt;multiple aspect ratios: support vertical, horizontal, and other video sizes;&lt;/li&gt;
&lt;li&gt;multiple models: connect to GPT, Qwen, DeepSeek, Ollama, and more;&lt;/li&gt;
&lt;li&gt;ComfyUI workflows: use built-in workflows or replace image, TTS, and video generation steps.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Recent updates also mention motion transfer, digital-human talking videos, image-to-video pipelines, multilingual TTS voices, RunningHub support, and a Windows all-in-one package. The project is clearly moving beyond a simple script toward a fuller creation tool.&lt;/p&gt;
&lt;h2 id=&#34;installation-and-launch&#34;&gt;Installation and Launch
&lt;/h2&gt;&lt;p&gt;Windows users can first look at the official all-in-one package. It is designed to reduce setup friction: no manual Python, uv, or ffmpeg installation is required. After extracting the package, run &lt;code&gt;start.bat&lt;/code&gt;, open the web interface, and configure the required APIs and image generation service.&lt;/p&gt;
&lt;p&gt;For source installation, the README gives this basic flow:&lt;/p&gt;
&lt;div class=&#34;highlight&#34;&gt;&lt;div class=&#34;chroma&#34;&gt;
&lt;table class=&#34;lntable&#34;&gt;&lt;tr&gt;&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code&gt;&lt;span class=&#34;lnt&#34;&gt;1
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;2
&lt;/span&gt;&lt;span class=&#34;lnt&#34;&gt;3
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;
&lt;td class=&#34;lntd&#34;&gt;
&lt;pre tabindex=&#34;0&#34; class=&#34;chroma&#34;&gt;&lt;code class=&#34;language-bash&#34; data-lang=&#34;bash&#34;&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;git clone https://github.com/AIDC-AI/Pixelle-Video.git
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;&lt;span class=&#34;nb&#34;&gt;cd&lt;/span&gt; Pixelle-Video
&lt;/span&gt;&lt;/span&gt;&lt;span class=&#34;line&#34;&gt;&lt;span class=&#34;cl&#34;&gt;uv run streamlit run web/app.py
&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;
&lt;/div&gt;
&lt;/div&gt;&lt;p&gt;The source route is suitable for macOS and Linux users, and for anyone who wants to modify templates, workflows, or service configuration. The main prerequisites are &lt;code&gt;uv&lt;/code&gt; and &lt;code&gt;ffmpeg&lt;/code&gt;.&lt;/p&gt;
&lt;h2 id=&#34;configuration-priorities&#34;&gt;Configuration Priorities
&lt;/h2&gt;&lt;p&gt;On first use, the key is not to click &amp;ldquo;generate&amp;rdquo; immediately. The important part is connecting the external capabilities properly.&lt;/p&gt;
&lt;p&gt;LLM configuration determines script quality. You can choose models such as Qwen, GPT, DeepSeek, or Ollama, then fill in the API Key, Base URL, and model name. If you want to minimize cost, local Ollama is one option. If you want more stable results, a cloud model is usually easier.&lt;/p&gt;
&lt;p&gt;Image and video generation configuration determines visual quality. The project supports local ComfyUI and RunningHub. Users who understand ComfyUI can place their own workflows under &lt;code&gt;workflows/&lt;/code&gt; to replace the default image, video, or TTS pipeline.&lt;/p&gt;
&lt;p&gt;Template configuration determines the final visual form. The project organizes video templates under &lt;code&gt;templates/&lt;/code&gt;, with naming rules for static templates, image templates, and video templates. For creators, this is more practical than generating raw assets only, because the output is a video that can be previewed and downloaded directly.&lt;/p&gt;
&lt;h2 id=&#34;who-it-is-for&#34;&gt;Who It Is For
&lt;/h2&gt;&lt;p&gt;Pixelle-Video is especially suitable for three groups:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;&lt;strong&gt;Short-video creators&lt;/strong&gt; who want to turn ideas into draft videos quickly.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;AIGC tool users&lt;/strong&gt; who want to connect LLMs, ComfyUI, TTS, and video composition.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Developers and automation users&lt;/strong&gt; who want to modify templates, workflows, or integrate their own materials and models.&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;If you only want to make one polished premium video, it may not replace manual editing. But if you want to generate many explainers, talking videos, or science and education videos with a consistent structure, its pipeline approach is valuable.&lt;/p&gt;
&lt;h2 id=&#34;things-to-note&#34;&gt;Things to Note
&lt;/h2&gt;&lt;p&gt;The ceiling of this kind of tool is determined by multiple links in the chain. A weak script model produces empty content; a weak image model gives scattered visuals; unnatural TTS makes the video feel rough; and a poor template weakens the final result.&lt;/p&gt;
&lt;p&gt;So it is better to start with one fixed scenario, such as a &amp;ldquo;60-second vertical science explainer.&amp;rdquo; Fix the LLM, visual style, TTS voice, BGM, and template first, then expand to more topics.&lt;/p&gt;
&lt;p&gt;The project supports a local free setup, but local setups often require a GPU, ComfyUI configuration, and model files. Users without a local inference environment can reduce setup difficulty by using a cloud LLM plus RunningHub, while keeping an eye on usage cost.&lt;/p&gt;
&lt;h2 id=&#34;short-take&#34;&gt;Short Take
&lt;/h2&gt;&lt;p&gt;Pixelle-Video is interesting not merely because it can &amp;ldquo;generate a video from one sentence.&amp;rdquo; Its real value is that it breaks short-video production into replaceable modules: script, visuals, voice, music, templates, and rendering. For ordinary users, it is a low-barrier AI video tool. For developers, it is closer to a hackable short-video automation framework.&lt;/p&gt;
&lt;p&gt;If you are studying AI short-video pipelines, or want to connect ComfyUI, TTS, LLMs, and template rendering into a usable product, Pixelle-Video is worth trying and dissecting.&lt;/p&gt;
</description>
        </item>
        
    </channel>
</rss>
