<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
    <channel>
        <title>InsightFace on KnightLi Blog</title>
        <link>https://www.knightli.com/en/tags/insightface/</link>
        <description>Recent content in InsightFace on KnightLi Blog</description>
        <generator>Hugo -- gohugo.io</generator>
        <language>en</language>
        <lastBuildDate>Sat, 11 Apr 2026 08:27:57 +0800</lastBuildDate><atom:link href="https://www.knightli.com/en/tags/insightface/index.xml" rel="self" type="application/rss+xml" /><item>
        <title>What Models Power fnOS AI Photos: Face, Object, and Semantic Search Stack</title>
        <link>https://www.knightli.com/en/2026/04/11/fnos-ai-photo-model-stack/</link>
        <pubDate>Sat, 11 Apr 2026 08:27:57 +0800</pubDate>
        
        <guid>https://www.knightli.com/en/2026/04/11/fnos-ai-photo-model-stack/</guid>
        <description>&lt;p&gt;The AI photo feature in Feiniu NAS (fnOS) is typically built by integrating mainstream open-source models, rather than training all core algorithms from scratch.&lt;/p&gt;
&lt;h2 id=&#34;1-face-recognition-insightface&#34;&gt;1) Face recognition: InsightFace
&lt;/h2&gt;&lt;p&gt;For face-related functions, InsightFace is usually the core.&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Common feature-learning method: ArcFace&lt;/li&gt;
&lt;li&gt;Main role: face detection, embedding extraction, clustering, and person recognition&lt;/li&gt;
&lt;/ul&gt;
&lt;h2 id=&#34;2-object-and-scene-understanding-yolo-family&#34;&gt;2) Object and scene understanding: YOLO family
&lt;/h2&gt;&lt;p&gt;Object detection in photos (for example cats, dogs, cars, computers) and part of scene-level understanding are generally handled by YOLO models (often YOLOv8 or lightweight variants).&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Strength: good speed/accuracy balance&lt;/li&gt;
&lt;li&gt;Fit: edge-like NAS environments with limited compute budgets&lt;/li&gt;
&lt;/ul&gt;
&lt;h2 id=&#34;3-semantic-search-clip--chinese-clip&#34;&gt;3) Semantic search: CLIP / Chinese-CLIP
&lt;/h2&gt;&lt;p&gt;A key capability is natural-language photo search, such as &amp;ldquo;a dog on the grass&amp;rdquo; or &amp;ldquo;a man wearing sunglasses.&amp;rdquo;&lt;/p&gt;
&lt;p&gt;Typical implementation uses CLIP:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;images and text are projected into the same embedding space&lt;/li&gt;
&lt;li&gt;Chinese deployments usually add Chinese-CLIP or similar localized variants&lt;/li&gt;
&lt;/ul&gt;
&lt;h2 id=&#34;summary&#34;&gt;Summary
&lt;/h2&gt;&lt;p&gt;A simple way to view the fnOS AI photo stack:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;InsightFace for faces&lt;/li&gt;
&lt;li&gt;YOLO for objects and scenes&lt;/li&gt;
&lt;li&gt;CLIP for text-image semantic alignment&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;The main engineering value is in integration quality, localization, and hardware acceleration, more than from-zero model invention.&lt;/p&gt;
</description>
        </item>
        
    </channel>
</rss>
