From Abstract Words to Cinematic Animation in One Click: A Complete Guide to Running Fogsight Locally
1. Why Fogsight Matters Today
Problem | Traditional Workflow | How Fogsight Changes the Game |
---|---|---|
Explaining a concept | 2–3 days of slide decks, screen recordings, storyboards | Type a word, wait 10–20 min, get a bilingual narrated short film |
Iterating on feedback | Re-render, re-record, version hell | Chat with AI in plain English, see changes in real time |
Required skills | After Effects, Premiere, scripting | Laptop + browser is all you need |
In short, Fogsight (雾象) turns prompt-writing into directing, letting anyone create cinema-grade micro-movies from a single sentence.
2. What Can Fogsight Actually Do?
2.1 Core Use Cases
-
Teachers: turn “Euler’s formula” into a 30-second clip for MOOCs. -
Science communicators: explain the second law of thermodynamics with an “entropy increase” animation that hit 100 k views on Bilibili. -
Product managers: demo “affordance in design” to stakeholders in 60 seconds instead of a 30-slide deck.
2.2 Public Examples (click to watch)
Input Phrase | Discipline | Length | Watch |
---|---|---|---|
entropy increase | Physics | 50 s | Bilibili BV1PXgKzBEyN |
Euler’s formula | Mathematics | 45 s | Bilibili BV1yXgKzqE42 |
bubble sort | Computer Science | 60 s | Bilibili BV1sQgKzMEox |
affordance in design | Design | 55 s | Bilibili BV1yQgKzMEo6 |
All clips include bilingual narration and are ready to embed in classrooms or social media.
3. How It Works (High-Level)
-
Large Language Model (LLM) parses your prompt and decomposes it into “shot list + narration.” -
Animation Orchestration Engine maps the list to visual assets, transitions, and sound. -
Language User Interface (LUI) lets you refine the result conversationally: “make the second scene warmer” updates everything in seconds.
4. Local Installation Walk-Through
Verified on macOS 14, Ubuntu 22, Windows 11.
4.1 Prerequisites
-
Python 3.9+ -
Modern browser (Chrome ≥ 120, Firefox ≥ 121, Edge ≥ 120) -
LLM API key (Gemini 2.5 Pro recommended)
4.2 Step-by-Step Commands
# 1. Clone the repo
git clone https://github.com/fogsightai/fogsight.git
cd fogsight
# 2. Install dependencies
pip install -r requirements.txt
# 3. Configure credentials
cp demo-credentials.json credentials.json
# Edit credentials.json
# "API_KEY": "your_Gemini_2.5_key",
# "BASE_URL": "https://generativelanguage.googleapis.com/v1beta/openai/"
# 4. Launch
python start_fogsight.py
Your browser will open automatically at http://127.0.0.1:8000
.
5. Your First Animation: “Bubble Sort” in Practice
-
Type bubble sort
in the input box. -
Click Generate. -
Progress bar: script → storyboard → voice-over → final render -
720p ~ 5–8 min, 1080p ~ 10–15 min on RTX 3060
-
-
Preview the clip. -
Chat refinement: -
“Slow down the third pass by 0.5 s and change the background to light blue.” -
New version ready in ~30 s.
-
6. Frequently Asked Questions
Question | Straight Answer |
---|---|
Does it support Chinese? | Yes—voice-over in Chinese, English, or bilingual. |
Why is the first run slow? | Initial model download; subsequent runs use cache and are 3–4× faster. |
Can I run it offline? | LLM still needs internet; rendering and caching are local. |
Output format? | MP4 (H.264), 720 p / 1080 p, 8 Mbps. |
Commercial use? | MIT license—commercial use allowed; attribution appreciated. |
No GPU? | CPU-only works; 720 p takes ~15 min. RTX 3060+ cuts time by 4–5×. |
How to join the community? | Visit the Feishu doc and click “Join Group.” |
7. Pro Tips for Better Clips
7.1 Prompt Tricks
-
Add domain keywords:
Euler’s formula mathematics polyhedron
→ more accurate shots. -
Specify length:
30-second short
vs90-second detailed version
. -
Language hint:
bilingual narration
orEnglish only
.
7.2 Ready-to-Copy LUI Commands
Goal | Command |
---|---|
Change color scheme | “Switch to cyber-punk palette, main colors #00ffff and #ff00ff.” |
Add subtitles | “Add white subtitles with black stroke, 48 pt, bottom center.” |
Adjust pacing | “Extend the second scene by 2 s and add fade-in/out.” |
Swap voice | “Use gentle female voice, speed 0.8×.” |
8. Contributors & Credits
Fogsight is part of the WaytoAGI open-source initiative.
-
Universities
-
Taited – PhD student, CUHK-Shenzhen -
yjydya – PhD student, NTU
-
-
WaytoAGI Community
-
richkatchen (陈财猫) -
kk
-
-
Index Future Lab
-
Lin He
-
-
AI Explorers & Independent Devs
-
Xiaodao Huang, shuyan-5200, Ruyue Wang, Jack-the-Builder, xiayurain95, Lixin Cai
-
Thank you to every bug-reporter and feedback-giver in the Feishu group.
9. Roadmap
-
Plugin system: upload custom visual assets. -
Local fine-tuning script: run 7 B LLM on 8 GB VRAM for offline generation. -
Community template library: one-click reuse of “Fermat’s Little Theorem,” “Quicksort,” etc.
10. Take Action Today
You now know:
-
Fogsight turns abstract terms into animated explainers. -
Local setup is five commands away. -
Iteration is as simple as chatting with AI.
Open your terminal and run:
git clone https://github.com/fogsightai/fogsight.git
Ten minutes from now, you could have a 45-second short on Schrödinger’s cat. Share your first clip in the Feishu group and let’s make knowledge visible together.