QuQu: The Free, Open-Source, and Privacy-First Alternative to Wispr Flow for Chinese Users

Are you tired of paying $12/month for voice dictation tools like Wispr Flow ? Concerned about your private voice data being processed in the cloud? Or maybe you’ve just found that mainstream tools don’t quite “get” Chinese the way you speak it?

If any of that sounds familiar, meet QuQu—a next-generation, open-source, and completely free voice-to-text workflow tool built specifically for Chinese speakers, with privacy and local processing at its core.

In this post, we’ll dive deep into what makes QuQu a compelling alternative to commercial options, how it works under the hood, and how you can get started in minutes.


Why QuQu Exists: Solving Real Pain Points

Tools like Wispr Flow have popularized AI-powered voice dictation, promising to let you “write at the speed of speech” . But they come with trade-offs:

  • Cost: Wispr Flow Pro starts at $12/month for unlimited words and editing commands [[21], [28]].
  • Privacy: Your voice is sent to remote servers for processing—raising legitimate concerns about data security.
  • Language Bias: While Wispr Flow supports over 100 languages , its models aren’t optimized for the nuances of Chinese internet slang, regional accents, or contextual corrections.

QuQu was created to address these gaps. It’s not just a clone—it’s a reimagined voice workflow that prioritizes local execution, Chinese linguistic intelligence, and open AI ecosystems.

QuQu vs. Wispr Flow: A Clear Comparison

Feature 🎯 QuQu 💰 Wispr Flow
Price Free & open-source $12/month subscription
Privacy 100% local processing Cloud-based transcription
Chinese Support Deeply optimized for Chinese ⚠️ Generic multilingual support
AI Model Flexibility Supports Chinese LLMs (Qwen, Kimi, etc.) Limited to Western models

This isn’t just about saving money—it’s about control, accuracy, and cultural relevance.


What Is QuQu? A Voice Assistant That Thinks Like You

QuQu is a desktop application (for macOS, Windows, and Linux) that turns your spoken words into polished, ready-to-use text—instantly and privately.

Here’s how it works in practice:

You say: “Change the function name to getUserProfileData—wait, no, make it fetchUserProfile.”

A basic speech-to-text tool would output the entire sentence, including the correction.
QuQu, however, uses a two-stage intelligent engine to deliver just:

fetchUserProfile

It’s like having a smart editor who listens, understands your intent, and outputs only what matters.


How QuQu Works: The Tech Behind the Magic

QuQu’s power comes from a smart fusion of local speech recognition and configurable large language models (LLMs).

1. State-of-the-Art Chinese ASR: FunASR Paraformer (Local & Private)

At its core, QuQu uses FunASR, an industrial-grade open-source speech recognition toolkit from Alibaba’s DAMO Academy. Specifically, it leverages the Paraformer-large model—a non-autoregressive end-to-end ASR system known for its high accuracy and speed [[2], [4]].

Key advantages:

  • Trained on tens of thousands of hours of Chinese audio data .
  • Supports real-time transcription with low latency .
  • Runs entirely on your machine—no data leaves your device.
  • Includes FSMN-VAD for precise voice activity detection and CT-Transformer for automatic punctuation .

This means QuQu understands not just standard Mandarin, but also colloquialisms, tech jargon, and even your “umms” and “ahhs”—which it can later clean up.

2. The “Two-Stage Engine”: ASR + LLM = Intelligent Output

QuQu doesn’t stop at transcription. It adds a second layer of intelligence:

  1. Stage 1 (ASR): FunASR converts your speech to raw text—locally and accurately.
  2. Stage 2 (LLM): That raw text is sent to an AI model of your choice for refinement.

The LLM can:

  • Correct self-interruptions (“Wednesday meeting—no, Thursday” → “Thursday meeting”)
  • Remove filler words (“you know,” “like,” “so…”)
  • Add proper punctuation and formatting
  • Adapt tone (formal email vs. casual chat)

And because you control the LLM, you control the output style.

3. Open AI Ecosystem: Built for China’s LLM Landscape

QuQu uses the OpenAI-compatible API standard, which means it works seamlessly with Chinese LLMs like:

  • Qwen (Tongyi Qianwen) from Alibaba
  • Kimi from Moonshot AI
  • GLM from Zhipu AI

This is more than convenience—it’s strategic. With OpenAI restricting API access for Chinese developers [[12], [13]], relying on domestic models ensures reliability, speed, and compliance.

You get:

  • Lower latency (servers in China)
  • Better Chinese understanding
  • Competitive pricing (or even free tiers)

4. Developer & Power User Features

QuQu shines for coders and productivity enthusiasts:

  • Accurate recognition of camelCase and snake_case—no more user name when you meant userName.
  • Context-aware output: Configure different LLM prompts based on your active app (e.g., code comments in VS Code, bullet points in Notion).
  • Global hotkey (F2): Start dictating anywhere, anytime.

Getting Started: Install QuQu in 4 Simple Steps

QuQu is easy to set up. Here’s how:

✅ Step 1: Check Requirements

  • OS: macOS 10.15+, Windows 10+, or Linux
  • Node.js 18+ and pnpm
  • Python 3.8+ (for FunASR)

✅ Step 2: Install & Launch

# Clone the repo
git clone https://github.com/yan5xu/ququ.git
cd ququ

# Install JS dependencies
pnpm install

# Install Python ASR engine
pip install funasr modelscope

# Start the app
pnpm run dev

On first launch, QuQu will download the FunASR models (~500MB–1GB). This happens once, and everything runs locally afterward.

✅ Step 3: Configure Your AI Model

In the settings panel, enter:

  • API Key (from Qwen, Kimi, etc.)
  • Base URL (e.g., https://dashscope.aliyuncs.com/compatible-mode/v1 for Qwen)
  • Model name (e.g., qwen-max, moonshot-v1-8k)

All config is stored locally—no cloud accounts needed.

✅ Step 4: Start Dictating!

Press F2, speak naturally, and watch your words appear—clean, corrected, and perfectly formatted.


Troubleshooting Common Issues (FAQ)

Q: FunASR model download is slow or fails.
A: Ensure a stable internet connection. The model is large but only downloads once.

Q: On macOS, I see SSL warnings slowing down startup.
A: Fix with:

python3 -m pip install "urllib3<2.0"

Q: Can I use QuQu without a Chinese LLM?
A: Yes! Any OpenAI-compatible API works (including OpenAI itself). But for best Chinese results, domestic models are strongly recommended.

Q: Is my data safe?
A: Absolutely. ASR runs locally. Only the transcribed text (not audio) is sent to your chosen LLM—and you control which one.


Tech Stack: Built for Performance & Extensibility

  • Frontend: React 19, TypeScript, Tailwind CSS, shadcn/ui
  • Desktop: Electron
  • Local ASR: FunASR (Paraformer-large + FSMN-VAD + CT-Transformer)
  • AI Backend: OpenAI-compatible API (supports Qwen, Kimi, GLM, etc.)
  • Storage: better-sqlite3 (local config & history)

Join the Movement: QuQu Is Open Source

QuQu is released under the Apache 2.0 License and welcomes contributions:

  • 🐞 Report bugs
  • 💡 Suggest features
  • 💻 Submit PRs

It’s built on the shoulders of giants like FunASR and OpenWhispr, and aims to give back to the open-source community.


Final Thoughts: The Future of Voice Is Local, Open, and Chinese-First

QuQu isn’t just a tool—it’s a statement. In an era of rising AI costs, cloud dependency, and language bias, it proves that privacy, affordability, and linguistic authenticity can coexist.

Whether you’re a developer, writer, student, or professional, QuQu offers a faster, smarter, and more respectful way to turn speech into action.

👉 Ready to try it?
Visit the QuQu GitHub repo and start dictating—your way.