diff --git a/LICENSE b/LICENSE
new file mode 100644
index 0000000..2329817
--- /dev/null
+++ b/LICENSE
@@ -0,0 +1,21 @@
+MIT License
+
+Copyright (c) 2025 kuraa
+
+Permission is hereby granted, free of charge, to any person obtaining a copy
+of this software and associated documentation files (the "Software"), to deal
+in the Software without restriction, including without limitation the rights
+to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
+copies of the Software, and to permit persons to whom the Software is
+furnished to do so, subject to the following conditions:
+
+The above copyright notice and this permission notice shall be included in all
+copies or substantial portions of the Software.
+
+THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
+IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
+FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
+AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
+LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
+OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
+SOFTWARE.
diff --git a/README.en.md b/README.en.md
new file mode 100644
index 0000000..3fd14cd
--- /dev/null
+++ b/README.en.md
@@ -0,0 +1,178 @@
+
+
+
+
+
+
+
+---
+
+## About
+
+CrossSubtitle-AI is a **local-first** audio/video subtitle processing tool. It uses [Whisper](https://github.com/ggerganov/whisper.cpp) for speech recognition, [Silero VAD](https://github.com/snakers4/silero-vad) for voice activity detection, and supports OpenAI-compatible APIs for intelligent translation — helping you quickly transcribe and translate media files into bilingual subtitles.
+
+All speech recognition runs locally on your machine. No audio or video files are ever uploaded to any server, ensuring your data privacy.
+
+## Features
+
+- **Speech Recognition** — High-accuracy speech-to-text powered by Whisper, supporting 17 source languages including Chinese, English, Japanese, Korean, French, and more
+- **Voice Activity Detection** — Silero VAD precisely splits speech segments and automatically filters out silence
+- **Smart Translation** — Connect to any OpenAI-compatible API (GLM, DeepSeek, ChatGPT, etc.) to translate transcripts into your target language
+- **Audio Extraction** — Built-in FFmpeg automatically extracts audio and converts to 16kHz mono WAV
+- **Multiple Export Formats** — Export subtitles in SRT, VTT, and ASS formats
+- **Bilingual Export** — Export side-by-side original + translated bilingual subtitles
+- **Subtitle Editor** — Built-in editor for modifying both source text and translations line by line
+- **Drag & Drop** — Drag and drop files to quickly create tasks
+- **Task Queue** — Batch process multiple media files with real-time progress tracking
+- **Bilingual UI** — Switch between Chinese and English interface languages
+- **Local-First** — Speech recognition runs entirely locally, no data upload required
+
+## Workflow
+
+1. **Choose Mode** — Select "Source" for transcription only, or "Translate" mode for automatic translation after transcription
+2. **Add Task** — Click "Add Task" or drag-and-drop media files onto the window
+3. **Wait for Processing** — Tasks go through: Audio Extraction → VAD Segmentation → Speech Recognition → (Optional) Translation
+4. **Review & Edit** — View and modify recognition results and translations in the subtitle editor
+5. **Export Subtitles** — Export as SRT, VTT, or ASS format
+
+## Screenshots
+
+| Subtitle | Subtitle Editor |
+|:---:|:---:|
+|  |  |
+
+## Installation
+
+Download the installer for your platform from [GitHub Releases](https://github.com/AndySkaura/crosssubtitle-ai/releases):
+
+| Platform | Package |
+|:---:|:---:|
+| macOS (Apple Silicon) | `.dmg` |
+| Windows | `.exe` (NSIS Installer) |
+
+> The Whisper model (~500MB) will be downloaded on first launch. An internet connection is required.
+
+## Usage
+
+### Quick Start
+
+1. Open the app and select a mode from the top toolbar:
+ - **Source** — Speech recognition only, outputs source language subtitles
+ - **Translate** — Transcribes then translates via an LLM API
+2. Click "Add Task" or drag-and-drop files onto the window
+3. Wait for processing to complete
+4. Review and edit results in the subtitle editor on the right
+5. Click "Export" to save subtitles in your preferred format
+
+### Translation Configuration
+
+Before using the translation feature, configure the LLM API:
+
+- Fill in the LLM API Base, API Key, and Model in "Advanced Settings"
+- Works with any OpenAI-compatible service, including:
+ - **GLM (Zhipu AI)** — GLM-4.7-Flash available for free
+ - **DeepSeek**
+ - **ChatGPT**
+ - **Self-hosted** — Ollama, vLLM, etc.
+
+### Advanced Settings
+
+- **Whisper Model Path** — Path to a local ggml model file
+- **VAD Model Path** — Path to a local Silero VAD ONNX model file
+- **Batch Size** — Number of segments to translate per batch (10-15)
+- **Context Size** — Number of preceding segments to include as context for translation (0-5)
+
+## Development
+
+### Prerequisites
+
+- [Rust](https://www.rust-lang.org/) toolchain
+- [Node.js](https://nodejs.org/) (18+)
+- [FFmpeg](https://ffmpeg.org/) (must be available on the command line)
+- [CMake](https://cmake.org/) (required for compiling whisper-rs)
+
+### Local Development
+
+```bash
+# Clone the repository
+git clone https://github.com/AndySkaura/crosssubtitle-ai.git
+cd crosssubtitle-ai
+
+# Install frontend dependencies
+npm install
+
+# Start development mode
+npm run tauri-dev
+```
+
+### Build
+
+```bash
+# macOS DMG build
+npm run tauri-build-dmg
+
+# Windows NSIS build
+npm run tauri-build-windows
+```
+
+## Tech Stack
+
+| Layer | Technology |
+|:---|:---|
+| Desktop Framework | [Tauri v2](https://v2.tauri.app/) |
+| Frontend | [Vue 3](https://vuejs.org/) + [TypeScript](https://www.typescriptlang.org/) |
+| State Management | [Pinia](https://pinia.vuejs.org/) |
+| Styling | [Tailwind CSS](https://tailwindcss.com/) |
+| Internationalization | [vue-i18n](https://vue-i18n.intlify.dev/) |
+| Speech Recognition | [whisper-rs](https://github.com/tazz4843/whisper-rs) (Whisper) |
+| Voice Detection | [ort](https://github.com/pykeio/ort) (Silero VAD ONNX) |
+| Audio Processing | FFmpeg |
+| LLM Translation | OpenAI-compatible API |
+
+## Project Structure
+
+```
+src/ Vue frontend
+ components/ UI components (TaskQueue, SubtitleEditor)
+ stores/ Pinia state management
+ locales/ i18n locale files (zh-CN, en)
+ lib/ Type definitions
+src-tauri/ Rust backend
+ src/
+ audio.rs Audio extraction & WAV reading
+ vad.rs Silero VAD voice activity detection
+ whisper.rs Whisper speech recognition interface
+ translate.rs OpenAI-compatible translation interface
+ subtitle.rs SRT / VTT / ASS export
+ task.rs Task orchestration & event broadcasting
+ state.rs Application state
+```
+
+## License
+
+This project is licensed under the [MIT](./LICENSE) License.
+
+## Acknowledgements
+
+- [whisper.cpp](https://github.com/ggerganov/whisper.cpp) — High-performance Whisper inference implementation
+- [Silero VAD](https://github.com/snakers4/silero-vad) — High-accuracy voice activity detection
+- [Tauri](https://tauri.app/) — Lightweight desktop application framework
+- All contributors and users
+
+---
+
+