Skip to content

web-infra-dev/midscene

Repository files navigation

Midscene.js

Midscene.js

Your AI Operator for Web, Android, Automation & Testing.

npm version huagging face model downloads License discord twitter

Midscene.js allows AI to serve as your web and Android operator ๐Ÿค–. Simply describe what you want to achieve in natural language, and it will assist you in operating the interface, validating content, and extracting data. Whether you seek a quick experience or in-depth development, you'll find it easy to get started.

Showcases

The following recorded example video is based on the UI-TARS-1.5-7B model, and the video has not been sped up at all~

Instruction Video
Post a Tweet
twitter-video-1080p.mp4
Use JS code to drive task orchestration, collect information about Jay Chou's concert, and write it into Google Docs
google-doc-1080p.mp4

๐Ÿ“ข 2025 Feb: New open-source model choice - UI-TARS and Qwen2.5-VL

Besides the default model GPT-4o, we have added two new recommended open-source models to Midscene.js: UI-TARS and Qwen2.5-VL. (Yes, Open Source models !) They are dedicated models for image recognition and UI automation, which are known for performing well in UI automation scenarios. Read more about it in Choose a model.

๐Ÿ’ก Features

  • Natural Language Interaction ๐Ÿ‘†: Just describe your goals and steps, and Midscene will plan and operate the user interface for you.
  • Chrome Extension Experience ๐Ÿ–ฅ๏ธ: Start in-browser experience immediately through the Chrome extension, no coding required.
  • Puppeteer/Playwright Integration ๐Ÿ”ง: Supports Puppeteer and Playwright integration, allowing you to combine AI capabilities with these powerful automation tools for easy automation.
  • Support Open-Source Models ๐Ÿค–: Supports private deployment of UI-TARS and Qwen2.5-VL, which outperforms closed-source models like GPT-4o and Claude in UI automation scenarios while better protecting data security.
  • Support General Models ๐ŸŒŸ: Supports general large models like GPT-4o and Claude, adapting to various scenario needs.
  • Visual Reports for Debugging ๐ŸŽž๏ธ: Through our test reports and Playground, you can easily understand, replay and debug the entire process.
  • Support Caching ๐Ÿ”„: The first time you execute a task through AI, it will be cached, and subsequent executions of the same task will significantly improve execution efficiency.
  • Completely Open Source ๐Ÿ”ฅ: Experience a whole new automation development experience, enjoy!
  • Understand UI, JSON Format Responses ๐Ÿ”: You can specify data format requirements and receive responses in JSON format.
  • Intuitive Assertions ๐Ÿค”: Express your assertions in natural language, and AI will understand and process them.

โœจ Model Choices

  • You can use general-purpose LLMs like gpt-4o, it works well for most cases. And also, gemini-1.5-pro, qwen-vl-max-latest are supported.
  • You can also use UI-TARS model, which is an open-source model dedicated for UI automation. You can deploy it on your own server, and it will dramatically improve the performance and data privacy.
  • Read more about Choose a model

๐Ÿ‘€ Comparing to ...

There are so many UI automation tools out there, and each one seems to be all-powerful. What's special about Midscene.js?

  • Debugging Experience: You will soon find that debugging and maintaining automation scripts is the real challenge point. No matter how magic the demo is, you still need to debug the process to make it stable over time. Midscene.js offers a visualized report file, a built-in playground, and a Chrome Extension to debug the entire process. This is what most developers really need. And we're continuing to work on improving the debugging experience.

  • Open Source, Free, Deploy as you want: Midscene.js is an open-source project. It's decoupled from any cloud service and model provider, you can choose either public or private deployment. There is always a suitable plan for your business.

  • Integrate with Javascript: You can always bet on Javascript ๐Ÿ˜Ž

๐Ÿ“„ Resources

๐Ÿค Community

๐Ÿ“ Credits

We would like to thank the following projects:

  • Rsbuild for the build tool.
  • UI-TARS for the open-source agent model UI-TARS.
  • Qwen2.5-VL for the open-source VL model Qwen2.5-VL.
  • scrcpy and yume-chan allow us to control Android devices with browser.
  • appium-adb for the javascript bridge of adb.
  • YADB for the yadb tool which improves the performance of text input.

Citation

If you use Midscene.js in your research or project, please cite:

@software{Midscene.js,
  author = {Zhou, Xiao and Yu, Tao},
  title = {Midscene.js: Let AI be your browser operator.},
  year = {2025},
  publisher = {GitHub},
  url = {https://github.com/web-infra-dev/midscene}
}

๐Ÿ“ License

Midscene.js is MIT licensed.


If this project helps you or inspires you, please give us a โญ๏ธ