
๐ค AI-powered Local Automation Tool | Make Your Computer Work For You
"Automate the tedious, reclaim your time for life"
automate.mp4
Special Note: The autoMate project is still in a very early stage and is currently more for learning purposes. We are constantly looking for breakthroughs and continuously integrating the latest technologies! If you have any questions, you can also add WeChat friends to join the group for communication.
Unlike the complexity of traditional RPA tools, autoMate leverages the power of large language models to complete complex automation processes with just natural language descriptions of tasks. Say goodbye to repetitive work and focus on what truly creates value!
Let automation create more possibilities for your life.
autoMate is a revolutionary AI+RPA automation tool built on OmniParser that can:
- ๐ Understand your needs and automatically plan tasks
- ๐ Intelligently comprehend screen content, simulating human vision and operations
- ๐ง Make autonomous decisions, judging and taking actions based on task requirements
- ๐ป Support local deployment to protect your data security and privacy
- ๐ฎ No-Code Automation - Describe tasks in natural language, no programming knowledge required
- ๐ฅ๏ธ Full Interface Control - Support operations on any visual interface, not limited to specific software
- ๐ Simplified Installation - Support for Chinese environment, one-click deployment
You can directly download the executable file from GitHub release to use.
It is strongly recommended to install miniConda first and use miniconda to install dependencies. There are many tutorials online; if you're still confused, you can ask AI. Then install the environment according to the following commands:
# Clone the project
git clone https://github.com/yuruotong1/autoMate.git
cd autoMate
# Create python3.12 environment
conda create -n "automate" python==3.12
# Activate environment
conda activate automate
# Install dependencies
python install.py
Launch Application
python main.py
Then open http://localhost:7888/
in your browser to configure your API key and basic settings.
The models that have been tested and are currently supported are as follows:
PS: The following are large language model providers that have been tested and confirmed to work. We have no business relationship with these providers, therefore we cannot guarantee after-sales service, feature assurance, or stability maintenance. Please consider carefully when using paid services.
Vendor | Model |
---|---|
yeka | gpt-4o,o1 |
openai | gpt-4o,gpt-4o-2024-08-06,gpt-4o-2024-11-20,o1,4.gpt-4.5-preview-2025-02-27 |
Currently, only OpenAI series models are supported. If you cannot access OpenAI from mainland China, it is recommended to use yeka as a proxy.
Why don't we support other models yet? We utilize multimodal + structured output capabilities, and few other model providers can simultaneously support these two capabilities. Adapting to other models would require significant changes to our underlying architecture, and effectiveness cannot be guaranteed. However, we are actively seeking solutions and will update as soon as progress is made.
If your computer doesn't have an NVIDIA dedicated GPU, it will run relatively slowly because we frequently call OCR to annotate visuals, which consumes a lot of GPU resources. We are actively optimizing and adapting. It is recommended to use an NVIDIA graphics card with no less than 4GB of memory, and ensure the version is compatible with your torch version:
- Run
pip list
to check your torch version; - Check supported CUDA versions from the official website;
- Uninstall existing torch and torchvision;
- Copy the official torch installation command and reinstall torch suitable for your CUDA version.
For example, if my CUDA version is 12.4, I need to install torch using the following command:
pip3 uninstall -y torch torchvision
pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu124
Every excellent open-source project crystallizes collective wisdom. The growth of autoMate depends on your participation and contribution. Whether fixing bugs, adding features, or improving documentation, each of your efforts will help thousands of people break free from repetitive work.
Join us in creating a smarter future.