Recently, OpenAI released an operator agent that can browse the web and perform actions based on prompts. I attempted to build a basic computer-use capability for learning purposes. My approach differs from these agents and browser-use libraries because it has the potential to automate various computer tasks. Claude's computer use approach and ByteDance's UI-TARS are more advanced versions compared to my attempt. As part of this project, I built a basic library that performs minimal computer tasks.
Disclaimer: It is important to note that this is a starting point and requires further development to handle more complex scenarios.
A specific use case I envision is web search. For example, if we need to perform a web search on a specific topic, we often use Selenium or Playwright for web scraping. However, we frequently get blocked, especially by Google sites. If we can search the way a human does, these sites may not detect scraping.
Disclaimer: Currently, the library can navigate sites based on instructions, but data scraping is not yet implemented. Design
At the heart of this library lies the Qwen 2.5 VL 7B Instruct model. This powerful, 7-billion-parameter vision-language model can understand both image and text content. It goes beyond simple object detection by analyzing the structure and content of images, including text, charts, and layouts. This capability makes it well-suited for automating computer tasks, where understanding the visual context of a screen or interface is crucial.
This required an Nvidia GPU with 12GB of VRAM to run the Hugging Face model locally. It also required the Gemini API.
-
Setup conda enviornment with python 3.12
-
Computer-interact and other software installation
conda config --add channels pytorch conda config --add channels conda-forge conda config --add channels nvidia python -m pip install --upgrade computer-interact python -m pip install git+https://github.com/huggingface/transformers@f3f6c86582611976e72be054675e2bf0abb5f775
-
Environment Setup
This guide explains how to set up and manage environment variables for your project using python-dotenv.
a. Install the python-dotenv library using pip:
pip install python-dotenv
b. Create a .env file in your project's root directory with the following structure:
GEMINI_API_KEY=your_gemini_api_key
c. Add .env to your .gitignore file to prevent accidentally committing sensitive information:
d. code for load environment variables
from dotenv import load_dotenv import os load_dotenv()
- print config
print(supervisor.config)
#Typical output
{'debug': False, 'step_creation_model': 'gemini-2.0-pro-exp-02-05', 'computer_use_model': 'Qwen/Qwen2.5-VL-7B-Instruct'}
- modify
supervisor.config["debug"] = True
print(supervisor.config)
#Typical output
{'debug': False, 'step_creation_model': 'gemini-2.0-pro-exp-02-05', 'computer_use_model': 'Qwen/Qwen2.5-VL-7B-Instruct'}
from computer_interact.supervisor import Supervisor
from dotenv import load_dotenv
load_dotenv()
query = """
action 1: open the firfox web browser
action 2: click on the address bar
action 3: type scholar.google.com
action 4: press enter for search
action 5: type openai in the search box of google scholar
action 6: press enter for search
action 7: close the browser
"""
user_query = query
supervisor = Supervisor()
supervisor.config["debug"] = True
# Make sure the config is changed before the configure function call.
supervisor.configure()
supervisor.run(user_query)
Here is the link for the medium article regarding the library: https://medium.com/intro-to-artificial-intelligence/my-attempt-to-build-a-computer-use-capability-using-an-llm-35441b61e535