mertunsal commited on
Commit
20bb7ec
·
verified ·
1 Parent(s): be4c3fb

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +109 -0
README.md ADDED
@@ -0,0 +1,109 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ pipeline_tag: image-text-to-text
3
+ library_name: transformers
4
+ language:
5
+ - en
6
+ base_model:
7
+ - Qwen/Qwen3-VL-30B-A3B-Instruct
8
+ tags:
9
+ - browser_use
10
+ ---
11
+
12
+
13
+ # BU-30B-A3B-Preview
14
+
15
+ <picture>
16
+ <source media="(prefers-color-scheme: light)" srcset="https://github.com/user-attachments/assets/2ccdb752-22fb-41c7-8948-857fc1ad7e24"">
17
+ <source media="(prefers-color-scheme: dark)" srcset="https://github.com/user-attachments/assets/774a46d5-27a0-490c-b7d0-e65fcbbfa358">
18
+ <img alt="Shows a black Browser Use Logo in light color mode and a white one in dark color mode." src="https://github.com/user-attachments/assets/2ccdb752-22fb-41c7-8948-857fc1ad7e24" width="full">
19
+ </picture>
20
+
21
+ Meet BU-30B-A3B-Preview — bringing SoTA Browser Use capabilities in a small model that can be hosted on a single GPU.
22
+
23
+ This model is heavily trained to be used with [browser-use OSS library](https://github.com/browser-use/browser-use) and provides comprehensive browsing capabilities with superior DOM understanding and visual reasoning.
24
+
25
+ ## Quickstart (BU Cloud)
26
+
27
+ You can directly use this model at BU Cloud. Simply
28
+
29
+ 1. Get your API key from [BU Cloud](https://cloud.browser-use.com/new-api-key)
30
+ 2. Set environment variable: export BROWSER_USE_API_KEY="your-key"
31
+ 3. Install the browser-use library following the instructions [here](https://github.com/browser-use/browser-use) and run
32
+
33
+ ```python
34
+ from dotenv import load_dotenv
35
+ from browser_use import Agent, ChatBrowserUse
36
+ load_dotenv()
37
+
38
+ llm = ChatBrowserUse(
39
+ model='browser-use/bu-30b-a3b-preview', # BU Open Source Model!!
40
+ )
41
+
42
+ agent = Agent(
43
+ task='Find the number of stars of browser-use and stagehand. Tell me which one has more stars :)',
44
+ llm=llm,
45
+ )
46
+ agent.run_sync()
47
+ ```
48
+
49
+ ## Quickstart (vLLM)
50
+
51
+ We recommend using this model with [vLLM](https://github.com/vllm-project/vllm).
52
+
53
+ #### Installation
54
+
55
+ Make sure to install **vllm >= 0.12.0**:
56
+
57
+ ```
58
+ pip install vllm --upgrade
59
+ ```
60
+
61
+ #### Serve
62
+
63
+ A simple launch command is:
64
+
65
+ ```bash
66
+ vllm serve browser-use/bu-30b-a3b-preview \
67
+ --max-model-len 32768 \
68
+ --host 0.0.0.0 \
69
+ --port 8000
70
+ ```
71
+
72
+ which will create an OpenAI compatible endpoint at localhost that you can use with.
73
+
74
+ ```python
75
+ from dotenv import load_dotenv
76
+ from browser_use import Agent, ChatOpenAI
77
+ load_dotenv()
78
+
79
+ llm = ChatOpenAI(
80
+ base_url='http://localhost:8000/v1',
81
+ model='browser-use/bu-30b-a3b-preview',
82
+ temperature=0.6,
83
+ top_p=0.95,
84
+ dont_force_structured_output=True, # speed up by disabling structured output
85
+ )
86
+
87
+ agent = Agent(
88
+ task='Find the number of stars of browser-use and stagehand. Tell me which one has more stars :)',
89
+ llm=llm,
90
+ )
91
+ agent.run_sync()
92
+ ```
93
+
94
+ ## Model Details
95
+
96
+ | Property | Value |
97
+ |----------|-------|
98
+ | **Base Model** | Qwen/Qwen3-VL-30B-A3B-Instruct |
99
+ | **Parameters** | 30B total, 3B active (MoE) |
100
+ | **Context Length** | 32,768 tokens |
101
+ | **Architecture** | Vision-Language Model (Mixture of Experts) |
102
+
103
+ ## Links
104
+
105
+ - 🌐 [Browser Use Cloud](https://cloud.browser-use.com)
106
+ - 📚 [Documentation](https://docs.browser-use.com)
107
+ - 💻 [GitHub](https://github.com/browser-use/browser-use)
108
+ - 💬 [Discord](https://link.browser-use.com/discord)
109
+