mirror of
https://github.com/browser-use/browser-use
synced 2026-05-06 17:52:15 +02:00
113 lines
2.9 KiB
Python
113 lines
2.9 KiB
Python
import asyncio
|
|
import http.client
|
|
import json
|
|
import os
|
|
import sys
|
|
|
|
sys.path.append(os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))))
|
|
|
|
from dotenv import load_dotenv
|
|
|
|
load_dotenv()
|
|
|
|
import logging
|
|
|
|
from pydantic import BaseModel
|
|
|
|
from browser_use import ActionResult, Agent, ChatOpenAI, Tools
|
|
from browser_use.browser.profile import BrowserProfile
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
class Person(BaseModel):
|
|
name: str
|
|
email: str | None = None
|
|
|
|
|
|
class PersonList(BaseModel):
|
|
people: list[Person]
|
|
|
|
|
|
SERP_API_KEY = os.getenv('SERPER_API_KEY')
|
|
if not SERP_API_KEY:
|
|
raise ValueError('SERPER_API_KEY is not set')
|
|
|
|
tools = Tools(exclude_actions=['search'], output_model=PersonList)
|
|
|
|
|
|
@tools.registry.action('Search the web for a specific query. Returns a short description and links of the results.')
|
|
async def search_web(query: str):
|
|
# do a serp search for the query
|
|
conn = http.client.HTTPSConnection('google.serper.dev')
|
|
payload = json.dumps({'q': query})
|
|
headers = {'X-API-KEY': SERP_API_KEY, 'Content-Type': 'application/json'}
|
|
conn.request('POST', '/search', payload, headers)
|
|
res = conn.getresponse()
|
|
data = res.read()
|
|
serp_data = json.loads(data.decode('utf-8'))
|
|
|
|
# exclude searchParameters and credits
|
|
serp_data = {k: v for k, v in serp_data.items() if k not in ['searchParameters', 'credits']}
|
|
|
|
# keep the value of the key "organic"
|
|
|
|
organic = serp_data.get('organic', [])
|
|
# remove the key "position"
|
|
organic = [{k: v for k, v in d.items() if k != 'position'} for d in organic]
|
|
|
|
# print the original data
|
|
logger.debug(json.dumps(organic, indent=2))
|
|
|
|
# to string
|
|
organic_str = json.dumps(organic)
|
|
|
|
return ActionResult(extracted_content=organic_str, include_in_memory=False, include_extracted_content_only_once=True)
|
|
|
|
|
|
names = [
|
|
'Ruedi Aebersold',
|
|
'Bernd Bodenmiller',
|
|
'Eugene Demler',
|
|
'Erich Fischer',
|
|
'Pietro Gambardella',
|
|
'Matthias Huss',
|
|
'Reto Knutti',
|
|
'Maksym Kovalenko',
|
|
'Antonio Lanzavecchia',
|
|
'Maria Lukatskaya',
|
|
'Jochen Markard',
|
|
'Javier Pérez-Ramírez',
|
|
'Federica Sallusto',
|
|
'Gisbert Schneider',
|
|
'Sonia I. Seneviratne',
|
|
'Michael Siegrist',
|
|
'Johan Six',
|
|
'Tanja Stadler',
|
|
'Shinichi Sunagawa',
|
|
'Michael Bruce Zimmermann',
|
|
]
|
|
|
|
|
|
async def main():
|
|
task = 'use search_web with "find email address of the following ETH professor:" for each of the following persons in a list of actions. Finally return the list with name and email if provided - do always 5 at once'
|
|
task += '\n' + '\n'.join(names)
|
|
model = ChatOpenAI(model='gpt-4.1-mini')
|
|
browser_profile = BrowserProfile()
|
|
agent = Agent(task=task, llm=model, tools=tools, browser_profile=browser_profile)
|
|
|
|
history = await agent.run()
|
|
|
|
result = history.final_result()
|
|
if result:
|
|
parsed: PersonList = PersonList.model_validate_json(result)
|
|
|
|
for person in parsed.people:
|
|
print(f'{person.name} - {person.email}')
|
|
else:
|
|
print('No result')
|
|
|
|
|
|
if __name__ == '__main__':
|
|
asyncio.run(main())
|