WebTools
The
WebTools
Class Methods OverviewCopied!
-
exa_search
-
scrape_urls
-
serper_search
-
scrape_url_with_serper
-
query_arxiv_api
-
get_weather_data
Let's dive into each method in detail, exploring their functionality, parameters, return values, and practical examples.
exa_searchCopied!
Method SignatureCopied!
@staticmethod def exa_search(queries: Union[str, List[str]], num_results: int = 10, search_type: str = "neural", num_sentences: int = 3, highlights_per_url: int = 3) -> dict:
ParametersCopied!
-
queries
-
num_results
-
search_type
-
num_sentences
-
highlights_per_url
Return ValueCopied!
-
dict
DescriptionCopied!
The
exa_search
The method first checks if the
EXA_API_KEY
ValueError
For each query, the method constructs a payload dictionary containing the search parameters, such as the query string, search type, number of results, and highlight settings. It then sends a POST request to the API endpoint with the payload.
If the request is successful, the method restructures the response data into a more user-friendly format. It extracts the relevant information, such as the title, URL, author, and highlights for each search result, and appends it to the
structured_data
If an error occurs during the request or JSON decoding, the method prints an error message and retries the request up to 3 times with a 1-second delay between attempts. If all attempts fail, it adds an empty result for that query to the
structured_data
Finally, the method returns the
structured_data
Example UsageCopied!
queries = ["Python programming", "Machine learning algorithms"] results = WebTools.exa_search(queries, num_results=5, search_type="neural", num_sentences=2, highlights_per_url=2) print(results)
scrape_urlsCopied!
Method SignatureCopied!
@staticmethod def scrape_urls(urls: Union[str, List[str]], include_html: bool = False, include_links: bool = False) -> List[Dict[str, Union[str, List[str]]]]:
ParametersCopied!
-
urls
-
include_html
-
include_links
Return ValueCopied!
-
List[Dict[str, Union[str, List[str]]]]
DescriptionCopied!
The
scrape_urls
For each URL, the method sends a GET request with a random user agent header to avoid being blocked by websites. It adds a small delay between requests to avoid overwhelming the server.
If the request is successful, the method parses the HTML content using BeautifulSoup. It removes any script, style, and SVG tags from the parsed content.
The method then constructs a dictionary for each URL, containing the URL and the scraped content. If
include_html
include_links
If an error occurs during the request or content processing, the method appends a dictionary to the results list with the URL and an error message describing the encountered issue.
Finally, the method returns the list of dictionaries containing the scraped data for each URL.
Example UsageCopied!
urls = ["https://www.example.com", "https://www.example.org"] scraped_data = WebTools.scrape_urls(urls, include_html=True, include_links=True) print(scraped_data)
get_weather_dataCopied!
Method SignatureCopied!
@staticmethod def get_weather_data( location: str, forecast_days: Optional[int] = None, include_current: bool = True, include_forecast: bool = True, include_astro: bool = False, include_hourly: bool = False, include_alerts: bool = False, ) -> str:
ParametersCopied!
-
location
-
forecast_days
-
include_current
-
include_forecast
-
include_astro
-
include_hourly
-
include_alerts
Return ValueCopied!
-
str
RaisesCopied!
-
ValueError
-
requests.RequestException
DescriptionCopied!
The
get_weather_data
The method first checks if the
WEATHER_API_KEY
ValueError
If
forecast_days
include_forecast
forecast_days
The method sends a GET request to the WeatherAPI endpoint with the constructed parameters. If the request is successful, it retrieves the JSON response data. If the API returns an error, it raises a
ValueError
The method then formats the weather data into a readable string report. It includes the location details, current conditions (if
include_current
include_forecast
include_astro
include_hourly
include_alerts
Finally, the method returns the formatted weather report as a string.
Example UsageCopied!
location = "New York" weather_report = WebTools.get_weather_data(location, forecast_days=3, include_current=True, include_forecast=True, include_astro=True, include_hourly=True, include_alerts=True) print(weather_report)
These are just a few examples of the powerful methods available in the
WebTools