web-unblocker


Nameweb-unblocker JSON
Version 0.1.0 PyPI version JSON
download
home_pagehttps://oxylabs.io/products/web-unblocker
SummaryFree trial Web Unblocker - an AI-powered proxy solution that can bypass even the most sophisticated anti-bot systems.
upload_time2023-11-13 14:11:19
maintainer
docs_urlNone
authorOxylabs
requires_python>=3.6
licenseMIT
keywords captcha web-scraper anti-bot bypass captcha-solving captcha-breaking captcha-bypass bypasscaptcha cloudflare-bypass rotatecaptcha scraper-api web-proxies amazon-captcha web-unblocker unblockers web-scraping-api
VCS
bugtrack_url
requirements requests beautifulsoup4 lxml
Travis-CI No Travis.
coveralls test coverage No coveralls.
            # Web Unblocker

[![Web unblocker](https://user-images.githubusercontent.com/129506779/249698988-62f810bb-fe99-4c46-be3d-d4f7e4bc27f8.png)](https://oxylabs.go2cloud.org/aff_c?offer_id=7&aff_id=877&url_id=14)


- [Introduction](#introduction)
- [Getting started](#getting-started)
  - [InsecureRequestWarning](#insecurerequestwarning)
- [Scraping Google Search Results](#scraping-google-search-results)
  - [Sample Python Script](#sample-python-script)
  - [Using Web Unblocker](#using-web-unblocker-1)
    - [Using the Same IP](#using-the-same-ip)
    - [Using the Same Location](#using-the-same-location)
    - [Sending Custom Headers](#sending-custom-headers)
    - [Rendering and Screenshots](#rendering-and-screenshots)
- [Scraping Amazon](#scraping-amazon)
    - [Sample Python Script](#sample-python-script-1)
  - [Using Web Unblocker](#using-web-unblocker-2)
    - [Sending Custom Headers](#sending-custom-headers-1)
    - [Rendering](#rendering)
    - [Setting the ZIP code or the location](#setting-the-zip-code-or-the-location)
- [Conclusion](#conclusion)

## Introduction

[Web Unblocker](https://oxylabs.io/products/web-unblocker) is an AI-powered proxy solution capable of bypassing sophisticated anti-bot systems.

Web Unblocker has the following features:

- ML-driven proxy management 
- Dynamic browser fingerprinting 
- ML-powered response recognition
- Auto-retry functionality
- JavaScript rendering 


## Getting started

Execute the following `curl` command from your terminal:

```shellW
curl --insecure --proxy unblock.oxylabs.io:60000 --proxy-user "USERNAME:PASSWORD" https://ip.oxylabs.io
```

The output should be a random IP.

Notice a few things here:

- `--proxy` or `x` is used for proxy. Here, we use it for the URL of Web Unblocker
- `--insecure` or the equivalent `-k` is required for Web Unblocker to work
- `--proxy-user` or `-U` is used for the proxy user and password. If you don't have one, sign up for a [free trial](https://oxylabs.io/products/web-unblocker).

If you are observing low success rates or retrieve empty content, please try adding additional `"x-oxylabs-render: html"` header with your request.

You can find the equivalent python code in [getting_started.py](src/gettting_started.py):

```python
import requests

proxies = {
    'http': 'http://USERNAME:PASSWORD@unblock.oxylabs.io:60000',
    'https': 'http://USERNAME:PASSWORD@unblock.oxylabs.io:60000'
}

url = 'https://ip.oxylabs.io'
response = requests.get(url,
        proxies=proxies,
        verify=False) # disable SSL certificate verification

print(response.text)
```

### InsecureRequestWarning

One side effect of using `verify=False` is that you may receive warnings for `InsecureRequestWarning`.

Add these two lines to suppress these warnings:

```python
from requests.packages.urllib3.exceptions import InsecureRequestWarning

requests.packages.urllib3.disable_warnings(InsecureRequestWarning)
```

## Scraping Google Search Results

### Sample Python Script

The following is a [sample script](src/google_blocked.py) that scrapes Google search engine results.

```python
import requests
import bs4

keyword = "shoes"
url = f'https://google.com/search?q={keyword}'


request_result = requests.get(url)
soup = bs4.BeautifulSoup(request_result.text, "lxml")
search_headings = soup.find_all('h3')
for info in search_headings:
    print(info.getText())
```

If you run this code with Python a few times, you’ll be blocked by Google.

### Using Web Unblocker

The easiest solution to bypass all possible bans is to use Web Unblocker. Add the following lines to use Web Unblocker, just like you would use proxies.

```python
proxies = {
    'http': 'http://USERNAME:PASSWORD@unblock.oxylabs.io:60000',
    'https': 'http://USERNAME:PASSWORD@unblock.oxylabs.io:60000'
}

# Sending the proxy information
response = requests.get(url,
    proxies=proxies,
    verify=False)
```

For the complete code, see [google_unblocked.py](src/google_unblocked.py).

#### Using the Same IP

If you use Web Unblocker to scrape multiple pages, you may want to maintain the same IP or geographical region.

To use the same IP, send a header `X-Oxylabs-Session-Id`  and use any random string its value.

```python
headers = {
    "X-Oxylabs-Session-Id": "aRandomString"
}

response = requests.get(
    'https://www.google.com/search?q=shoes',
    verify=False, 
    proxies=proxies,
    headers=headers,
)
```

#### Using the Same Location

Instead of using the same IP, you can also rotate the IP from a specific location. To do that, send the header `x-oxylabs-geo-location` and set its value as a country, state, city, coordinates, or radius.

```python
headers = {
    "x-oxylabs-geo-location": "New York,New York,United States"
}

response = requests.get(
    'https://www.google.com/search?q=shoes',
    verify=False, 
    proxies=proxies,
    headers=headers,
)
```

For a complete example, see [google_location.py](src/google_location.py)

See [documentation](https://developers.oxylabs.io/advanced-proxy-solutions/web-unblocker/making-requests/geo-location#google) for more details.

#### Sending Custom Headers

You can add standard or custom headers to the request. 

One of the possible use cases is getting device-specific search results.

```python
headers = {
    "User-Agent": "Mozilla/5.0 (iPhone; CPU iPhone OS 16_3_1 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/16.3 Mobile/15E148 Safari/604.1",
    "Accept-Language": "en-US",
}

response = requests.get(url, 
    proxies=proxies, 
    verify=False, 
    headers=headers)
```

#### Rendering and Screenshots

Web Unblockers fully supports rendering. If you want to render a page, add the customer header as shown below:

```python
headers = {
    "X-Oxylabs-Render": "html"
}
```

In this particular example, Google doesn't need rendering. However, this can be used to take a screenshot. Send the same header, but change the value to `png`.

```python
headers = {
    "X-Oxylabs-Render": "png"
}

response = requests.get(
    url, verify=False, proxies=proxies, headers=headers,
)

# Save screenshot as PNG file
with open("google_rendered.png", 'wb') as f:
    f.write(response.content)
```

See [google_screenshots.py](src/google_screenshots.py) for the complete source code.

## Scraping Amazon

This section shows you how to scrape Amazon with Web Unblocker.

We start with a simple script and will add more features.

#### Sample Python Script

Examine the code in  [basic_script.py](src/amazon/basic_script.py). First, you will notice that Amazon will not even return a response without a valid user agent. 

Second, upon executing the following script a few times, Amazon will block you.

```python
import requests
url = 'https://www.amazon.com/Bose-QuietComfort-45-Bluetooth-Canceling-Headphones/dp/B098FKXT8L'
custom_headers = {
    'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/108.0.0.0 Safari/537.36',
    'accept-language': 'en-US,en;q=0.9',
}

response = requests.get(url, headers=custom_headers)

print(response.text)
```

### Using Web Unblocker

The easiest solution to bypass all the bans is to use the Web Unblocker.

Add the following lines to use the Web Unlcoker, just like you would use proxies.

```python
proxies = {
    'http': 'http://USERNAME:PASSWORD@unblock.oxylabs.io:60000',
    'https': 'http://USERNAME:PASSWORD@unblock.oxylabs.io:60000'
}

response = requests.get(url,
    proxies=proxies,
    verify=False,
    headers=custom_headers
)
```

Once you have the response, you can use BeautifulSoup to extract the product title and price as follows:

```python
soup = bs4.BeautifulSoup(response.text, "lxml")

product_title = soup.find('span', id="productTitle")
price_element = soup.select_one("div[role='radio'] [class='a-price-whole']")

print(product_title.getText())
if price_element:
    print(price_element.getText())

```

For the complete code, see [amazon_unblocked.py](src/amazon/amazon_unblocked.py).

#### Sending Custom Headers

As shown in the Sample Python Script, you cannot scrape Amazon without at least one header—user agent.

Similarly, you can send any other custom header, which will be forwarded to Amazon by Web Unblocker.

```python
url = 'https://www.amazon.com/Bose-QuietComfort-45-Bluetooth-Canceling-Headphones/dp/B098FKXT8L'

custom_headers = {
    'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/108.0.0.0 Safari/537.36',
    'accept-language': 'en-US,en;q=0.9',
}
proxies = {
    'http': 'http://USERNAME:PASSWORD@unblock.oxylabs.io:60000',
    'https': 'http://USERNAME:PASSWORD@unblock.oxylabs.io:60000'
}


response = requests.get(url,
                        proxies=proxies,
                        verify=False,
                        headers=custom_headers
                        )
```

For the complete code, see [amazon_unblocked.py](src/amazon/amazon_unblocked.py).

#### Rendering

Usually, sending the user agent is enough for web scraping Amazon, as shown in the previous example. If you are still facing problems, you can render the page by sending the header `X-Oxylabs-Render`:

```python
custom_headers = {
    "X-Oxylabs-Render": "html"
}
#...
response = requests.get(url,
                        proxies=proxies,
                        verify=False,
                        headers=custom_headers
                        )
```

The rest of the code remains the same. 

See [amazon_rendering.py](src/amazon/amazon_rendering.py) for the complete code.

#### Setting the ZIP code or the location

Using the `x-oxylabs-geo-location` parameter value for Amazon pages will yield a result with a corresponding delivery preference setting.

You can use this parameter to get correctly-localized Amazon results in a few ways. For most Amazon domains, you can send a zip/postcode or a [2-letter ISO 3166-1 alpha-2 country code](https://en.wikipedia.org/wiki/ISO_3166-1_alpha-2).

Note that this is different from scraping Google, where you cannot specify zip but instead can send a City. See [documentation](https://developers.oxylabs.io/advanced-proxy-solutions/web-unblocker/making-requests/geo-location#amazon) for more details.

```python
headers = {
    "x-oxylabs-geo-location": "11001" #New York ZIP
}

response = requests.get(
    url,
    verify=False, 
    proxies=proxies,
    headers=headers,
)
```

## Conclusion

You should now be able to scrape Google and Amazon with Web Unblocker. We have provided many examples in the [amazon](src/amazon) and [google](src/google) folders. 

Learn more about [Web Unblocker](https://oxylabs.io/products/web-unblocker).

If you face any problems, reach out to support.


            

Raw data

            {
    "_id": null,
    "home_page": "https://oxylabs.io/products/web-unblocker",
    "name": "web-unblocker",
    "maintainer": "",
    "docs_url": null,
    "requires_python": ">=3.6",
    "maintainer_email": "",
    "keywords": "captcha,web-scraper,anti-bot bypass,captcha-solving,captcha-breaking,captcha-bypass,bypasscaptcha,cloudflare-bypass,rotatecaptcha,scraper-api,web-proxies,amazon-captcha,web-unblocker,unblockers,web-scraping-api",
    "author": "Oxylabs",
    "author_email": "marketing@oxylabs.io",
    "download_url": "https://files.pythonhosted.org/packages/41/96/35d5fbadcf242bd8c32964f7f042fc59d2e96d3ea7c30f7f172b4f4fa611/web-unblocker-0.1.0.tar.gz",
    "platform": null,
    "description": "# Web Unblocker\n\n[![Web unblocker](https://user-images.githubusercontent.com/129506779/249698988-62f810bb-fe99-4c46-be3d-d4f7e4bc27f8.png)](https://oxylabs.go2cloud.org/aff_c?offer_id=7&aff_id=877&url_id=14)\n\n\n- [Introduction](#introduction)\n- [Getting started](#getting-started)\n  - [InsecureRequestWarning](#insecurerequestwarning)\n- [Scraping Google Search Results](#scraping-google-search-results)\n  - [Sample Python Script](#sample-python-script)\n  - [Using Web Unblocker](#using-web-unblocker-1)\n    - [Using the Same IP](#using-the-same-ip)\n    - [Using the Same Location](#using-the-same-location)\n    - [Sending Custom Headers](#sending-custom-headers)\n    - [Rendering and Screenshots](#rendering-and-screenshots)\n- [Scraping Amazon](#scraping-amazon)\n    - [Sample Python Script](#sample-python-script-1)\n  - [Using Web Unblocker](#using-web-unblocker-2)\n    - [Sending Custom Headers](#sending-custom-headers-1)\n    - [Rendering](#rendering)\n    - [Setting the ZIP code or the location](#setting-the-zip-code-or-the-location)\n- [Conclusion](#conclusion)\n\n## Introduction\n\n[Web Unblocker](https://oxylabs.io/products/web-unblocker) is an AI-powered proxy solution capable of bypassing sophisticated anti-bot systems.\n\nWeb Unblocker has the following features:\n\n- ML-driven proxy management \n- Dynamic browser fingerprinting \n- ML-powered response recognition\n- Auto-retry functionality\n- JavaScript rendering \n\n\n## Getting started\n\nExecute the following `curl` command from your terminal:\n\n```shellW\ncurl --insecure --proxy unblock.oxylabs.io:60000 --proxy-user \"USERNAME:PASSWORD\" https://ip.oxylabs.io\n```\n\nThe output should be a random IP.\n\nNotice a few things here:\n\n- `--proxy` or `x` is used for proxy. Here, we use it for the URL of Web Unblocker\n- `--insecure` or the equivalent `-k` is required for Web Unblocker to work\n- `--proxy-user` or `-U` is used for the proxy user and password. If you don't have one, sign up for a [free trial](https://oxylabs.io/products/web-unblocker).\n\nIf you are observing low success rates or retrieve empty content, please try adding additional `\"x-oxylabs-render: html\"` header with your request.\n\nYou can find the equivalent python code in [getting_started.py](src/gettting_started.py):\n\n```python\nimport requests\n\nproxies = {\n    'http': 'http://USERNAME:PASSWORD@unblock.oxylabs.io:60000',\n    'https': 'http://USERNAME:PASSWORD@unblock.oxylabs.io:60000'\n}\n\nurl = 'https://ip.oxylabs.io'\nresponse = requests.get(url,\n        proxies=proxies,\n        verify=False) # disable SSL certificate verification\n\nprint(response.text)\n```\n\n### InsecureRequestWarning\n\nOne side effect of using `verify=False` is that you may receive warnings for `InsecureRequestWarning`.\n\nAdd these two lines to suppress these warnings:\n\n```python\nfrom requests.packages.urllib3.exceptions import InsecureRequestWarning\n\nrequests.packages.urllib3.disable_warnings(InsecureRequestWarning)\n```\n\n## Scraping Google Search Results\n\n### Sample Python Script\n\nThe following is a [sample script](src/google_blocked.py) that scrapes Google search engine results.\n\n```python\nimport requests\nimport bs4\n\nkeyword = \"shoes\"\nurl = f'https://google.com/search?q={keyword}'\n\n\nrequest_result = requests.get(url)\nsoup = bs4.BeautifulSoup(request_result.text, \"lxml\")\nsearch_headings = soup.find_all('h3')\nfor info in search_headings:\n    print(info.getText())\n```\n\nIf you run this code with Python a few times, you\u2019ll be blocked by Google.\n\n### Using Web Unblocker\n\nThe easiest solution to bypass all possible bans is to use Web Unblocker. Add the following lines to use Web Unblocker, just like you would use proxies.\n\n```python\nproxies = {\n    'http': 'http://USERNAME:PASSWORD@unblock.oxylabs.io:60000',\n    'https': 'http://USERNAME:PASSWORD@unblock.oxylabs.io:60000'\n}\n\n# Sending the proxy information\nresponse = requests.get(url,\n    proxies=proxies,\n    verify=False)\n```\n\nFor the complete code, see [google_unblocked.py](src/google_unblocked.py).\n\n#### Using the Same IP\n\nIf you use Web Unblocker to scrape multiple pages, you may want to maintain the same IP or geographical region.\n\nTo use the same IP, send a header `X-Oxylabs-Session-Id`  and use any random string its value.\n\n```python\nheaders = {\n    \"X-Oxylabs-Session-Id\": \"aRandomString\"\n}\n\nresponse = requests.get(\n    'https://www.google.com/search?q=shoes',\n    verify=False, \n    proxies=proxies,\n    headers=headers,\n)\n```\n\n#### Using the Same Location\n\nInstead of using the same IP, you can also rotate the IP from a specific location. To do that, send the header `x-oxylabs-geo-location` and set its value as a country, state, city, coordinates, or radius.\n\n```python\nheaders = {\n    \"x-oxylabs-geo-location\": \"New York,New York,United States\"\n}\n\nresponse = requests.get(\n    'https://www.google.com/search?q=shoes',\n    verify=False, \n    proxies=proxies,\n    headers=headers,\n)\n```\n\nFor a complete example, see [google_location.py](src/google_location.py)\n\nSee [documentation](https://developers.oxylabs.io/advanced-proxy-solutions/web-unblocker/making-requests/geo-location#google) for more details.\n\n#### Sending Custom Headers\n\nYou can add standard or custom headers to the request. \n\nOne of the possible use cases is getting device-specific search results.\n\n```python\nheaders = {\n    \"User-Agent\": \"Mozilla/5.0 (iPhone; CPU iPhone OS 16_3_1 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/16.3 Mobile/15E148 Safari/604.1\",\n    \"Accept-Language\": \"en-US\",\n}\n\nresponse = requests.get(url, \n    proxies=proxies, \n    verify=False, \n    headers=headers)\n```\n\n#### Rendering and Screenshots\n\nWeb Unblockers fully supports rendering. If you want to render a page, add the customer header as shown below:\n\n```python\nheaders = {\n    \"X-Oxylabs-Render\": \"html\"\n}\n```\n\nIn this particular example, Google doesn't need rendering. However, this can be used to take a screenshot. Send the same header, but change the value to `png`.\n\n```python\nheaders = {\n    \"X-Oxylabs-Render\": \"png\"\n}\n\nresponse = requests.get(\n    url, verify=False, proxies=proxies, headers=headers,\n)\n\n# Save screenshot as PNG file\nwith open(\"google_rendered.png\", 'wb') as f:\n    f.write(response.content)\n```\n\nSee [google_screenshots.py](src/google_screenshots.py) for the complete source code.\n\n## Scraping Amazon\n\nThis section shows you how to scrape Amazon with Web Unblocker.\n\nWe start with a simple script and will add more features.\n\n#### Sample Python Script\n\nExamine the code in  [basic_script.py](src/amazon/basic_script.py). First, you will notice that Amazon will not even return a response without a valid user agent. \n\nSecond, upon executing the following script a few times, Amazon will block you.\n\n```python\nimport requests\nurl = 'https://www.amazon.com/Bose-QuietComfort-45-Bluetooth-Canceling-Headphones/dp/B098FKXT8L'\ncustom_headers = {\n    'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/108.0.0.0 Safari/537.36',\n    'accept-language': 'en-US,en;q=0.9',\n}\n\nresponse = requests.get(url, headers=custom_headers)\n\nprint(response.text)\n```\n\n### Using Web Unblocker\n\nThe easiest solution to bypass all the bans is to use the Web Unblocker.\n\nAdd the following lines to use the Web Unlcoker, just like you would use proxies.\n\n```python\nproxies = {\n    'http': 'http://USERNAME:PASSWORD@unblock.oxylabs.io:60000',\n    'https': 'http://USERNAME:PASSWORD@unblock.oxylabs.io:60000'\n}\n\nresponse = requests.get(url,\n    proxies=proxies,\n    verify=False,\n    headers=custom_headers\n)\n```\n\nOnce you have the response, you can use BeautifulSoup to extract the product title and price as follows:\n\n```python\nsoup = bs4.BeautifulSoup(response.text, \"lxml\")\n\nproduct_title = soup.find('span', id=\"productTitle\")\nprice_element = soup.select_one(\"div[role='radio'] [class='a-price-whole']\")\n\nprint(product_title.getText())\nif price_element:\n    print(price_element.getText())\n\n```\n\nFor the complete code, see [amazon_unblocked.py](src/amazon/amazon_unblocked.py).\n\n#### Sending Custom Headers\n\nAs shown in the Sample Python Script, you cannot scrape Amazon without at least one header\u2014user agent.\n\nSimilarly, you can send any other custom header, which will be forwarded to Amazon by Web Unblocker.\n\n```python\nurl = 'https://www.amazon.com/Bose-QuietComfort-45-Bluetooth-Canceling-Headphones/dp/B098FKXT8L'\n\ncustom_headers = {\n    'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/108.0.0.0 Safari/537.36',\n    'accept-language': 'en-US,en;q=0.9',\n}\nproxies = {\n    'http': 'http://USERNAME:PASSWORD@unblock.oxylabs.io:60000',\n    'https': 'http://USERNAME:PASSWORD@unblock.oxylabs.io:60000'\n}\n\n\nresponse = requests.get(url,\n                        proxies=proxies,\n                        verify=False,\n                        headers=custom_headers\n                        )\n```\n\nFor the complete code, see [amazon_unblocked.py](src/amazon/amazon_unblocked.py).\n\n#### Rendering\n\nUsually, sending the user agent is enough for web scraping Amazon, as shown in the previous example. If you are still facing problems, you can render the page by sending the header `X-Oxylabs-Render`:\n\n```python\ncustom_headers = {\n    \"X-Oxylabs-Render\": \"html\"\n}\n#...\nresponse = requests.get(url,\n                        proxies=proxies,\n                        verify=False,\n                        headers=custom_headers\n                        )\n```\n\nThe rest of the code remains the same. \n\nSee [amazon_rendering.py](src/amazon/amazon_rendering.py) for the complete code.\n\n#### Setting the ZIP code or the location\n\nUsing the `x-oxylabs-geo-location` parameter value for Amazon pages will yield a result with a corresponding delivery preference setting.\n\nYou can use this parameter to get correctly-localized Amazon results in a few ways. For most Amazon domains, you can send a zip/postcode or a [2-letter ISO 3166-1 alpha-2 country code](https://en.wikipedia.org/wiki/ISO_3166-1_alpha-2).\n\nNote that this is different from scraping Google, where you cannot specify zip but instead can send a City. See [documentation](https://developers.oxylabs.io/advanced-proxy-solutions/web-unblocker/making-requests/geo-location#amazon) for more details.\n\n```python\nheaders = {\n    \"x-oxylabs-geo-location\": \"11001\" #New York ZIP\n}\n\nresponse = requests.get(\n    url,\n    verify=False, \n    proxies=proxies,\n    headers=headers,\n)\n```\n\n## Conclusion\n\nYou should now be able to scrape Google and Amazon with Web Unblocker. We have provided many examples in the [amazon](src/amazon) and [google](src/google) folders. \n\nLearn more about [Web Unblocker](https://oxylabs.io/products/web-unblocker).\n\nIf you face any problems, reach out to support.\n\n",
    "bugtrack_url": null,
    "license": "MIT",
    "summary": "Free trial Web Unblocker - an AI-powered proxy solution that can bypass even the most sophisticated anti-bot systems.",
    "version": "0.1.0",
    "project_urls": {
        "Bug Reports": "https://github.com/oxylabs/web-unblocker/issues",
        "Documentation": "https://developers.oxylabs.io/advanced-proxy-solutions/web-unblocker",
        "Homepage": "https://oxylabs.io/products/web-unblocker",
        "Source": "https://github.com/oxylabs/web-unblocker"
    },
    "split_keywords": [
        "captcha",
        "web-scraper",
        "anti-bot bypass",
        "captcha-solving",
        "captcha-breaking",
        "captcha-bypass",
        "bypasscaptcha",
        "cloudflare-bypass",
        "rotatecaptcha",
        "scraper-api",
        "web-proxies",
        "amazon-captcha",
        "web-unblocker",
        "unblockers",
        "web-scraping-api"
    ],
    "urls": [
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "7b0b1f12b1169ee419faa085e087ab3e909be3a3c2f1f8f80d32c5152bcbc8a5",
                "md5": "6dfb00e2296211f5f2e35c75f4aa79c6",
                "sha256": "df2060ec4819a13a4e469da3383c0242c0a3d2e5636d4cfed3dc7a78617fefa3"
            },
            "downloads": -1,
            "filename": "web_unblocker-0.1.0-py3-none-any.whl",
            "has_sig": false,
            "md5_digest": "6dfb00e2296211f5f2e35c75f4aa79c6",
            "packagetype": "bdist_wheel",
            "python_version": "py3",
            "requires_python": ">=3.6",
            "size": 4797,
            "upload_time": "2023-11-13T14:11:17",
            "upload_time_iso_8601": "2023-11-13T14:11:17.143531Z",
            "url": "https://files.pythonhosted.org/packages/7b/0b/1f12b1169ee419faa085e087ab3e909be3a3c2f1f8f80d32c5152bcbc8a5/web_unblocker-0.1.0-py3-none-any.whl",
            "yanked": false,
            "yanked_reason": null
        },
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "419635d5fbadcf242bd8c32964f7f042fc59d2e96d3ea7c30f7f172b4f4fa611",
                "md5": "c5a4aa13050269c53a44eaf807baaea9",
                "sha256": "9bf865bbbb021442e9bfa253e5ad084f9854391d3a820e5140672cb76fffea9e"
            },
            "downloads": -1,
            "filename": "web-unblocker-0.1.0.tar.gz",
            "has_sig": false,
            "md5_digest": "c5a4aa13050269c53a44eaf807baaea9",
            "packagetype": "sdist",
            "python_version": "source",
            "requires_python": ">=3.6",
            "size": 5077,
            "upload_time": "2023-11-13T14:11:19",
            "upload_time_iso_8601": "2023-11-13T14:11:19.033297Z",
            "url": "https://files.pythonhosted.org/packages/41/96/35d5fbadcf242bd8c32964f7f042fc59d2e96d3ea7c30f7f172b4f4fa611/web-unblocker-0.1.0.tar.gz",
            "yanked": false,
            "yanked_reason": null
        }
    ],
    "upload_time": "2023-11-13 14:11:19",
    "github": true,
    "gitlab": false,
    "bitbucket": false,
    "codeberg": false,
    "github_user": "oxylabs",
    "github_project": "web-unblocker",
    "travis_ci": false,
    "coveralls": false,
    "github_actions": false,
    "requirements": [
        {
            "name": "requests",
            "specs": []
        },
        {
            "name": "beautifulsoup4",
            "specs": []
        },
        {
            "name": "lxml",
            "specs": []
        }
    ],
    "lcname": "web-unblocker"
}
        
Elapsed time: 0.31811s