Datadome is an anti-bot and anti-scraping service used by websites like Leboncoin, Vinted, Deezer etc. to block non-human visitors.
In this article, we'll be taking a look at how to bypass Datadome anti-scraping protection. We'll start by taking a quick look at what Datadome is, how to identify it and how is it identifying web scrapers. Then, we'll take a look at existing techniques and tools for bypassing Datadome bot protection. Let's dive in!
Legal Disclaimer and Precautions
This tutorial covers popular web scraping techniques for education. Interacting with public servers requires diligence and respect and here's a good summary of what not to do:
Do not scrape at rates that could damage the website.
Do not scrape data that's not available publicly.
Do not store PII of EU citizens who are protected by GDPR.
Do not repurpose the entire public datasets which can be illegal in some countries.
Scrapfly does not offer legal advice but these are good general rules to follow in web scraping
and for more you should consult a lawyer.
What is Datadome?
Datadome is a paid WAF service that protects websites from automated requests. In the context of security, it's used to block malicious bots and scripts, which causes several critical issues such as DDoS attacks and fraud activities.
In the context of web scraping, it's used to protect the public data on websites, which is particularly popular with European websites.
Datadome Block Page Examples
These errors are mostly encountered on the first request to the website. However, Datadome utilizes AI behavior analysis, making it able to block requests after a few successful requests.
How does Datadome Detect Web Scrapers?
To identify web scrapers, Datadome employs various techniques to get an estimate on whether the connecting client is a bot or a real user.
Based on the final trust score, Datadome either lets the user in, blocks them or requests a captcha challenge to be solved.
This complex process is done in real-time, making web scraping difficult as many factors can influence the trust score. However, by understanding each step of this process we have a good chance of bypassing Datadome bot protection. Let's take a look at each step in detail.
TLS (or SSL) is the first step in the HTTP connection. When using encrypted connections, like the https instead of http, both the server and client have to negotiate an encryption method. With the availability of various encryption methods and ciphers, the negotiation process can reveal significant information about the client.
This is generally referred to as JA3 fingerprinting. Different operating systems, web browsers or programming libraries have varying access to TLS encryption which results in different JA3 fingerprints.
If a scraper uses a library that has different TLS capabilities of a usual web browser, it can be identified using this method.
So, use web scraping libraries and tools that are resistant to JA3 fingerprinting. There are many online tools like ScrapFly's JA3 fingerprint web tool that can be used to validate your tools for JA3 fingerprinting.
The next step is IP address analysis. Datadome can access many different IP databases and look up the connecting client's IP address. This can be used to identify the client's location, ISP, reputation and other related information.
The most critical metric used here is IP address type, as there are three different types of IP addresses:
Residential are home addresses assigned by internet provides to average people. So, residential IP addresses provide a positive trust score as these are mostly used by humans and are expensive to acquire.
Mobile addresses are assigned by mobile phone towers and mobile users. So, mobile IPs also provide a positive trust score as these are mostly used by humans. In addition, since mobile towers might share and recycle IP addresses it makes it much more difficult to rely on IP addresses for identification.
Datacenter addresses are assigned to various data centers and server platforms like Amazon's AWS, Google Cloud etc. So, datacenter IPs provide a significant negative trust score as they are likely to be used by scripts.
Using IP analysis Datadome can have a rough estimate of how likely the connecting client is a human or a bot. For example, very few people browse the web from IPs owned by data centers.
Moreover, Datadome can block a client if the requesting rate is extensive in a short time window.
So, rotate high-quality residential or mobile IP addresses to bypass Datadome while scraping.
The next step is to analyze the HTTP connection details. HTTP protocol is becoming increasingly complex, making it easier to identify connections from web scrapers.
Most of the web operates through HTTP2 or HTTP3 while most web scraping libraries still use HTTP1.1. So, if a connecting client uses HTTP1.1, it's likely that this is a bot. That being said, many modern libraries like Python's httpx and cURL support HTTP2 but it's not enabled by default.
HTTP2 is also susceptible to HTTP2 fingerprinting which can be used to identify web scrapers. See our http2 fingerprint test page for more info.
Then, request headers and header order plays an important role in identifying web scrapers. Since most web browsers have strict header value and order rules, any mismatch like missing Origin or User-Agent header can leak the fact that the request sender is a bot.
So, make sure to use HTTP2 and match header values and order of real web browser.
Hardware and operating system details
Web browser information and capabilities
A more practical approach is to use a real web browser for web scraping. This can be done using browser automation libraries like Selenium, Puppeteer or Playwright that can start a real headless browser and navigate it for web scraping.
Many advanced scraping tools can even combine browser and HTTP scraping capabilities for optimal performance. Using resource-heavy browsers to establish a trust score and continue scraping using fast HTTP clients like httpx in Python - this feature is also available using Scrapfly sessions.
Datadome is using AI to analyze connection patterns and user profiles. So, even with the above steps passed, Datadome can still block the client if it detects suspicious behavior.
This means the trust score is not a static number but is constantly being adjusted based on the client's behavior.
So, it's important to distribute web scraper traffic through multiple different agents using proxies and different fingerprinting configurations tp bypass Datadome. For example, when scraping using browser automation tools, it's important to use different browser profiles like screen size, operating systems and rendering capabilities.
How to Bypass Datadome Bot Protection?
We can see what a complex process Datadome is using to identify web scrapers. Fortunately, this can work to our advantage as by avoiding common pitfalls and web scraper agent details it is possible to bypass Datadome bot protection. Here's a quick summary:
Use high-quality residential or mobile IP addresses
Use HTTP2 and match header values and order of real web browser
Introduce browser automation using tools like Selenium, Puppeteer or Playwright
Distribute web scraper traffic through multiple different agents
Note that as Datadome develops it introduces more techniques to identify web scrapers. So, it's important to keep up with the latest developments and use the latest web scraping tools.
For example, recently Datadome was updated with the capability to detecting headless browser use. So, plugins like Puppeteer stealth can be a valuable addition to bypass Datadome in 2024 when web scraping.
Bypass Datadome with Scrapfly
While bypassing Datadome is possible, maintaining bypass strategies can be very time-consuming. This is where services like ScrapFly come in!
Using ScrapFly web scraping API we can hand over all of the web scraping complexity and bypass logic to an API!
Scrapfly is not only a Datadome bypasser but also offers many other web scraping features:
from scrapfly import ScrapflyClient, ScrapeConfig
scrapfly = ScrapflyClient(key="YOUR API KEY")
result = scrapfly.scrape(ScrapeConfig(
# and set proxies by country like France
# and proxy type like residential:
To wrap this article let's take a look at some frequently asked questions regarding web scraping Datadome protected pages:
Is it legal to scrape Datadome protected pages?
Yes. Web scraping publicly available data is perfectly legal around the world as long as the scrapers do not cause damage to the website.
Is it possible to bypass Datadome using cache services?
Yes, public page caching services like Google Cache or Archive.org can be used to bypass Datadome protected pages as Google and Archive tend to be whitelisted. However, not all pages are cached and the ones that are are often outdated making them unsuitable for web scraping. Cached pages can also be missing parts of content that are loaded dynamically.
Is it possible to bypass Datadome entirely and scrape the website directly?
This is more of an internet security problem as that would be possible only by taking advantage of a vulnerability. This can be illegal in some countries and is often very difficult to do either way.
In this article, we took a deep dive into Datadome anti-bot protection when web scraping.
Finally, we've taken a look at some frequently asked questions like alternative bypass methods and the legality of it all.
For an easier way to handle web scraper blocking and power up your web scrapers check out ScrapFly for free!
Learn how to prevent TLS fingerprinting by impersonating normal web browser configurations. We'll start by explaining what the Curl Impersonate is, how it works, how to install and use it. Finally, we'll explore using it with Python to avoid web scraping blocking.
In this article, we'll explore the FlareSolverr tool and how to use it to get around Cloudflare while scraping. We'll start by explaining what FlareSolverr is, how it works, how to install and use it. Let's get started!