How to Solve the cURL (60) Error When Using Proxy?

The cURL (60) error is a common error encountered when using proxies with cURL. The full error traceback looks like the following:

curl: (60) schannel: SEC_E_UNTRUSTED_ROOT (0x80090325) - The certificate chain was issued by an authority that is not trusted.
More details here:

curl failed to verify the legitimacy of the server and therefore could not
establish a secure connection to it.

This error happens when cURL attempts to request a web page over the HTTPs protocol with a proxy server that doesn't have a valid TLS certificate. Hence, it fails to establish a secure connection.

To solve the cURL (60) error, you can use a proxy that has a valid TLS certificate to support the HTTPs protocol. Alternatively, you can omit establishing the secure connection by adding the -k cURL option:

curl -x -k

For further details on proxies, including their types and best practices when choosing a proxy, refer to our introduction guide.

Introduction To Proxies in Web Scraping

Discover the different types of proxies, how they compare, and the best practices for web scraping while tackling common proxy challenges.

proxies article banner
Question tagged: cURL, Proxies

Related Posts

Sending HTTP Requests With Curlie: A better cURL

In this guide, we'll explore Curlie, a better cURL version. We'll start by defining what Curlie is and how it compares to cURL. We'll also go over a step-by-step guide on using and configuring Curlie to send HTTP requests.

How to Use cURL For Web Scraping

In this article, we'll go over a step-by-step guide on sending and configuring HTTP requests with cURL. We'll also explore advanced usages of cURL for web scraping, such as scraping dynamic pages and avoiding getting blocked.

Use Curl Impersonate to scrape as Chrome or Firefox

Learn how to prevent TLS fingerprinting by impersonating normal web browser configurations. We'll start by explaining what the Curl Impersonate is, how it works, how to install and use it. Finally, we'll explore using it with Python to avoid web scraping blocking.