Welcome! This guide assumes that you have a working 4get instance. This will help you configure your instance to the best it can be!
data/config.php
data/proxies/*.txt
data/captcha/your_image_set/*.png
data/fonts/captcha.ttf
Note: this only allows you to bypass the browser integrity checks. Captchas & javascript challenges will not be bypassed.
Configuring this lets you fetch images sitting behind Cloudflare and allows you to scrape the Yep & the Mwmbl search engines. Please be aware that APT will fight against you and will re-install the openSSL-version of curl constantly when updating.
First, follow these instructions. Only install the Firefox modules:
https://github.com/lwthiker/curl-impersonate/blob/main/INSTALL.md#native-build
Once you did this, you should be able to run the following inside your terminal:
$ curl_ff117 --version
curl 8.1.1 (x86_64-pc-linux-gnu) libcurl/8.1.1 NSS/3.92 zlib/1.2.13 brotli/1.0.9 zstd/1.5.4 libidn2/2.3.3 nghttp2/1.56.0
Release-Date: 2023-05-23
Protocols: dict file ftp ftps gopher gophers http https imap imaps mqtt pop3 pop3s rtsp smb smbs smtp smtps telnet tftp ws wss
Features: alt-svc AsynchDNS brotli HSTS HTTP2 HTTPS-proxy IDN IPv6 Largefile libz NTLM NTLM_WB SSL threadsafe UnixSockets zstd
Now, after compiling, you should have a libcurl-impersonate-ff.so
sitting somewhere. Mine (on my debian install) is located at /usr/local/lib/libcurl-impersonate-ff.so
.
Find the libcurl.so.4
file used by your current installation of curl. For me, this file is located at /usr/lib/x86_64-linux-gnu/libcurl.so.4
Now comes the sketchy part: replace libcurl.so.4
with libcurl-impersonate-ff.so
. You can do this in the following way:
sudo rm /usr/lib/x86_64-linux-gnu/libcurl.so.4
sudo cp /usr/local/lib/libcurl-impersonate-ff.so /usr/lib/x86_64-linux-gnu/libcurl.so.4
Make sure to restart your webserver and/or PHP daemon, otherwise it will keep using the old library. You should now be able to bypass Cloudflare's shitty checks!!
Make sure you configure this right to optimize your search engine presence! Head over to /robots.txt
and change the 4get.ca domain to your own domain.
To be listed on https://4get.ca/instances , you must contact any of the people in the server list and ask them to add you to their list of instances in their configuration. The instance list is distributed, and I don't have control over it.
If you see spammy entries in your instances list, simply remove the instance from your list that pushes the offending entries.
4get supports rotating proxies for scrapers! Configuring one is really easy.
myproxy
, but make sure it has the txt
extension.Add your proxies to the file. Examples:
# format -> <protocol>:<address>:<port>:<username>:<password>
# protocol list:
# raw_ip, http, https, socks4, socks5, socks4a, socks5_hostname
socks5:1.1.1.1:juicy:cloaca00
http:1.3.3.7::
raw_ip::::
Go to the main configuration file. Then, find which website you want to setup a proxy for.
Modify the value false
with "myproxy"
, with quotes included and the semicolon at the end.
Done! The scraper you chose should now be using the rotating proxies. When asking for the next page of results, it will use the same proxy to avoid detection!
If you ever test out a socks5
proxy locally on your machine and find out it works but doesn't on your server, try supplying the socks5_hostname
protocol instead. Hopefully this tip can save you 3 hours of your life!