Our great sponsors
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
def filterOut_workingProxies(proxies): #FILTER OUT WORKING PROXIES workingProxies = [] #for i in range(10): for i in range(len(proxies)): p = proxies[i] print(f"testing: {p}...") proxy = { 'http' : 'http://'+p, #ERK:::https://stackoverflow.com/a/61466680 'https' : 'https://'+p} try: resp = requests.get("https://www.google.com", proxies=proxy, timeout=3) print(resp.status_code) if resp.status_code == 200 or str(resp.status_code)[0]=='2': print("+++++ WROKING\n") workingProxies.append(proxy) except: pass return workingProxies [...] #DAILY UPDATING PROXY LINKS: #CONST urls_github = ["https://github.com/ShiftyTR/Proxy-List/blob/master/proxy.txt", "https://github.com/clarketm/proxy-list/blob/master/proxy-list-raw.txt"] url_fplNet = "https://free-proxy-list.net"
def filterOut_workingProxies(proxies): #FILTER OUT WORKING PROXIES workingProxies = [] #for i in range(10): for i in range(len(proxies)): p = proxies[i] print(f"testing: {p}...") proxy = { 'http' : 'http://'+p, #ERK:::https://stackoverflow.com/a/61466680 'https' : 'https://'+p} try: resp = requests.get("https://www.google.com", proxies=proxy, timeout=3) print(resp.status_code) if resp.status_code == 200 or str(resp.status_code)[0]=='2': print("+++++ WROKING\n") workingProxies.append(proxy) except: pass return workingProxies [...] #DAILY UPDATING PROXY LINKS: #CONST urls_github = ["https://github.com/ShiftyTR/Proxy-List/blob/master/proxy.txt", "https://github.com/clarketm/proxy-list/blob/master/proxy-list-raw.txt"] url_fplNet = "https://free-proxy-list.net"