| """ |
| Proxy Handler Class |
| """ |
| import json |
| from queue import Queue |
| import time |
| import urllib.parse |
| import threading |
| import requests |
|
|
| class ThreadSafeDict(dict): |
| """ |
| Thread safe dict |
| """ |
| def __init__(self, *args, **kwargs): |
| super().__init__(*args, **kwargs) |
| self.lock = threading.Lock() |
| def __getitem__(self, key): |
| with self.lock: |
| return super().__getitem__(key) |
| def __setitem__(self, key, value): |
| with self.lock: |
| return super().__setitem__(key, value) |
| def __delitem__(self, key): |
| with self.lock: |
| return super().__delitem__(key) |
| def __contains__(self, key): |
| with self.lock: |
| return super().__contains__(key) |
| def __len__(self): |
| with self.lock: |
| return super().__len__() |
| def __iter__(self): |
| with self.lock: |
| return super().__iter__() |
| def __repr__(self): |
| with self.lock: |
| return super().__repr__() |
| def __str__(self): |
| with self.lock: |
| return super().__str__() |
|
|
| class ProxyHandler: |
| """ |
| Sends request to http://{ip}:{port}/get_response_raw?url={url} with auth |
| """ |
| def __init__(self, proxy_list_file,proxy_auth="user:pass",port=80, wait_time=0.1,timeouts=10): |
| self.proxy_auth = proxy_auth |
| self.port = port |
| self.proxy_list = [] |
| self.commit_time = ThreadSafeDict() |
| self.timeouts = timeouts |
| self.wait_time = wait_time |
| self.lock = threading.Lock() |
| self.last_logged_activities = Queue(maxsize=100) |
| with open(proxy_list_file, 'r', encoding='utf-8') as f: |
| for line in f: |
| self.proxy_list.append(line.strip()) |
| for i, proxy in enumerate(self.proxy_list): |
| if not proxy.startswith("http"): |
| proxy = "http://" + proxy |
| if ":" not in proxy: |
| proxy += f":{self.port}" |
| if not proxy.endswith("/"): |
| proxy += "/" |
| self.proxy_list[i] = proxy |
| self.proxy_index = -1 |
| def log_time(self): |
| """ |
| Logs the time |
| """ |
| self.last_logged_activities.put(time.time()) |
| if self.last_logged_activities.full(): |
| |
| self.last_logged_activities.get() |
| def get_average_time(self): |
| """ |
| Returns the average time |
| """ |
| |
| if len(self.last_logged_activities.queue) > 1: |
| return (self.last_logged_activities.queue[-1] - self.last_logged_activities.queue[0]) / self.last_logged_activities.qsize() |
| return 0 |
| def wait_until_commit(self, proxy_index=None): |
| """ |
| Waits until the commit time |
| """ |
| if proxy_index is None: |
| proxy_index = self.proxy_index |
| if proxy_index not in self.commit_time: |
| self.commit_time[proxy_index] = 0 |
| while time.time() < self.commit_time[proxy_index] + self.wait_time: |
| time.sleep(0.01) |
| self.commit_time[proxy_index] = time.time() |
| def _update_proxy_index(self): |
| """ |
| Updates the proxy index |
| """ |
| with self.lock: |
| self.proxy_index = (self.proxy_index + 1) % len(self.proxy_list) |
| return self.proxy_index |
| def get_response(self, url): |
| """ |
| Returns the response of the url |
| """ |
| url = urllib.parse.quote(url, safe='') |
| try: |
| index = self._update_proxy_index() |
| self.wait_until_commit(index) |
| self.log_time() |
| response = requests.get(self.proxy_list[index] + f"get_response?url={url}", timeout=self.timeouts, auth=tuple(self.proxy_auth.split(":"))) |
| if response.status_code == 200: |
| json_response = response.json() |
| if json_response["success"]: |
| return json.loads(json_response["response"]) |
| else: |
| if "429" in json_response["response"]: |
| self.commit_time[index] = time.time() + self.timeouts |
| print(f"Error: {json_response['response']}, waiting {self.timeouts} seconds") |
| print(f"Failed in proxy side: {json_response['response']}") |
| return None |
| elif response.status_code == 429: |
| self.commit_time[index] = time.time() + self.timeouts |
| print(f"Error: {response.status_code}, waiting {self.timeouts} seconds") |
| else: |
| print(f"Failed in proxy side: {response.status_code}") |
| return None |
| except Exception as e: |
| print(f"Error while processing response from proxy: {e}") |
| return None |
| def get(self, url): |
| """ |
| Returns the response of the url |
| """ |
| url = urllib.parse.quote(url, safe='') |
| try: |
| index = self._update_proxy_index() |
| self.wait_until_commit(index) |
| response = requests.get(self.proxy_list[index] + f"get_response_raw?url={url}", timeout=self.timeouts, auth=tuple(self.proxy_auth.split(":"))) |
| if response.status_code == 200: |
| return response |
| else: |
| print(f"Error: {response.status_code}") |
| return None |
| except Exception as e: |
| print(f"Exception: {e}") |
| return None |
| def filesize(self, url): |
| """ |
| Returns the filesize of the url |
| """ |
| url = urllib.parse.quote(url, safe='') |
| try: |
| index = self._update_proxy_index() |
| self.wait_until_commit(index) |
| response = requests.get(self.proxy_list[index] + f"file_size?url={url}", timeout=self.timeouts, auth=tuple(self.proxy_auth.split(":"))) |
| if response.status_code == 200: |
| return int(response.text) |
| else: |
| print(f"Error: {response.status_code} when getting filesize from {url}") |
| return None |
| except Exception as e: |
| print(f"Exception: {e}") |
| return None |
| def get_filepart(self, url, start, end): |
| """ |
| Returns the response of the url with range |
| """ |
| url = urllib.parse.quote(url, safe='') |
| try: |
| index = self._update_proxy_index() |
| self.wait_until_commit(index) |
| response = requests.get(self.proxy_list[index] + f"filepart?url={url}&start={start}&end={end}", timeout=self.timeouts, auth=tuple(self.proxy_auth.split(":"))) |
| if response.status_code == 200: |
| return response |
| else: |
| print(f"Error: {response.status_code}") |
| return None |
| except Exception as e: |
| print(f"Exception: {e}") |
| return None |
| def check(self,raise_exception=False): |
| """ |
| Checks if the proxies are working |
| """ |
| failed_proxies = [] |
| for i, proxy in enumerate(self.proxy_list): |
| try: |
| response = requests.get(proxy, auth=tuple(self.proxy_auth.split(":")), timeout=2) |
| if response.status_code == 200: |
| continue |
| else: |
| print(f"Proxy {proxy} is not working") |
| failed_proxies.append(i) |
| except Exception as e: |
| print(f"Proxy {proxy} is not working: {e}") |
| failed_proxies.append(i) |
| if len(failed_proxies) > 0: |
| if raise_exception: |
| raise Exception(f"Proxies {failed_proxies} are not working") |
| else: |
| print(f"Proxies {failed_proxies} are not working, total {len(failed_proxies)} proxies of {len(self.proxy_list)} are not working") |
| |
| for i in failed_proxies[::-1]: |
| del self.proxy_list[i] |
| if len(self.proxy_list) == 0: |
| raise Exception("No proxies available") |
| else: |
| print(f"All {len(self.proxy_list)} proxies are working") |
|
|
| class SingleProxyHandler(ProxyHandler): |
| """ |
| Sends request to http://{ip}:{port}/get_response_raw?url={url} with auth |
| """ |
| def __init__(self, proxy_url, proxy_auth="user:pass",port=80, wait_time=0.1,timeouts=10): |
| self.proxy_auth = proxy_auth |
| self.port = port |
| self.proxy_list = [proxy_url] |
| self.proxy_index = -1 |
| self.commit_time = ThreadSafeDict() |
| self.timeouts = timeouts |
| self.wait_time = wait_time |
| self.lock = threading.Lock() |
|
|