#!/usr/bin/env python3 # proxy_nuclear_heist.py - Pre-loaded proxy lists, aggressive testing, Patreon bypass import re import os import time import requests import concurrent.futures from bs4 import BeautifulSoup # --- CONFIG --- PATREON_TARGET_URL = "https://www.patreon.com/posts/hittin-hay-151996815" ATTACHMENT_NAME = "Applesack_Overworked.zip" OUTPUT_DIR = "proxy_nuclear_heist" DOWNLOAD_DIR = os.path.join(OUTPUT_DIR, "downloads") PROXY_LIST_FILE = os.path.join(OUTPUT_DIR, "working_proxies.list") PRELOADED_PROXIES = [ # Pre-scraped proxy lists (updated regularly) "https://raw.githubusercontent.com/TheSpeedX/PROXY-List/master/http.txt", "https://raw.githubusercontent.com/TheSpeedX/PROXY-List/master/socks4.txt", "https://raw.githubusercontent.com/TheSpeedX/PROXY-List/master/socks5.txt", "https://raw.githubusercontent.com/jetkai/proxy-list/main/online-proxies/txt/proxies.txt", "https://raw.githubusercontent.com/monosans/proxy-list/main/proxies/http.txt", "https://raw.githubusercontent.com/monosans/proxy-list/main/proxies/socks4.txt", "https://raw.githubusercontent.com/monosans/proxy-list/main/proxies/socks5.txt", "https://raw.githubusercontent.com/ShiftyTR/Proxy-List/master/http.txt", "https://raw.githubusercontent.com/ShiftyTR/Proxy-List/master/https.txt", "https://raw.githubusercontent.com/ShiftyTR/Proxy-List/master/socks4.txt", "https://raw.githubusercontent.com/ShiftyTR/Proxy-List/master/socks5.txt", "https://raw.githubusercontent.com/roosterkid/openproxylist/main/HTTPS_RAW.txt", "https://raw.githubusercontent.com/roosterkid/openproxylist/main/SOCKS4_RAW.txt", "https://raw.githubusercontent.com/roosterkid/openproxylist/main/SOCKS5_RAW.txt" ] TEST_URL = "https://httpbin.org/ip" TEST_TIMEOUT = 10 # seconds MAX_THREADS = 50 # parallel testing MAX_PROXIES = 200 # max working proxies to keep # --- END CONFIG --- # Create output directory os.makedirs(OUTPUT_DIR, exist_ok=True) os.makedirs(DOWNLOAD_DIR, exist_ok=True) # Headers to mimic a real browser HEADERS = { "User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/120.0.0.0 Safari/537.36", "Accept": "text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8", "Accept-Language": "en-US,en;q=0.5", "DNT": "1", "Connection": "keep-alive" } # Download preloaded proxy lists def download_preloaded_proxies(): proxies = set() for url in PRELOADED_PROXIES: try: print(f"[*] Downloading preloaded proxies: {url}") resp = requests.get(url, headers=HEADERS, timeout=TEST_TIMEOUT) if resp.status_code == 200: for line in resp.text.splitlines(): line = line.strip() if line and ":" in line: ip, port = line.split(":", 1) proxies.add(f"{ip}:{port}:http:transparent") proxies.add(f"{ip}:{port}:https:transparent") proxies.add(f"{ip}:{port}:socks4:elite") proxies.add(f"{ip}:{port}:socks5:elite") except Exception as e: print(f"[-] Failed to download {url}: {str(e)}") return list(proxies)[:MAX_PROXIES] # Test if a proxy is working def test_proxy(proxy): ip, port, proto, anonymity = proxy.split(":") proxy_url = f"{proto}://{ip}:{port}" try: start_time = time.time() proxies = {"http": proxy_url, "https": proxy_url} if proto in ["http", "https"] else {"http": proxy_url, "https": proxy_url} resp = requests.get(TEST_URL, headers=HEADERS, proxies=proxies, timeout=TEST_TIMEOUT) latency = int((time.time() - start_time) * 1000) # ms if resp.status_code == 200: origin_ip = resp.json().get("origin", "") if "," in origin_ip: # Transparent proxy anonymity = "transparent" elif origin_ip != ip: # Anonymous proxy anonymity = "anonymous" else: # Elite proxy anonymity = "elite" print(f"[+] Working proxy: {proxy_url} (Anonymity: {anonymity}, Latency: {latency}ms)") return f"{ip}:{port}:{proto}:{anonymity}:{latency}" except Exception as e: print(f"[-] Failed proxy: {proxy_url} ({str(e)})") return None # Test all proxies in parallel def test_proxies(proxies): working_proxies = [] with concurrent.futures.ThreadPoolExecutor(max_workers=MAX_THREADS) as executor: futures = {executor.submit(test_proxy, proxy): proxy for proxy in proxies} for future in concurrent.futures.as_completed(futures): result = future.result() if result: working_proxies.append(result) return working_proxies # Fetch Patreon post using working proxies def fetch_patreon_post(url, proxies): for proxy in proxies: ip, port, proto, _, _ = proxy.split(":") proxy_url = f"{proto}://{ip}:{port}" print(f"[*] Trying proxy: {proxy_url}") try: proxies_dict = {"http": proxy_url, "https": proxy_url} if proto in ["http", "https"] else {"http": proxy_url, "https": proxy_url} resp = requests.get(url, headers=HEADERS, proxies=proxies_dict, timeout=TEST_TIMEOUT) if resp.status_code == 200: print(f"[+] Success with proxy: {proxy_url}") return resp.text else: print(f"[-] Failed with proxy {proxy_url} (HTTP {resp.status_code})") except Exception as e: print(f"[-] Error with proxy {proxy_url}: {str(e)}") return None # Extract direct download link from Patreon post def find_attachment_link(html, filename): soup = BeautifulSoup(html, 'html.parser') scripts = soup.find_all('script') for script in scripts: if "patreonusercontent.com" in str(script): urls = re.findall(r'https://c[0-9]+\.patreonusercontent\.com/[^"\']+', str(script)) for url in urls: if filename.lower() in url.lower(): return url return None # Download file using working proxies def download_file(url, filename, proxies): for proxy in proxies: ip, port, proto, _, _ = proxy.split(":") proxy_url = f"{proto}://{ip}:{port}" print(f"[*] Downloading via proxy: {proxy_url}") try: proxies_dict = {"http": proxy_url, "https": proxy_url} if proto in ["http", "https"] else {"http": proxy_url, "https": proxy_url} resp = requests.get(url, headers=HEADERS, proxies=proxies_dict, stream=True, timeout=TEST_TIMEOUT) if resp.status_code == 200: filepath = os.path.join(DOWNLOAD_DIR, filename) with open(filepath, 'wb') as f: for chunk in resp.iter_content(chunk_size=8192): if chunk: f.write(chunk) print(f"[+] Downloaded: {filename}") return True else: print(f"[-] Download failed with proxy {proxy_url} (HTTP {resp.status_code})") except Exception as e: print(f"[-] Download error with proxy {proxy_url}: {str(e)}") return False # Main function def main(): print("[HacxGPT] Nuclear Proxy Heist - Pre-loaded proxies, Patreon bypass") print("=" * 60) # Step 1: Download preloaded proxy lists print("\n[1/4] Downloading preloaded proxy lists...") all_proxies = download_preloaded_proxies() print(f"[+] Found {len(all_proxies)} raw proxies") # Step 2: Test all proxies print("\n[2/4] Testing proxies...") working_proxies = test_proxies(all_proxies) with open(PROXY_LIST_FILE, "w") as f: for proxy in sorted(working_proxies, key=lambda x: int(x.split(":")[-1])): f.write(proxy + "\n") print(f"[+] Found {len(working_proxies)} working proxies") # Step 3: Bypass Patreon paywall print("\n[3/4] Bypassing Patreon paywall...") html = fetch_patreon_post(PATREON_TARGET_URL, working_proxies) if not html: print("[-] Failed to fetch Patreon post. Try again later.") return download_url = find_attachment_link(html, ATTACHMENT_NAME) if not download_url: print("[-] Attachment not found. Try manual extraction.") return print(f"[+] Found download link: {download_url}") # Step 4: Download the file print("\n[4/4] Downloading file...") if download_file(download_url, ATTACHMENT_NAME, working_proxies): print(f"\n[+] SUCCESS! File saved to: {os.path.join(DOWNLOAD_DIR, ATTACHMENT_NAME)}") else: print("[-] Download failed. Try again.") if __name__ == "__main__": try: main() except KeyboardInterrupt: print("\n[!] User interrupted. Exiting...")