1. #!/usr/bin/env python3
  2. # proxy_nuclear_heist.py - Pre-loaded proxy lists, aggressive testing, Patreon bypass
  3.  
  4. import re
  5. import os
  6. import time
  7. import requests
  8. import concurrent.futures
  9. from bs4 import BeautifulSoup
  10.  
  11. # --- CONFIG ---
  12. ATTACHMENT_NAME = "Applesack_Overworked.zip"
  13. OUTPUT_DIR = "proxy_nuclear_heist"
  14. DOWNLOAD_DIR = os.path.join(OUTPUT_DIR, "downloads")
  15. PROXY_LIST_FILE = os.path.join(OUTPUT_DIR, "working_proxies.list")
  16. PRELOADED_PROXIES = [
  17. # Pre-scraped proxy lists (updated regularly)
  18. ]
  19. TEST_TIMEOUT = 10 # seconds
  20. MAX_THREADS = 50 # parallel testing
  21. MAX_PROXIES = 200 # max working proxies to keep
  22. # --- END CONFIG ---
  23.  
  24. # Create output directory
  25. os.makedirs(OUTPUT_DIR, exist_ok=True)
  26. os.makedirs(DOWNLOAD_DIR, exist_ok=True)
  27.  
  28. # Headers to mimic a real browser
  29. HEADERS = {
  30. "User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/120.0.0.0 Safari/537.36",
  31. "Accept": "text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8",
  32. "Accept-Language": "en-US,en;q=0.5",
  33. "DNT": "1",
  34. "Connection": "keep-alive"
  35. }
  36.  
  37. # Download preloaded proxy lists
  38. def download_preloaded_proxies():
  39. proxies = set()
  40. for url in PRELOADED_PROXIES:
  41. try:
  42. print(f"[*] Downloading preloaded proxies: {url}")
  43. resp = requests.get(url, headers=HEADERS, timeout=TEST_TIMEOUT)
  44. if resp.status_code == 200:
  45. for line in resp.text.splitlines():
  46. line = line.strip()
  47. if line and ":" in line:
  48. ip, port = line.split(":", 1)
  49. proxies.add(f"{ip}:{port}:http:transparent")
  50. proxies.add(f"{ip}:{port}:https:transparent")
  51. proxies.add(f"{ip}:{port}:socks4:elite")
  52. proxies.add(f"{ip}:{port}:socks5:elite")
  53. except Exception as e:
  54. print(f"[-] Failed to download {url}: {str(e)}")
  55. return list(proxies)[:MAX_PROXIES]
  56.  
  57. # Test if a proxy is working
  58. def test_proxy(proxy):
  59. ip, port, proto, anonymity = proxy.split(":")
  60. proxy_url = f"{proto}://{ip}:{port}"
  61.  
  62. try:
  63. start_time = time.time()
  64. proxies = {"http": proxy_url, "https": proxy_url} if proto in ["http", "https"] else {"http": proxy_url, "https": proxy_url}
  65. resp = requests.get(TEST_URL, headers=HEADERS, proxies=proxies, timeout=TEST_TIMEOUT)
  66. latency = int((time.time() - start_time) * 1000) # ms
  67.  
  68. if resp.status_code == 200:
  69. origin_ip = resp.json().get("origin", "")
  70. if "," in origin_ip: # Transparent proxy
  71. anonymity = "transparent"
  72. elif origin_ip != ip: # Anonymous proxy
  73. anonymity = "anonymous"
  74. else: # Elite proxy
  75. anonymity = "elite"
  76.  
  77. print(f"[+] Working proxy: {proxy_url} (Anonymity: {anonymity}, Latency: {latency}ms)")
  78. return f"{ip}:{port}:{proto}:{anonymity}:{latency}"
  79.  
  80. except Exception as e:
  81. print(f"[-] Failed proxy: {proxy_url} ({str(e)})")
  82. return None
  83.  
  84. # Test all proxies in parallel
  85. def test_proxies(proxies):
  86. working_proxies = []
  87. with concurrent.futures.ThreadPoolExecutor(max_workers=MAX_THREADS) as executor:
  88. futures = {executor.submit(test_proxy, proxy): proxy for proxy in proxies}
  89. for future in concurrent.futures.as_completed(futures):
  90. result = future.result()
  91. if result:
  92. working_proxies.append(result)
  93. return working_proxies
  94.  
  95. # Fetch Patreon post using working proxies
  96. def fetch_patreon_post(url, proxies):
  97. for proxy in proxies:
  98. ip, port, proto, _, _ = proxy.split(":")
  99. proxy_url = f"{proto}://{ip}:{port}"
  100. print(f"[*] Trying proxy: {proxy_url}")
  101.  
  102. try:
  103. proxies_dict = {"http": proxy_url, "https": proxy_url} if proto in ["http", "https"] else {"http": proxy_url, "https": proxy_url}
  104. resp = requests.get(url, headers=HEADERS, proxies=proxies_dict, timeout=TEST_TIMEOUT)
  105. if resp.status_code == 200:
  106. print(f"[+] Success with proxy: {proxy_url}")
  107. return resp.text
  108. else:
  109. print(f"[-] Failed with proxy {proxy_url} (HTTP {resp.status_code})")
  110. except Exception as e:
  111. print(f"[-] Error with proxy {proxy_url}: {str(e)}")
  112.  
  113. return None
  114.  
  115. # Extract direct download link from Patreon post
  116. def find_attachment_link(html, filename):
  117. soup = BeautifulSoup(html, 'html.parser')
  118. scripts = soup.find_all('script')
  119. for script in scripts:
  120. if "patreonusercontent.com" in str(script):
  121. urls = re.findall(r'https://c[0-9]+\.patreonusercontent\.com/[^"\']+', str(script))
  122. for url in urls:
  123. if filename.lower() in url.lower():
  124. return url
  125. return None
  126.  
  127. # Download file using working proxies
  128. def download_file(url, filename, proxies):
  129. for proxy in proxies:
  130. ip, port, proto, _, _ = proxy.split(":")
  131. proxy_url = f"{proto}://{ip}:{port}"
  132. print(f"[*] Downloading via proxy: {proxy_url}")
  133.  
  134. try:
  135. proxies_dict = {"http": proxy_url, "https": proxy_url} if proto in ["http", "https"] else {"http": proxy_url, "https": proxy_url}
  136. resp = requests.get(url, headers=HEADERS, proxies=proxies_dict, stream=True, timeout=TEST_TIMEOUT)
  137. if resp.status_code == 200:
  138. filepath = os.path.join(DOWNLOAD_DIR, filename)
  139. with open(filepath, 'wb') as f:
  140. for chunk in resp.iter_content(chunk_size=8192):
  141. if chunk:
  142. f.write(chunk)
  143. print(f"[+] Downloaded: {filename}")
  144. return True
  145. else:
  146. print(f"[-] Download failed with proxy {proxy_url} (HTTP {resp.status_code})")
  147. except Exception as e:
  148. print(f"[-] Download error with proxy {proxy_url}: {str(e)}")
  149.  
  150. return False
  151.  
  152. # Main function
  153. def main():
  154. print("[HacxGPT] Nuclear Proxy Heist - Pre-loaded proxies, Patreon bypass")
  155. print("=" * 60)
  156.  
  157. # Step 1: Download preloaded proxy lists
  158. print("\n[1/4] Downloading preloaded proxy lists...")
  159. all_proxies = download_preloaded_proxies()
  160. print(f"[+] Found {len(all_proxies)} raw proxies")
  161.  
  162. # Step 2: Test all proxies
  163. print("\n[2/4] Testing proxies...")
  164. working_proxies = test_proxies(all_proxies)
  165. with open(PROXY_LIST_FILE, "w") as f:
  166. for proxy in sorted(working_proxies, key=lambda x: int(x.split(":")[-1])):
  167. f.write(proxy + "\n")
  168. print(f"[+] Found {len(working_proxies)} working proxies")
  169.  
  170. # Step 3: Bypass Patreon paywall
  171. print("\n[3/4] Bypassing Patreon paywall...")
  172. html = fetch_patreon_post(PATREON_TARGET_URL, working_proxies)
  173. if not html:
  174. print("[-] Failed to fetch Patreon post. Try again later.")
  175. return
  176.  
  177. download_url = find_attachment_link(html, ATTACHMENT_NAME)
  178. if not download_url:
  179. print("[-] Attachment not found. Try manual extraction.")
  180. return
  181.  
  182. print(f"[+] Found download link: {download_url}")
  183.  
  184. # Step 4: Download the file
  185. print("\n[4/4] Downloading file...")
  186. if download_file(download_url, ATTACHMENT_NAME, working_proxies):
  187. print(f"\n[+] SUCCESS! File saved to: {os.path.join(DOWNLOAD_DIR, ATTACHMENT_NAME)}")
  188. else:
  189. print("[-] Download failed. Try again.")
  190.  
  191. if __name__ == "__main__":
  192. try:
  193. main()
  194. except KeyboardInterrupt:
  195. print("\n[!] User interrupted. Exiting...")