# Reversed by CrackerMain.
Net reversers iseeu && Example
import requests
import uuid
import os
from bs4 import BeautifulSoup
import threading
from tqdm import tqdm
import time
from colorama import Fore, Style, init
init(autoreset=True)
PREMIUM_GROUP_IDS = {'12', '101', '94', '4', '96', '93', '100', '3', '6', '92',
'11', '99', '97'}
RED = [Link] + [Link]
YELLOW = [Link] + [Link]
GREY = '[38;5;246m'
engadget_domains = ['[Link]', '[Link]', '[Link]', '[Link]',
'[Link]', '[Link]', '[Link]', '[Link]', '[Link]',
'[Link]', '[Link]', '[Link]', '[Link]', '[Link]',
'[Link]', '[Link]', '[Link]', '[Link]', '[Link]', '[Link]',
'[Link]', '[Link]', '[Link]', '[Link]', '[Link]',
'[Link]', '[Link]', '[Link]', '[Link]', '[Link]', '[Link]',
'[Link]', '[Link]', '[Link]', '[Link]', '[Link]',
'[Link]', '[Link]', '[Link]', '[Link]',
'[Link]', '[Link]', '[Link]', '[Link]', '[Link]',
'[Link]', '[Link]', '[Link]', '[Link]', '[Link]',
'[Link]']
def get_hwid():
return str([Link]())
def get_proxies_from_geonode():
url = '[Link]
limit=500&page=1&sort_by=lastChecked&sort_type=desc&speed=slow&
protocols=http'
try:
response = [Link](url)
response.raise_for_status()
data = [Link]()
proxies = []
for proxy in data['data']:
ip = proxy['ip']
port = proxy['port']
if 'http' in proxy['protocols']:
proxy_address =
f'[Link]
[Link](proxy_address)
return proxies
except [Link] as e:
print(f'Error fetching proxies from Geonode: {e}')
return []
def get_proxies_from_proxyscrape():
url = '[Link]
request=display_proxies&proxy_format=ipport&format=text&timeout=10000'
try:
response = [Link](url)
response.raise_for_status()
proxies = [Link]()
return proxies
except [Link] as e:
print(f'Error fetching proxies from ProxyScrape: {e}')
return []
def get_all_proxies():
proxies_from_geonode = get_proxies_from_geonode()
proxies_from_proxyscrape = get_proxies_from_proxyscrape()
return proxies_from_geonode + proxies_from_proxyscrape
def load_dorks(file_path):
with open(file_path, 'r') as file:
return [[Link]() for line in [Link]()]
def process_dork(dork, output_file, proxy, mode, failed_dorks, progress_bar,
retries_counter):
url = f'[Link]
try:
response = [Link](url, proxies={'http': proxy,
'https': proxy}, timeout=10)
soup = BeautifulSoup([Link], '[Link]')
results = soup.find_all('a')
success = False
with open(output_file, 'a') as f:
for link in results:
href = [Link]('href')
if href and
[Link]('/url?q='):
clean_link =
[Link]('/url?q=')[1].split('&')[0]
domain =
clean_link.split('/')[2] if '://' in clean_link else clean_link.split('/')[0]
if mode ==
'Full URL Mode':
[Link](clean_link + '\n')
elif mode ==
'Only Parameter Mode' and '?' in clean_link:
[Link](clean_link + '\n')
success =
True
if success:
with open('processed_dorks.txt', 'a') as
processed_file:
processed_file.write(dork +
'\n')
else:
raise Exception('No results')
progress_bar.update(1)
except Exception:
failed_dorks.append(dork)
retries_counter[0] += 1
def start_threads(dorks, num_threads, output_file, proxies, mode, progress_bar,
retries_counter):
threads = []
failed_dorks = []
for dork in dorks:
proxy = proxies[len(threads) % len(proxies)]
thread = [Link](target=process_dork,
args=(dork, output_file, proxy, mode, failed_dorks, progress_bar, retries_counter))
[Link](thread)
[Link]()
if len(threads) >= num_threads:
for thread in threads:
[Link]()
threads = []
if failed_dorks:
start_threads(failed_dorks, num_threads, output_file,
proxies, mode, progress_bar, retries_counter)
def main():
dork_file = '[Link]'
output_file = '[Link]'
processed_dorks_file = 'processed_dorks.txt'
try:
with open(processed_dorks_file, 'r') as f:
processed_dorks =
set([Link]().splitlines())
except FileNotFoundError:
processed_dorks = set()
dorks = [dork for dork in load_dorks(dork_file) if dork not in
processed_dorks]
if not dorks:
print('No dorks to process. Exiting.')
return
proxies = get_all_proxies()
if not proxies:
print('No proxies found. Exiting.')
return
print([Link] + 'Reversed by [Link] reversers iseeu
&& Example')
print('')
print([Link] + 'Note: Please make sure that the Parser and the
[Link] file are the same!')
print([Link] + 'The program is in Beta Version. There may be
errors!')
mode = input('Select a mode (1 - Full URL Mode, 2 - Only Parameter
Mode): ')
mode = 'Full URL Mode' if mode == '1' else 'Only Parameter Mode'
num_threads = int(input('How many threads to use? '))
retries_counter = [0]
progress_bar = tqdm(total=len(dorks), desc=GREY + 'Processing',
dynamic_ncols=True)
[Link](target=start_threads, args=(dorks, num_threads,
output_file, proxies, mode, progress_bar, retries_counter)).start()
while progress_bar.n < progress_bar.total or
threading.active_count() > 1:
progress_bar.set_postfix_str(YELLOW + f'Retries:
{retries_counter[0]}')
[Link](0.1)
progress_bar.close()
print(f'Completed! Total retries: {retries_counter[0]}')
if __name__ == '__main__':