Advertisement
Guest User

Trojan Kil3r Amazigh

a guest
Mar 6th, 2019
1,555
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
Python 17.31 KB | None | 0 0
  1. #!/usr/bin/env python
  2. # -*- coding: utf-8 -*-
  3.  
  4.  
  5. #Ne touche pas le script -_-
  6. #Don't Edit Logo -_-
  7.  
  8.  
  9. import requests, httplib, urllib
  10. import socket
  11. from platform import system
  12. import os
  13. import sys, time
  14. import re
  15. import threading
  16. from multiprocessing.dummy import Pool
  17. from colorama import Fore                              
  18. from colorama import Style                             
  19. from colorama import init                                              
  20. init(autoreset=True)
  21. fr  =   Fore.RED
  22. fh  =   Fore.RED
  23. fc  =   Fore.CYAN
  24. fo  =   Fore.MAGENTA
  25. fw  =   Fore.WHITE
  26. fy  =   Fore.YELLOW
  27. fbl =   Fore.BLUE
  28. fg  =   Fore.GREEN                                         
  29. sd  =   Style.DIM
  30. fb  =   Fore.RESET
  31. sn  =   Style.NORMAL                                       
  32. sb  =   Style.BRIGHT
  33.  
  34. user = {"User-Agent": "Mozilla/5.0 (Windows NT 6.1; rv:57.0) Gecko/20100101 Firefox/57.0"}
  35.  
  36. url = "http://www.zone-h.org/archive/notifier="
  37. urll = "http://zone-h.org/archive/published=0"
  38. url2 = "http://www.defacers.org/onhold!"
  39. url4 = "http://www.defacers.org/gold!"
  40. my_cook = {
  41.     "ZHE" : "4e272d65712d4668b216595185f47c7d",
  42.     "PHPSESSID" : "0pg7okq5lbhj7o699c5amut5a1"
  43.     }
  44.  
  45.  
  46. def zonehh():
  47.     print("""
  48.         |---| Grabb Sites From Zone-h |--|
  49.         \033[91m[1] \033[95mGrabb Sites By Notifier
  50.         \033[91m[2] \033[95mGrabb Sites By Onhold
  51.         """)
  52.     sec = int(raw_input("Choose Section: "))
  53.     if sec == 1:
  54.         notf = raw_input("\033[95mEntre notifier: \033[92m")
  55.  
  56.         for i in range(1, 51):
  57.             dz = requests.get(url + notf +"/page=" + str(i), cookies=my_cook)
  58.             dzz = dz.content
  59.             print(url + notf +"/page=" + str(i))
  60.             if '<html><body>-<script type="text/javascript"' in dzz:
  61.                 print("Change Cookies Please")
  62.                 sys.exit()
  63.             elif '<input type="text" name="captcha" value=""><input type="submit">' in dzz:
  64.                 print("Entre Captcha In Zone-h From Ur Browser :/")
  65.                 sys.exit() 
  66.             else:
  67.                 Hunt_urls = re.findall('<td>(.*)\n                          </td>', dzz)
  68.                 if '/mirror/id/' in dzz:
  69.                     for xx in Hunt_urls:
  70.                         qqq = xx.replace('...','')
  71.                         print '    ['  + '*' + '] ' + qqq.split('/')[0]
  72.                         with open( notf + '.txt', 'a') as rr:
  73.                             rr.write("http://" + qqq.split('/')[0] + '\n')
  74.                 else:
  75.                     print("Grabb Sites Completed !!")
  76.                     sys.exit()
  77.                    
  78.     elif sec == 2:
  79.         print(":* __Grabb Sites By Onhold__ ^_^")
  80.         for qwd in range(1, 51):
  81.             rb = requests.get(urll + "/page=" + str(qwd) , cookies=my_cook)
  82.             dzq = rb.content
  83.  
  84.             if '<html><body>-<script type="text/javascript"' in dzq:
  85.                 print("Change Cookies Plz")
  86.                 sys.exit()
  87.                
  88.             elif "captcha" in dzq:
  89.                 print("Entre captcha In Your Browser Of Site [zone-h.org]")
  90.             else:
  91.                 Hunt_urlss = re.findall('<td>(.*)\n                         </td>', dzq)
  92.                 for xxx in Hunt_urlss:
  93.                     qqqq = xxx.replace('...','')
  94.                     print '    ['  + '*' + '] ' + qqqq.split('/')[0]
  95.                     with open('onhold_zone.txt', 'a') as rrr:
  96.                         rrr.write("http://" + qqqq.split('/')[0] + '\n')
  97.     else:
  98.         print("Fuck You Men")
  99.  
  100. def defacers():
  101.     print("""
  102.         |---| Grabb Sites From Defacers.org |--|
  103.             \033[91m[1] \033[95mGrabb Sites By Onhold
  104.             \033[91m[2] \033[95mGrabb Sites By Archive
  105.         """)
  106.     sec = int(raw_input("Choose Section: "))
  107.     if sec == 1:
  108.         for i in range(1, 380):
  109.             print("Page: "), str(i) + "\033[91m Waiting Grabbed Sites .....  <3"
  110.             rb = requests.get(url2 + str(i),headers=user)
  111.             okbb = rb.content
  112.             domains = re.findall(r'title=".*" tar.?', okbb)
  113.             for iii in domains:
  114.                 iii = iii.replace('" target="_blank" reel="nofollow">', "")
  115.                 iii = iii.replace('title="', "")
  116.                 iii = iii.replace('" targ', "")
  117.                 print("\033[95mhttp://" + iii + "/")
  118.                 with open("Onhold_defacer.txt", "a") as by:
  119.                     by.writelines("http://" + iii + "/")
  120.                     by.writelines("\n")
  121.             print ("\t\t[+] Page Saved_"),str(i) +(" done [+]\n")
  122.     elif sec == 2:
  123.         for i in range(1, 25):
  124.             print("Page: "), str(i) + " \033[91mWaiting Grabbed Sites Governement .....  <3"
  125.             rb = requests.get(url4 + str(i),headers=user)
  126.             okbb = rb.content
  127.             domains = re.findall(r'title=".*" tar.?', okbb)
  128.             for iii in domains:
  129.                 iii = iii.replace('" target="_blank" reel="nofollow">', "")
  130.                 iii = iii.replace('title="', "")
  131.                 iii = iii.replace('" targ', "")
  132.                 print("\033[95mhttp://" + iii + "/")
  133.                 with open("govSites_defacer.txt", "a") as by:
  134.                     by.writelines("http://" + iii + "/")
  135.                     by.writelines("\n")
  136.             print ("\t\t[+] Page Saved_"),str(i) +(" done [+]\n")
  137.     else:
  138.         print("Fuck You Men 2")
  139.  
  140.  
  141. def mirroirh():
  142.     print("""
  143.            |---| Grabb Sites From Mirror-h.org |--|
  144.             \033[91m[1] \033[95mGrabb Sites By Onhold
  145.             \033[91m[2] \033[95mGrabb Sites By Auto_Notifier
  146.         """)
  147.     sec = int(raw_input("Choose Section: "))
  148.     if sec == 1:
  149.         url = "https://mirror-h.org/archive/page/"
  150.         try:
  151.             for pp in range(1, 40254):
  152.                 dz = requests.get(url + str(pp))
  153.                 dzz = dz.content
  154.                 qwd = re.findall(r'/zone/(.*)</a></td>', dzz)
  155.                 print(" \033[91m[*] Please Wait To Grabb Sites ...... Page: "), pp
  156.                 for ii in qwd:
  157.                     ii = ii.replace('<i class="icon-search"></i>', "")
  158.                     ii = ii.replace(ii[:10], "")
  159.                     ii = ii.replace("\r\n\r\n", "\r\n")
  160.                     ii = ii.strip()
  161.                     #iio = ii.replace('<i class="icon-search"></i>', "hhhhhhhhhhhhh")
  162.                     print("\033[95m" + ii)
  163.                     with open( 'onzeb_mirror.txt', 'a') as rr:
  164.                         rr.write(ii + '\n')
  165.         except:
  166.             pass
  167.     elif sec == 2:
  168.         url = "https://mirror-h.org/search/hacker/"
  169.         try:
  170.             for ha in range(1, 2000):
  171.                 print("\033[91mWait To Grabb From Hacker: "), ha
  172.                 dz = requests.get(url + str(ha) + "/pages/1")
  173.                 dzz = dz.content
  174.                 qwd = re.findall(r'/pages/\d" title="Last"', dzz)
  175.                 for i in qwd:
  176.                     i = i.rstrip()
  177.                     sss = i.replace("/pages/","")
  178.                     ss = sss.replace('" title="Last"',"")
  179.                     ssf = int(ss) + 1
  180.                     for ii in range(1, ssf):
  181.                         print(" \033[91m[*] Please Wait To Grabb Sites ...... Page: "), ii
  182.                         dd = requests.get(url + str(ha) + "/pages/"+ str(ii))
  183.                         op = dd.content
  184.                         qwdd = re.findall(r'/zone/(.*)</a></td>', op)
  185.                         for idi in qwdd:
  186.                             idi = idi.replace('<i class="icon-search"></i>', "")
  187.                             idi = idi.replace(idi[:10], "")
  188.                             idi = idi.replace("\r\n\r\n", "\r\n")
  189.                             idi = idi.strip()
  190.                             #iio = ii.replace('<i class="icon-search"></i>', "hhhhhhhhhhhhh")
  191.                             print("\033[95m" + idi)
  192.                             with open( 'top_mirror.txt', 'a') as rr:
  193.                                 rr.write(idi + '\n')
  194.         except:
  195.             pass
  196.  
  197.  
  198. def overflowzone():
  199.     print("""
  200.         |---| Grabb Sites From overflowzone.com |--|
  201.             \033[91m[1] \033[95mGrabb Sites By Onhold
  202.             \033[91m[2] \033[95mGrabb Sites By AutoNotifier
  203.         """)
  204.     sec = int(raw_input("Choose Section: "))
  205.     if sec == 1:
  206.         url = "http://attacker.work/onhold/onhold/page/"
  207.         dz = requests.get(url + "1")
  208.         dzz = dz.content
  209.         tn = re.findall(r'<a href="/onhold/page/(.*)" title="Last">', dzz)
  210.         for ii in tn:
  211.             qwd = ii.split('/')[-1]
  212.             for ok in range(1, int(qwd)):
  213.                 okk = requests.get(url + str(ok))
  214.                 print("`\t\t\t" + url + str(ok))
  215.                 fel = okk.content
  216.                 okkk = re.findall(r'">http://(.*)</a></td>', fel)
  217.                 for iii in okkk:
  218.                     iii = iii.rstrip()
  219.                     print("\033[95mhttp://" + iii.split('/')[0])
  220.                     with open( 'onhold_attackerwork.txt', 'a') as rr:
  221.                         rr.write("http://" + iii.split('/')[0] + '\n')
  222.     elif sec == 2:
  223.         url = "http://attacker.work/archive/page/"
  224.         dz = requests.get(url + "1")
  225.         dzz = dz.content
  226.         tn = re.findall(r'<a href="/archive/page/(.*)" title="Last">', dzz)
  227.         for ii in tn:
  228.             qwd = ii.split('/')[-1]
  229.             for ok in range(1, int(qwd)):
  230.                 okk = requests.get(url + str(ok))
  231.                 print("`\t\t\t" + url + str(ok))
  232.                 fel = okk.content
  233.                 okkk = re.findall(r'">http://(.*)</a></td>', fel)
  234.                 for iii in okkk:
  235.                     iii = iii.rstrip()
  236.  
  237.                     print("\033[95mhttp://" + iii.split('/')[0])
  238.                     with open( 'archive_attackerwork.txt', 'a') as rr:
  239.                         rr.write("http://" + iii.split('/')[0] + '\n')
  240.     else:
  241.         print("hhhhhhhh tnkt")
  242. def bYPAS():
  243.     exploit = ["/member/","/admin/login.php","/admin/panel.php","/admin/","/login.php","/admin.html","/admin.php","/admin-login.php"]
  244.     try:
  245.         q = raw_input('\033[96m Entre Liste Site: \033[90m ')
  246.         q = open(q, 'r')
  247.     except:
  248.         print("\033[91mEntre List Sites -_- #Noob ")
  249.         sys.exit()
  250.     for lst in q:
  251.         lst = lst.rstrip()
  252.         print("\033[94m     Wait Scaning ....... \033[94m"), lst
  253.         for exploits in exploit:
  254.             exploits.rstrip()
  255.             try:
  256.                 if lst[:7] == "http://":
  257.                     lst = lst.replace("http://","")
  258.                 if lst[:8] == "https://":
  259.                     lst = lst.replace("https://", "")
  260.                 if lst[-1] == "/":
  261.                     lst = lst.replace("/","")
  262.                 socket.setdefaulttimeout(5)
  263.                 conn = httplib.HTTPConnection(lst)
  264.                 conn.request("POST", exploits)
  265.                 conn = conn.getresponse()
  266.                 htmlconn = conn.read()
  267.                 if conn.status == 200 and ('type="password"') in htmlconn:
  268.                     print("\033[92m [+] Admin Panel [+] ======\033[96m=======> \033[96m ") , lst + exploits
  269.                     with open("admin_panels.txt", "a") as by:
  270.                         by.writelines(lst + exploits + "\n")
  271.                 else:
  272.                     print("\033[91m [-] Not Found : [-]"),lst + exploits
  273.             except:
  274.                 pass
  275.  
  276. def add_http():
  277.     dz = raw_input("Entre List Site: ")
  278.     dz = open(dz, "r")
  279.     for i in dz:
  280.         i = i.rstrip()
  281.         print("http://"+i)
  282.         with open( 'aziz.txt', 'a') as rr:
  283.             rr.write("http://" + i + '\n')
  284.     print("Text Saved !!")
  285.  
  286. def binger():
  287.     qwd = {"User-Agent": "Mozilla/5.0 (Windows NT 6.1; rv:57.0) Gecko/20100101 Firefox/57.0"}
  288.  
  289.     print("""
  290.         \033[91m[1] \033[95mGrabb Sites By Ip List
  291.         \033[91m[1] \033[95mGrabb Sites Fox_Contact And Bypass By Ip List
  292.         """)
  293.     o = int(raw_input("Choose Section: "))
  294.     if o == 1:
  295.         gr = raw_input('Give me List Ip: ')
  296.         gr = open(gr,'r')
  297.         for done in gr:
  298.             remo = []
  299.             page = 1
  300.             while page < 251:
  301.                 bing = "http://www.bing.com/search?q=ip%3A"+done+"+&count=50&first="+str(page)
  302.                 opene = requests.get(bing,verify=False,headers=qwd)
  303.                 read = opene.content
  304.                 findwebs = re.findall('<h2><a href="(.*?)"', read)
  305.                 for i in findwebs:
  306.                     o = i.split('/')
  307.                     if (o[0]+'//'+o[2]) in remo:
  308.                         pass
  309.                     else:
  310.                         remo.append(o[0]+'//'+o[2])
  311.                         print '{}[XxX] '.format(fg,sb),(o[0]+'//'+o[2])
  312.                         with open('Grabbed.txt','a') as s:
  313.                             s.writelines((o[0]+'//'+o[2])+'\n')
  314.                 page = page+5
  315.     elif o == 2:
  316.         qwd = {"User-Agent": "Mozilla/5.0 (Windows NT 6.1; rv:57.0) Gecko/20100101 Firefox/57.0"}
  317.         gr = raw_input('Give me List Ip: ')
  318.         gr = open(gr,'r')
  319.         for done in gr:
  320.             remo = []
  321.             page = 1
  322.             print("Wait Grabb Sites From iP: "), done
  323.             while page < 251:
  324.                 bing = "http://www.bing.com/search?q=ip%3A"+done + " powered by fox_contact"+"+&count=50&first="+str(page)
  325.                 opene = requests.get(bing,verify=False,headers=qwd)
  326.                 read = opene.content
  327.                 findwebs = re.findall('<h2><a href="(.*?)"', read)
  328.                 for i in findwebs:
  329.                     o = i.split('/')
  330.                     if (o[0]+'//'+o[2]) in remo:
  331.                         pass
  332.                     else:
  333.                         remo.append(o[0]+'//'+o[2])
  334.                         print '[XxX] ' + (o[0]+'//'+o[2])
  335.                         with open('foxcontact.txt','a') as s:
  336.                             s.writelines((o[0]+'//'+o[2])+'\n')
  337.                 page = page+5
  338.  
  339.                 bing = "http://www.bing.com/search?q=ip%3A"+done + " admin/login.php"+"+&count=50&first="+str(page)
  340.                 opene = requests.get(bing,verify=False,headers=qwd)
  341.                 read = opene.content
  342.                 findwebs = re.findall('<h2><a href="(.*?)"', read)
  343.                 for i in findwebs:
  344.                     o = i.split('/')
  345.                     if (o[0]+'//'+o[2]) in remo:
  346.                         pass
  347.                     else:
  348.                         remo.append(o[0]+'//'+o[2])
  349.                         dd = requests.get(o[0]+'//'+o[2] + "/admin/login.php")
  350.                         ddd = dd.content
  351.                         if 'type="password"' in ddd:
  352.                             print("\033[92mAdmin_Panel Site: >>>>>>\033[91m"),o[0]+'//'+o[2] + "/admin/login.php"
  353.                             with open('admin panel.txt','a') as s:
  354.                                 s.writelines((o[0]+'//'+o[2])+'\n')
  355.                 page = page+5
  356.     else:
  357.         print("dir numero azbi nooooooob")
  358.  
  359.  
  360. def cms_detected():
  361.     lst = raw_input("Entre List Site: ")
  362.     lst = open(lst, 'r')
  363.     for i in lst:
  364.         i = i.rstrip()
  365.         print("\033[91m[+] \033[95mPlease Waiting To Scaning ... "), "\033[94m" + i + " \033[91m[+]"
  366.         try:
  367.             dz = requests.get(i)
  368.             ok = dz.content
  369.             #-------------WP---------------------------
  370.            
  371.             if "wp-content" in ok:
  372.                 print("\033[92mWp Site : >>>>>>>>>>>>>>\033[91m"), i + "/wp-login.php"
  373.                 with open("wp sites.txt", "a") as wpp:
  374.                     wpp.writelines(i + "/wp-login.php"+ "\n")
  375.                     #-------JM--------------------------
  376.             elif "com_content" in ok:
  377.                 print("\033[92mJm Site: >>>>>>>>>>>>>>\033[91m"), i + "/administrator/"
  378.                 with open("joomla sites.txt", "a") as jmm:
  379.                     jmm.writelines(i + "/administrator/"+ "\n")
  380.                     #---------OPENCARTE-----------------------
  381.             elif "index.php?route" in ok:
  382.                 print("\033[92mOpenCart Site: >>>>>>>>>>>>>>\033[91m"), i + "/admin/"
  383.                 with open("OpenCart sites.txt", "a") as opncrt:
  384.                     opncrt.writelines(i + "/admin/"+ "\n")
  385.                     #---------------------------------
  386.             elif "/node/" in ok:
  387.                 print("\033[92mDrupal Site: >>>>>>>>>>>>>>\033[91m"), i + "/user/login"
  388.                 with open("Drupal sites.txt", "a") as drbl:
  389.                     drbl.writelines(i + "/user/login"+ "\n")
  390.             else:
  391.                 bypass = ["/admin/login.php","/admin/","/login.php","/admin.html","/admin.php","/member/"]
  392.                 for byp in bypass:
  393.                     byp = byp.rstrip()
  394.                     dd = requests.get(i + byp)
  395.                     ddd = dd.content
  396.                     if 'type="password"' in ddd:
  397.                         print("\033[92mAdmin_Panel Site: >>>>>>\033[91m"),i +  byp
  398.                         with open("Admin Sites.txt", "a") as by:
  399.                             by.writelines(i + byp + "\n")
  400.                     else:
  401.                         pass
  402.  
  403.                 print("\033[91m[-] Not Found Cms: [-]"), "\033[91m" + i
  404.         except:
  405.             pass
  406.  
  407.  
  408.  
  409. def spotii():
  410.     #url =  "https://www.spotify.com"
  411.  
  412.     rl = "http://www.spotify.com/us/xhr/json/isEmailAvailable.php?signup_form[email]="
  413.  
  414.     try:
  415.         ok = raw_input("{}root@kil3r~# Entre List Email: ".format(fy,sn))
  416.         okd = open(ok, 'r')
  417.     except:
  418.         print("{}zebi  entre list  email -_- nooob").format(fh,sn)
  419.     for i in okd:
  420.         i = i.rstrip()
  421.         qwd = url + i + "&email=" + i
  422.         dz = requests.get(qwd, headers=user)
  423.         dzz = dz.content
  424.         if 'false' in dzz:
  425.             print("{}   [LIVE]     ".format(fg,sn)), "{}".format(fg,sn) + i
  426.             with open("spotify checked.txt", "a") as zebi:
  427.                 zebi.writelines(i + '\n')
  428.         else:
  429.             print("{}   [DEAD]     ").format(fh,sn), "{}".format(fh,sn) + i
  430.  
  431.  
  432.  
  433. def clearscrn():
  434.     if system() == 'Linux':
  435.         os.system('clear')
  436.     if system() == 'Windows':
  437.         os.system('cls')
  438.         os.system('color a')
  439. clearscrn()
  440.  
  441. def slowprint(s):
  442.     for c in s + '\n':
  443.         sys.stdout.write(c)
  444.         sys.stdout.flush()
  445.         time.sleep(4. / 100)
  446.  
  447. def helper4():
  448.     clearscrn()
  449.     banner = """\033[94m
  450.                             _       _       _  ___ _ _____      
  451.   / \  _ __ ___   __ _ ___(_) __ _| |__   | |/ (_) |___ / _ __
  452.  / _ \ | '_ ` _ \ / _` |_  / |/ _` | '_ \ | ' /| | | |_ \| '__|
  453. / ___ \| | | | | | (_| |/ /| | (_| | | | | | . \| | |___) | |  
  454. /_/   \_\_| |_| |_|\__,_/___|_|\__, |_| |_| |_|\_\_|_|____/|_|  
  455.                               |___/                            
  456.          
  457.     """
  458.     print("""\033[95m
  459.                             _       _       _  ___ _ _____      
  460.   / \  _ __ ___   __ _ ___(_) __ _| |__   | |/ (_) |___ / _ __
  461.  / _ \ | '_ ` _ \ / _` |_  / |/ _` | '_ \ | ' /| | | |_ \| '__|
  462. / ___ \| | | | | | (_| |/ /| | (_| | | | | | . \| | |___) | |  
  463. /_/   \_\_| |_| |_|\__,_/___|_|\__, |_| |_| |_|\_\_|_|____/|_|  
  464.                               |___/                            
  465.          
  466.                                  Script Name : Mnanok Tester ^_^
  467.                     Greetz To : \033[93mKabyles \033[92mHackers  \033[91m|D\033[92mz| \033[91mAlgeria \033[92mHackers
  468.                                     """)
  469.     slowprint("\n\t\t\t\t\tPowered By : Trojan Kil3r Amazigh / Aziz" + "\n\t\t\t\t\t\t            Facebook : fb.com/amazigh.kil3r")
  470.     print("")
  471.     print("""
  472.         \033[91m[1] \033[95mGrabb Sites  \033[92m From Zone-h.org   |  \033[91m[3] \033[95mGrabb Sites  \033[92m From mirror-h.org     |
  473.         \033[91m[2] \033[95mGrabb Sites  \033[92m From Defacers.org |  \033[91m[4] \033[95mGrabb Sites  \033[92m From overflowzone.com |
  474.        
  475.         \033[91m[5] \033[95mGet Sites bypass With List [Bypass Finder]
  476.         \033[91m[6] \033[95mMass Add (http://) To List ^_^
  477.         \033[91m[7] \033[95mGrabber Sites From Bing :D
  478.         \033[91m[8] \033[95mCms Filtre
  479.         \033[91m[9] \033[95mEmail Valid Spotify
  480.  
  481.  
  482.             #######################################################
  483.             #     Love 4 Palestine |\033[91m| Live 4 Palestine        #
  484.             #######################################################
  485.             """)       
  486.     try:
  487.         qq = int(raw_input("\033[91m[-] \033[90mroot@kil3r~# \033[92mChoose Section !!\033[95m : \033[90m"))
  488.         if qq == 1:
  489.             clearscrn()
  490.             print(banner)
  491.             zonehh()
  492.         if qq == 2:
  493.             clearscrn()
  494.             print(banner)
  495.             defacers()
  496.         if qq == 3:
  497.             clearscrn()
  498.             print(banner)
  499.             mirroirh()
  500.         if qq == 4:
  501.             clearscrn()
  502.             print(banner)
  503.             overflowzone()
  504.         if qq == 5:
  505.             clearscrn()
  506.             print(banner)
  507.             bYPAS()
  508.         if qq == 6:
  509.             clearscrn()
  510.             print(banner)
  511.             add_http()
  512.         if qq == 7:
  513.             clearscrn()
  514.             print(banner)
  515.             binger()
  516.         if qq == 8:
  517.             clearscrn()
  518.             print(banner)
  519.             cms_detected()
  520.         if qq == 9:
  521.             clearscrn()
  522.             print(banner)
  523.             spotii()
  524.  
  525.     except:
  526.         pass
  527. helper4()
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement