12 Years of Service
16%
Code:
>
[center]#!/usr/bin/perl
use strict;
use warnings;
use HTTP::Request;
use LWP::UserAgent;
######
my $dork;
my $url;
my $i;
my $request;
my $useragent;
my $response;
my $start;
my $end;
my $result;
my $fl;
my $link;
my $req;
my $ua;
my $result2;
my $res;
my $save;
my $pages;
my $page;
my $choice;
######
my @z;
print q{
_ ____ _
| | _ \ | |
__| | |_) | ___ | |_
/ _` | _ < / _ \| __|
| (_| | |_) | (_) | |_
\__,_|____/ \___/ \__|
######
## / SQLi Crawler / ##
## Private Edition ##
## ~Coded by dbx~ ##
######
};
MainMenu:
print "------\n";
print "Enter [1] To Begin SQLi.\n";
print "Enter [2] To Exit.\n";
print "------\n\n";
print "Your Choice: ";
chomp ($choice = );
print "\n";
if ($choice eq 1) {&sql_scan}
if ($choice eq 5) {die;}
sub sql_scan
{
print "[+] Enter Bing! dork: ";
chomp ($dork = );
print "\n";
print "[+] How Many Pages To Leech?: ";
chomp ($pages = );
print "\n";
$page = $pages.'1';
print "[~] Crawling...\n\n";
for ($i = 0; $i <= $page; $i=$i+11)
{
$url = "http://www.bing.com/search?q=$dork&go=&qs=n&sk=&sc=8-13&first=$i";
$request = HTTP::Request->new(GET => $url);
$useragent = LWP::UserAgent->new();
$response = $useragent->request($request);
$result = $response->content;
$start = '{
print "Saving File...\n\n";
open(vuln_file, ">>Vulns.txt");
foreach (@z)
{
print vuln_file "$_ \n";
}
close(vuln_file);
print "File Saved!\n\n";
}
goto MainMenu;
}[/center]