mirror of
https://github.com/laramies/theHarvester.git
synced 2025-02-25 15:03:01 +08:00
81 lines
2.6 KiB
Python
81 lines
2.6 KiB
Python
from discovery.constants import *
|
|
from lib.core import *
|
|
from parsers import myparser
|
|
import re
|
|
import requests
|
|
import time
|
|
|
|
|
|
class search_exalead:
|
|
|
|
def __init__(self, word, limit, start):
|
|
self.word = word
|
|
self.files = 'pdf'
|
|
self.results = ""
|
|
self.totalresults = ""
|
|
self.server = 'www.exalead.com'
|
|
self.hostname = 'www.exalead.com'
|
|
self.limit = limit
|
|
self.counter = start
|
|
|
|
def do_search(self):
|
|
url = 'http:// ' + self.server + '/search/web/results/?q=%40' + self.word \
|
|
+ '&elements_per_page=50&start_index=' + str(self.counter)
|
|
headers = {
|
|
'Host': self.hostname,
|
|
'Referer': ('http://' + self.hostname + '/search/web/results/?q=%40' + self.word),
|
|
'User-agent': Core.get_user_agent()
|
|
}
|
|
h = requests.get(url=url, headers=headers)
|
|
self.results = h.text
|
|
self.totalresults += self.results
|
|
|
|
def do_search_files(self, files):
|
|
url = 'http:// ' + self.server + '/search/web/results/?q=%40' + self.word \
|
|
+ 'filetype:' + self.files + '&elements_per_page=50&start_index=' + str(self.counter)
|
|
headers = {
|
|
'Host': self.hostname,
|
|
'Referer': ('http://' + self.hostname + '/search/web/results/?q=%40' + self.word),
|
|
'User-agent': Core.get_user_agent()
|
|
}
|
|
h = requests.get(url=url, headers=headers)
|
|
self.results = h.text
|
|
self.totalresults += self.results
|
|
|
|
def check_next(self):
|
|
renext = re.compile('topNextUrl')
|
|
nextres = renext.findall(self.results)
|
|
if nextres != []:
|
|
nexty = '1'
|
|
print(str(self.counter))
|
|
else:
|
|
nexty = '0'
|
|
return nexty
|
|
|
|
def get_emails(self):
|
|
rawres = myparser.Parser(self.totalresults, self.word)
|
|
return rawres.emails()
|
|
|
|
def get_hostnames(self):
|
|
rawres = myparser.Parser(self.totalresults, self.word)
|
|
return rawres.hostnames()
|
|
|
|
def get_files(self):
|
|
rawres = myparser.Parser(self.totalresults, self.word)
|
|
return rawres.fileurls(self.files)
|
|
|
|
def process(self):
|
|
while self.counter <= self.limit:
|
|
self.do_search()
|
|
self.counter += 50
|
|
print(f'\tSearching {self.counter} results.')
|
|
|
|
def process_files(self, files):
|
|
while self.counter < self.limit:
|
|
self.do_search_files(files)
|
|
time.sleep(getDelay())
|
|
more = self.check_next()
|
|
if more == '1':
|
|
self.counter += 50
|
|
else:
|
|
break
|