mirror of
https://github.com/laramies/theHarvester.git
synced 2024-09-20 15:26:31 +08:00
Merge pull request #25 from duckspeaker/remove-123people
remove now defunct 123people
This commit is contained in:
commit
a5e76ea790
|
@ -8,7 +8,6 @@
|
|||
"googlesets",
|
||||
"dnssearch",
|
||||
"shodansearch",
|
||||
"people123",
|
||||
"jigsaw",
|
||||
"twittersearch",
|
||||
"dogpilesearch",
|
||||
|
|
|
@ -1,46 +0,0 @@
|
|||
import string
|
||||
import httplib
|
||||
import sys
|
||||
import myparser
|
||||
import re
|
||||
import time
|
||||
|
||||
|
||||
class search_123people:
|
||||
|
||||
def __init__(self, word, limit):
|
||||
self.word = word.replace(' ', '%20')
|
||||
self.results = ""
|
||||
self.totalresults = ""
|
||||
self.server = "www.google.com"
|
||||
self.hostname = "www.google.com"
|
||||
self.userAgent = "(Mozilla/5.0 (Windows; U; Windows NT 6.0;en-US; rv:1.9.2) Gecko/20100115 Firefox/3.6"
|
||||
self.quantity = "100"
|
||||
self.limit = int(limit)
|
||||
self.counter = 0
|
||||
|
||||
def do_search(self):
|
||||
h = httplib.HTTP(self.server)
|
||||
h.putrequest('GET', "/search?num=100&start=" + str(self.counter)
|
||||
+ "&hl=en&meta=&q=site%3A123people.com%20" + self.word)
|
||||
h.putheader('User-agent', self.userAgent)
|
||||
h.endheaders()
|
||||
returncode, returnmsg, headers = h.getreply()
|
||||
self.results = h.getfile().read()
|
||||
self.totalresults += self.results
|
||||
|
||||
def get_people(self):
|
||||
rawres = myparser.parser(self.totalresults, self.word)
|
||||
return rawres.people_123people()
|
||||
|
||||
def get_emails(self):
|
||||
rawres = myparser.parser(self.totalresults, self.word)
|
||||
return rawres.emails()
|
||||
|
||||
def process(self):
|
||||
while self.counter <= self.limit and self.counter <= 1000:
|
||||
self.do_search()
|
||||
#more = self.check_next()
|
||||
time.sleep(1)
|
||||
print "\tSearching " + str(self.counter) + " results..."
|
||||
self.counter += 100
|
14
myparser.py
14
myparser.py
|
@ -116,20 +116,6 @@ def profiles(self):
|
|||
resul.append(y)
|
||||
return resul
|
||||
|
||||
def people_123people(self):
|
||||
#reg_people = re.compile('www\.123people\.com/s/[a-zA-Z0-9.-_]*\+[a-zA-Z0-9.-_]*\+?[a-zA-Z0-9.-_]*\"')
|
||||
reg_people = re.compile('www.123people.com/e/[a-zA-Z0-9.-_+]*</cite>')
|
||||
self.temp = reg_people.findall(self.results)
|
||||
self.temp2 = []
|
||||
for x in self.temp:
|
||||
y = x.replace("www.123people.com/e/", "")
|
||||
y = y.replace('"', '')
|
||||
y = y.replace('+', ' ')
|
||||
y = y.replace('</cite>', '')
|
||||
|
||||
self.temp2.append(y)
|
||||
return self.temp2
|
||||
|
||||
def people_jigsaw(self):
|
||||
res = []
|
||||
#reg_people = re.compile("'tblrow' title='[a-zA-Z0-9.-]*'><span class='nowrap'/>")
|
||||
|
|
|
@ -41,9 +41,8 @@ def usage():
|
|||
|
||||
print "Usage: theharvester options \n"
|
||||
print " -d: Domain to search or company name"
|
||||
print """ -b: data source: google, googleCSE, bing, bingapi, pgp
|
||||
linkedin, google-profiles, people123, jigsaw,
|
||||
twitter, googleplus, all\n"""
|
||||
print """ -b: data source: google, googleCSE, bing, bingapi, pgp, linkedin,
|
||||
google-profiles, jigsaw, twitter, googleplus, all\n"""
|
||||
print " -s: Start in result number X (default: 0)"
|
||||
print " -v: Verify host name via dns resolution and search for virtual hosts"
|
||||
print " -f: Save the results into an HTML and XML file"
|
||||
|
@ -104,10 +103,10 @@ def start(argv):
|
|||
dnstld = True
|
||||
elif opt == '-b':
|
||||
engine = arg
|
||||
if engine not in ("google","googleCSE" , "linkedin", "pgp", "all", "google-profiles", "bing", "bingapi",
|
||||
"yandex", "people123", "jigsaw", "dogpilesearch", "twitter", "googleplus", "yahoo", "baidu"):
|
||||
if engine not in ("google","googleCSE" , "linkedin", "pgp", "all", "google-profiles", "bing", "bingapi",
|
||||
"yandex", "jigsaw", "dogpilesearch", "twitter", "googleplus", "yahoo", "baidu"):
|
||||
usage()
|
||||
print "Invalid search engine, try with: bing, google, linkedin, pgp, jigsaw, bingapi, people123, google-profiles, dogpilesearch, twitter, googleplus, yahoo, baidu"
|
||||
print "Invalid search engine, try with: bing, google, linkedin, pgp, jigsaw, bingapi, google-profiles, dogpilesearch, twitter, googleplus, yahoo, baidu"
|
||||
sys.exit()
|
||||
else:
|
||||
pass
|
||||
|
@ -159,17 +158,6 @@ def start(argv):
|
|||
all_emails = search.get_emails()
|
||||
all_hosts = search.get_hostnames()
|
||||
|
||||
elif engine == "people123":
|
||||
print "[-] Searching in 123People.."
|
||||
search = people123.search_123people(word, limit)
|
||||
search.process()
|
||||
people = search.get_people()
|
||||
all_emails = search.get_emails()
|
||||
print "Users from 123People:"
|
||||
print "====================="
|
||||
for user in people:
|
||||
print user
|
||||
|
||||
elif engine == "jigsaw":
|
||||
print "[-] Searching in Jigsaw.."
|
||||
search = jigsaw.search_jigsaw(word, limit)
|
||||
|
@ -212,7 +200,7 @@ def start(argv):
|
|||
for user in people:
|
||||
print user
|
||||
sys.exit()
|
||||
|
||||
|
||||
elif engine == "twitter":
|
||||
print "[-] Searching in Twitter .."
|
||||
search = twittersearch.search_twitter(word, limit)
|
||||
|
@ -223,7 +211,7 @@ def start(argv):
|
|||
for user in people:
|
||||
print user
|
||||
sys.exit()
|
||||
|
||||
|
||||
elif engine == "linkedin":
|
||||
print "[-] Searching in Linkedin.."
|
||||
search = linkedinsearch.search_linkedin(word, limit)
|
||||
|
|
Loading…
Reference in a new issue