mirror of
https://github.com/ail-project/ail-framework.git
synced 2024-11-10 08:38:28 +00:00
Fix the onion module, log the valid onions.
This commit is contained in:
parent
f4b89669fc
commit
0c6b09f379
3 changed files with 57 additions and 4 deletions
35
bin/DumpValidOnion.py
Executable file
35
bin/DumpValidOnion.py
Executable file
|
@ -0,0 +1,35 @@
|
||||||
|
#!/usr/bin/env python2
|
||||||
|
# -*-coding:UTF-8 -*
|
||||||
|
|
||||||
|
from pubsublogger import publisher
|
||||||
|
from Helper import Process
|
||||||
|
import datetime
|
||||||
|
import time
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
publisher.port = 6380
|
||||||
|
publisher.channel = "Script"
|
||||||
|
|
||||||
|
config_section = 'DumpValidOnion'
|
||||||
|
dump_file = 'dump.out'
|
||||||
|
|
||||||
|
p = Process(config_section)
|
||||||
|
|
||||||
|
# FUNCTIONS #
|
||||||
|
publisher.info("Script subscribed to channel ValidOnion")
|
||||||
|
|
||||||
|
while True:
|
||||||
|
message = p.get_from_set()
|
||||||
|
if message is not None:
|
||||||
|
f = open(dump_file, 'a')
|
||||||
|
while message is not None:
|
||||||
|
print message
|
||||||
|
date = datetime.datetime.now()
|
||||||
|
if message is not None:
|
||||||
|
f.write(date.isoformat() + ' ' + message + '\n')
|
||||||
|
else:
|
||||||
|
break
|
||||||
|
message = p.get_from_set()
|
||||||
|
f.close()
|
||||||
|
else:
|
||||||
|
time.sleep(20)
|
21
bin/Onion.py
21
bin/Onion.py
|
@ -29,12 +29,15 @@ import datetime
|
||||||
import os
|
import os
|
||||||
import base64
|
import base64
|
||||||
import subprocess
|
import subprocess
|
||||||
|
import redis
|
||||||
|
|
||||||
from Helper import Process
|
from Helper import Process
|
||||||
|
|
||||||
|
|
||||||
def fetch(p, urls, domains, path):
|
def fetch(p, r_cache, urls, domains, path):
|
||||||
for url, domain in zip(urls, domains):
|
for url, domain in zip(urls, domains):
|
||||||
|
if r_cache.exists(url):
|
||||||
|
continue
|
||||||
to_fetch = base64.standard_b64encode(url)
|
to_fetch = base64.standard_b64encode(url)
|
||||||
process = subprocess.Popen(["python", './tor_fetcher.py', to_fetch],
|
process = subprocess.Popen(["python", './tor_fetcher.py', to_fetch],
|
||||||
stdout=subprocess.PIPE)
|
stdout=subprocess.PIPE)
|
||||||
|
@ -42,6 +45,8 @@ def fetch(p, urls, domains, path):
|
||||||
time.sleep(1)
|
time.sleep(1)
|
||||||
|
|
||||||
if process.returncode == 0:
|
if process.returncode == 0:
|
||||||
|
r_cache.setbit(url, 0, 1)
|
||||||
|
r_cache.expire(url, 3600)
|
||||||
tempfile = process.stdout.read().strip()
|
tempfile = process.stdout.read().strip()
|
||||||
with open(tempfile, 'r') as f:
|
with open(tempfile, 'r') as f:
|
||||||
filename = path + domain
|
filename = path + domain
|
||||||
|
@ -54,7 +59,9 @@ def fetch(p, urls, domains, path):
|
||||||
os.makedirs(dirname)
|
os.makedirs(dirname)
|
||||||
with open(save_path, 'w') as ff:
|
with open(save_path, 'w') as ff:
|
||||||
ff.write(content)
|
ff.write(content)
|
||||||
p.populate_set_out(save_path)
|
p.populate_set_out(save_path, 'Global')
|
||||||
|
p.populate_set_out(url, 'ValidOnion')
|
||||||
|
yield url
|
||||||
os.unlink(tempfile)
|
os.unlink(tempfile)
|
||||||
else:
|
else:
|
||||||
print 'Failed at downloading', url
|
print 'Failed at downloading', url
|
||||||
|
@ -71,6 +78,10 @@ if __name__ == "__main__":
|
||||||
config_section = 'Onion'
|
config_section = 'Onion'
|
||||||
|
|
||||||
p = Process(config_section)
|
p = Process(config_section)
|
||||||
|
r_cache = redis.StrictRedis(
|
||||||
|
host=p.config.get("Redis_Cache", "host"),
|
||||||
|
port=p.config.getint("Redis_Cache", "port"),
|
||||||
|
db=p.config.getint("Redis_Cache", "db"))
|
||||||
|
|
||||||
# FUNCTIONS #
|
# FUNCTIONS #
|
||||||
publisher.info("Script subscribed to channel onion_categ")
|
publisher.info("Script subscribed to channel onion_categ")
|
||||||
|
@ -121,7 +132,11 @@ if __name__ == "__main__":
|
||||||
str(now.month).zfill(2),
|
str(now.month).zfill(2),
|
||||||
str(now.day).zfill(2),
|
str(now.day).zfill(2),
|
||||||
str(int(time.mktime(now.utctimetuple()))))
|
str(int(time.mktime(now.utctimetuple()))))
|
||||||
fetch(p, urls, domains_list, path)
|
to_print = 'Onion;{};{};{};'.format(PST.p_source,
|
||||||
|
PST.p_date,
|
||||||
|
PST.p_name)
|
||||||
|
for url in fetch(p, r_cache, urls, domains_list, path):
|
||||||
|
publisher.warning('{}Valid: {}'.format(to_print, url))
|
||||||
else:
|
else:
|
||||||
publisher.info('{}Onion related'.format(to_print))
|
publisher.info('{}Onion related'.format(to_print))
|
||||||
|
|
||||||
|
|
|
@ -34,7 +34,10 @@ subscribe = Redis_Mail
|
||||||
|
|
||||||
[Onion]
|
[Onion]
|
||||||
subscribe = Redis_Onion
|
subscribe = Redis_Onion
|
||||||
publish = Redis_Global
|
publish = Redis_Global,Redis_ValidOnion
|
||||||
|
|
||||||
|
[DumpValidOnion]
|
||||||
|
subscribe = Redis_ValidOnion
|
||||||
|
|
||||||
[Web]
|
[Web]
|
||||||
subscribe = Redis_Web
|
subscribe = Redis_Web
|
||||||
|
|
Loading…
Reference in a new issue