Merge branch 'ail-project:dev' into dev

This commit is contained in:
Steve Clement 2024-02-01 09:59:39 +01:00 committed by GitHub
commit 1ac2fd026d
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
35 changed files with 1686 additions and 205 deletions

View file

@ -607,7 +607,7 @@ function launch_all {
function menu_display { function menu_display {
options=("Redis" "Ardb" "Kvrocks" "Logs" "Scripts" "Flask" "Killall" "Update" "Update-config" "Update-thirdparty") options=("Redis" "Kvrocks" "Logs" "Scripts" "Flask" "Killall" "Update" "Update-config" "Update-thirdparty")
menu() { menu() {
echo "What do you want to Launch?:" echo "What do you want to Launch?:"
@ -635,9 +635,6 @@ function menu_display {
Redis) Redis)
launch_redis; launch_redis;
;; ;;
Ardb)
launch_ardb;
;;
Kvrocks) Kvrocks)
launch_kvrocks; launch_kvrocks;
;; ;;

View file

@ -34,16 +34,20 @@ class D4Client(AbstractModule):
self.d4_client = d4.create_d4_client() self.d4_client = d4.create_d4_client()
self.last_refresh = time.time() self.last_refresh = time.time()
self.last_config_check = time.time()
# Send module state to logs # Send module state to logs
self.logger.info(f'Module {self.module_name} initialized') self.logger.info(f'Module {self.module_name} initialized')
def compute(self, dns_record): def compute(self, dns_record):
# Refresh D4 Client # Refresh D4 Client
if self.last_config_check < int(time.time()) - 30:
print('refresh rrrr')
if self.last_refresh < d4.get_config_last_update_time(): if self.last_refresh < d4.get_config_last_update_time():
self.d4_client = d4.create_d4_client() self.d4_client = d4.create_d4_client()
self.last_refresh = time.time() self.last_refresh = time.time()
print('D4 Client: config updated') print('D4 Client: config updated')
self.last_config_check = time.time()
if self.d4_client: if self.d4_client:
# Send DNS Record to D4Server # Send DNS Record to D4Server

View file

@ -142,10 +142,17 @@ class Crawler(AbstractModule):
return capture return capture
elif status == crawlers.CaptureStatus.UNKNOWN: elif status == crawlers.CaptureStatus.UNKNOWN:
capture_start = capture.get_start_time(r_str=False) capture_start = capture.get_start_time(r_str=False)
if capture_start == 0:
task = capture.get_task()
task.delete()
capture.delete()
self.logger.warning(f'capture UNKNOWN ERROR STATE, {task.uuid} Removed from queue')
return None
if int(time.time()) - capture_start > 600: # TODO ADD in new crawler config if int(time.time()) - capture_start > 600: # TODO ADD in new crawler config
task = capture.get_task() task = capture.get_task()
task.reset() task.reset()
capture.delete() capture.delete()
self.logger.warning(f'capture UNKNOWN Timeout, {task.uuid} Send back in queue')
else: else:
capture.update(status) capture.update(status)
else: else:
@ -154,7 +161,7 @@ class Crawler(AbstractModule):
except ConnectionError: except ConnectionError:
print(capture.uuid) print(capture.uuid)
capture.update(self, -1) capture.update(-1)
self.refresh_lacus_status() self.refresh_lacus_status()
time.sleep(self.pending_seconds) time.sleep(self.pending_seconds)

View file

@ -319,10 +319,6 @@ class MISPExporterAutoDaily(MISPExporter):
def __init__(self, url='', key='', ssl=False): def __init__(self, url='', key='', ssl=False):
super().__init__(url=url, key=key, ssl=ssl) super().__init__(url=url, key=key, ssl=ssl)
# create event if don't exists
try:
self.event_id = self.get_daily_event_id()
except MISPConnectionError:
self.event_id = - 1 self.event_id = - 1
self.date = datetime.date.today() self.date = datetime.date.today()
@ -345,6 +341,7 @@ class MISPExporterAutoDaily(MISPExporter):
self.add_event_object(self.event_id, obj) self.add_event_object(self.event_id, obj)
except MISPConnectionError: except MISPConnectionError:
self.event_id = - 1
return -1 return -1

View file

@ -100,6 +100,7 @@ class FeederImporter(AbstractImporter):
else: else:
objs = set() objs = set()
if data_obj:
objs.add(data_obj) objs.add(data_obj)
for obj in objs: for obj in objs:
@ -107,7 +108,7 @@ class FeederImporter(AbstractImporter):
gzip64_content = feeder.get_gzip64_content() gzip64_content = feeder.get_gzip64_content()
return obj, f'{feeder_name} {gzip64_content}' return obj, f'{feeder_name} {gzip64_content}'
else: # Messages save on DB else: # Messages save on DB
if obj.exists(): if obj.exists() and obj.type != 'chat':
return obj, f'{feeder_name}' return obj, f'{feeder_name}'
@ -136,4 +137,5 @@ class FeederModuleImporter(AbstractModule):
# Launch Importer # Launch Importer
if __name__ == '__main__': if __name__ == '__main__':
module = FeederModuleImporter() module = FeederModuleImporter()
# module.debug = True
module.run() module.run()

View file

@ -93,7 +93,10 @@ class AbstractChatFeeder(DefaultFeeder, ABC):
return self.json_data['meta'].get('reactions', []) return self.json_data['meta'].get('reactions', [])
def get_message_timestamp(self): def get_message_timestamp(self):
return self.json_data['meta']['date']['timestamp'] # TODO CREATE DEFAULT TIMESTAMP if not self.json_data['meta'].get('date'):
return None
else:
return self.json_data['meta']['date']['timestamp']
# if self.json_data['meta'].get('date'): # if self.json_data['meta'].get('date'):
# date = datetime.datetime.fromtimestamp( self.json_data['meta']['date']['timestamp']) # date = datetime.datetime.fromtimestamp( self.json_data['meta']['date']['timestamp'])
# date = date.strftime('%Y/%m/%d') # date = date.strftime('%Y/%m/%d')
@ -115,17 +118,29 @@ class AbstractChatFeeder(DefaultFeeder, ABC):
def get_message_reply_id(self): def get_message_reply_id(self):
return self.json_data['meta'].get('reply_to', {}).get('message_id') return self.json_data['meta'].get('reply_to', {}).get('message_id')
def get_message_forward(self):
return self.json_data['meta'].get('forward')
def get_message_content(self): def get_message_content(self):
decoded = base64.standard_b64decode(self.json_data['data']) decoded = base64.standard_b64decode(self.json_data['data'])
return _gunzip_bytes_obj(decoded) return _gunzip_bytes_obj(decoded)
def get_obj(self): # TODO handle others objects -> images, pdf, ... def get_obj(self):
#### TIMESTAMP #### #### TIMESTAMP ####
timestamp = self.get_message_timestamp() timestamp = self.get_message_timestamp()
#### Create Object ID #### #### Create Object ID ####
chat_id = self.get_chat_id() chat_id = self.get_chat_id()
try:
message_id = self.get_message_id() message_id = self.get_message_id()
except KeyError:
if chat_id:
self.obj = Chat(chat_id, self.get_chat_instance_uuid())
return self.obj
else:
self.obj = None
return None
thread_id = self.get_thread_id() thread_id = self.get_thread_id()
# channel id # channel id
# thread id # thread id
@ -236,7 +251,10 @@ class AbstractChatFeeder(DefaultFeeder, ABC):
# # ADD NEW MESSAGE REF (used by discord) # # ADD NEW MESSAGE REF (used by discord)
def process_sender(self, new_objs, obj, date, timestamp): def process_sender(self, new_objs, obj, date, timestamp):
meta = self.json_data['meta']['sender'] meta = self.json_data['meta'].get('sender')
if not meta:
return None
user_account = UsersAccount.UserAccount(meta['id'], self.get_chat_instance_uuid()) user_account = UsersAccount.UserAccount(meta['id'], self.get_chat_instance_uuid())
# date stat + correlation # date stat + correlation
@ -286,8 +304,6 @@ class AbstractChatFeeder(DefaultFeeder, ABC):
# REPLY # REPLY
reply_id = self.get_message_reply_id() reply_id = self.get_message_reply_id()
# TODO Translation
print(self.obj.type) print(self.obj.type)
# TODO FILES + FILES REF # TODO FILES + FILES REF
@ -295,7 +311,7 @@ class AbstractChatFeeder(DefaultFeeder, ABC):
# get object by meta object type # get object by meta object type
if self.obj.type == 'message': if self.obj.type == 'message':
# Content # Content
obj = Messages.create(self.obj.id, self.get_message_content()) # TODO translation obj = Messages.create(self.obj.id, self.get_message_content())
# FILENAME # FILENAME
media_name = self.get_media_name() media_name = self.get_media_name()
@ -305,7 +321,8 @@ class AbstractChatFeeder(DefaultFeeder, ABC):
for reaction in self.get_reactions(): for reaction in self.get_reactions():
obj.add_reaction(reaction['reaction'], int(reaction['count'])) obj.add_reaction(reaction['reaction'], int(reaction['count']))
elif self.obj.type == 'chat':
pass
else: else:
chat_id = self.get_chat_id() chat_id = self.get_chat_id()
thread_id = self.get_thread_id() thread_id = self.get_thread_id()
@ -341,9 +358,26 @@ class AbstractChatFeeder(DefaultFeeder, ABC):
# CHAT # CHAT
chat_objs = self.process_chat(new_objs, obj, date, timestamp, reply_id=reply_id) chat_objs = self.process_chat(new_objs, obj, date, timestamp, reply_id=reply_id)
# Message forward
# if self.get_json_meta().get('forward'):
# forward_from = self.get_message_forward()
# print('-----------------------------------------------------------')
# print(forward_from)
# if forward_from:
# forward_from_type = forward_from['from']['type']
# if forward_from_type == 'channel' or forward_from_type == 'chat':
# chat_forward_id = forward_from['from']['id']
# chat_forward = Chat(chat_forward_id, self.get_chat_instance_uuid())
# if chat_forward.exists():
# for chat_obj in chat_objs:
# if chat_obj.type == 'chat':
# chat_forward.add_relationship(chat_obj.get_global_id(), 'forward')
# # chat_forward.add_relationship(obj.get_global_id(), 'forward')
# SENDER # TODO HANDLE NULL SENDER # SENDER # TODO HANDLE NULL SENDER
user_account = self.process_sender(new_objs, obj, date, timestamp) user_account = self.process_sender(new_objs, obj, date, timestamp)
if user_account:
# UserAccount---ChatObjects # UserAccount---ChatObjects
for obj_chat in chat_objs: for obj_chat in chat_objs:
user_account.add_correlation(obj_chat.type, obj_chat.get_subtype(r_str=True), obj_chat.id) user_account.add_correlation(obj_chat.type, obj_chat.get_subtype(r_str=True), obj_chat.id)

View file

@ -16,6 +16,7 @@ sys.path.append(os.environ['AIL_BIN'])
from lib.ConfigLoader import ConfigLoader from lib.ConfigLoader import ConfigLoader
config_loader = ConfigLoader() config_loader = ConfigLoader()
r_cache = config_loader.get_redis_conn("Redis_Cache")
TRANSLATOR_URL = config_loader.get_config_str('Translation', 'libretranslate') TRANSLATOR_URL = config_loader.get_config_str('Translation', 'libretranslate')
config_loader = None config_loader = None
@ -298,6 +299,25 @@ def _clean_text_to_translate(content, html=False, keys_blocks=True):
content = content.replace(it, '') content = content.replace(it, '')
return content return content
#### AIL Objects ####
def get_obj_translation(obj_global_id, content, field='', source=None, target='en'):
"""
Returns translated content
"""
translation = r_cache.get(f'translation:{target}:{obj_global_id}:{field}')
if translation:
# DEBUG
# print('cache')
# r_cache.expire(f'translation:{target}:{obj_global_id}:{field}', 0)
return translation
translation = LanguageTranslator().translate(content, source=source, target=target)
if translation:
r_cache.set(f'translation:{target}:{obj_global_id}:{field}', translation)
r_cache.expire(f'translation:{target}:{obj_global_id}:{field}', 300)
return translation
## --AIL Objects-- ##
class LanguagesDetector: class LanguagesDetector:
@ -388,7 +408,7 @@ class LanguageTranslator:
return language return language
def translate(self, content, source=None, target="en"): # TODO source target def translate(self, content, source=None, target="en"): # TODO source target
if target not in LIST_LANGUAGES: if target not in get_translation_languages():
return None return None
translation = None translation = None
if content: if content:

View file

@ -178,7 +178,7 @@ class ChatServiceInstance:
if 'chats' in options: if 'chats' in options:
meta['chats'] = [] meta['chats'] = []
for chat_id in self.get_chats(): for chat_id in self.get_chats():
meta['chats'].append(Chats.Chat(chat_id, self.uuid).get_meta({'created_at', 'icon', 'nb_subchannels'})) meta['chats'].append(Chats.Chat(chat_id, self.uuid).get_meta({'created_at', 'icon', 'nb_subchannels', 'nb_messages'}))
return meta return meta
def get_nb_chats(self): def get_nb_chats(self):
@ -280,7 +280,6 @@ def create_chat_service_instance(protocol, network=None, address=None):
####################################################################################### #######################################################################################
def get_obj_chat(chat_type, chat_subtype, chat_id): def get_obj_chat(chat_type, chat_subtype, chat_id):
print(chat_type, chat_subtype, chat_id)
if chat_type == 'chat': if chat_type == 'chat':
return Chats.Chat(chat_id, chat_subtype) return Chats.Chat(chat_id, chat_subtype)
elif chat_type == 'chat-subchannel': elif chat_type == 'chat-subchannel':
@ -300,12 +299,12 @@ def get_obj_chat_meta(obj_chat, new_options=set()):
options.add(option) options.add(option)
return obj_chat.get_meta(options=options) return obj_chat.get_meta(options=options)
def get_subchannels_meta_from_global_id(subchannels): def get_subchannels_meta_from_global_id(subchannels, translation_target=None):
meta = [] meta = []
for sub in subchannels: for sub in subchannels:
_, instance_uuid, sub_id = sub.split(':', 2) _, instance_uuid, sub_id = sub.split(':', 2)
subchannel = ChatSubChannels.ChatSubChannel(sub_id, instance_uuid) subchannel = ChatSubChannels.ChatSubChannel(sub_id, instance_uuid)
meta.append(subchannel.get_meta({'nb_messages'})) meta.append(subchannel.get_meta({'nb_messages', 'created_at', 'icon', 'translation'}, translation_target=translation_target))
return meta return meta
def get_chat_meta_from_global_id(chat_global_id): def get_chat_meta_from_global_id(chat_global_id):
@ -336,13 +335,13 @@ def api_get_chat(chat_id, chat_instance_uuid, translation_target=None, nb=-1, pa
chat = Chats.Chat(chat_id, chat_instance_uuid) chat = Chats.Chat(chat_id, chat_instance_uuid)
if not chat.exists(): if not chat.exists():
return {"status": "error", "reason": "Unknown chat"}, 404 return {"status": "error", "reason": "Unknown chat"}, 404
meta = chat.get_meta({'created_at', 'icon', 'info', 'nb_participants', 'subchannels', 'threads', 'username'}) meta = chat.get_meta({'created_at', 'icon', 'info', 'nb_participants', 'subchannels', 'threads', 'translation', 'username'}, translation_target=translation_target)
if meta['username']: if meta['username']:
meta['username'] = get_username_meta_from_global_id(meta['username']) meta['username'] = get_username_meta_from_global_id(meta['username'])
if meta['subchannels']: if meta['subchannels']:
meta['subchannels'] = get_subchannels_meta_from_global_id(meta['subchannels']) meta['subchannels'] = get_subchannels_meta_from_global_id(meta['subchannels'], translation_target=translation_target)
else: else:
if translation_target not in Language.LIST_LANGUAGES: if translation_target not in Language.get_translation_languages():
translation_target = None translation_target = None
meta['messages'], meta['pagination'], meta['tags_messages'] = chat.get_messages(translation_target=translation_target, nb=nb, page=page) meta['messages'], meta['pagination'], meta['tags_messages'] = chat.get_messages(translation_target=translation_target, nb=nb, page=page)
return meta, 200 return meta, 200
@ -374,7 +373,7 @@ def api_get_subchannel(chat_id, chat_instance_uuid, translation_target=None, nb=
subchannel = ChatSubChannels.ChatSubChannel(chat_id, chat_instance_uuid) subchannel = ChatSubChannels.ChatSubChannel(chat_id, chat_instance_uuid)
if not subchannel.exists(): if not subchannel.exists():
return {"status": "error", "reason": "Unknown subchannel"}, 404 return {"status": "error", "reason": "Unknown subchannel"}, 404
meta = subchannel.get_meta({'chat', 'created_at', 'icon', 'nb_messages', 'nb_participants', 'threads'}) meta = subchannel.get_meta({'chat', 'created_at', 'icon', 'nb_messages', 'nb_participants', 'threads', 'translation'}, translation_target=translation_target)
if meta['chat']: if meta['chat']:
meta['chat'] = get_chat_meta_from_global_id(meta['chat']) meta['chat'] = get_chat_meta_from_global_id(meta['chat'])
if meta.get('threads'): if meta.get('threads'):
@ -394,21 +393,18 @@ def api_get_thread(thread_id, thread_instance_uuid, translation_target=None, nb=
meta['messages'], meta['pagination'], meta['tags_messages'] = thread.get_messages(translation_target=translation_target, nb=nb, page=page) meta['messages'], meta['pagination'], meta['tags_messages'] = thread.get_messages(translation_target=translation_target, nb=nb, page=page)
return meta, 200 return meta, 200
def api_get_message(message_id): def api_get_message(message_id, translation_target=None):
message = Messages.Message(message_id) message = Messages.Message(message_id)
if not message.exists(): if not message.exists():
return {"status": "error", "reason": "Unknown uuid"}, 404 return {"status": "error", "reason": "Unknown uuid"}, 404
meta = message.get_meta({'chat', 'content', 'icon', 'images', 'link', 'parent', 'parent_meta', 'user-account'}) meta = message.get_meta({'chat', 'content', 'files-names', 'icon', 'images', 'link', 'parent', 'parent_meta', 'reactions', 'thread', 'translation', 'user-account'}, translation_target=translation_target)
# if meta['chat']:
# print(meta['chat'])
# # meta['chat'] =
return meta, 200 return meta, 200
def api_get_user_account(user_id, instance_uuid): def api_get_user_account(user_id, instance_uuid, translation_target=None):
user_account = UsersAccount.UserAccount(user_id, instance_uuid) user_account = UsersAccount.UserAccount(user_id, instance_uuid)
if not user_account.exists(): if not user_account.exists():
return {"status": "error", "reason": "Unknown user-account"}, 404 return {"status": "error", "reason": "Unknown user-account"}, 404
meta = user_account.get_meta({'chats', 'icon', 'info', 'subchannels', 'threads', 'username', 'username_meta'}) meta = user_account.get_meta({'chats', 'icon', 'info', 'subchannels', 'threads', 'translation', 'username', 'username_meta'}, translation_target=translation_target)
return meta, 200 return meta, 200
# # # # # # # # # # LATER # # # # # # # # # # LATER

View file

@ -1331,6 +1331,8 @@ class CrawlerCapture:
start_time = self.get_task().get_start_time() start_time = self.get_task().get_start_time()
if r_str: if r_str:
return start_time return start_time
elif not start_time:
return 0
else: else:
start_time = datetime.strptime(start_time, "%Y/%m/%d - %H:%M.%S").timestamp() start_time = datetime.strptime(start_time, "%Y/%m/%d - %H:%M.%S").timestamp()
return int(start_time) return int(start_time)

View file

@ -76,14 +76,15 @@ class ChatSubChannel(AbstractChatObject):
# TODO TIME LAST MESSAGES # TODO TIME LAST MESSAGES
def get_meta(self, options=set()): def get_meta(self, options=set(), translation_target=None):
meta = self._get_meta(options=options) meta = self._get_meta(options=options)
meta['tags'] = self.get_tags(r_list=True) meta['tags'] = self.get_tags(r_list=True)
meta['name'] = self.get_name() meta['name'] = self.get_name()
if 'chat' in options: if 'chat' in options:
meta['chat'] = self.get_chat() meta['chat'] = self.get_chat()
if 'img' in options: if 'icon' in options:
meta['img'] = self.get_img() meta['icon'] = self.get_icon()
meta['img'] = meta['icon']
if 'nb_messages' in options: if 'nb_messages' in options:
meta['nb_messages'] = self.get_nb_messages() meta['nb_messages'] = self.get_nb_messages()
if 'created_at' in options: if 'created_at' in options:
@ -94,6 +95,8 @@ class ChatSubChannel(AbstractChatObject):
meta['participants'] = self.get_participants() meta['participants'] = self.get_participants()
if 'nb_participants' in options: if 'nb_participants' in options:
meta['nb_participants'] = self.get_nb_participants() meta['nb_participants'] = self.get_nb_participants()
if 'translation' in options and translation_target:
meta['translation_name'] = self.translate(meta['name'], field='name', target=translation_target)
return meta return meta
def get_misp_object(self): def get_misp_object(self):

View file

@ -70,18 +70,23 @@ class Chat(AbstractChatObject):
icon = '\uf086' icon = '\uf086'
return {'style': style, 'icon': icon, 'color': '#4dffff', 'radius': 5} return {'style': style, 'icon': icon, 'color': '#4dffff', 'radius': 5}
def get_meta(self, options=set()): def get_meta(self, options=set(), translation_target=None):
meta = self._get_meta(options=options) meta = self._get_meta(options=options)
meta['name'] = self.get_name() meta['name'] = self.get_name()
meta['tags'] = self.get_tags(r_list=True) meta['tags'] = self.get_tags(r_list=True)
if 'icon' in options: if 'icon' in options:
meta['icon'] = self.get_icon() meta['icon'] = self.get_icon()
meta['img'] = meta['icon']
if 'info' in options: if 'info' in options:
meta['info'] = self.get_info() meta['info'] = self.get_info()
if 'translation' in options and translation_target:
meta['translation_info'] = self.translate(meta['info'], field='info', target=translation_target)
if 'participants' in options: if 'participants' in options:
meta['participants'] = self.get_participants() meta['participants'] = self.get_participants()
if 'nb_participants' in options: if 'nb_participants' in options:
meta['nb_participants'] = self.get_nb_participants() meta['nb_participants'] = self.get_nb_participants()
if 'nb_messages' in options:
meta['nb_messages'] = self.get_nb_messages()
if 'username' in options: if 'username' in options:
meta['username'] = self.get_username() meta['username'] = self.get_username()
if 'subchannels' in options: if 'subchannels' in options:
@ -92,7 +97,8 @@ class Chat(AbstractChatObject):
meta['created_at'] = self.get_created_at(date=True) meta['created_at'] = self.get_created_at(date=True)
if 'threads' in options: if 'threads' in options:
meta['threads'] = self.get_threads() meta['threads'] = self.get_threads()
print(meta['threads']) if 'tags_safe' in options:
meta['tags_safe'] = self.is_tags_safe(meta['tags'])
return meta return meta
def get_misp_object(self): def get_misp_object(self):

View file

@ -148,20 +148,38 @@ class Message(AbstractObject):
def add_reaction(self, reactions, nb_reaction): def add_reaction(self, reactions, nb_reaction):
r_object.hset(f'meta:reactions:{self.type}::{self.id}', reactions, nb_reaction) r_object.hset(f'meta:reactions:{self.type}::{self.id}', reactions, nb_reaction)
# Update value on import # Interactions between users -> use replies
# reply to -> parent ?
# reply/comment - > children ?
# nb views # nb views
# reactions # MENTIONS -> Messages + Chats
# nb fowards # # relationship -> mention - Chat -> Chat
# room ??? # - Message -> Chat
# message from channel ??? # - Message -> Message ??? fetch mentioned messages
# FORWARDS
# TODO Create forward CHAT -> message
# message (is forwarded) -> message (is forwarded from) ???
# # TODO get source message timestamp
#
# # is forwarded
# # forwarded from -> check if relationship
# # nb forwarded -> scard relationship
#
# Messages -> CHATS -> NB forwarded
# CHAT -> NB forwarded by chats -> NB messages -> parse full set ????
#
#
#
#
#
#
# show users chats
# message media # message media
# flag is deleted -> event or missing from feeder pass ???
def get_translation(self, content=None, source=None, target='fr'): def get_translation(self, content=None, source=None, target='fr'):
""" """
Returns translated content Returns translated content
""" """
# return self._get_field('translated') # return self._get_field('translated')
global_id = self.get_global_id() global_id = self.get_global_id()
translation = r_cache.get(f'translation:{target}:{global_id}') translation = r_cache.get(f'translation:{target}:{global_id}')
@ -272,7 +290,7 @@ class Message(AbstractObject):
if 'reactions' in options: if 'reactions' in options:
meta['reactions'] = self.get_reactions() meta['reactions'] = self.get_reactions()
if 'translation' in options and translation_target: if 'translation' in options and translation_target:
meta['translation'] = self.get_translation(content=meta.get('content'), target=translation_target) meta['translation'] = self.translate(content=meta.get('content'), target=translation_target)
# meta['encoding'] = None # meta['encoding'] = None
return meta return meta

View file

@ -5,6 +5,7 @@ import os
import sys import sys
# import re # import re
# from datetime import datetime
from flask import url_for from flask import url_for
from pymisp import MISPObject from pymisp import MISPObject
@ -88,6 +89,13 @@ class UserAccount(AbstractSubtypeObject):
def set_info(self, info): def set_info(self, info):
return self._set_field('info', info) return self._set_field('info', info)
# def get_created_at(self, date=False):
# created_at = self._get_field('created_at')
# if date and created_at:
# created_at = datetime.fromtimestamp(float(created_at))
# created_at = created_at.isoformat(' ')
# return created_at
# TODO MESSAGES: # TODO MESSAGES:
# 1) ALL MESSAGES + NB # 1) ALL MESSAGES + NB
# 2) ALL MESSAGES TIMESTAMP # 2) ALL MESSAGES TIMESTAMP
@ -122,7 +130,7 @@ class UserAccount(AbstractSubtypeObject):
def get_messages_by_chat_obj(self, chat_obj): def get_messages_by_chat_obj(self, chat_obj):
return self.get_correlation_iter_obj(chat_obj, 'message') return self.get_correlation_iter_obj(chat_obj, 'message')
def get_meta(self, options=set()): # TODO Username timeline def get_meta(self, options=set(), translation_target=None): # TODO Username timeline
meta = self._get_meta(options=options) meta = self._get_meta(options=options)
meta['id'] = self.id meta['id'] = self.id
meta['subtype'] = self.subtype meta['subtype'] = self.subtype
@ -141,6 +149,10 @@ class UserAccount(AbstractSubtypeObject):
meta['icon'] = self.get_icon() meta['icon'] = self.get_icon()
if 'info' in options: if 'info' in options:
meta['info'] = self.get_info() meta['info'] = self.get_info()
if 'translation' in options and translation_target:
meta['translation_info'] = self.translate(meta['info'], field='info', target=translation_target)
# if 'created_at':
# meta['created_at'] = self.get_created_at(date=True)
if 'chats' in options: if 'chats' in options:
meta['chats'] = self.get_chats() meta['chats'] = self.get_chats()
if 'subchannels' in options: if 'subchannels' in options:

View file

@ -24,6 +24,8 @@ from lib.ConfigLoader import ConfigLoader
from lib import Duplicate from lib import Duplicate
from lib.correlations_engine import get_nb_correlations, get_correlations, add_obj_correlation, delete_obj_correlation, delete_obj_correlations, exists_obj_correlation, is_obj_correlated, get_nb_correlation_by_correl_type, get_obj_inter_correlation from lib.correlations_engine import get_nb_correlations, get_correlations, add_obj_correlation, delete_obj_correlation, delete_obj_correlations, exists_obj_correlation, is_obj_correlated, get_nb_correlation_by_correl_type, get_obj_inter_correlation
from lib.Investigations import is_object_investigated, get_obj_investigations, delete_obj_investigations from lib.Investigations import is_object_investigated, get_obj_investigations, delete_obj_investigations
from lib.relationships_engine import get_obj_nb_relationships, add_obj_relationship
from lib.Language import get_obj_translation
from lib.Tracker import is_obj_tracked, get_obj_trackers, delete_obj_trackers from lib.Tracker import is_obj_tracked, get_obj_trackers, delete_obj_trackers
logging.config.dictConfig(ail_logger.get_config(name='ail')) logging.config.dictConfig(ail_logger.get_config(name='ail'))
@ -284,6 +286,32 @@ class AbstractObject(ABC):
## -Correlation- ## ## -Correlation- ##
## Relationship ##
def get_nb_relationships(self, filter=[]):
return get_obj_nb_relationships(self.get_global_id())
def add_relationship(self, obj2_global_id, relationship, source=True):
# is source
if source:
print(self.get_global_id(), obj2_global_id, relationship)
add_obj_relationship(self.get_global_id(), obj2_global_id, relationship)
# is target
else:
add_obj_relationship(obj2_global_id, self.get_global_id(), relationship)
## -Relationship- ##
## Translation ##
def translate(self, content=None, field='', source=None, target='en'):
global_id = self.get_global_id()
if not content:
content = self.get_content()
return get_obj_translation(global_id, content, field=field, source=source, target=target)
## -Translation- ##
## Parent ## ## Parent ##
def is_parent(self): def is_parent(self):

View file

@ -10,6 +10,7 @@ sys.path.append(os.environ['AIL_BIN'])
from lib.ConfigLoader import ConfigLoader from lib.ConfigLoader import ConfigLoader
from lib.ail_core import get_all_objects, get_object_all_subtypes from lib.ail_core import get_all_objects, get_object_all_subtypes
from lib import correlations_engine from lib import correlations_engine
from lib import relationships_engine
from lib import btc_ail from lib import btc_ail
from lib import Tag from lib import Tag
@ -468,6 +469,24 @@ def get_correlations_graph_node(obj_type, subtype, obj_id, filter_types=[], max_
# --- CORRELATION --- # # --- CORRELATION --- #
def get_obj_nb_relationships(obj_type, subtype, obj_id, filter_types=[]):
obj = get_object(obj_type, subtype, obj_id)
return obj.get_nb_relationships(filter=filter_types)
def get_relationships_graph_node(obj_type, subtype, obj_id, filter_types=[], max_nodes=300, level=1,
objs_hidden=set(),
flask_context=False):
obj_global_id = get_obj_global_id(obj_type, subtype, obj_id)
nodes, links, meta = relationships_engine.get_relationship_graph(obj_global_id,
filter_types=filter_types,
max_nodes=max_nodes, level=level,
objs_hidden=objs_hidden)
# print(meta)
meta['objs'] = list(meta['objs'])
return {"nodes": create_correlation_graph_nodes(nodes, obj_global_id, flask_context=flask_context),
"links": links,
"meta": meta}
# if __name__ == '__main__': # if __name__ == '__main__':
# r = get_objects([{'lvl': 1, 'type': 'item', 'subtype': '', 'id': 'crawled/2020/09/14/circl.lu0f4976a4-dda4-4189-ba11-6618c4a8c951'}]) # r = get_objects([{'lvl': 1, 'type': 'item', 'subtype': '', 'id': 'crawled/2020/09/14/circl.lu0f4976a4-dda4-4189-ba11-6618c4a8c951'}])

111
bin/lib/relationships_engine.py Executable file
View file

@ -0,0 +1,111 @@
#!/usr/bin/env python3
# -*-coding:UTF-8 -*
import os
import sys
sys.path.append(os.environ['AIL_BIN'])
##################################
# Import Project packages
##################################
from lib.ConfigLoader import ConfigLoader
config_loader = ConfigLoader()
r_rel = config_loader.get_db_conn("Kvrocks_Relationships")
config_loader = None
RELATIONSHIPS = {
"forward",
"mention"
}
def get_relationships():
return RELATIONSHIPS
def get_obj_relationships_by_type(obj_global_id, relationship):
return r_rel.smembers(f'rel:{relationship}:{obj_global_id}')
def get_obj_nb_relationships_by_type(obj_global_id, relationship):
return r_rel.scard(f'rel:{relationship}:{obj_global_id}')
def get_obj_relationships(obj_global_id):
relationships = []
for relationship in get_relationships():
for rel in get_obj_relationships_by_type(obj_global_id, relationship):
meta = {'relationship': relationship}
direction, obj_id = rel.split(':', 1)
if direction == 'i':
meta['source'] = obj_id
meta['target'] = obj_global_id
else:
meta['target'] = obj_id
meta['source'] = obj_global_id
if not obj_id.startswith('chat'):
continue
meta['id'] = obj_id
# meta['direction'] = direction
relationships.append(meta)
return relationships
def get_obj_nb_relationships(obj_global_id):
nb = {}
for relationship in get_relationships():
nb[relationship] = get_obj_nb_relationships_by_type(obj_global_id, relationship)
return nb
# TODO Filter by obj type ???
def add_obj_relationship(source, target, relationship):
r_rel.sadd(f'rel:{relationship}:{source}', f'o:{target}')
r_rel.sadd(f'rel:{relationship}:{target}', f'i:{source}')
# r_rel.sadd(f'rels:{source}', relationship)
# r_rel.sadd(f'rels:{target}', relationship)
def get_relationship_graph(obj_global_id, filter_types=[], max_nodes=300, level=1, objs_hidden=set()):
links = []
nodes = set()
meta = {'complete': True, 'objs': set()}
done = set()
done_link = set()
_get_relationship_graph(obj_global_id, links, nodes, meta, level, max_nodes, filter_types=filter_types, objs_hidden=objs_hidden, done=done, done_link=done_link)
return nodes, links, meta
def _get_relationship_graph(obj_global_id, links, nodes, meta, level, max_nodes, filter_types=[], objs_hidden=set(), done=set(), done_link=set()):
meta['objs'].add(obj_global_id)
nodes.add(obj_global_id)
for rel in get_obj_relationships(obj_global_id):
meta['objs'].add(rel['id'])
if rel['id'] in done:
continue
if len(nodes) > max_nodes != 0:
meta['complete'] = False
break
nodes.add(rel['id'])
str_link = f"{rel['source']}{rel['target']}{rel['relationship']}"
if str_link not in done_link:
links.append({"source": rel['source'], "target": rel['target'], "relationship": rel['relationship']})
done_link.add(str_link)
if level > 0:
next_level = level - 1
_get_relationship_graph(rel['id'], links, nodes, meta, next_level, max_nodes, filter_types=filter_types, objs_hidden=objs_hidden, done=done, done_link=done_link)
# done.add(rel['id'])
if __name__ == '__main__':
source = ''
target = ''
add_obj_relationship(source, target, 'forward')
# print(get_obj_relationships(source))

View file

@ -41,7 +41,13 @@ class DomClassifier(AbstractModule):
addr_dns = config_loader.get_config_str("DomClassifier", "dns") addr_dns = config_loader.get_config_str("DomClassifier", "dns")
self.c = DomainClassifier.domainclassifier.Extract(rawtext="", nameservers=[addr_dns]) redis_host = config_loader.get_config_str('Redis_Cache', 'host')
redis_port = config_loader.get_config_int('Redis_Cache', 'port')
redis_db = config_loader.get_config_int('Redis_Cache', 'db')
self.dom_classifier = DomainClassifier.domainclassifier.Extract(rawtext="", nameservers=[addr_dns],
redis_host=redis_host,
redis_port=redis_port, redis_db=redis_db,
re_timeout=30)
self.cc = config_loader.get_config_str("DomClassifier", "cc") self.cc = config_loader.get_config_str("DomClassifier", "cc")
self.cc_tld = config_loader.get_config_str("DomClassifier", "cc_tld") self.cc_tld = config_loader.get_config_str("DomClassifier", "cc_tld")
@ -58,34 +64,34 @@ class DomClassifier(AbstractModule):
item_source = item.get_source() item_source = item.get_source()
try: try:
self.c.text(rawtext=host) self.dom_classifier.text(rawtext=host)
if not self.c.domain: if not self.dom_classifier.domain:
return return
print(self.c.domain) print(self.dom_classifier.domain)
self.c.validdomain(passive_dns=True, extended=False) self.dom_classifier.validdomain(passive_dns=True, extended=False)
# self.logger.debug(self.c.vdomain) # self.logger.debug(self.dom_classifier.vdomain)
print(self.c.vdomain) print(self.dom_classifier.vdomain)
print() print()
if self.c.vdomain and d4.is_passive_dns_enabled(): if self.dom_classifier.vdomain and d4.is_passive_dns_enabled():
for dns_record in self.c.vdomain: for dns_record in self.dom_classifier.vdomain:
self.add_message_to_queue(obj=None, message=dns_record) self.add_message_to_queue(obj=None, message=dns_record)
if self.cc_tld: if self.cc_tld:
localizeddomains = self.c.include(expression=self.cc_tld) localizeddomains = self.dom_classifier.include(expression=self.cc_tld)
if localizeddomains: if localizeddomains:
print(localizeddomains) print(localizeddomains)
self.redis_logger.warning(f"DomainC;{item_source};{item_date};{item_basename};Checked {localizeddomains} located in {self.cc_tld};{item.get_id()}") self.redis_logger.warning(f"DomainC;{item_source};{item_date};{item_basename};Checked {localizeddomains} located in {self.cc_tld};{item.get_id()}")
if self.cc: if self.cc:
localizeddomains = self.c.localizedomain(cc=self.cc) localizeddomains = self.dom_classifier.localizedomain(cc=self.cc)
if localizeddomains: if localizeddomains:
print(localizeddomains) print(localizeddomains)
self.redis_logger.warning(f"DomainC;{item_source};{item_date};{item_basename};Checked {localizeddomains} located in {self.cc};{item.get_id()}") self.redis_logger.warning(f"DomainC;{item_source};{item_date};{item_basename};Checked {localizeddomains} located in {self.cc};{item.get_id()}")
if r_result: if r_result:
return self.c.vdomain return self.dom_classifier.vdomain
except IOError as err: except IOError as err:
self.redis_logger.error(f"Duplicate;{item_source};{item_date};{item_basename};CRC Checksum Failed") self.redis_logger.error(f"Duplicate;{item_source};{item_date};{item_basename};CRC Checksum Failed")

View file

@ -18,13 +18,14 @@ import os
import re import re
import sys import sys
import DomainClassifier.domainclassifier
sys.path.append(os.environ['AIL_BIN']) sys.path.append(os.environ['AIL_BIN'])
################################## ##################################
# Import Project packages # Import Project packages
################################## ##################################
from modules.abstract_module import AbstractModule from modules.abstract_module import AbstractModule
from lib.ConfigLoader import ConfigLoader from lib.ConfigLoader import ConfigLoader
from lib.objects.Items import Item
class Hosts(AbstractModule): class Hosts(AbstractModule):
""" """
@ -43,28 +44,29 @@ class Hosts(AbstractModule):
# Waiting time in seconds between to message processed # Waiting time in seconds between to message processed
self.pending_seconds = 1 self.pending_seconds = 1
self.host_regex = r'\b([a-zA-Z\d-]{,63}(?:\.[a-zA-Z\d-]{,63})+)\b' redis_host = config_loader.get_config_str('Redis_Cache', 'host')
re.compile(self.host_regex) redis_port = config_loader.get_config_int('Redis_Cache', 'port')
redis_db = config_loader.get_config_int('Redis_Cache', 'db')
self.dom_classifier = DomainClassifier.domainclassifier.Extract(rawtext="",
redis_host=redis_host,
redis_port=redis_port,
redis_db=redis_db,
re_timeout=30)
self.logger.info(f"Module: {self.module_name} Launched") self.logger.info(f"Module: {self.module_name} Launched")
def compute(self, message): def compute(self, message):
item = self.get_obj() obj = self.get_obj()
# mimetype = item_basic.get_item_mimetype(item.get_id()) content = obj.get_content()
# if mimetype.split('/')[0] == "text": self.dom_classifier.text(content)
if self.dom_classifier.domain:
content = item.get_content() print(f'{len(self.dom_classifier.domain)} host {obj.get_id()}')
hosts = self.regex_findall(self.host_regex, item.get_id(), content, r_set=True) # print(self.dom_classifier.domain)
if hosts: for domain in self.dom_classifier.domain:
print(f'{len(hosts)} host {item.get_id()}') if domain:
for host in hosts: self.add_message_to_queue(message=domain, queue='Host')
# print(host)
if not host.endswith('.onion'):
self.add_message_to_queue(message=str(host), queue='Host')
if __name__ == '__main__': if __name__ == '__main__':
module = Hosts() module = Hosts()
module.run() module.run()

View file

@ -1,14 +1,14 @@
################################ GENERAL ##################################### ################################ GENERAL #####################################
# By default kvrocks listens for connections from all the network interfaces # By default kvrocks listens for connections from localhost interface.
# available on the server. It is possible to listen to just one or multiple # It is possible to listen to just one or multiple interfaces using
# interfaces using the "bind" configuration directive, followed by one or # the "bind" configuration directive, followed by one or more IP addresses.
# more IP addresses.
# #
# Examples: # Examples:
# #
# bind 192.168.1.100 10.0.0.1 # bind 192.168.1.100 10.0.0.1
# bind 127.0.0.1 ::1 # bind 127.0.0.1 ::1
# bind 0.0.0.0
bind 127.0.0.1 bind 127.0.0.1
# Unix socket. # Unix socket.
@ -26,32 +26,52 @@ port 6383
# Close the connection after a client is idle for N seconds (0 to disable) # Close the connection after a client is idle for N seconds (0 to disable)
timeout 0 timeout 0
# The number of worker's threads, increase or decrease it would effect the performance. # The number of worker's threads, increase or decrease would affect the performance.
workers 8 workers 8
# By default kvrocks does not run as a daemon. Use 'yes' if you need it. # By default, kvrocks does not run as a daemon. Use 'yes' if you need it.
# Note that kvrocks will write a pid file in /var/run/kvrocks.pid when daemonized. # Note that kvrocks will write a PID file in /var/run/kvrocks.pid when daemonized
daemonize no daemonize no
# Kvrocks implements cluster solution that is similar with redis cluster solution. # Kvrocks implements the cluster solution that is similar to the Redis cluster solution.
# You can get cluster information by CLUSTER NODES|SLOTS|INFO command, it also is # You can get cluster information by CLUSTER NODES|SLOTS|INFO command, it also is
# adapted to redis-cli, redis-benchmark, redis cluster SDK and redis cluster proxy. # adapted to redis-cli, redis-benchmark, Redis cluster SDK, and Redis cluster proxy.
# But kvrocks doesn't support to communicate with each others, so you must set # But kvrocks doesn't support communicating with each other, so you must set
# cluster topology by CLUSTER SETNODES|SETNODEID commands, more details: #219. # cluster topology by CLUSTER SETNODES|SETNODEID commands, more details: #219.
# #
# PLEASE NOTE: # PLEASE NOTE:
# If you enable cluster, kvrocks will encode key with its slot id calculated by # If you enable cluster, kvrocks will encode key with its slot id calculated by
# CRC16 and modulo 16384, endoding key with its slot id makes it efficient to # CRC16 and modulo 16384, encoding key with its slot id makes it efficient to
# migrate keys based on slot. So if you enabled at first time, cluster mode must # migrate keys based on the slot. So if you enabled at first time, cluster mode must
# not be disabled after restarting, and vice versa. That is to say, data is not # not be disabled after restarting, and vice versa. That is to say, data is not
# compatible between standalone mode with cluster mode, you must migrate data # compatible between standalone mode with cluster mode, you must migrate data
# if you want to change mode, otherwise, kvrocks will make data corrupt. # if you want to change mode, otherwise, kvrocks will make data corrupt.
# #
# Default: no # Default: no
cluster-enabled no cluster-enabled no
# By default, namespaces are stored in the configuration file and won't be replicated
# to replicas. This option allows to change this behavior, so that namespaces are also
# propagated to slaves. Note that:
# 1) it won't replicate the 'masterauth' to prevent breaking master/replica replication
# 2) it will overwrite replica's namespace with master's namespace, so be careful of in-using namespaces
# 3) cannot switch off the namespace replication once it's enabled
#
# Default: no
repl-namespace-enabled no
# Persist the cluster nodes topology in local file($dir/nodes.conf). This configuration
# takes effect only if the cluster mode was enabled.
#
# If yes, it will try to load the cluster topology from the local file when starting,
# and dump the cluster nodes into the file if it was changed.
#
# Default: yes
persist-cluster-nodes-enabled yes
# Set the max number of connected clients at the same time. By default # Set the max number of connected clients at the same time. By default
# this limit is set to 10000 clients, however if the server is not # this limit is set to 10000 clients. However, if the server is not
# able to configure the process file limit to allow for the specified limit # able to configure the process file limit to allow for the specified limit
# the max number of allowed clients is set to the current file limit # the max number of allowed clients is set to the current file limit
# #
@ -71,18 +91,17 @@ maxclients 10000
# 150k passwords per second against a good box. This means that you should # 150k passwords per second against a good box. This means that you should
# use a very strong password otherwise it will be very easy to break. # use a very strong password otherwise it will be very easy to break.
# #
# requirepass foobared
requirepass ail requirepass ail
# If the master is password protected (using the "masterauth" configuration # If the master is password protected (using the "masterauth" configuration
# directive below) it is possible to tell the slave to authenticate before # directive below) it is possible to tell the slave to authenticate before
# starting the replication synchronization process, otherwise the master will # starting the replication synchronization process. Otherwise, the master will
# refuse the slave request. # refuse the slave request.
# #
# masterauth foobared # masterauth foobared
# Master-Salve replication would check db name is matched. if not, the slave should # Master-Salve replication would check db name is matched. if not, the slave should
# refuse to sync the db from master. Don't use default value, set the db-name to identify # refuse to sync the db from master. Don't use the default value, set the db-name to identify
# the cluster. # the cluster.
db-name change.me.db db-name change.me.db
@ -98,7 +117,22 @@ dir DATA_KVROCKS
# #
# log-dir stdout # log-dir stdout
# When running daemonized, kvrocks writes a pid file in ${CONFIG_DIR}/kvrocks.pid by # Log level
# Possible values: info, warning, error, fatal
# Default: info
log-level info
# You can configure log-retention-days to control whether to enable the log cleaner
# and the maximum retention days that the INFO level logs will be kept.
#
# if set to -1, that means to disable the log cleaner.
# if set to 0, all previous INFO level logs will be immediately removed.
# if set to between 0 to INT_MAX, that means it will retent latest N(log-retention-days) day logs.
# By default the log-retention-days is -1.
log-retention-days -1
# When running in daemonize mode, kvrocks writes a PID file in ${CONFIG_DIR}/kvrocks.pid by
# default. You can specify a custom pid file location here. # default. You can specify a custom pid file location here.
# pidfile /var/run/kvrocks.pid # pidfile /var/run/kvrocks.pid
pidfile DATA_KVROCKS/kvrocks.pid pidfile DATA_KVROCKS/kvrocks.pid
@ -146,7 +180,7 @@ tcp-backlog 511
master-use-repl-port no master-use-repl-port no
# Currently, master only checks sequence number when replica asks for PSYNC, # Currently, master only checks sequence number when replica asks for PSYNC,
# that is not enough since they may have different replication history even # that is not enough since they may have different replication histories even
# the replica asking sequence is in the range of the master current WAL. # the replica asking sequence is in the range of the master current WAL.
# #
# We design 'Replication Sequence ID' PSYNC, we add unique replication id for # We design 'Replication Sequence ID' PSYNC, we add unique replication id for
@ -180,11 +214,11 @@ use-rsid-psync no
# is still in progress, the slave can act in two different ways: # is still in progress, the slave can act in two different ways:
# #
# 1) if slave-serve-stale-data is set to 'yes' (the default) the slave will # 1) if slave-serve-stale-data is set to 'yes' (the default) the slave will
# still reply to client requests, possibly with out of date data, or the # still reply to client requests, possibly with out-of-date data, or the
# data set may just be empty if this is the first synchronization. # data set may just be empty if this is the first synchronization.
# #
# 2) if slave-serve-stale-data is set to 'no' the slave will reply with # 2) if slave-serve-stale-data is set to 'no' the slave will reply with
# an error "SYNC with master in progress" to all the kind of commands # an error "SYNC with master in progress" to all kinds of commands
# but to INFO and SLAVEOF. # but to INFO and SLAVEOF.
# #
slave-serve-stale-data yes slave-serve-stale-data yes
@ -203,6 +237,35 @@ slave-serve-stale-data yes
# Default: no # Default: no
slave-empty-db-before-fullsync no slave-empty-db-before-fullsync no
# A Kvrocks master is able to list the address and port of the attached
# replicas in different ways. For example the "INFO replication" section
# offers this information, which is used, among other tools, by
# Redis Sentinel in order to discover replica instances.
# Another place where this info is available is in the output of the
# "ROLE" command of a master.
#
# The listed IP address and port normally reported by a replica is
# obtained in the following way:
#
# IP: The address is auto detected by checking the peer address
# of the socket used by the replica to connect with the master.
#
# Port: The port is communicated by the replica during the replication
# handshake, and is normally the port that the replica is using to
# listen for connections.
#
# However when port forwarding or Network Address Translation (NAT) is
# used, the replica may actually be reachable via different IP and port
# pairs. The following two options can be used by a replica in order to
# report to its master a specific set of IP and port, so that both INFO
# and ROLE will report those values.
#
# There is no need to use both the options if you need to override just
# the port or the IP address.
#
# replica-announce-ip 5.5.5.5
# replica-announce-port 1234
# If replicas need full synchronization with master, master need to create # If replicas need full synchronization with master, master need to create
# checkpoint for feeding replicas, and replicas also stage a checkpoint of # checkpoint for feeding replicas, and replicas also stage a checkpoint of
# the master. If we also keep the backup, it maybe occupy extra disk space. # the master. If we also keep the backup, it maybe occupy extra disk space.
@ -212,7 +275,7 @@ slave-empty-db-before-fullsync no
# Default: no # Default: no
purge-backup-on-fullsync no purge-backup-on-fullsync no
# The maximum allowed rate (in MB/s) that should be used by Replication. # The maximum allowed rate (in MB/s) that should be used by replication.
# If the rate exceeds max-replication-mb, replication will slow down. # If the rate exceeds max-replication-mb, replication will slow down.
# Default: 0 (i.e. no limit) # Default: 0 (i.e. no limit)
max-replication-mb 0 max-replication-mb 0
@ -220,8 +283,8 @@ max-replication-mb 0
# The maximum allowed aggregated write rate of flush and compaction (in MB/s). # The maximum allowed aggregated write rate of flush and compaction (in MB/s).
# If the rate exceeds max-io-mb, io will slow down. # If the rate exceeds max-io-mb, io will slow down.
# 0 is no limit # 0 is no limit
# Default: 500 # Default: 0
max-io-mb 500 max-io-mb 0
# The maximum allowed space (in GB) that should be used by RocksDB. # The maximum allowed space (in GB) that should be used by RocksDB.
# If the total size of the SST files exceeds max_allowed_space, writes to RocksDB will fail. # If the total size of the SST files exceeds max_allowed_space, writes to RocksDB will fail.
@ -231,7 +294,7 @@ max-db-size 0
# The maximum backup to keep, server cron would run every minutes to check the num of current # The maximum backup to keep, server cron would run every minutes to check the num of current
# backup, and purge the old backup if exceed the max backup num to keep. If max-backup-to-keep # backup, and purge the old backup if exceed the max backup num to keep. If max-backup-to-keep
# is 0, no backup would be keep. But now, we only support 0 or 1. # is 0, no backup would be kept. But now, we only support 0 or 1.
max-backup-to-keep 1 max-backup-to-keep 1
# The maximum hours to keep the backup. If max-backup-keep-hours is 0, wouldn't purge any backup. # The maximum hours to keep the backup. If max-backup-keep-hours is 0, wouldn't purge any backup.
@ -243,6 +306,115 @@ max-backup-keep-hours 24
# Default: 16 # Default: 16
max-bitmap-to-string-mb 16 max-bitmap-to-string-mb 16
# Whether to enable SCAN-like cursor compatible with Redis.
# If enabled, the cursor will be unsigned 64-bit integers.
# If disabled, the cursor will be a string.
# Default: no
redis-cursor-compatible yes
# Whether to enable the RESP3 protocol.
# NOTICE: RESP3 is still under development, don't enable it in production environment.
#
# Default: no
# resp3-enabled no
# Maximum nesting depth allowed when parsing and serializing
# JSON documents while using JSON commands like JSON.SET.
# Default: 1024
json-max-nesting-depth 1024
# The underlying storage format of JSON data type
# NOTE: This option only affects newly written/updated key-values
# The CBOR format may reduce the storage size and speed up JSON commands
# Available values: json, cbor
# Default: json
json-storage-format json
################################## TLS ###################################
# By default, TLS/SSL is disabled, i.e. `tls-port` is set to 0.
# To enable it, `tls-port` can be used to define TLS-listening ports.
# tls-port 0
# Configure a X.509 certificate and private key to use for authenticating the
# server to connected clients, masters or cluster peers.
# These files should be PEM formatted.
#
# tls-cert-file kvrocks.crt
# tls-key-file kvrocks.key
# If the key file is encrypted using a passphrase, it can be included here
# as well.
#
# tls-key-file-pass secret
# Configure a CA certificate(s) bundle or directory to authenticate TLS/SSL
# clients and peers. Kvrocks requires an explicit configuration of at least one
# of these, and will not implicitly use the system wide configuration.
#
# tls-ca-cert-file ca.crt
# tls-ca-cert-dir /etc/ssl/certs
# By default, clients on a TLS port are required
# to authenticate using valid client side certificates.
#
# If "no" is specified, client certificates are not required and not accepted.
# If "optional" is specified, client certificates are accepted and must be
# valid if provided, but are not required.
#
# tls-auth-clients no
# tls-auth-clients optional
# By default, only TLSv1.2 and TLSv1.3 are enabled and it is highly recommended
# that older formally deprecated versions are kept disabled to reduce the attack surface.
# You can explicitly specify TLS versions to support.
# Allowed values are case insensitive and include "TLSv1", "TLSv1.1", "TLSv1.2",
# "TLSv1.3" (OpenSSL >= 1.1.1) or any combination.
# To enable only TLSv1.2 and TLSv1.3, use:
#
# tls-protocols "TLSv1.2 TLSv1.3"
# Configure allowed ciphers. See the ciphers(1ssl) manpage for more information
# about the syntax of this string.
#
# Note: this configuration applies only to <= TLSv1.2.
#
# tls-ciphers DEFAULT:!MEDIUM
# Configure allowed TLSv1.3 ciphersuites. See the ciphers(1ssl) manpage for more
# information about the syntax of this string, and specifically for TLSv1.3
# ciphersuites.
#
# tls-ciphersuites TLS_CHACHA20_POLY1305_SHA256
# When choosing a cipher, use the server's preference instead of the client
# preference. By default, the server follows the client's preference.
#
# tls-prefer-server-ciphers yes
# By default, TLS session caching is enabled to allow faster and less expensive
# reconnections by clients that support it. Use the following directive to disable
# caching.
#
# tls-session-caching no
# Change the default number of TLS sessions cached. A zero value sets the cache
# to unlimited size. The default size is 20480.
#
# tls-session-cache-size 5000
# Change the default timeout of cached TLS sessions. The default timeout is 300
# seconds.
#
# tls-session-cache-timeout 60
# By default, a replica does not attempt to establish a TLS connection
# with its master.
#
# Use the following directive to enable TLS on replication links.
#
# tls-replication yes
################################## SLOW LOG ################################### ################################## SLOW LOG ###################################
# The Kvrocks Slow Log is a mechanism to log queries that exceeded a specified # The Kvrocks Slow Log is a mechanism to log queries that exceeded a specified
@ -301,8 +473,8 @@ supervised no
# Default: empty # Default: empty
# profiling-sample-commands "" # profiling-sample-commands ""
# Ratio of the samples would be recorded. We simply use the rand to determine # Ratio of the samples would be recorded. It is a number between 0 and 100.
# whether to record the sample or not. # We simply use the rand to determine whether to record the sample or not.
# #
# Default: 0 # Default: 0
profiling-sample-ratio 0 profiling-sample-ratio 0
@ -331,7 +503,19 @@ profiling-sample-record-threshold-ms 100
# 0-7am every day. # 0-7am every day.
compaction-checker-range 0-7 compaction-checker-range 0-7
# Bgsave scheduler, auto bgsave at schedule time # When the compaction checker is triggered, the db will periodically pick the SST file
# with the highest "deleted percentage" (i.e. the percentage of deleted keys in the SST
# file) to compact, in order to free disk space.
# However, if a specific SST file was created more than "force-compact-file-age" seconds
# ago, and its percentage of deleted keys is higher than
# "force-compact-file-min-deleted-percentage", it will be forcely compacted as well.
# Default: 172800 seconds; Range: [60, INT64_MAX];
# force-compact-file-age 172800
# Default: 10 %; Range: [1, 100];
# force-compact-file-min-deleted-percentage 10
# Bgsave scheduler, auto bgsave at scheduled time
# time expression format is the same as crontab(currently only support * and int) # time expression format is the same as crontab(currently only support * and int)
# e.g. bgsave-cron 0 3 * * * 0 4 * * * # e.g. bgsave-cron 0 3 * * * 0 4 * * *
# would bgsave the db at 3am and 4am every day # would bgsave the db at 3am and 4am every day
@ -339,7 +523,7 @@ compaction-checker-range 0-7
# Command renaming. # Command renaming.
# #
# It is possible to change the name of dangerous commands in a shared # It is possible to change the name of dangerous commands in a shared
# environment. For instance the KEYS command may be renamed into something # environment. For instance, the KEYS command may be renamed into something
# hard to guess so that it will still be available for internal-use tools # hard to guess so that it will still be available for internal-use tools
# but not available for general clients. # but not available for general clients.
# #
@ -352,39 +536,26 @@ compaction-checker-range 0-7
# #
# rename-command KEYS "" # rename-command KEYS ""
# The key-value size may so be quite different in many scenes, and use 256MiB as SST file size
# may cause data loading(large index/filter block) ineffective when the key-value was too small.
# kvrocks supports user-defined SST file in config(rocksdb.target_file_size_base),
# but it still too trivial and inconvenient to adjust the different sizes for different instances.
# so we want to periodic auto-adjust the SST size in-flight with user avg key-value size.
#
# If enabled, kvrocks will auto resize rocksdb.target_file_size_base
# and rocksdb.write_buffer_size in-flight with user avg key-value size.
# Please see #118.
#
# Default: yes
auto-resize-block-and-sst yes
################################ MIGRATE ##################################### ################################ MIGRATE #####################################
# If the network bandwidth is completely consumed by the migration task, # If the network bandwidth is completely consumed by the migration task,
# it will affect the availability of kvrocks. To avoid this situation, # it will affect the availability of kvrocks. To avoid this situation,
# migrate-speed is adpoted to limit the migrating speed. # migrate-speed is adopted to limit the migrating speed.
# Migrating speed is limited by controling the duraiton between sending data, # Migrating speed is limited by controlling the duration between sending data,
# the duation is calculated by: 1000000 * migrate-pipeline-size / migrate-speed (us). # the duration is calculated by: 1000000 * migrate-pipeline-size / migrate-speed (us).
# Value: [0,INT_MAX], 0 means no limit # Value: [0,INT_MAX], 0 means no limit
# #
# Default: 4096 # Default: 4096
migrate-speed 4096 migrate-speed 4096
# In order to reduce data transimission times and improve the efficiency of data migration, # In order to reduce data transmission times and improve the efficiency of data migration,
# pipeline is adopted to send multiple data at once. Pipeline size can be set by this option. # pipeline is adopted to send multiple data at once. Pipeline size can be set by this option.
# Value: [1, INT_MAX], it can't be 0 # Value: [1, INT_MAX], it can't be 0
# #
# Default: 16 # Default: 16
migrate-pipeline-size 16 migrate-pipeline-size 16
# In order to reduce the write forbidden time during migrating slot, we will migrate the incremetal # In order to reduce the write forbidden time during migrating slot, we will migrate the incremental
# data sevral times to reduce the amount of incremetal data. Until the quantity of incremetal # data several times to reduce the amount of incremental data. Until the quantity of incremental
# data is reduced to a certain threshold, slot will be forbidden write. The threshold is set by # data is reduced to a certain threshold, slot will be forbidden write. The threshold is set by
# this option. # this option.
# Value: [1, INT_MAX], it can't be 0 # Value: [1, INT_MAX], it can't be 0
@ -394,22 +565,21 @@ migrate-sequence-gap 10000
################################ ROCKSDB ##################################### ################################ ROCKSDB #####################################
# Specify the capacity of metadata column family block cache. Larger block cache # Specify the capacity of column family block cache. A larger block cache
# may make request faster while more keys would be cached. Max Size is 200*1024. # may make requests faster while more keys would be cached. Max Size is 400*1024.
# Default: 2048MB # Default: 4096MB
rocksdb.metadata_block_cache_size 2048 rocksdb.block_cache_size 4096
# Specify the capacity of subkey column family block cache. Larger block cache # Specify the type of cache used in the block cache.
# may make request faster while more keys would be cached. Max Size is 200*1024. # Accept value: "lru", "hcc"
# Default: 2048MB # "lru" stands for the cache with the LRU(Least Recently Used) replacement policy.
rocksdb.subkey_block_cache_size 2048
# Metadata column family and subkey column family will share a single block cache
# if set 'yes'. The capacity of shared block cache is
# metadata_block_cache_size + subkey_block_cache_size
# #
# Default: yes # "hcc" stands for the Hyper Clock Cache, a lock-free cache alternative
rocksdb.share_metadata_and_subkey_block_cache yes # that offers much improved CPU efficiency vs. LRU cache under high parallel
# load or high contention.
#
# default lru
rocksdb.block_cache_type lru
# A global cache for table-level rows in RocksDB. If almost always point # A global cache for table-level rows in RocksDB. If almost always point
# lookups, enlarging row cache may improve read performance. Otherwise, # lookups, enlarging row cache may improve read performance. Otherwise,
@ -423,7 +593,7 @@ rocksdb.row_cache_size 0
# files opened are always kept open. You can estimate number of files based # files opened are always kept open. You can estimate number of files based
# on target_file_size_base and target_file_size_multiplier for level-based # on target_file_size_base and target_file_size_multiplier for level-based
# compaction. For universal-style compaction, you can usually set it to -1. # compaction. For universal-style compaction, you can usually set it to -1.
# Default: 4096 # Default: 8096
rocksdb.max_open_files 8096 rocksdb.max_open_files 8096
# Amount of data to build up in memory (backed by an unsorted log # Amount of data to build up in memory (backed by an unsorted log
@ -442,7 +612,7 @@ rocksdb.max_open_files 8096
# default is 64MB # default is 64MB
rocksdb.write_buffer_size 64 rocksdb.write_buffer_size 64
# Target file size for compaction, target file size for Leve N can be caculated # Target file size for compaction, target file size for Level N can be calculated
# by target_file_size_base * (target_file_size_multiplier ^ (L-1)) # by target_file_size_base * (target_file_size_multiplier ^ (L-1))
# #
# Default: 128MB # Default: 128MB
@ -457,20 +627,29 @@ rocksdb.target_file_size_base 128
# allowed. # allowed.
rocksdb.max_write_buffer_number 4 rocksdb.max_write_buffer_number 4
# Maximum number of concurrent background jobs (compactions and flushes).
# For backwards compatibility we will set `max_background_jobs =
# max_background_compactions + max_background_flushes` in the case where user
# sets at least one of `max_background_compactions` or `max_background_flushes`
# (we replace -1 by 1 in case one option is unset).
rocksdb.max_background_jobs 4
# DEPRECATED: it is automatically decided based on the value of rocksdb.max_background_jobs
# Maximum number of concurrent background compaction jobs, submitted to # Maximum number of concurrent background compaction jobs, submitted to
# the default LOW priority thread pool. # the default LOW priority thread pool.
rocksdb.max_background_compactions 4 rocksdb.max_background_compactions -1
# DEPRECATED: it is automatically decided based on the value of rocksdb.max_background_jobs
# Maximum number of concurrent background memtable flush jobs, submitted by # Maximum number of concurrent background memtable flush jobs, submitted by
# default to the HIGH priority thread pool. If the HIGH priority thread pool # default to the HIGH priority thread pool. If the HIGH priority thread pool
# is configured to have zero threads, flush jobs will share the LOW priority # is configured to have zero threads, flush jobs will share the LOW priority
# thread pool with compaction jobs. # thread pool with compaction jobs.
rocksdb.max_background_flushes 4 rocksdb.max_background_flushes -1
# This value represents the maximum number of threads that will # This value represents the maximum number of threads that will
# concurrently perform a compaction job by breaking it into multiple, # concurrently perform a compaction job by breaking it into multiple,
# smaller ones that are run simultaneously. # smaller ones that are run simultaneously.
# Default: 2 (i.e. no subcompactions) # Default: 2
rocksdb.max_sub_compactions 2 rocksdb.max_sub_compactions 2
# In order to limit the size of WALs, RocksDB uses DBOptions::max_total_wal_size # In order to limit the size of WALs, RocksDB uses DBOptions::max_total_wal_size
@ -494,7 +673,7 @@ rocksdb.max_sub_compactions 2
# default is 512MB # default is 512MB
rocksdb.max_total_wal_size 512 rocksdb.max_total_wal_size 512
# We impl the repliction with rocksdb WAL, it would trigger full sync when the seq was out of range. # We implement the replication with rocksdb WAL, it would trigger full sync when the seq was out of range.
# wal_ttl_seconds and wal_size_limit_mb would affect how archived logs will be deleted. # wal_ttl_seconds and wal_size_limit_mb would affect how archived logs will be deleted.
# If WAL_ttl_seconds is not 0, then WAL files will be checked every WAL_ttl_seconds / 2 and those that # If WAL_ttl_seconds is not 0, then WAL files will be checked every WAL_ttl_seconds / 2 and those that
# are older than WAL_ttl_seconds will be deleted# # are older than WAL_ttl_seconds will be deleted#
@ -514,17 +693,17 @@ rocksdb.wal_size_limit_mb 16384
# actual size of the unit read from disk may be smaller if # actual size of the unit read from disk may be smaller if
# compression is enabled. # compression is enabled.
# #
# Default: 4KB # Default: 16KB
rocksdb.block_size 16384 rocksdb.block_size 16384
# Indicating if we'd put index/filter blocks to the block cache # Indicating if we'd put index/filter blocks to the block cache
# #
# Default: no # Default: yes
rocksdb.cache_index_and_filter_blocks yes rocksdb.cache_index_and_filter_blocks yes
# Specify the compression to use. Only compress level greater # Specify the compression to use. Only compress level greater
# than 2 to improve performance. # than 2 to improve performance.
# Accept value: "no", "snappy" # Accept value: "no", "snappy", "lz4", "zstd", "zlib"
# default snappy # default snappy
rocksdb.compression snappy rocksdb.compression snappy
@ -623,16 +802,16 @@ rocksdb.enable_blob_garbage_collection yes
rocksdb.blob_garbage_collection_age_cutoff 25 rocksdb.blob_garbage_collection_age_cutoff 25
# The purpose of following three options are to dynamically adjust the upper limit of # The purpose of the following three options are to dynamically adjust the upper limit of
# the data that each layer can store according to the size of the different # the data that each layer can store according to the size of the different
# layers of the LSM. Enabling this option will bring some improvements in # layers of the LSM. Enabling this option will bring some improvements in
# deletion efficiency and space amplification, but it will lose a certain # deletion efficiency and space amplification, but it will lose a certain
# amount of read performance. # amount of read performance.
# If you want know more details about Levels' Target Size, you can read RocksDB wiki: # If you want to know more details about Levels' Target Size, you can read RocksDB wiki:
# https://github.com/facebook/rocksdb/wiki/Leveled-Compaction#levels-target-size # https://github.com/facebook/rocksdb/wiki/Leveled-Compaction#levels-target-size
# #
# Default: no # Default: yes
rocksdb.level_compaction_dynamic_level_bytes no rocksdb.level_compaction_dynamic_level_bytes yes
# The total file size of level-1 sst. # The total file size of level-1 sst.
# #
@ -641,38 +820,92 @@ rocksdb.max_bytes_for_level_base 268435456
# Multiplication factor for the total file size of L(n+1) layers. # Multiplication factor for the total file size of L(n+1) layers.
# This option is a double type number in RocksDB, but kvrocks is # This option is a double type number in RocksDB, but kvrocks is
# not support double data type number yet, so we use int data # not support the double data type number yet, so we use integer
# number instead of double currently. # number instead of double currently.
# #
# Default: 10 # Default: 10
rocksdb.max_bytes_for_level_multiplier 10 rocksdb.max_bytes_for_level_multiplier 10
# This feature only takes effect in Iterators and MultiGet.
# If yes, RocksDB will try to read asynchronously and in parallel as much as possible to hide IO latency.
# In iterators, it will prefetch data asynchronously in the background for each file being iterated on.
# In MultiGet, it will read the necessary data blocks from those files in parallel as much as possible.
# Default no
rocksdb.read_options.async_io no
# If yes, the write will be flushed from the operating system
# buffer cache before the write is considered complete.
# If this flag is enabled, writes will be slower.
# If this flag is disabled, and the machine crashes, some recent
# rites may be lost. Note that if it is just the process that
# crashes (i.e., the machine does not reboot), no writes will be
# lost even if sync==false.
#
# Default: no
rocksdb.write_options.sync no
# If yes, writes will not first go to the write ahead log,
# and the write may get lost after a crash.
# You must keep wal enabled if you use replication.
#
# Default: no
rocksdb.write_options.disable_wal no
# If enabled and we need to wait or sleep for the write request, fails
# immediately.
#
# Default: no
rocksdb.write_options.no_slowdown no
# If enabled, write requests are of lower priority if compaction is
# behind. In this case, no_slowdown = true, the request will be canceled
# immediately. Otherwise, it will be slowed down.
# The slowdown value is determined by RocksDB to guarantee
# it introduces minimum impacts to high priority writes.
#
# Default: no
rocksdb.write_options.low_pri no
# If enabled, this writebatch will maintain the last insert positions of each
# memtable as hints in concurrent write. It can improve write performance
# in concurrent writes if keys in one writebatch are sequential.
#
# Default: no
rocksdb.write_options.memtable_insert_hint_per_batch no
# Support RocksDB auto-tune rate limiter for the background IO
# if enabled, Rate limiter will limit the compaction write if flush write is high
# Please see https://rocksdb.org/blog/2017/12/18/17-auto-tuned-rate-limiter.html
#
# Default: yes
rocksdb.rate_limiter_auto_tuned yes
# Enable this option will schedule the deletion of obsolete files in a background thread
# on iterator destruction. It can reduce the latency if there are many files to be removed.
# see https://github.com/facebook/rocksdb/wiki/IO#avoid-blocking-io
#
# Default: yes
# rocksdb.avoid_unnecessary_blocking_io yes
################################ NAMESPACE ##################################### ################################ NAMESPACE #####################################
# namespace.test change.me # namespace.test change.me
-# investigation -> db ????
-# ail2ail -> a2a ????
backup-dir DATA_KVROCKS/backup backup-dir DATA_KVROCKS/backup
fullsync-recv-file-delay 0
log-dir DATA_KVROCKS log-dir DATA_KVROCKS
unixsocketperm 26
namespace.cor ail_correls namespace.cor ail_correls
namespace.crawl ail_crawlers namespace.crawl ail_crawlers
namespace.db ail_datas namespace.db ail_datas
namespace.dup ail_dups namespace.dup ail_dups
namespace.obj ail_objs namespace.obj ail_objs
namespace.tl ail_tls namespace.rel ail_rels
namespace.stat ail_stats namespace.stat ail_stats
namespace.tag ail_tags namespace.tag ail_tags
namespace.tl ail_tls
namespace.track ail_trackers namespace.track ail_trackers
# investigation -> db ????
# ail2ail -> a2a ?????

View file

@ -196,6 +196,11 @@ host = localhost
port = 6383 port = 6383
password = ail_objs password = ail_objs
[Kvrocks_Relationships]
host = localhost
port = 6383
password = ail_rels
[Kvrocks_Timeline] [Kvrocks_Timeline]
host = localhost host = localhost
port = 6383 port = 6383

View file

@ -88,7 +88,7 @@ DEFAULT_HOME=$(pwd)
#### KVROCKS #### #### KVROCKS ####
test ! -d kvrocks/ && git clone https://github.com/apache/incubator-kvrocks.git kvrocks test ! -d kvrocks/ && git clone https://github.com/apache/incubator-kvrocks.git kvrocks
pushd kvrocks pushd kvrocks
./x.py build ./x.py build -j 4
popd popd
DEFAULT_KVROCKS_DATA=$DEFAULT_HOME/DATA_KVROCKS DEFAULT_KVROCKS_DATA=$DEFAULT_HOME/DATA_KVROCKS

24
update/v5.3/Update.py Executable file
View file

@ -0,0 +1,24 @@
#!/usr/bin/env python3
# -*-coding:UTF-8 -*
import os
import sys
sys.path.append(os.environ['AIL_HOME'])
##################################
# Import Project packages
##################################
from update.bin.ail_updater import AIL_Updater
from lib import ail_updates
class Updater(AIL_Updater):
"""default Updater."""
def __init__(self, version):
super(Updater, self).__init__(version)
if __name__ == '__main__':
updater = Updater('v5.3')
updater.run_update()

50
update/v5.3/Update.sh Executable file
View file

@ -0,0 +1,50 @@
#!/bin/bash
[ -z "$AIL_HOME" ] && echo "Needs the env var AIL_HOME. Run the script from the virtual environment." && exit 1;
[ -z "$AIL_REDIS" ] && echo "Needs the env var AIL_REDIS. Run the script from the virtual environment." && exit 1;
[ -z "$AIL_BIN" ] && echo "Needs the env var AIL_ARDB. Run the script from the virtual environment." && exit 1;
[ -z "$AIL_FLASK" ] && echo "Needs the env var AIL_FLASK. Run the script from the virtual environment." && exit 1;
export PATH=$AIL_HOME:$PATH
export PATH=$AIL_REDIS:$PATH
export PATH=$AIL_BIN:$PATH
export PATH=$AIL_FLASK:$PATH
GREEN="\\033[1;32m"
DEFAULT="\\033[0;39m"
echo -e $GREEN"Shutting down AIL ..."$DEFAULT
bash ${AIL_BIN}/LAUNCH.sh -k
wait
# SUBMODULES #
git submodule update
echo ""
echo -e $GREEN"Updating python packages ..."$DEFAULT
echo ""
pip install -U gcld3
pip install -U libretranslatepy
pip install -U xxhash
pip install -U DomainClassifier
echo ""
echo -e $GREEN"Updating KVROCKS ..."$DEFAULT
echo ""
pushd ${AIL_HOME}/kvrocks
git pull
./x.py build -j 4
popd
bash ${AIL_BIN}/LAUNCH.sh -lrv
bash ${AIL_BIN}/LAUNCH.sh -lkv
echo ""
echo -e $GREEN"Updating AIL VERSION ..."$DEFAULT
echo ""
python ${AIL_HOME}/update/v5.3/Update.py
wait
echo ""
echo ""
exit 0

View file

@ -63,7 +63,7 @@ def chats_explorer_networks():
return render_template('chats_networks.html', protocol=protocol, networks=networks) return render_template('chats_networks.html', protocol=protocol, networks=networks)
@chats_explorer.route("chats/explorer/instance", methods=['GET']) @chats_explorer.route("chats/explorer/instances", methods=['GET'])
@login_required @login_required
@login_read_only @login_read_only
def chats_explorer_instance(): def chats_explorer_instance():
@ -163,13 +163,17 @@ def chats_explorer_chat_participants():
@login_read_only @login_read_only
def objects_message(): def objects_message():
message_id = request.args.get('id') message_id = request.args.get('id')
message = chats_viewer.api_get_message(message_id) target = request.args.get('target')
if target == "Don't Translate":
target = None
message = chats_viewer.api_get_message(message_id, translation_target=target)
if message[1] != 200: if message[1] != 200:
return create_json_response(message[0], message[1]) return create_json_response(message[0], message[1])
else: else:
message = message[0] message = message[0]
languages = Language.get_translation_languages() languages = Language.get_translation_languages()
return render_template('ChatMessage.html', meta=message, bootstrap_label=bootstrap_label, return render_template('ChatMessage.html', meta=message, bootstrap_label=bootstrap_label,
translation_languages=languages, translation_target=target,
modal_add_tags=Tag.get_modal_add_tags(message['id'], object_type='message')) modal_add_tags=Tag.get_modal_add_tags(message['id'], object_type='message'))
@chats_explorer.route("/objects/user-account", methods=['GET']) @chats_explorer.route("/objects/user-account", methods=['GET'])
@ -178,9 +182,14 @@ def objects_message():
def objects_user_account(): def objects_user_account():
instance_uuid = request.args.get('subtype') instance_uuid = request.args.get('subtype')
user_id = request.args.get('id') user_id = request.args.get('id')
user_account = chats_viewer.api_get_user_account(user_id, instance_uuid) target = request.args.get('target')
if target == "Don't Translate":
target = None
user_account = chats_viewer.api_get_user_account(user_id, instance_uuid, translation_target=target)
if user_account[1] != 200: if user_account[1] != 200:
return create_json_response(user_account[0], user_account[1]) return create_json_response(user_account[0], user_account[1])
else: else:
user_account = user_account[0] user_account = user_account[0]
return render_template('user_account.html', meta=user_account, bootstrap_label=bootstrap_label) languages = Language.get_translation_languages()
return render_template('user_account.html', meta=user_account, bootstrap_label=bootstrap_label,
translation_languages=languages, translation_target=target)

View file

@ -203,7 +203,7 @@ def get_description():
return Response(json.dumps({"status": "error", "reason": "404 Not Found"}, indent=2, sort_keys=True), mimetype='application/json'), 404 return Response(json.dumps({"status": "error", "reason": "404 Not Found"}, indent=2, sort_keys=True), mimetype='application/json'), 404
# object exist # object exist
else: else:
res = ail_objects.get_object_meta(obj_type, subtype, obj_id, options={'tags', 'tags_safe'}, res = ail_objects.get_object_meta(obj_type, subtype, obj_id, options={'icon', 'tags', 'tags_safe'},
flask_context=True) flask_context=True)
if 'tags' in res: if 'tags' in res:
res['tags'] = list(res['tags']) res['tags'] = list(res['tags'])
@ -292,3 +292,64 @@ def correlation_tags_add():
max_nodes=nb_max, max_nodes=nb_max,
hidden=hidden, hidden_str=",".join(hidden), hidden=hidden, hidden_str=",".join(hidden),
filter=",".join(filter_types))) filter=",".join(filter_types)))
#####################################################################################
@correlation.route('/relationships/graph_node_json')
@login_required
@login_read_only
def relationships_graph_node_json():
obj_id = request.args.get('id')
subtype = request.args.get('subtype')
obj_type = request.args.get('type')
max_nodes = sanitise_nb_max_nodes(request.args.get('max_nodes'))
level = sanitise_level(request.args.get('level'))
json_graph = ail_objects.get_relationships_graph_node(obj_type, subtype, obj_id, max_nodes=max_nodes, level=level, flask_context=True)
return jsonify(json_graph)
@correlation.route('/relationship/show', methods=['GET', 'POST'])
@login_required
@login_read_only
def show_relationship():
if request.method == 'POST':
object_type = request.form.get('obj_type')
subtype = request.form.get('subtype')
obj_id = request.form.get('obj_id')
max_nodes = request.form.get('max_nb_nodes_in')
level = sanitise_level(request.form.get('level'))
# redirect to keep history and bookmark
return redirect(url_for('correlation.show_relationship', type=object_type, subtype=subtype, id=obj_id,
max_nodes=max_nodes, level=level))
# request.method == 'GET'
else:
obj_type = request.args.get('type')
subtype = request.args.get('subtype', '')
obj_id = request.args.get('id')
max_nodes = sanitise_nb_max_nodes(request.args.get('max_nodes'))
level = sanitise_level(request.args.get('level'))
# check if obj_id exist
if not ail_objects.exists_obj(obj_type, subtype, obj_id):
return abort(404)
# object exist
else: # TODO remove old dict key
dict_object = {"type": obj_type,
"id": obj_id,
"object_type": obj_type,
"max_nodes": max_nodes, "level": level,
"correlation_id": obj_id,
"metadata": ail_objects.get_object_meta(obj_type, subtype, obj_id, options={'tags', 'info', 'icon', 'username'}, flask_context=True),
"nb_relation": ail_objects.get_obj_nb_relationships(obj_type, subtype, obj_id)
}
if subtype:
dict_object["subtype"] = subtype
dict_object["metadata"]['type_id'] = subtype
else:
dict_object["subtype"] = ''
dict_object["metadata_card"] = ail_objects.get_object_card_meta(obj_type, subtype, obj_id)
return render_template("show_relationship.html", dict_object=dict_object, bootstrap_label=bootstrap_label,
tags_selector_data=Tag.get_tags_selector_data())

View file

@ -50,7 +50,12 @@ const create_heatmap_week_hour = (container_id, data, options) => {
.style("stroke", "black") .style("stroke", "black")
//.style("stroke-opacity", 1) //.style("stroke-opacity", 1)
var xPosition = d3.mouse(this)[0] + margin.left;
var yPosition = d3.mouse(this)[1] + margin.top + window.scrollY + 100;
tooltip.html(d.date + " " + d.hour + "-" + (d.hour + 1) + "h: <b>" + d.count + "</b> messages") tooltip.html(d.date + " " + d.hour + "-" + (d.hour + 1) + "h: <b>" + d.count + "</b> messages")
.style("left", xPosition + "px")
.style("top", yPosition + "px");
} }
const mouseleave = function(d) { const mouseleave = function(d) {
tooltip.style("opacity", 0) tooltip.style("opacity", 0)

View file

@ -135,6 +135,9 @@
{% include 'objects/image/block_blur_img_slider.html' %} {% include 'objects/image/block_blur_img_slider.html' %}
{% with translate_url=url_for('chats_explorer.objects_message', id=meta['id']), obj_id=meta['id'] %}
{% include 'chats_explorer/block_translation.html' %}
{% endwith %}
<div class="position-relative"> <div class="position-relative">

View file

@ -75,6 +75,9 @@
<tr> <tr>
<td> <td>
{{ subchannel['name'] }} {{ subchannel['name'] }}
{% if subchannel['translation_name'] %}
<div class="text-secondary">{{ subchannel['translation_name'] }}</div>
{% endif %}
</td> </td>
<td>{{ subchannel["created_at"] }}</td> <td>{{ subchannel["created_at"] }}</td>
<td> <td>

View file

@ -39,9 +39,9 @@
</div> </div>
</div> </div>
<link href="{{ url_for('static', filename='css/daterangepicker.min.css') }}" rel="stylesheet"> {#<link href="{{ url_for('static', filename='css/daterangepicker.min.css') }}" rel="stylesheet">#}
<script src="{{ url_for('static', filename='js/moment.min.js') }}"></script> {#<script src="{{ url_for('static', filename='js/moment.min.js') }}"></script>#}
<script src="{{ url_for('static', filename='js/jquery.daterangepicker.min.js') }}"></script> {#<script src="{{ url_for('static', filename='js/jquery.daterangepicker.min.js') }}"></script>#}
<script> <script>
function filter_by_time() { function filter_by_time() {
@ -55,7 +55,7 @@
} }
$('#date-range-from').dateRangePicker({ /*$('#date-range-from').dateRangePicker({
separator : ' to ', separator : ' to ',
getValue: function(){ getValue: function(){
if ($('#date-range-from-input').val() && $('#date-range-to-input').val() ) if ($('#date-range-from-input').val() && $('#date-range-to-input').val() )
@ -80,6 +80,6 @@
$('#date-range-from-input').val(s1); $('#date-range-from-input').val(s1);
$('#date-range-to-input').val(s2); $('#date-range-to-input').val(s2);
} }
}); });*/
</script> </script>

View file

@ -70,7 +70,8 @@
<th>Created at</th> <th>Created at</th>
<th>First Seen</th> <th>First Seen</th>
<th>Last Seen</th> <th>Last Seen</th>
<th>NB SubChannels</th> <th>SubChannels</th>
<th><i class="fas fa-comment-dots"></i></th>
</tr> </tr>
</thead> </thead>
<tbody style="font-size: 15px;"> <tbody style="font-size: 15px;">
@ -94,6 +95,7 @@
{% endif %} {% endif %}
</td> </td>
<td>{{ chat['nb_subchannels'] }}</td> <td>{{ chat['nb_subchannels'] }}</td>
<td>{{ chat['nb_messages'] }}</td>
</tr> </tr>
{% endfor %} {% endfor %}
</tbody> </tbody>

View file

@ -100,6 +100,10 @@
{% if chat['info'] %} {% if chat['info'] %}
<li class="list-group-item py-0"> <li class="list-group-item py-0">
<pre class="my-0">{{ chat['info'] }}</pre> <pre class="my-0">{{ chat['info'] }}</pre>
{% if chat['translation_info'] %}
<hr class="m-1">
<pre class="my-0 text-secondary">{{ chat['translation_info'] }}</pre>
{% endif %}
</li> </li>
{% endif %} {% endif %}
</li> </li>
@ -112,8 +116,12 @@
<span class="badge badge-{{ bootstrap_label[loop.index0 % 5] }}">{{ tag }} <span class="badge badge-light">{{ chat['tags_messages'][tag] }}</span></span> <span class="badge badge-{{ bootstrap_label[loop.index0 % 5] }}">{{ tag }} <span class="badge badge-light">{{ chat['tags_messages'][tag] }}</span></span>
{% endfor %} {% endfor %}
{% with translate_url=url_for('chats_explorer.chats_explorer_chat', uuid=chat['subtype']), obj_id=chat['id'], pagination=chat['pagination'] %}
{% include 'chats_explorer/block_translation.html' %}
{% endwith %}
{% if chat['subchannels'] %} {% if chat['subchannels'] %}
<h4>Sub-Channels:</h4> <h4 class="mt-2">Sub-Channels:</h4>
<table id="tablesubchannels" class="table"> <table id="tablesubchannels" class="table">
<thead class="bg-dark text-white"> <thead class="bg-dark text-white">
<tr> <tr>
@ -123,7 +131,7 @@
<th>Created at</th> <th>Created at</th>
<th>First Seen</th> <th>First Seen</th>
<th>Last Seen</th> <th>Last Seen</th>
<th>NB Messages</th> <th><i class="fas fa-comment-dots"></i></th>
</tr> </tr>
</thead> </thead>
<tbody style="font-size: 15px;"> <tbody style="font-size: 15px;">
@ -132,7 +140,12 @@
<td> <td>
<img src="{{ url_for('static', filename='image/ail-icon.png') }}" class="rounded-circle mr-1" alt="{{ meta['id'] }}" width="40" height="40"> <img src="{{ url_for('static', filename='image/ail-icon.png') }}" class="rounded-circle mr-1" alt="{{ meta['id'] }}" width="40" height="40">
</td> </td>
<td><b>{{ meta['name'] }}</b></td> <td>
<b>{{ meta['name'] }}</b>
{% if meta['translation_name'] %}
<div class="text-secondary">{{ meta['translation_name'] }}</div>
{% endif %}
</td>
<td><a href="{{ url_for('chats_explorer.objects_subchannel_messages') }}?uuid={{ meta['subtype'] }}&id={{ meta['id'] }}">{{ meta['id'] }}</a></td> <td><a href="{{ url_for('chats_explorer.objects_subchannel_messages') }}?uuid={{ meta['subtype'] }}&id={{ meta['id'] }}">{{ meta['id'] }}</a></td>
<td>{{ meta['created_at'] }}</td> <td>{{ meta['created_at'] }}</td>
<td> <td>
@ -161,9 +174,6 @@
{% include 'objects/image/block_blur_img_slider.html' %} {% include 'objects/image/block_blur_img_slider.html' %}
</span> </span>
{% with translate_url=url_for('chats_explorer.chats_explorer_chat', uuid=chat['subtype']), obj_id=chat['id'], pagination=chat['pagination'] %}
{% include 'chats_explorer/block_translation.html' %}
{% endwith %}
{% with obj_subtype=chat['subtype'], obj_id=chat['id'], url_endpoint=url_for("chats_explorer.chats_explorer_chat"), nb=chat['pagination']['nb'] %} {% with obj_subtype=chat['subtype'], obj_id=chat['id'], url_endpoint=url_for("chats_explorer.chats_explorer_chat"), nb=chat['pagination']['nb'] %}
{% set date_from=chat['first_seen'] %} {% set date_from=chat['first_seen'] %}
{% set date_to=chat['last_seen'] %} {% set date_to=chat['last_seen'] %}

View file

@ -46,7 +46,6 @@
<tr> <tr>
<th>username</th> <th>username</th>
<th>ID</th> <th>ID</th>
<th>Created at</th>
<th>First Seen</th> <th>First Seen</th>
<th>Last Seen</th> <th>Last Seen</th>
<th>NB Chats</th> <th>NB Chats</th>
@ -56,7 +55,6 @@
<tr> <tr>
<td>{{ meta['username']['id'] }}</td> <td>{{ meta['username']['id'] }}</td>
<td>{{ meta['id'] }}</td> <td>{{ meta['id'] }}</td>
<td>{{ meta['created_at'] }}</td>
<td> <td>
{% if meta['first_seen'] %} {% if meta['first_seen'] %}
{{ meta['first_seen'][0:4] }}-{{ meta['first_seen'][4:6] }}-{{ meta['first_seen'][6:8] }} {{ meta['first_seen'][0:4] }}-{{ meta['first_seen'][4:6] }}-{{ meta['first_seen'][6:8] }}
@ -74,6 +72,10 @@
{% if meta['info'] %} {% if meta['info'] %}
<li class="list-group-item py-0"> <li class="list-group-item py-0">
<pre class="my-0">{{ meta['info'] }}</pre> <pre class="my-0">{{ meta['info'] }}</pre>
{% if meta['translation_info'] %}
<hr>
<pre class="my-0 text-secondary">{{ meta['translation_info'] }}</pre>
{% endif %}
</li> </li>
{% endif %} {% endif %}
</li> </li>
@ -100,6 +102,10 @@
</div> </div>
</div> </div>
{% with translate_url=url_for('chats_explorer.objects_user_account', subtype=meta['subtype']), obj_id=meta['id'] %}
{% include 'chats_explorer/block_translation.html' %}
{% endwith %}
{# {% if meta['subchannels'] %}#} {# {% if meta['subchannels'] %}#}
{# <h4>Sub-Channels:</h4>#} {# <h4>Sub-Channels:</h4>#}

View file

@ -0,0 +1,78 @@
<link href="{{ url_for('static', filename='css/tags.css') }}" rel="stylesheet" type="text/css" />
<script src="{{ url_for('static', filename='js/tags.js') }}"></script>
{#{% with modal_add_tags=dict_object['metadata_card']['add_tags_modal']%}#}
{# {% include 'modals/add_tags.html' %}#}
{#{% endwith %}#}
{% include 'modals/edit_tag.html' %}
<div class="card my-3">
<div class="card-header" style="background-color:#d9edf7;font-size: 15px">
<h4>{{ dict_object["correlation_id"] }}</h4>
{{ dict_object }}
<div class="text-secondary">{{ dict_object["correlation_id"] }}</div>
<ul class="list-group mb-2">
<li class="list-group-item py-0">
<div class="row">
<div class="col-md-10">
<table class="table">
<thead>
<tr>
<th>Object type</th>
<th>First seen</th>
<th>Last seen</th>
<th>Nb seen</th>
</tr>
</thead>
<tbody>
<tr>
<td>
<svg height="26" width="26">
<g class="nodes">
<circle cx="13" cy="13" r="13" fill="orange"></circle>
<text x="13" y="13" text-anchor="middle" dominant-baseline="central" class="graph_node_icon {{ dict_object["metadata"]["icon"]["style"] }}" font-size="16px">{{ dict_object["metadata"]["icon"]["icon"] }}</text>
</g>
</svg>
{{ dict_object["object_type"] }}
</td>
<td>{{ dict_object["metadata"]['first_seen'] }}</td>
<td>{{ dict_object["metadata"]['last_seen'] }}</td>
<td>{{ dict_object["metadata"]['nb_seen'] }}</td>
</tr>
</tbody>
</table>
</div>
<div class="col-md-1">
<div id="sparkline"></div>
</div>
</div>
</li>
<li class="list-group-item py-0">
<br>
<div class="mb-3">
Tags:
{% for tag in dict_object["metadata"]['tags'] %}
<button class="btn btn-{{ bootstrap_label[loop.index0 % 5] }}" data-toggle="modal" data-target="#edit_tags_modal"
data-tagid="{{ tag }}" data-objtype="cookie-name" data-objsubtype="" data-objid="{{ dict_object["correlation_id"] }}">
{{ tag }}
</button>
{% endfor %}
<button type="button" class="btn btn-light" data-toggle="modal" data-target="#add_tags_modal">
<i class="far fa-plus-square"></i>
</button>
</div>
</li>
</ul>
{% with obj_type='cookie-name', obj_id=dict_object['correlation_id'], obj_subtype='' %}
{% include 'modals/investigations_register_obj.html' %}
{% endwith %}
<button type="button" class="btn btn-primary" data-toggle="modal" data-target="#investigations_register_obj_modal">
<i class="fas fa-microscope"></i> Investigations
</button>
</div>
</div>

View file

@ -541,6 +541,15 @@ d3.json(url)
.on("drag", dragged) .on("drag", dragged)
.on("end", drag_end)); .on("end", drag_end));
/*
node.append("image")
.attr("xlink:href", "https://circl.lu/assets/images/circl-logo.png")
.attr("height", 20)
.attr("width", 20)
.attr("x", -10)
.attr("y", -10);
*/
node.append("circle") node.append("circle")
.attr("r", function(d) { .attr("r", function(d) {

View file

@ -0,0 +1,719 @@
<!DOCTYPE html>
<html>
<head>
<meta charset="utf-8">
<title>AIL - framework</title>
<link rel="icon" href="{{ url_for('static', filename='image/ail-icon.png') }}">
<!-- Core CSS -->
<link href="{{ url_for('static', filename='css/bootstrap4.min.css') }}" rel="stylesheet">
<link href="{{ url_for('static', filename='css/font-awesome.min.css') }}" rel="stylesheet">
<!-- JS -->
<script src="{{ url_for('static', filename='js/jquery.js')}}"></script>
<script src="{{ url_for('static', filename='js/helper.js')}}"></script>
<script src="{{ url_for('static', filename='js/popper.min.js')}}"></script>
<script src="{{ url_for('static', filename='js/bootstrap4.min.js')}}"></script>
<script src="{{ url_for('static', filename='js/d3.min.js') }}"></script>
<style>
.icon_legend {
color: #0c5460;
background-color: #bee5eb;
}
line.link {
stroke: #666;
}
line.link:hover{
stroke: red;
stroke-width: 2px;
}
.node {
pointer-events: all;
}
circle {
stroke: none;
}
.graph_text_node {
font: 8px sans-serif;
pointer-events: none;
}
.graph_node_icon {
pointer-events: none;
}
.node text {
font: 8px sans-serif;
pointer-events: auto;
}
div.tooltip_graph {
position: absolute;
text-align: center;
padding: 2px;
font: 12px sans-serif;
background: #ebf4fb;
border: 2px solid #b7ddf2;
border-radius: 8px;
pointer-events: none;
color: #000000;
}
.pixelated {
image-rendering: pixelated;
}
.blured {
filter: blur(5px);
}
.graph_panel {
padding: unset;
}
.line_graph {
fill: none;
stroke: steelblue;
stroke-width: 2px;
stroke-linejoin: round;
stroke-linecap: round;
stroke-width: 1.5;
/*attr('stroke', '#bcbd22').*/
}
</style>
</head>
<body>
{% include 'nav_bar.html' %}
<div class="container-fluid">
<div class="row">
{% include 'sidebars/sidebar_objects.html' %}
<div class="col-12 col-lg-10" id="core_content">
{% if dict_object["object_type"] == "pgp" %}
{% include 'correlation/metadata_card_pgp.html' %}
{% elif dict_object["object_type"] == "cryptocurrency" %}
{% include 'correlation/metadata_card_cryptocurrency.html' %}
{% elif dict_object["object_type"] == "username" %}
{% include 'correlation/metadata_card_username.html' %}
{% elif dict_object["object_type"] == "decoded" %}
{% include 'correlation/metadata_card_decoded.html' %}
{% elif dict_object["object_type"] == "chat" %}
{% include 'correlation/metadata_card_chat.html' %}
{% elif dict_object["object_type"] == "cve" %}
{% include 'correlation/metadata_card_cve.html' %}
{% elif dict_object["object_type"] == "domain" %}
{% include 'correlation/metadata_card_domain.html' %}
{% elif dict_object["object_type"] == "screenshot" %}
{% include 'correlation/metadata_card_screenshot.html' %}
{% elif dict_object["object_type"] == "title" %}
{% include 'correlation/metadata_card_title.html' %}
{% elif dict_object["object_type"] == "cookie-name" %}
{% include 'correlation/metadata_card_cookie_name.html' %}
{% elif dict_object["object_type"] == "etag" %}
{% include 'correlation/metadata_card_etag.html' %}
{% elif dict_object["object_type"] == "hhhash" %}
{% include 'correlation/metadata_card_hhhash.html' %}
{% elif dict_object["object_type"] == "item" %}
{% include 'correlation/metadata_card_item.html' %}
{% endif %}
<div class="row">
<div class="col-xl-10">
<div class="card mb-3">
<div class="card-header">
<i class="fas fa-project-diagram"></i> Graph
<span class="float-right">
<div class="card border-secondary">
<div class="card-body py-2">
<div class="row">
<div class="col-md-3 text-center px-0">
<button class="btn btn-sm btn-secondary" onclick="blur_slider_correlation.val(0);blur_tooltip();">
<i class="fas fa-eye-slash"></i>
<span class="label-icon">Hide</span>
</button>
</div>
<div class="col-md-6 text-center pl-0 pt-1">
<input type="range" min="0" max="15" step="0.1" value="10" id="blur-slider-correlation">
</div>
<div class="col-md-3 text-center">
<button class="btn btn-sm btn-secondary" onclick="blur_slider_correlation.val(15);blur_tooltip();">
<i class="fas fa-image"></i>
<span class="label-icon">Full</span>
</button>
</div>
</div>
</div>
</div>
</span>
{# <span class="float-right mx-1">#}
{# {% if dict_object["object_type"] != "" %}#}
{# {% with obj_type=dict_object["object_type"], obj_id=dict_object["correlation_id"], obj_subtype=dict_object["metadata"]["type_id"],obj_lvl=1%}#}
{# {% include 'import_export/block_add_user_object_to_export.html' %}#}
{# {% endwith %}#}
{# {% endif %}#}
{# </span>#}
<span class="float-right mt-2">
<button class="btn btn-primary py-1" onclick="resize_graph();">
<i class="fas fa-sync"></i>&nbsp;Resize Graph
</button>
</span>
<span class="float-right mt-2 mx-1">
<button class="btn btn-primary py-1" onclick="reset_graph();">
<i class="fas fa-undo"></i>&nbsp;Reset Graph
</button>
</span>
<div id="incomplete_graph" class="text-danger mt-3">
<i class="fas fa-exclamation-triangle"></i>&nbsp;Graph Incomplete, Max Nodes Reached.
</div>
</div>
<div class="card-body graph_panel">
<div id="graph_loading" class="ml-3 mt-3">
<div class="spinner-border text-secondary" role="status">
<span class="sr-only">Loading...</span>
</div>
<span>Loading...</span>
</div>
<div id="graph">
</div>
</div>
</div>
{# <p>Press <b>H</b> on an object / node to hide it.</p>#}
{# {% if dict_object["hidden"] %}#}
{# <h5>Hidden objects:</h5>#}
{# {% for obj_hidden in dict_object["hidden"] %}#}
{# {{ obj_hidden }} <br>#}
{# {% endfor %}#}
{# {% endif %}#}
</div>
<div class="col-xl-2">
<div class="card mb-3">
<div class="card-body text-center px-0 py-0">
<ul class="list-group">
<li class="list-group-item list-group-item-info">Relationship</li>
<form action="{{ url_for('correlation.show_relationship') }}" method="post">
<input type="hidden" id="obj_type" name="obj_type" value="{{ dict_object["object_type"] }}">
<input type="hidden" id="subtype" name="subtype" value="{{ dict_object["metadata"]["type_id"] }}">
<input type="hidden" id="obj_id" name="obj_id" value="{{ dict_object["correlation_id"] }}">
{# <li class="list-group-item text-left">#}
{# <div class="form-check">#}
{# <input class="form-check-input" type="checkbox" value="True" id="forwardCheck" name="forwardCheck" {%if "forward" in dict_object["filter"]%}checked{%endif%}>#}
{# <label class="form-check-label" for="forwardCheck">Forwards</label>#}
{# </div>#}
{# <div class="form-check">#}
{# <input class="form-check-input" type="checkbox" value="True" id="mentionCheck" name="mentionCheck" {%if "mention" in dict_object["filter"]%}checked{%endif%}>#}
{# <label class="form-check-label" for="mentionCheck">Mentions</label>#}
{# </div>#}
{##}
{# </li>#}
<li class="list-group-item text-left">
<div class="form-group">
<label for="max_nb_nodes_in">Correlation Depth:</label>
<input class="form-control" type="number" value="{{dict_object["level"]}}" min="0" id="level" name="level">
</div>
</li>
<li class="list-group-item text-left">
<div class="form-group">
<label for="max_nb_nodes_in">Max number of nodes:</label>
<input class="form-control" type="number" value="{{dict_object["max_nodes"]}}" min="0" id="max_nb_nodes_in" name="max_nb_nodes_in">
</div>
<div class="text-center">
<input class="btn btn-primary" type="submit" value="Redraw Graph">
</div>
</li>
</form>
</ul>
<ul class="list-group">
<li class="list-group-item list-group-item-info"><i class="fas fa-info-circle fa-2x"></i></li>
<li class="list-group-item text-left">
<p>Double click on a node to open this object<br><br>
<svg height="26" width="26">
<g class="nodes">
<circle cx="13" cy="13" r="13" fill="orange"></circle>
</g>
</svg>
Current Object<br>
</p>
</li>
</ul>
<ul class="list-group">
<li class="list-group-item list-group-item-secondary"><i class="fas fa-project-diagram"></i> Direct Relationships</li>
<li class="list-group-item text-left">
{% for relationship in dict_object['nb_relation'] %}
<div class="row">
<div class="col-8">
{{ relationship }}
</div>
<div class="col-4">
<span class="badge badge-primary">{{ dict_object['nb_relation'][relationship] }}</span>
</div>
</div>
{% endfor %}
</li>
</ul>
</div>
</div>
</div>
</div>
<div class="row">
<div class="col-xl-10">
{% include 'correlation/legend_graph_correlation.html' %}
</div>
</div>
<div class="card my-2">
<div class="card-header bg-light">
<h4><i class="fas fa-tags"></i> Tags All Objects</h4>
</div>
<div class="card-body">
<form action="{{ url_for('correlation.correlation_tags_add') }}" method='post' onsubmit="SubmitAddTags();">
<input type="hidden" id="tag_obj_type" name="tag_obj_type" value="{{ dict_object["object_type"] }}">
<input type="hidden" id="tag_subtype" name="tag_subtype" value="{{ dict_object["metadata"]["type_id"] }}">
<input type="hidden" id="tag_obj_id" name="tag_obj_id" value="{{ dict_object["correlation_id"] }}">
<input type="hidden" id="tag_level" name="tag_level" value="{{dict_object["level"]}}">
<input type="hidden" id="tag_nb_max" name="tag_nb_max" value="{{dict_object["max_nodes"]}}">
<input type="hidden" id="filter" name="tag_filter" value="{{dict_object["filter_str"]}}">
<input type="hidden" id="tag_hidden" name="tag_hidden" value="{{dict_object["hidden_str"]}}">
{% include 'tags/block_tags_selector.html' %}
<button class="btn btn-primary mt-2">
<i class="fas fa-tag"></i> Add Tags
</button>
</form>
</div>
</div>
</div>
</div>
</div>
<script>
var all_graph = {};
$(document).ready(function(){
$("#incomplete_graph").hide();
$("#page-Decoded").addClass("active");
all_graph.node_graph = create_graph("{{ url_for('correlation.relationships_graph_node_json') }}?id={{ dict_object["correlation_id"] }}&type={{ dict_object["object_type"] }}&level={{ dict_object["level"] }}&filter={{ dict_object["filter_str"] }}&max_nodes={{dict_object["max_nodes"]}}{% if 'type_id' in dict_object["metadata"] %}&subtype={{ dict_object["metadata"]["type_id"] }}{% endif %}&hidden={{ dict_object["hidden_str"] }}");
all_graph.onResize();
});
const blur_slider_correlation = $('#blur-slider-correlation');
function blur_tooltip(){
var image = $('#tooltip_screenshot_correlation')[0];
if (image) {
var blurValue = $('#blur-slider-correlation').val();
blurValue = 15 - blurValue;
image.style.filter = "blur(" + blurValue + "px)";
}
}
$(window).on("resize", function() {
all_graph.onResize();
});
function toggle_sidebar(){
if($('#nav_menu').is(':visible')){
$('#nav_menu').hide();
$('#side_menu').removeClass('border-right')
$('#side_menu').removeClass('col-lg-2')
$('#core_content').removeClass('col-lg-10')
}else{
$('#nav_menu').show();
$('#side_menu').addClass('border-right')
$('#side_menu').addClass('col-lg-2')
$('#core_content').addClass('col-lg-10')
}
}
function SubmitAddTags() {
var tags = ltags.getValue();
var tagsgalaxy = ltagsgalaxies.getValue();
$('#ltags').val(tags);
$('#ltagsgalaxies').val(tagsgalaxy);
return true;
}
</script>
<script>
function resize_graph() {
zoom.translateTo(svg_node, 200, 200);
zoom.scaleTo(svg_node, 2);
}
function reset_graph() {
window.location.href = "{{ url_for('correlation.show_relationship') }}?type={{ dict_object["type"] }}&subtype={{ dict_object["subtype"] }}&id={{ dict_object["id"] }}"
}
</script>
<script>
var correl_link = "{{ url_for('correlation.show_relationship') }}?type={{ dict_object["type"] }}&subtype={{ dict_object["subtype"] }}&id={{ dict_object["id"] }}&max_nodes={{ dict_object["max_nodes"] }}&level={{ dict_object["level"] }}&filter={{ dict_object["filter_str"] }}"
{% if 'hidden_str' in dict_object %}
correl_link = correl_link + "&hidden={{ dict_object["hidden_str"] }}"
{% endif %}
var width = 400,
height = 400;
var link;
var zoom = d3.zoom()
.scaleExtent([.2, 10])
.on("zoom", zoomed);
// Loading:
//$("#graph_loading").show()
//var transform = d3.zoomIdentity;
var color = d3.scaleOrdinal(d3.schemeCategory10);
var div = d3.select("body").append("div")
.attr("class", "tooltip_graph")
.style("opacity", 0);
var simulation = d3.forceSimulation()
.force("link", d3.forceLink().id(function(d) { return d.id; }))
.force("charge", d3.forceManyBody())
.force("center", d3.forceCenter(width / 2, height / 2));
//.on("tick", ticked);
var currentObject = null;
var svg_node = d3.select("#graph").append("svg")
.attr("id", "graph_div")
.attr("width", width)
.attr("height", height)
.call(d3.zoom().scaleExtent([1, 8]).on("zoom", zoomed))
.on("dblclick.zoom", null)
var container_graph = svg_node.append("g");
//.attr("transform", "translate(40,0)")
//.attr("transform", "scale(2)");
function create_graph(url){
d3.json(url)
.then(function(data){
// ----------------------------------------------------------------------------------
svg_node.append("defs").selectAll("marker")
.data(["end"]) // Different link/path types can be defined here
.enter().append("marker")
.attr("id", String)
.attr("viewBox", "0 -5 10 10")
.attr("refX", 25) // Controls the distance of the marker from the node
.attr("refY", 0)
.attr("markerWidth", 6)
.attr("markerHeight", 6)
.attr("orient", "auto")
.append("path")
.attr("d", "M0,-5L10,0L0,5");
// ----------------------------------------------------------------------------------
var link = container_graph.append("g")
/*.selectAll("line")
.data(data.links)
.enter().append("line")
.attr("class", "link")
.attr("marker-end", "url(#end)");*/
.selectAll("g")
.data(data.links)
.enter().append("g");
link.append("line")
.attr("class", "link")
.attr("marker-end", "url(#end)")
link.append("text")
.attr("class", "link-label")
.style("font-size", "6px")
.text(function(d) { return d.relationship; })
//.attr("stroke-width", function(d) { return Math.sqrt(d.value); })
var node = container_graph.selectAll(".node")
.data(data.nodes)
.enter().append("g")
.attr("class", "nodes")
.on("dblclick", doubleclick)
.on("click", click)
.on("mouseover", mouseovered)
.on("mouseout", mouseouted)
.call(d3.drag()
.on("start", drag_start)
.on("drag", dragged)
.on("end", drag_end));
/*
node.append("image")
.attr("xlink:href", "https://circl.lu/assets/images/circl-logo.png")
.attr("height", 20)
.attr("width", 20)
.attr("x", -10)
.attr("y", -10);
*/
node.append("circle")
.attr("r", function(d) {
return (d.style.node_radius);})
.attr("fill", function(d) {
return d.style.node_color;});
node.append('text')
.attr('text-anchor', 'middle')
.attr('dominant-baseline', 'central')
.attr("class", function(d) {
return "graph_node_icon " + d.style.icon_class
})
.attr('font-size', '8px' )
.attr('pointer-events', 'none')
.text(function(d) {
//if(d.hash){
return d.style.icon_text
//}
});
zoom.translateTo(svg_node, 200, 200);
zoom.scaleTo(svg_node, 2);
/* node.append("text")
.attr("dy", 3)
.attr("dx", 7)
.attr("class", "graph_text_node")
//.style("text-anchor", function(d) { return d.children ? "end" : "start"; })
.text(function(d) { return d.id; });*/
simulation
.nodes(data.nodes)
.on("tick", ticked);
simulation.force("link")
.links(data.links);
function ticked() {
link.select("line")
.attr("x1", function(d) { return d.source.x; })
.attr("y1", function(d) { return d.source.y; })
.attr("x2", function(d) { return d.target.x; })
.attr("y2", function(d) { return d.target.y; });
link.select("text")
.attr("x", function(d) {
return (d.source.x + d.target.x) / 2;
})
.attr("y", function(d) {
return (d.source.y + d.target.y) / 2;
});
/*node
.attr("cx", function(d) { return d.x; })
.attr("cy", function(d) { return d.y; });*/
node.attr("transform", function(d) { return "translate(" + d.x + "," + d.y + ")"; });
}
// Loading ...
$("#graph_loading").remove();
if (!data.meta.complete){
$("#incomplete_graph").show();
}
d3.select("body").on("keypress", keypressed)
})
.catch(function(error) {
$("#graph_loading").remove()
svg_node.remove();
d3.select("#graph").append("div")
.text(error);
});
}
function zoomed() {
container_graph.attr("transform", d3.event.transform);
}
function doubleclick (d) {
window.open(d.url, '_blank');
}
function keypressed () {
//console.log(d3.event.keyCode)
//console.log(currentObject.id)
// hide node, H or h key
if ((d3.event.keyCode === 72 || d3.event.keyCode === 104) && currentObject) {
window.location.href = correl_link + "&hide=" + currentObject.id
}
}
function click (d) {
console.log('clicked')
}
function drag_start(d) {
if (!d3.event.active) simulation.alphaTarget(0.3).restart();
d.fx = d.x;
d.fy = d.y;
}
function dragged(d) {
d.fx = d3.event.x;
d.fy = d3.event.y;
}
function drag_end(d) {
if (!d3.event.active) simulation.alphaTarget(0);
d.fx = d.x;
d.fy = d.y;
}
function mouseovered(d) {
currentObject = d;
var d3_pageX = d3.event.pageX;
var d3_pageY = d3.event.pageY;
if (d.popover) {
div.html(d.popover)
.style("left", (d3_pageX) + "px")
.style("top", (d3_pageY - 28) + "px");
div.transition()
.duration(200)
.style("opacity", 1);
blur_tooltip();
} else {
var pop_header = "<div class=\"card text-white\" style=\"max-width: 25rem;\"><div class=\"card-header bg-dark pb-0 border-white\"><h6>"+ sanitize_text(d.text) +"</h6></div>"
var spinner = "<div class=\"card-body bg-dark pt-0\"><div class=\"spinner-border text-warning\" role=\"status\"></div> Loading...</div>"
div.html(pop_header + spinner)
.style("left", (d3_pageX) + "px")
.style("top", (d3_pageY - 28) + "px");
div.transition()
.duration(200)
.style("opacity", 1);
$.getJSON("{{ url_for('correlation.get_description') }}?object_id="+ d.id,
function(data){
var desc = pop_header + "<div class=\"card-body bg-dark pb-1 pt-2\"><dl class=\"row py-0 my-0\">"
Object.keys(data).forEach(function(key) {
if (key=="status") {
desc = desc + "<dt class=\"col-sm-3 px-0\">status</dt><dd class=\"col-sm-9 px-0\"><div class=\"badge badge-pill badge-light flex-row-reverse\" style=\"color:"
if (data["status"]) {
desc = desc + "Green"
} else {
desc = desc + "Red"
}
desc = desc + ";\"><i class=\"fas "
if (data["status"]) {
desc = desc + "fa-check-circle\"></i>UP"
} else {
desc = desc + "fa-times-circle\"></i>DOWN"
}
desc = desc + "</div></dd>"
} else if (key!="tags" && key!="id" && key!="img" && key!="icon" && key!="link" && key!="type") {
desc = desc + "<dt class=\"col-sm-3 px-0\">" + sanitize_text(key) + "</dt><dd class=\"col-sm-9 px-0\">" + sanitize_text(data[key]) + "</dd>"
}
});
desc = desc + "</dl>"
if (data["tags"]) {
data["tags"].forEach(function(tag) {
desc = desc + "<span class=\"badge badge-warning\">"+ sanitize_text(tag) +"</span>";
});
}
if (data["img"]) {
if (data["tags_safe"]) {
if (data["type"] === "screenshot") {
desc = desc + "<img src={{ url_for('objects_item.screenshot', filename="") }}"
} else {
desc = desc + "<img src={{ url_for('objects_image.image', filename="") }}"
}
desc = desc + data["img"] +" class=\"img-thumbnail blured\" id=\"tooltip_screenshot_correlation\" style=\"\"/>";
} else {
desc = desc + "<span class=\"my-2 fa-stack fa-4x\"><i class=\"fas fa-stack-1x fa-image\"></i><i class=\"fas fa-stack-2x fa-ban\" style=\"color:Red\"></i></span>";
}
}
desc = desc + "</div></div>"
div.html(desc)
.style("left", (d3_pageX) + "px")
.style("top", (d3_pageY - 28) + "px");
d.popover = desc
if (data["img"]) {
blur_tooltip();
}
})
.fail(function(error) {
desc = pop_header + "<div class=\"card-body bg-dark pt-0\"><i class=\"fas fa-3x fa-times text-danger\"></i>"+ error.statusText +"</div>"
div.html(desc)
.style("left", (d3_pageX) + "px")
.style("top", (d3_pageY - 28) + "px");
});
}
}
function mouseouted() {
currentObject = null;
div.transition()
.duration(500)
.style("opacity", 0);
//link.select("line")
}
all_graph.onResize = function () {
var aspect = 1000 / 500, all_graph = $("#graph_div");
var targetWidth = all_graph.parent().width();
all_graph.attr("width", targetWidth);
all_graph.attr("height", targetWidth / aspect);
}
window.all_graph = all_graph;
</script>
</body>
</html>