mirror of
https://github.com/ail-project/ail-framework.git
synced 2024-11-30 01:37:17 +00:00
chg: [Item] change parent-child key
This commit is contained in:
parent
d295c084c6
commit
b0891f45ec
3 changed files with 53 additions and 55 deletions
|
@ -369,11 +369,11 @@ def get_item_duplicates_dict(item_id):
|
|||
def items_migration():
|
||||
print('ITEMS MIGRATION...')
|
||||
# MIGRATE IMPORTED URLEXTRACT Father
|
||||
# for item_id in Items.get_items_by_source('urlextract'):
|
||||
# father_id = get_item_father(item_id)
|
||||
# if father_id:
|
||||
# item = Items.Item(item_id)
|
||||
# item.set_father(father_id)
|
||||
for item_id in Items.get_items_by_source('urlextract'):
|
||||
father_id = get_item_father(item_id)
|
||||
if father_id:
|
||||
item = Items.Item(item_id)
|
||||
item.set_parent(father_id)
|
||||
|
||||
# DUPLICATES
|
||||
for tag in ['infoleak:automatic-detection="credential"']: # Creditcards, Mail, Keys ???????????????????????????????
|
||||
|
@ -871,10 +871,10 @@ if __name__ == '__main__':
|
|||
#core_migration()
|
||||
#user_migration()
|
||||
#tags_migration()
|
||||
# items_migration()
|
||||
items_migration()
|
||||
#crawler_migration()
|
||||
# domain_migration() # TO TEST ###########################
|
||||
decodeds_migration()
|
||||
# decodeds_migration()
|
||||
# screenshots_migration()
|
||||
# subtypes_obj_migration()
|
||||
# ail_2_ail_migration()
|
||||
|
|
|
@ -87,7 +87,7 @@ def get_item_mimetype(item_id):
|
|||
|
||||
# # # # TREE CHILD/FATHER # # # #
|
||||
def is_parent(item_id):
|
||||
return r_object.exists(f'obj:child:item::{item_id}')
|
||||
return r_object.exists(f'child:item::{item_id}')
|
||||
|
||||
def is_children(item_id):
|
||||
return r_object.hexists(f'meta:item::{item_id}' 'parent')
|
||||
|
@ -145,7 +145,7 @@ def get_item_parent(item_id):
|
|||
return r_object.hget(f'meta:item::{item_id}', 'parent')
|
||||
|
||||
def get_item_children(item_id):
|
||||
return list(r_object.smembers(f'obj:child:item::{item_id}'))
|
||||
return list(r_object.smembers(f'child:item::{item_id}'))
|
||||
|
||||
# # TODO: handle domain last origin in domain lib
|
||||
# def _delete_node(item_id):
|
||||
|
|
|
@ -40,10 +40,6 @@ baseurl = config_loader.get_config_str("Notifications", "ail_domain")
|
|||
config_loader = None
|
||||
|
||||
|
||||
################################################################################
|
||||
################################################################################
|
||||
################################################################################
|
||||
|
||||
class Item(AbstractObject):
|
||||
"""
|
||||
AIL Item Object. (strings)
|
||||
|
@ -127,23 +123,27 @@ class Item(AbstractObject):
|
|||
return item_basic.get_item_parent(self.id)
|
||||
|
||||
def set_parent(self, parent_id):
|
||||
r_object.sadd(f'obj:child:item::{parent_id}', self.id) # TODO
|
||||
r_object.sadd(f'child:item::{parent_id}', self.id)
|
||||
r_object.hset(f'meta:item::{self.id}', 'parent', parent_id)
|
||||
|
||||
def add_children(self, child_id):
|
||||
r_object.sadd(f'obj:child:item::{self.id}', child_id) # TODO
|
||||
r_object.sadd(f'child:item::{self.id}', child_id)
|
||||
r_object.hset(f'meta:item::{child_id}', 'parent', self.id)
|
||||
|
||||
####################################################################################
|
||||
####################################################################################
|
||||
|
||||
def sanitize_id(self):
|
||||
pass
|
||||
|
||||
|
||||
# # TODO: sanitize_id
|
||||
# # TODO: check if already exists ?
|
||||
# # TODO: check if duplicate
|
||||
def save_on_disk(self, content, binary=True, compressed=False, base64=False):
|
||||
def save_on_disk(self, content, binary=True, compressed=False, b64=False):
|
||||
if not binary:
|
||||
content = content.encode()
|
||||
if base64:
|
||||
if b64:
|
||||
content = base64.standard_b64decode(content)
|
||||
if not compressed:
|
||||
content = gzip.compress(content)
|
||||
|
@ -156,9 +156,8 @@ class Item(AbstractObject):
|
|||
with open(filename, 'wb') as f:
|
||||
f.write(content)
|
||||
|
||||
|
||||
# # TODO: correlations
|
||||
#
|
||||
# # TODO:
|
||||
# correlations
|
||||
# content
|
||||
# tags
|
||||
# origin
|
||||
|
@ -180,10 +179,6 @@ class Item(AbstractObject):
|
|||
self.add_duplicate(obj_id, dup['algo'], dup['similarity'])
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
# # WARNING: UNCLEAN DELETE /!\ TEST ONLY /!\
|
||||
# TODO: DELETE ITEM CORRELATION + TAGS + METADATA + ...
|
||||
def delete(self):
|
||||
|
@ -194,6 +189,9 @@ class Item(AbstractObject):
|
|||
except FileNotFoundError:
|
||||
return False
|
||||
|
||||
####################################################################################
|
||||
####################################################################################
|
||||
|
||||
def get_link(self, flask_context=False):
|
||||
if flask_context:
|
||||
url = url_for('objects_item.showItem', id=self.id)
|
||||
|
@ -213,10 +211,9 @@ class Item(AbstractObject):
|
|||
obj = MISPObject('ail-leak', standalone=True)
|
||||
obj.first_seen = obj_date
|
||||
|
||||
obj_attrs = []
|
||||
obj_attrs.append( obj.add_attribute('first-seen', value=obj_date) )
|
||||
obj_attrs.append( obj.add_attribute('raw-data', value=self.id, data=self.get_raw_content()) )
|
||||
obj_attrs.append( obj.add_attribute('sensor', value=get_ail_uuid()) )
|
||||
obj_attrs = [obj.add_attribute('first-seen', value=obj_date),
|
||||
obj.add_attribute('raw-data', value=self.id, data=self.get_raw_content()),
|
||||
obj.add_attribute('sensor', value=get_ail_uuid())]
|
||||
for obj_attr in obj_attrs:
|
||||
for tag in self.get_tags():
|
||||
obj_attr.add_tag(tag)
|
||||
|
@ -253,12 +250,16 @@ class Item(AbstractObject):
|
|||
self.set_parent(parent_id)
|
||||
|
||||
# options: set of optional meta fields
|
||||
def get_meta(self, options=set()):
|
||||
meta = {}
|
||||
meta['id'] = self.id
|
||||
meta['date'] = self.get_date(separator=True) ############################ # TODO:
|
||||
meta['source'] = self.get_source()
|
||||
meta['tags'] = self.get_tags(r_list=True)
|
||||
def get_meta(self, options=None):
|
||||
"""
|
||||
:type options: set
|
||||
"""
|
||||
if options is None:
|
||||
options = set()
|
||||
meta = {'id': self.id,
|
||||
'date': self.get_date(separator=True),
|
||||
'source': self.get_source(),
|
||||
'tags': self.get_tags(r_list=True)}
|
||||
# optional meta fields
|
||||
if 'content' in options:
|
||||
meta['content'] = self.get_content()
|
||||
|
@ -282,7 +283,12 @@ class Item(AbstractObject):
|
|||
# meta['encoding'] = None
|
||||
return meta
|
||||
|
||||
def get_meta_crawler(self, tags=[]):
|
||||
def get_meta_crawler(self, tags=None):
|
||||
"""
|
||||
:type tags: list
|
||||
"""
|
||||
if tags is None:
|
||||
tags = []
|
||||
crawler = {}
|
||||
if self.is_crawled():
|
||||
crawler['domain'] = self.get_domain()
|
||||
|
@ -330,13 +336,16 @@ class Item(AbstractObject):
|
|||
############################################################################
|
||||
############################################################################
|
||||
|
||||
def _get_dir_source_name(dir, source_name=None, l_sources_name=set(), filter_dir=False):
|
||||
def _get_dir_source_name(directory, source_name=None, l_sources_name=None, filter_dir=False):
|
||||
"""
|
||||
:type l_sources_name: set
|
||||
"""
|
||||
if not l_sources_name:
|
||||
l_sources_name = set()
|
||||
if source_name:
|
||||
l_dir = os.listdir(os.path.join(dir, source_name))
|
||||
l_dir = os.listdir(os.path.join(directory, source_name))
|
||||
else:
|
||||
l_dir = os.listdir(dir)
|
||||
l_dir = os.listdir(directory)
|
||||
# empty directory
|
||||
if not l_dir:
|
||||
return l_sources_name.add(source_name)
|
||||
|
@ -355,7 +364,7 @@ def _get_dir_source_name(dir, source_name=None, l_sources_name=set(), filter_dir
|
|||
# pass
|
||||
if source_name:
|
||||
src_name = os.path.join(source_name, src_name)
|
||||
l_sources_name = _get_dir_source_name(dir, source_name=src_name, l_sources_name=l_sources_name, filter_dir=filter_dir)
|
||||
l_sources_name = _get_dir_source_name(directory, source_name=src_name, l_sources_name=l_sources_name, filter_dir=filter_dir)
|
||||
return l_sources_name
|
||||
|
||||
def get_items_sources(filter_dir=False, r_list=False):
|
||||
|
@ -397,6 +406,7 @@ def _manual_set_items_date_first_last():
|
|||
################################################################################
|
||||
################################################################################
|
||||
################################################################################
|
||||
# TODO
|
||||
|
||||
def exist_item(item_id):
|
||||
return item_basic.exist_item(item_id)
|
||||
|
@ -455,12 +465,6 @@ def get_item_metadata(item_id, item_content=None):
|
|||
}
|
||||
return item_metadata
|
||||
|
||||
def get_item_parent(item_id):
|
||||
return item_basic.get_item_parent(item_id)
|
||||
|
||||
def add_item_parent(item_parent, item_id):
|
||||
return item_basic.add_item_parent(item_parent, item_id)
|
||||
|
||||
def get_item_content(item_id):
|
||||
return item_basic.get_item_content(item_id)
|
||||
|
||||
|
@ -646,6 +650,7 @@ def get_domain(item_id):
|
|||
item_id = item_id[-1]
|
||||
return item_id[:-36]
|
||||
|
||||
# TODO MOVE ME
|
||||
def get_item_har_name(item_id):
|
||||
har_path = os.path.join(har_directory, item_id) + '.json'
|
||||
if os.path.isfile(har_path):
|
||||
|
@ -653,9 +658,6 @@ def get_item_har_name(item_id):
|
|||
else:
|
||||
return None
|
||||
|
||||
def get_item_har(har_path):
|
||||
pass
|
||||
|
||||
def get_item_filename(item_id):
|
||||
# Creating the full filepath
|
||||
filename = os.path.join(ITEMS_FOLDER, item_id)
|
||||
|
@ -717,10 +719,6 @@ def create_item(obj_id, obj_metadata, io_content):
|
|||
# Item not created
|
||||
return False
|
||||
|
||||
# # TODO:
|
||||
def delete_item(obj_id):
|
||||
pass
|
||||
|
||||
# # check if item exists
|
||||
# if not exist_item(obj_id):
|
||||
# return False
|
||||
|
|
Loading…
Reference in a new issue