fix: get bigger chunks of variot data at a time
Some checks failed
Models Tests / Python 3.10 sample (push) Has been cancelled
Models Tests / Python 3.11 sample (push) Has been cancelled
Models Tests / Python 3.12 sample (push) Has been cancelled
CodeQL / Analyze (push) Has been cancelled
Deploy Jekyll with GitHub Pages dependencies preinstalled / build (push) Has been cancelled
MyPy / Python 3.10 sample (push) Has been cancelled
MyPy / Python 3.11 sample (push) Has been cancelled
MyPy / Python 3.12 sample (push) Has been cancelled
API Test / Python 3.10 sample (push) Has been cancelled
API Test / Python 3.11 sample (push) Has been cancelled
API Test / Python 3.12 sample (push) Has been cancelled
Deploy Jekyll with GitHub Pages dependencies preinstalled / deploy (push) Has been cancelled

This commit is contained in:
Raphaël Vinot 2024-10-15 13:28:14 +02:00
parent ed04c9a48d
commit e9e6fce05a

View file

@ -29,7 +29,7 @@ class VARIoTDB(AbstractFeeder):
import_complete: bool = True
p = self.storage.pipeline()
pyvariotids: dict[str, float] = {}
for vuln in self.client.get_vulnerabilities_iter(since=last_update):
for vuln in self.client.get_vulnerabilities_iter(since=last_update, limit=1000):
vuln_id = vuln['id'].lower()
last_modified = fromisoformat_wrapper(vuln['last_update_date'])
pyvariotids[vuln_id] = last_modified.timestamp()