Compare commits
1 Commits
pixeebot/d
...
pixeebot/d
Author | SHA1 | Date | |
---|---|---|---|
6c2da10eaa |
@ -1,149 +1,149 @@
|
||||
#!/usr/bin/env python3
|
||||
# -*- coding: utf-8 -*-
|
||||
|
||||
from app.utils import read_config
|
||||
from app.classes.iocs import IOCs
|
||||
from app.classes.whitelist import WhiteList
|
||||
from app.classes.misp import MISP
|
||||
|
||||
import requests
|
||||
import json
|
||||
import urllib3
|
||||
import time
|
||||
from multiprocessing import Process
|
||||
|
||||
"""
|
||||
This file is parsing the watchers present
|
||||
in the configuration file. This in order to get
|
||||
automatically new iocs / elements from remote
|
||||
sources without user interaction.
|
||||
"""
|
||||
|
||||
urllib3.disable_warnings(urllib3.exceptions.InsecureRequestWarning)
|
||||
|
||||
|
||||
def watch_iocs():
|
||||
"""
|
||||
Retrieve IOCs from the remote URLs defined in config/watchers.
|
||||
For each IOC, add it to the DB.
|
||||
"""
|
||||
|
||||
# Retrieve the URLs from the configuration
|
||||
urls = read_config(("watchers", "iocs"))
|
||||
watchers = [{"url": url, "status": False} for url in urls]
|
||||
|
||||
while True:
|
||||
for w in watchers:
|
||||
if w["status"] == False:
|
||||
iocs = IOCs()
|
||||
iocs_list = []
|
||||
to_delete = []
|
||||
try:
|
||||
res = requests.get(w["url"], verify=True)
|
||||
if res.status_code == 200:
|
||||
content = json.loads(res.content)
|
||||
iocs_list = content["iocs"] if "iocs" in content else []
|
||||
to_delete = content["to_delete"] if "to_delete" in content else []
|
||||
else:
|
||||
w["status"] = False
|
||||
except:
|
||||
w["status"] = False
|
||||
|
||||
for ioc in iocs_list:
|
||||
try:
|
||||
iocs.add(ioc["type"], ioc["tag"],
|
||||
ioc["tlp"], ioc["value"], "watcher")
|
||||
w["status"] = True
|
||||
except:
|
||||
continue
|
||||
|
||||
for ioc in to_delete:
|
||||
try:
|
||||
iocs.delete_by_value(ioc["value"])
|
||||
w["status"] = True
|
||||
except:
|
||||
continue
|
||||
|
||||
# If at least one URL haven't be parsed, let's retry in 1min.
|
||||
if False in [w["status"] for w in watchers]:
|
||||
time.sleep(60)
|
||||
else:
|
||||
break
|
||||
|
||||
|
||||
def watch_whitelists():
|
||||
"""
|
||||
Retrieve whitelist elements from the remote URLs
|
||||
defined in config/watchers. For each (new ?) element,
|
||||
add it to the DB.
|
||||
"""
|
||||
|
||||
urls = read_config(("watchers", "whitelists"))
|
||||
watchers = [{"url": url, "status": False} for url in urls]
|
||||
|
||||
while True:
|
||||
for w in watchers:
|
||||
if w["status"] == False:
|
||||
whitelist = WhiteList()
|
||||
elements = []
|
||||
to_delete = []
|
||||
try:
|
||||
res = requests.get(w["url"], verify=True)
|
||||
if res.status_code == 200:
|
||||
content = json.loads(res.content)
|
||||
elements = content["elements"] if "elements" in content else []
|
||||
to_delete = content["to_delete"] if "to_delete" in content else []
|
||||
else:
|
||||
w["status"] = False
|
||||
except:
|
||||
w["status"] = False
|
||||
|
||||
for elem in elements:
|
||||
try:
|
||||
whitelist.add(elem["type"], elem["element"], "watcher")
|
||||
w["status"] = True
|
||||
except:
|
||||
continue
|
||||
|
||||
for elem in to_delete:
|
||||
try:
|
||||
whitelist.delete_by_value(elem["element"])
|
||||
w["status"] = True
|
||||
except:
|
||||
continue
|
||||
|
||||
if False in [w["status"] for w in watchers]:
|
||||
time.sleep(60)
|
||||
else:
|
||||
break
|
||||
|
||||
|
||||
def watch_misp():
|
||||
"""
|
||||
Retrieve IOCs from misp instances. Each new element is
|
||||
tested and then added to the database.
|
||||
"""
|
||||
iocs, misp = IOCs(), MISP()
|
||||
instances = [i for i in misp.get_instances()]
|
||||
|
||||
while instances:
|
||||
for i, ist in enumerate(instances):
|
||||
status = misp.test_instance(ist["url"],
|
||||
ist["apikey"],
|
||||
ist["verifycert"])
|
||||
if status:
|
||||
for ioc in misp.get_iocs(ist["id"]):
|
||||
iocs.add(ioc["type"], ioc["tag"], ioc["tlp"],
|
||||
ioc["value"], "misp-{}".format(ist["id"]))
|
||||
misp.update_sync(ist["id"])
|
||||
instances.pop(i)
|
||||
if instances: time.sleep(60)
|
||||
|
||||
|
||||
p1 = Process(target=watch_iocs)
|
||||
p2 = Process(target=watch_whitelists)
|
||||
p3 = Process(target=watch_misp)
|
||||
|
||||
p1.start()
|
||||
p2.start()
|
||||
p3.start()
|
||||
#!/usr/bin/env python3
|
||||
# -*- coding: utf-8 -*-
|
||||
|
||||
from app.utils import read_config
|
||||
from app.classes.iocs import IOCs
|
||||
from app.classes.whitelist import WhiteList
|
||||
from app.classes.misp import MISP
|
||||
|
||||
import requests
|
||||
import json
|
||||
import urllib3
|
||||
import time
|
||||
from multiprocessing import Process
|
||||
|
||||
"""
|
||||
This file is parsing the watchers present
|
||||
in the configuration file. This in order to get
|
||||
automatically new iocs / elements from remote
|
||||
sources without user interaction.
|
||||
"""
|
||||
|
||||
urllib3.disable_warnings(urllib3.exceptions.InsecureRequestWarning)
|
||||
|
||||
|
||||
def watch_iocs():
|
||||
"""
|
||||
Retrieve IOCs from the remote URLs defined in config/watchers.
|
||||
For each IOC, add it to the DB.
|
||||
"""
|
||||
|
||||
# Retrieve the URLs from the configuration
|
||||
urls = read_config(("watchers", "iocs"))
|
||||
watchers = [{"url": url, "status": False} for url in urls]
|
||||
|
||||
while True:
|
||||
for w in watchers:
|
||||
if w["status"] == False:
|
||||
iocs = IOCs()
|
||||
iocs_list = []
|
||||
to_delete = []
|
||||
try:
|
||||
res = requests.get(w["url"], verify=False)
|
||||
if res.status_code == 200:
|
||||
content = json.loads(res.content)
|
||||
iocs_list = content["iocs"] if "iocs" in content else []
|
||||
to_delete = content["to_delete"] if "to_delete" in content else []
|
||||
else:
|
||||
w["status"] = False
|
||||
except:
|
||||
w["status"] = False
|
||||
|
||||
for ioc in iocs_list:
|
||||
try:
|
||||
iocs.add(ioc["type"], ioc["tag"],
|
||||
ioc["tlp"], ioc["value"], "watcher")
|
||||
w["status"] = True
|
||||
except:
|
||||
continue
|
||||
|
||||
for ioc in to_delete:
|
||||
try:
|
||||
iocs.delete_by_value(ioc["value"])
|
||||
w["status"] = True
|
||||
except:
|
||||
continue
|
||||
|
||||
# If at least one URL haven't be parsed, let's retry in 1min.
|
||||
if False in [w["status"] for w in watchers]:
|
||||
time.sleep(60)
|
||||
else:
|
||||
break
|
||||
|
||||
|
||||
def watch_whitelists():
|
||||
"""
|
||||
Retrieve whitelist elements from the remote URLs
|
||||
defined in config/watchers. For each (new ?) element,
|
||||
add it to the DB.
|
||||
"""
|
||||
|
||||
urls = read_config(("watchers", "whitelists"))
|
||||
watchers = [{"url": url, "status": False} for url in urls]
|
||||
|
||||
while True:
|
||||
for w in watchers:
|
||||
if w["status"] == False:
|
||||
whitelist = WhiteList()
|
||||
elements = []
|
||||
to_delete = []
|
||||
try:
|
||||
res = requests.get(w["url"], verify=False)
|
||||
if res.status_code == 200:
|
||||
content = json.loads(res.content)
|
||||
elements = content["elements"] if "elements" in content else []
|
||||
to_delete = content["to_delete"] if "to_delete" in content else []
|
||||
else:
|
||||
w["status"] = False
|
||||
except:
|
||||
w["status"] = False
|
||||
|
||||
for elem in elements:
|
||||
try:
|
||||
whitelist.add(elem["type"], elem["element"], "watcher")
|
||||
w["status"] = True
|
||||
except:
|
||||
continue
|
||||
|
||||
for elem in to_delete:
|
||||
try:
|
||||
whitelist.delete_by_value(elem["element"])
|
||||
w["status"] = True
|
||||
except:
|
||||
continue
|
||||
|
||||
if False in [w["status"] for w in watchers]:
|
||||
time.sleep(60)
|
||||
else:
|
||||
break
|
||||
|
||||
|
||||
def watch_misp():
|
||||
"""
|
||||
Retrieve IOCs from misp instances. Each new element is
|
||||
tested and then added to the database.
|
||||
"""
|
||||
iocs, misp = IOCs(), MISP()
|
||||
instances = [i for i in misp.get_instances()]
|
||||
|
||||
while instances:
|
||||
for i, ist in enumerate(instances):
|
||||
status = misp.test_instance(ist["url"],
|
||||
ist["apikey"],
|
||||
ist["verifycert"])
|
||||
if status:
|
||||
for ioc in misp.get_iocs(ist["id"]):
|
||||
iocs.add(ioc["type"], ioc["tag"], ioc["tlp"],
|
||||
ioc["value"], "misp-{}".format(ist["id"]))
|
||||
misp.update_sync(ist["id"])
|
||||
instances.pop(i)
|
||||
if instances: time.sleep(60)
|
||||
|
||||
|
||||
p1 = Process(target=watch_iocs)
|
||||
p2 = Process(target=watch_whitelists)
|
||||
p3 = Process(target=watch_misp)
|
||||
|
||||
p1.start()
|
||||
p2.start()
|
||||
p3.start()
|
||||
|
@ -1,128 +1,128 @@
|
||||
#!/usr/bin/env python3
|
||||
# -*- coding: utf-8 -*-
|
||||
|
||||
import subprocess as sp
|
||||
from app.utils import terminate_process, read_config
|
||||
from os import mkdir, path
|
||||
from flask import send_file, jsonify
|
||||
import datetime
|
||||
import shutil
|
||||
import json
|
||||
import random
|
||||
import sys
|
||||
import re
|
||||
|
||||
|
||||
class Capture(object):
|
||||
|
||||
def __init__(self):
|
||||
self.random_choice_alphabet = "ABCDEF1234567890"
|
||||
|
||||
def start_capture(self):
|
||||
"""
|
||||
Start a tshark capture on the created AP interface and save
|
||||
it in a temporary directory under /tmp/.
|
||||
|
||||
:return: dict containing capture token and status.
|
||||
"""
|
||||
|
||||
# Kill potential tshark zombies instances, if any.
|
||||
terminate_process("tshark")
|
||||
|
||||
# Few context variable assignment
|
||||
self.capture_token = "".join(
|
||||
[random.choice(self.random_choice_alphabet) for i in range(8)])
|
||||
self.capture_dir = "/tmp/{}/".format(self.capture_token)
|
||||
self.assets_dir = "/tmp/{}/assets/".format(self.capture_token)
|
||||
self.pcap = self.capture_dir + "capture.pcap"
|
||||
self.iface = read_config(("network", "in"))
|
||||
|
||||
# For packets monitoring
|
||||
self.list_pkts = []
|
||||
self.last_pkts = 0
|
||||
|
||||
# Make the capture and the assets directory
|
||||
mkdir(self.capture_dir)
|
||||
mkdir(self.assets_dir)
|
||||
|
||||
try:
|
||||
sp.Popen(["tshark", "-i", self.iface, "-w",
|
||||
self.pcap, "-f", "tcp or udp"])
|
||||
return {"status": True,
|
||||
"message": "Capture started",
|
||||
"capture_token": self.capture_token}
|
||||
except:
|
||||
return {"status": False,
|
||||
"message": "Unexpected error: %s" % sys.exc_info()[0]}
|
||||
|
||||
def get_capture_stats(self):
|
||||
"""
|
||||
Get some dirty capture statistics in order to have a sparkline
|
||||
in the background of capture view.
|
||||
|
||||
:return: dict containing stats associated to the capture
|
||||
"""
|
||||
with open("/sys/class/net/{}/statistics/tx_packets".format(self.iface)) as f:
|
||||
tx_pkts = int(f.read())
|
||||
with open("/sys/class/net/{}/statistics/rx_packets".format(self.iface)) as f:
|
||||
rx_pkts = int(f.read())
|
||||
|
||||
if self.last_pkts == 0:
|
||||
self.last_pkts = tx_pkts + rx_pkts
|
||||
return {"status": True,
|
||||
"packets": [0*400]}
|
||||
else:
|
||||
curr_pkts = (tx_pkts + rx_pkts) - self.last_pkts
|
||||
self.last_pkts = tx_pkts + rx_pkts
|
||||
self.list_pkts.append(curr_pkts)
|
||||
return {"status": True,
|
||||
"packets": self.beautify_stats(self.list_pkts)}
|
||||
|
||||
@staticmethod
|
||||
def beautify_stats(data):
|
||||
"""
|
||||
Add 0 at the end of the array if the len of the array is less
|
||||
than max_len. Else, get the last 100 stats. This allows to
|
||||
show a kind of "progressive chart" in the background for
|
||||
the first packets.
|
||||
|
||||
:return: a list of integers.
|
||||
"""
|
||||
max_len = 400
|
||||
if len(data) >= max_len:
|
||||
return data[-max_len:]
|
||||
else:
|
||||
return data + [1] * (max_len - len(data))
|
||||
|
||||
def stop_capture(self):
|
||||
"""
|
||||
Stop tshark if any instance present & ask create_capinfos.
|
||||
:return: dict as a small confirmation.
|
||||
"""
|
||||
if terminate_process("tshark"):
|
||||
self.create_capinfos()
|
||||
return {"status": True,
|
||||
"message": "Capture stopped"}
|
||||
else:
|
||||
return {"status": False,
|
||||
"message": "No active capture"}
|
||||
|
||||
def create_capinfos(self):
|
||||
"""
|
||||
Creates a capinfo json file.
|
||||
:return: dict as a small confirmation.
|
||||
"""
|
||||
infos = sp.Popen(["capinfos", self.pcap],
|
||||
stdout=sp.PIPE, stderr=sp.PIPE)
|
||||
infos = infos.communicate()[0]
|
||||
data = {}
|
||||
for l in infos.decode().splitlines():
|
||||
try:
|
||||
l = l.split(": ") if ": " in l else l.split("= ")
|
||||
if len(l[0]) and len(l[1]):
|
||||
data[l[0].strip()] = l[1].strip()
|
||||
except:
|
||||
continue
|
||||
with open("{}capinfos.json".format(self.assets_dir), 'w') as f:
|
||||
json.dump(data, f)
|
||||
return True
|
||||
#!/usr/bin/env python3
|
||||
# -*- coding: utf-8 -*-
|
||||
|
||||
import subprocess as sp
|
||||
from app.utils import terminate_process, read_config
|
||||
from os import mkdir, path
|
||||
from flask import send_file, jsonify
|
||||
import datetime
|
||||
import shutil
|
||||
import json
|
||||
import sys
|
||||
import re
|
||||
import secrets
|
||||
|
||||
|
||||
class Capture(object):
|
||||
|
||||
def __init__(self):
|
||||
self.random_choice_alphabet = "ABCDEF1234567890"
|
||||
|
||||
def start_capture(self):
|
||||
"""
|
||||
Start a tshark capture on the created AP interface and save
|
||||
it in a temporary directory under /tmp/.
|
||||
|
||||
:return: dict containing capture token and status.
|
||||
"""
|
||||
|
||||
# Kill potential tshark zombies instances, if any.
|
||||
terminate_process("tshark")
|
||||
|
||||
# Few context variable assignment
|
||||
self.capture_token = "".join(
|
||||
[secrets.choice(self.random_choice_alphabet) for i in range(8)])
|
||||
self.capture_dir = "/tmp/{}/".format(self.capture_token)
|
||||
self.assets_dir = "/tmp/{}/assets/".format(self.capture_token)
|
||||
self.pcap = self.capture_dir + "capture.pcap"
|
||||
self.iface = read_config(("network", "in"))
|
||||
|
||||
# For packets monitoring
|
||||
self.list_pkts = []
|
||||
self.last_pkts = 0
|
||||
|
||||
# Make the capture and the assets directory
|
||||
mkdir(self.capture_dir)
|
||||
mkdir(self.assets_dir)
|
||||
|
||||
try:
|
||||
sp.Popen(["tshark", "-i", self.iface, "-w",
|
||||
self.pcap, "-f", "tcp or udp"])
|
||||
return {"status": True,
|
||||
"message": "Capture started",
|
||||
"capture_token": self.capture_token}
|
||||
except:
|
||||
return {"status": False,
|
||||
"message": "Unexpected error: %s" % sys.exc_info()[0]}
|
||||
|
||||
def get_capture_stats(self):
|
||||
"""
|
||||
Get some dirty capture statistics in order to have a sparkline
|
||||
in the background of capture view.
|
||||
|
||||
:return: dict containing stats associated to the capture
|
||||
"""
|
||||
with open("/sys/class/net/{}/statistics/tx_packets".format(self.iface)) as f:
|
||||
tx_pkts = int(f.read())
|
||||
with open("/sys/class/net/{}/statistics/rx_packets".format(self.iface)) as f:
|
||||
rx_pkts = int(f.read())
|
||||
|
||||
if self.last_pkts == 0:
|
||||
self.last_pkts = tx_pkts + rx_pkts
|
||||
return {"status": True,
|
||||
"packets": [0*400]}
|
||||
else:
|
||||
curr_pkts = (tx_pkts + rx_pkts) - self.last_pkts
|
||||
self.last_pkts = tx_pkts + rx_pkts
|
||||
self.list_pkts.append(curr_pkts)
|
||||
return {"status": True,
|
||||
"packets": self.beautify_stats(self.list_pkts)}
|
||||
|
||||
@staticmethod
|
||||
def beautify_stats(data):
|
||||
"""
|
||||
Add 0 at the end of the array if the len of the array is less
|
||||
than max_len. Else, get the last 100 stats. This allows to
|
||||
show a kind of "progressive chart" in the background for
|
||||
the first packets.
|
||||
|
||||
:return: a list of integers.
|
||||
"""
|
||||
max_len = 400
|
||||
if len(data) >= max_len:
|
||||
return data[-max_len:]
|
||||
else:
|
||||
return data + [1] * (max_len - len(data))
|
||||
|
||||
def stop_capture(self):
|
||||
"""
|
||||
Stop tshark if any instance present & ask create_capinfos.
|
||||
:return: dict as a small confirmation.
|
||||
"""
|
||||
if terminate_process("tshark"):
|
||||
self.create_capinfos()
|
||||
return {"status": True,
|
||||
"message": "Capture stopped"}
|
||||
else:
|
||||
return {"status": False,
|
||||
"message": "No active capture"}
|
||||
|
||||
def create_capinfos(self):
|
||||
"""
|
||||
Creates a capinfo json file.
|
||||
:return: dict as a small confirmation.
|
||||
"""
|
||||
infos = sp.Popen(["capinfos", self.pcap],
|
||||
stdout=sp.PIPE, stderr=sp.PIPE)
|
||||
infos = infos.communicate()[0]
|
||||
data = {}
|
||||
for l in infos.decode().splitlines():
|
||||
try:
|
||||
l = l.split(": ") if ": " in l else l.split("= ")
|
||||
if len(l[0]) and len(l[1]):
|
||||
data[l[0].strip()] = l[1].strip()
|
||||
except:
|
||||
continue
|
||||
with open("{}capinfos.json".format(self.assets_dir), 'w') as f:
|
||||
json.dump(data, f)
|
||||
return True
|
||||
|
@ -9,13 +9,13 @@ import sys
|
||||
import time
|
||||
import qrcode
|
||||
import base64
|
||||
import random
|
||||
import requests
|
||||
|
||||
from wifi import Cell
|
||||
from os import path, remove
|
||||
from io import BytesIO
|
||||
from app.utils import terminate_process, read_config
|
||||
import secrets
|
||||
|
||||
|
||||
class Network(object):
|
||||
@ -170,14 +170,14 @@ class Network(object):
|
||||
|
||||
# Generate the hostapd configuration
|
||||
if read_config(("network", "tokenized_ssids")):
|
||||
token = "".join([random.choice(self.random_choice_alphabet)
|
||||
token = "".join([secrets.choice(self.random_choice_alphabet)
|
||||
for i in range(4)])
|
||||
self.AP_SSID = random.choice(read_config(
|
||||
self.AP_SSID = secrets.choice(read_config(
|
||||
("network", "ssids"))) + "-" + token
|
||||
else:
|
||||
self.AP_SSID = random.choice(read_config(("network", "ssids")))
|
||||
self.AP_SSID = secrets.choice(read_config(("network", "ssids")))
|
||||
self.AP_PASS = "".join(
|
||||
[random.choice(self.random_choice_alphabet) for i in range(8)])
|
||||
[secrets.choice(self.random_choice_alphabet) for i in range(8)])
|
||||
|
||||
# Launch hostapd
|
||||
if self.write_hostapd_config():
|
||||
|
Reference in New Issue
Block a user