Re-structuring on new model
This commit is contained in:
@@ -3,8 +3,9 @@ from utils import run_func
|
||||
from modules.nfproxy.models import Service, PyFilter
|
||||
import os
|
||||
import asyncio
|
||||
import socket
|
||||
import shutil
|
||||
from utils import DEBUG
|
||||
import traceback
|
||||
from fastapi import HTTPException
|
||||
|
||||
nft = FiregexTables()
|
||||
|
||||
@@ -13,22 +14,16 @@ class FiregexInterceptor:
|
||||
def __init__(self):
|
||||
self.srv:Service
|
||||
self._stats_updater_cb:callable
|
||||
self.filter_map_lock:asyncio.Lock
|
||||
self.filter_map: dict[str, PyFilter]
|
||||
self.pyfilters: set[PyFilter]
|
||||
self.update_config_lock:asyncio.Lock
|
||||
self.process:asyncio.subprocess.Process
|
||||
self.base_dir = os.path.join(
|
||||
os.path.dirname(os.path.abspath(__file__)),
|
||||
"socks", self.srv.id
|
||||
)
|
||||
self.n_threads = int(os.getenv("NTHREADS","1"))
|
||||
|
||||
self.connection_socket = os.path.join(self.base_dir, "connection.sock")
|
||||
self.vedict_sockets = [os.path.join(self.base_dir, f"vedict{i}.sock") for i in range(self.n_threads)]
|
||||
self.socks = []
|
||||
|
||||
def add_sock(self, path):
|
||||
sock = socket.socket(socket.AF_UNIX, socket.SOCK_DGRAM)
|
||||
sock.bind(path)
|
||||
self.socks.append(sock)
|
||||
return sock
|
||||
self.update_task: asyncio.Task
|
||||
self.ack_arrived = False
|
||||
self.ack_status = None
|
||||
self.ack_fail_what = ""
|
||||
self.ack_lock = asyncio.Lock()
|
||||
|
||||
async def _call_stats_updater_callback(self, filter: PyFilter):
|
||||
if self._stats_updater_cb:
|
||||
@@ -37,25 +32,24 @@ class FiregexInterceptor:
|
||||
@classmethod
|
||||
async def start(cls, srv: Service, stats_updater_cb:callable):
|
||||
self = cls()
|
||||
self.srv = srv
|
||||
self._stats_updater_cb = stats_updater_cb
|
||||
os.makedirs(self.base_dir, exist_ok=True)
|
||||
self.add_sock(self.connection_socket)
|
||||
for path in self.vedict_sockets:
|
||||
self.add_sock(path)
|
||||
self.srv = srv
|
||||
self.filter_map_lock = asyncio.Lock()
|
||||
self.update_config_lock = asyncio.Lock()
|
||||
queue_range = await self._start_binary()
|
||||
# TODO starts python workers
|
||||
self.update_task = asyncio.create_task(self.update_stats())
|
||||
nft.add(self.srv, queue_range)
|
||||
if not self.ack_lock.locked():
|
||||
await self.ack_lock.acquire()
|
||||
return self
|
||||
|
||||
async def _start_binary(self):
|
||||
proxy_binary_path = os.path.join(os.path.dirname(os.path.abspath(__file__)),"../cppproxy")
|
||||
proxy_binary_path = os.path.join(os.path.dirname(os.path.abspath(__file__)),"../cpproxy")
|
||||
self.process = await asyncio.create_subprocess_exec(
|
||||
proxy_binary_path,
|
||||
stdout=asyncio.subprocess.PIPE, stdin=asyncio.subprocess.PIPE,
|
||||
env={"NTHREADS": os.getenv("NTHREADS","1")},
|
||||
)
|
||||
self.process.stdin.write(self.base_dir.encode().hex().encode()+b" 3\n")
|
||||
await self.process.stdin.drain()
|
||||
line_fut = self.process.stdout.readuntil()
|
||||
try:
|
||||
line_fut = await asyncio.wait_for(line_fut, timeout=3)
|
||||
@@ -70,14 +64,58 @@ class FiregexInterceptor:
|
||||
self.process.kill()
|
||||
raise Exception("Invalid binary output")
|
||||
|
||||
async def update_stats(self):
|
||||
try:
|
||||
while True:
|
||||
line = (await self.process.stdout.readuntil()).decode()
|
||||
if DEBUG:
|
||||
print(line)
|
||||
if line.startswith("BLOCKED "):
|
||||
filter_id = line.split()[1]
|
||||
async with self.filter_map_lock:
|
||||
if filter_id in self.filter_map:
|
||||
self.filter_map[filter_id].blocked_packets+=1
|
||||
await self.filter_map[filter_id].update()
|
||||
if line.startswith("EDITED "):
|
||||
filter_id = line.split()[1]
|
||||
async with self.filter_map_lock:
|
||||
if filter_id in self.filter_map:
|
||||
self.filter_map[filter_id].edited_packets+=1
|
||||
await self.filter_map[filter_id].update()
|
||||
if line.startswith("ACK "):
|
||||
self.ack_arrived = True
|
||||
self.ack_status = line.split()[1].upper() == "OK"
|
||||
if not self.ack_status:
|
||||
self.ack_fail_what = " ".join(line.split()[2:])
|
||||
self.ack_lock.release()
|
||||
except asyncio.CancelledError:
|
||||
pass
|
||||
except asyncio.IncompleteReadError:
|
||||
pass
|
||||
except Exception:
|
||||
traceback.print_exc()
|
||||
|
||||
async def stop(self):
|
||||
self.update_task.cancel()
|
||||
if self.process and self.process.returncode is None:
|
||||
self.process.kill()
|
||||
for sock in self.socks:
|
||||
sock.close()
|
||||
shutil.rmtree(self.base_dir)
|
||||
|
||||
async def _update_config(self, filters_codes):
|
||||
async with self.update_config_lock:
|
||||
# TODO write compiled code correctly
|
||||
# self.process.stdin.write((" ".join(filters_codes)+"\n").encode())
|
||||
await self.process.stdin.drain()
|
||||
try:
|
||||
async with asyncio.timeout(3):
|
||||
await self.ack_lock.acquire()
|
||||
except TimeoutError:
|
||||
pass
|
||||
if not self.ack_arrived or not self.ack_status:
|
||||
raise HTTPException(status_code=500, detail=f"NFQ error: {self.ack_fail_what}")
|
||||
|
||||
async def reload(self, filters:list[PyFilter]):
|
||||
# filters are the functions to use in the workers (other functions are disabled or not flagged as filters)
|
||||
# TODO update filters in python workers (prob for new filters added) (reading from file????)
|
||||
pass
|
||||
async with self.filter_map_lock:
|
||||
self.filter_map = self.compile_filters(filters)
|
||||
# TODO COMPILE CODE
|
||||
#await self._update_config(filters_codes) TODO pass the compiled code
|
||||
|
||||
|
||||
Reference in New Issue
Block a user