diff --git a/front/php/templates/language/merge_translations.py b/front/php/templates/language/merge_translations.py
index d985ccb3..ee35575d 100755
--- a/front/php/templates/language/merge_translations.py
+++ b/front/php/templates/language/merge_translations.py
@@ -1,6 +1,6 @@
import json
import os
-import sys
+
def merge_translations(main_file, other_files):
# Load main file
@@ -30,10 +30,14 @@ def merge_translations(main_file, other_files):
json.dump(data, f, indent=4, ensure_ascii=False)
f.truncate()
+
if __name__ == "__main__":
current_path = os.path.dirname(os.path.abspath(__file__))
# language codes can be found here: http://www.lingoes.net/en/translator/langcode.htm
- # "en_us.json" has to be first!
- json_files = [ "en_us.json", "ar_ar.json", "ca_ca.json", "cs_cz.json", "de_de.json", "es_es.json", "fa_fa.json", "fr_fr.json", "it_it.json", "ja_jp.json", "nb_no.json", "pl_pl.json", "pt_br.json", "pt_pt.json", "ru_ru.json", "sv_sv.json", "tr_tr.json", "uk_ua.json", "zh_cn.json"]
+ # ⚠ "en_us.json" has to be first!
+ json_files = ["en_us.json", "ar_ar.json", "ca_ca.json", "cs_cz.json", "de_de.json",
+ "es_es.json", "fa_fa.json", "fr_fr.json", "it_it.json", "ja_jp.json",
+ "nb_no.json", "pl_pl.json", "pt_br.json", "pt_pt.json", "ru_ru.json",
+ "sv_sv.json", "tr_tr.json", "uk_ua.json", "zh_cn.json"]
file_paths = [os.path.join(current_path, file) for file in json_files]
merge_translations(file_paths[0], file_paths[1:])
diff --git a/front/plugins/__template/rename_me.py b/front/plugins/__template/rename_me.py
index 6530846b..6fe0bd61 100755
--- a/front/plugins/__template/rename_me.py
+++ b/front/plugins/__template/rename_me.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import sys
@@ -8,12 +8,12 @@ from pytz import timezone
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from const import logPath
-from plugin_helper import Plugin_Objects
-from logger import mylog, Logger
-from helper import get_setting_value
+from const import logPath # noqa: E402, E261 [flake8 lint suppression]
+from plugin_helper import Plugin_Objects # noqa: E402, E261 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402, E261 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402, E261 [flake8 lint suppression]
-import conf
+import conf # noqa: E402, E261 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -32,9 +32,8 @@ RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
plugin_objects = Plugin_Objects(RESULT_FILE)
-
def main():
- mylog('verbose', [f'[{pluginName}] In script'])
+ mylog('verbose', [f'[{pluginName}] In script'])
# Retrieve configuration settings
some_setting = get_setting_value('SYNC_plugins')
@@ -47,14 +46,14 @@ def main():
# Process the data into native application tables
if len(device_data) > 0:
- # insert devices into the lats_result.log
- # make sure the below mapping is mapped in config.json, for example:
+ # insert devices into the lats_result.log
+ # make sure the below mapping is mapped in config.json, for example:
# "database_column_definitions": [
# {
# "column": "Object_PrimaryID", <--------- the value I save into primaryId
# "mapped_to_column": "cur_MAC", <--------- gets inserted into the CurrentScan DB
# table column cur_MAC
- #
+ #
for device in device_data:
plugin_objects.add_object(
primaryId = device['mac_address'],
@@ -65,11 +64,11 @@ def main():
watched4 = device['last_seen'],
extra = '',
foreignKey = device['mac_address']
- # helpVal1 = "Something1", # Optional Helper values to be passed for mapping into the app
- # helpVal2 = "Something1", # If you need to use even only 1, add the remaining ones too
+ # helpVal1 = "Something1", # Optional Helper values to be passed for mapping into the app
+ # helpVal2 = "Something1", # If you need to use even only 1, add the remaining ones too
# helpVal3 = "Something1", # and set them to 'null'. Check the the docs for details:
# helpVal4 = "Something1", # https://github.com/jokob-sk/NetAlertX/blob/main/docs/PLUGINS_DEV.md
- )
+ )
mylog('verbose', [f'[{pluginName}] New entries: "{len(device_data)}"'])
@@ -78,14 +77,15 @@ def main():
return 0
+
# retrieve data
def get_device_data(some_setting):
-
+
device_data = []
# do some processing, call exteranl APIs, and return a device_data list
# ...
- #
+ #
# Sample data for testing purposes, you can adjust the processing in main() as needed
# ... before adding it to the plugin_objects.add_object(...)
device_data = [
@@ -113,8 +113,9 @@ def get_device_data(some_setting):
}
]
- # Return the data to be detected by the main application
+ # Return the data to be detected by the main application
return device_data
+
if __name__ == '__main__':
main()
diff --git a/front/plugins/__test/test.py b/front/plugins/__test/test.py
index 9639f346..555a6bd6 100755
--- a/front/plugins/__test/test.py
+++ b/front/plugins/__test/test.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
# Just a testing library plugin for development purposes
import os
import sys
@@ -11,10 +11,10 @@ INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
# NetAlertX modules
-from const import logPath
-from plugin_helper import Plugin_Objects
-from logger import mylog
-from helper import get_setting_value
+from const import logPath # noqa: E402 [flake8 lint suppression]
+from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
+from logger import mylog # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
pluginName = 'TESTONLY'
@@ -28,14 +28,11 @@ plugin_objects = Plugin_Objects(RESULT_FILE)
md5_hash = hashlib.md5()
-
# globals
-
-
def main():
# START
- mylog('verbose', [f'[{pluginName}] In script'])
-
+ mylog('verbose', [f'[{pluginName}] In script'])
+
# SPACE FOR TESTING 🔽
str = "ABC-MBP._another.localdomain."
@@ -43,28 +40,23 @@ def main():
# result = cleanDeviceName(str, True)
regexes = get_setting_value('NEWDEV_NAME_CLEANUP_REGEX')
-
print(regexes)
subnets = get_setting_value('SCAN_SUBNETS')
-
+
print(subnets)
- for rgx in regexes:
+ for rgx in regexes:
mylog('trace', ["[cleanDeviceName] applying regex : " + rgx])
mylog('trace', ["[cleanDeviceName] name before regex : " + str])
-
+
str = re.sub(rgx, "", str)
mylog('trace', ["[cleanDeviceName] name after regex : " + str])
mylog('debug', ["[cleanDeviceName] output: " + str])
-
-
# SPACE FOR TESTING 🔼
# END
- mylog('verbose', [f'[{pluginName}] result "{str}"'])
-
-
+ mylog('verbose', [f'[{pluginName}] result "{str}"'])
# -------------INIT---------------------
diff --git a/front/plugins/_publisher_apprise/apprise.py b/front/plugins/_publisher_apprise/apprise.py
index 7119133f..f84c069d 100755
--- a/front/plugins/_publisher_apprise/apprise.py
+++ b/front/plugins/_publisher_apprise/apprise.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import json
import subprocess
@@ -9,15 +9,15 @@ import sys
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-import conf
-from const import confFileName, logPath
-from utils.datetime_utils import timeNowDB
-from plugin_helper import Plugin_Objects
-from logger import mylog, Logger
-from helper import get_setting_value
-from models.notification_instance import NotificationInstance
-from database import DB
-from pytz import timezone
+import conf # noqa: E402 [flake8 lint suppression]
+from const import confFileName, logPath # noqa: E402 [flake8 lint suppression]
+from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
+from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from models.notification_instance import NotificationInstance # noqa: E402 [flake8 lint suppression]
+from database import DB # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value("TIMEZONE"))
@@ -35,7 +35,7 @@ def main():
mylog("verbose", [f"[{pluginName}](publisher) In script"])
# Check if basic config settings supplied
- if check_config() == False:
+ if check_config() is False:
mylog(
"none",
[
@@ -65,9 +65,9 @@ def main():
# Log result
plugin_objects.add_object(
primaryId = pluginName,
- secondaryId = timeNowDB(),
+ secondaryId = timeNowDB(),
watched1 = notification["GUID"],
- watched2 = result,
+ watched2 = result,
watched3 = 'null',
watched4 = 'null',
extra = 'null',
@@ -80,8 +80,7 @@ def main():
# -------------------------------------------------------------------------------
def check_config():
if get_setting_value("APPRISE_HOST") == "" or (
- get_setting_value("APPRISE_URL") == ""
- and get_setting_value("APPRISE_TAG") == ""
+ get_setting_value("APPRISE_URL") == "" and get_setting_value("APPRISE_TAG") == ""
):
return False
else:
diff --git a/front/plugins/_publisher_email/email_smtp.py b/front/plugins/_publisher_email/email_smtp.py
index c0017d8c..df18cb6a 100755
--- a/front/plugins/_publisher_email/email_smtp.py
+++ b/front/plugins/_publisher_email/email_smtp.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import sys
import re
@@ -16,15 +16,15 @@ INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
# NetAlertX modules
-import conf
-from const import confFileName, logPath
-from plugin_helper import Plugin_Objects
-from utils.datetime_utils import timeNowDB
-from logger import mylog, Logger
-from helper import get_setting_value, hide_email
-from models.notification_instance import NotificationInstance
-from database import DB
-from pytz import timezone
+import conf # noqa: E402 [flake8 lint suppression]
+from const import confFileName, logPath # noqa: E402 [flake8 lint suppression]
+from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
+from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value, hide_email # noqa: E402 [flake8 lint suppression]
+from models.notification_instance import NotificationInstance # noqa: E402 [flake8 lint suppression]
+from database import DB # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -38,13 +38,12 @@ LOG_PATH = logPath + '/plugins'
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
-
def main():
-
- mylog('verbose', [f'[{pluginName}](publisher) In script'])
-
+
+ mylog('verbose', [f'[{pluginName}](publisher) In script'])
+
# Check if basic config settings supplied
- if check_config() == False:
+ if check_config() is False:
mylog('none', [f'[{pluginName}] ⚠ ERROR: Publisher notification gateway not set up correctly. Check your {confFileName} {pluginName}_* variables.'])
return
@@ -61,7 +60,7 @@ def main():
# Retrieve new notifications
new_notifications = notifications.getNew()
- # mylog('verbose', [f'[{pluginName}] new_notifications: ', new_notifications])
+ # mylog('verbose', [f'[{pluginName}] new_notifications: ', new_notifications])
mylog('verbose', [f'[{pluginName}] SMTP_SERVER: ', get_setting_value("SMTP_SERVER")])
mylog('verbose', [f'[{pluginName}] SMTP_PORT: ', get_setting_value("SMTP_PORT")])
mylog('verbose', [f'[{pluginName}] SMTP_SKIP_LOGIN: ', get_setting_value("SMTP_SKIP_LOGIN")])
@@ -72,19 +71,18 @@ def main():
# mylog('verbose', [f'[{pluginName}] SMTP_REPORT_TO: ', get_setting_value("SMTP_REPORT_TO")])
# mylog('verbose', [f'[{pluginName}] SMTP_REPORT_FROM: ', get_setting_value("SMTP_REPORT_FROM")])
-
# Process the new notifications (see the Notifications DB table for structure or check the /php/server/query_json.php?file=table_notifications.json endpoint)
for notification in new_notifications:
# Send notification
- result = send(notification["HTML"], notification["Text"])
+ result = send(notification["HTML"], notification["Text"])
# Log result
plugin_objects.add_object(
primaryId = pluginName,
- secondaryId = timeNowDB(),
+ secondaryId = timeNowDB(),
watched1 = notification["GUID"],
- watched2 = result,
+ watched2 = result,
watched3 = 'null',
watched4 = 'null',
extra = 'null',
@@ -93,25 +91,33 @@ def main():
plugin_objects.write_result_file()
-#-------------------------------------------------------------------------------
-def check_config ():
+
+# -------------------------------------------------------------------------------
+def check_config():
server = get_setting_value('SMTP_SERVER')
report_to = get_setting_value("SMTP_REPORT_TO")
report_from = get_setting_value("SMTP_REPORT_FROM")
-
+
if server == '' or report_from == '' or report_to == '':
mylog('none', [f'[Email Check Config] ⚠ ERROR: Email service not set up correctly. Check your {confFileName} SMTP_*, SMTP_REPORT_FROM and SMTP_REPORT_TO variables.'])
return False
else:
return True
-
-#-------------------------------------------------------------------------------
+
+
+# -------------------------------------------------------------------------------
def send(pHTML, pText):
mylog('debug', [f'[{pluginName}] SMTP_REPORT_TO: {hide_email(str(get_setting_value("SMTP_REPORT_TO")))} SMTP_USER: {hide_email(str(get_setting_value("SMTP_USER")))}'])
- subject, from_email, to_email, message_html, message_text = sanitize_email_content(str(get_setting_value("SMTP_SUBJECT")), get_setting_value("SMTP_REPORT_FROM"), get_setting_value("SMTP_REPORT_TO"), pHTML, pText)
+ subject, from_email, to_email, message_html, message_text = sanitize_email_content(
+ str(get_setting_value("SMTP_SUBJECT")),
+ get_setting_value("SMTP_REPORT_FROM"),
+ get_setting_value("SMTP_REPORT_TO"),
+ pHTML,
+ pText
+ )
emails = []
@@ -132,10 +138,10 @@ def send(pHTML, pText):
msg['Subject'] = subject
msg['From'] = from_email
msg['To'] = mail_addr
- msg['Date'] = formatdate(localtime=True)
+ msg['Date'] = formatdate(localtime=True)
- msg.attach (MIMEText (message_text, 'plain'))
- msg.attach (MIMEText (message_html, 'html'))
+ msg.attach(MIMEText(message_text, 'plain'))
+ msg.attach(MIMEText(message_html, 'html'))
# Set a timeout for the SMTP connection (in seconds)
smtp_timeout = 30
@@ -144,30 +150,31 @@ def send(pHTML, pText):
if get_setting_value("LOG_LEVEL") == 'debug':
- send_email(msg,smtp_timeout)
+ send_email(msg, smtp_timeout)
else:
try:
- send_email(msg,smtp_timeout)
-
- except smtplib.SMTPAuthenticationError as e:
+ send_email(msg, smtp_timeout)
+
+ except smtplib.SMTPAuthenticationError as e:
mylog('none', [' ERROR: Couldn\'t connect to the SMTP server (SMTPAuthenticationError)'])
mylog('none', [' ERROR: Double-check your SMTP_USER and SMTP_PASS settings.)'])
mylog('none', [' ERROR: ', str(e)])
- except smtplib.SMTPServerDisconnected as e:
+ except smtplib.SMTPServerDisconnected as e:
mylog('none', [' ERROR: Couldn\'t connect to the SMTP server (SMTPServerDisconnected)'])
mylog('none', [' ERROR: ', str(e)])
- except socket.gaierror as e:
+ except socket.gaierror as e:
mylog('none', [' ERROR: Could not resolve hostname (socket.gaierror)'])
- mylog('none', [' ERROR: ', str(e)])
- except ssl.SSLError as e:
+ mylog('none', [' ERROR: ', str(e)])
+ except ssl.SSLError as e:
mylog('none', [' ERROR: Could not establish SSL connection (ssl.SSLError)'])
mylog('none', [' ERROR: Are you sure you need SMTP_FORCE_SSL enabled? Check your SMTP provider docs.'])
- mylog('none', [' ERROR: ', str(e)])
+ mylog('none', [' ERROR: ', str(e)])
+
# ----------------------------------------------------------------------------------
-def send_email(msg,smtp_timeout):
+def send_email(msg, smtp_timeout):
# Send mail
if get_setting_value('SMTP_FORCE_SSL'):
mylog('debug', ['SMTP_FORCE_SSL == True so using .SMTP_SSL()'])
@@ -182,10 +189,10 @@ def send_email(msg,smtp_timeout):
mylog('debug', ['SMTP_FORCE_SSL == False so using .SMTP()'])
if get_setting_value("SMTP_PORT") == 0:
mylog('debug', ['SMTP_PORT == 0 so sending .SMTP(SMTP_SERVER)'])
- smtp_connection = smtplib.SMTP (get_setting_value('SMTP_SERVER'))
+ smtp_connection = smtplib.SMTP(get_setting_value('SMTP_SERVER'))
else:
mylog('debug', ['SMTP_PORT == 0 so sending .SMTP(SMTP_SERVER, SMTP_PORT)'])
- smtp_connection = smtplib.SMTP (get_setting_value('SMTP_SERVER'), get_setting_value('SMTP_PORT'))
+ smtp_connection = smtplib.SMTP(get_setting_value('SMTP_SERVER'), get_setting_value('SMTP_PORT'))
mylog('debug', ['Setting SMTP debug level'])
@@ -193,7 +200,7 @@ def send_email(msg,smtp_timeout):
if get_setting_value('LOG_LEVEL') == 'debug':
smtp_connection.set_debuglevel(1)
- mylog('debug', [ 'Sending .ehlo()'])
+ mylog('debug', ['Sending .ehlo()'])
smtp_connection.ehlo()
if not get_setting_value('SMTP_SKIP_TLS'):
@@ -203,12 +210,13 @@ def send_email(msg,smtp_timeout):
smtp_connection.ehlo()
if not get_setting_value('SMTP_SKIP_LOGIN'):
mylog('debug', ['SMTP_SKIP_LOGIN == False so sending .login()'])
- smtp_connection.login (get_setting_value('SMTP_USER'), get_setting_value('SMTP_PASS'))
+ smtp_connection.login(get_setting_value('SMTP_USER'), get_setting_value('SMTP_PASS'))
mylog('debug', ['Sending .sendmail()'])
- smtp_connection.sendmail (get_setting_value("SMTP_REPORT_FROM"), get_setting_value("SMTP_REPORT_TO"), msg.as_string())
+ smtp_connection.sendmail(get_setting_value("SMTP_REPORT_FROM"), get_setting_value("SMTP_REPORT_TO"), msg.as_string())
smtp_connection.quit()
+
# ----------------------------------------------------------------------------------
def sanitize_email_content(subject, from_email, to_email, message_html, message_text):
# Validate and sanitize subject
@@ -229,6 +237,7 @@ def sanitize_email_content(subject, from_email, to_email, message_html, message_
return subject, from_email, to_email, message_html, message_text
+
# ----------------------------------------------------------------------------------
if __name__ == '__main__':
sys.exit(main())
diff --git a/front/plugins/_publisher_mqtt/mqtt.py b/front/plugins/_publisher_mqtt/mqtt.py
index aeb25ec0..9d7a6ee8 100755
--- a/front/plugins/_publisher_mqtt/mqtt.py
+++ b/front/plugins/_publisher_mqtt/mqtt.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import json
import os
@@ -18,15 +18,14 @@ INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
# NetAlertX modules
-import conf
-from const import confFileName, logPath
-from utils.plugin_utils import getPluginObject
-from plugin_helper import Plugin_Objects
-from logger import mylog, Logger
+import conf # noqa: E402 [flake8 lint suppression]
+from const import confFileName, logPath # noqa: E402 [flake8 lint suppression]
+from utils.plugin_utils import getPluginObject # noqa: E402 [flake8 lint suppression]
+from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
from helper import get_setting_value, bytes_to_string, \
- sanitize_string, normalize_string
-from utils.datetime_utils import timeNowDB
-from database import DB, get_device_stats
+ sanitize_string, normalize_string # noqa: E402 [flake8 lint suppression]
+from database import DB, get_device_stats # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
@@ -287,11 +286,11 @@ def publish_mqtt(mqtt_client, topic, message):
# mylog('verbose', [f"[{pluginName}] mqtt_client.is_connected(): {mqtt_client.is_connected()} "])
result = mqtt_client.publish(
- topic=topic,
- payload=message,
- qos=qos,
- retain=True,
- )
+ topic=topic,
+ payload=message,
+ qos=qos,
+ retain=True,
+ )
status = result[0]
@@ -303,6 +302,7 @@ def publish_mqtt(mqtt_client, topic, message):
time.sleep(0.1)
return True
+
# ------------------------------------------------------------------------------
# Create a generic device for overal stats
def create_generic_device(mqtt_client, deviceId, deviceName):
@@ -434,7 +434,6 @@ def mqtt_start(db):
if not mqtt_connected_to_broker:
mqtt_client = mqtt_create_client()
-
deviceName = get_setting_value('MQTT_DEVICE_NAME')
deviceId = get_setting_value('MQTT_DEVICE_ID')
@@ -449,16 +448,18 @@ def mqtt_start(db):
row = get_device_stats(db)
# Publish (wrap into {} and remove last ',' from above)
- publish_mqtt(mqtt_client, f"{topic_root}/sensor/{deviceId}/state",
- {
- "online": row[0],
- "down": row[1],
- "all": row[2],
- "archived": row[3],
- "new": row[4],
- "unknown": row[5]
- }
- )
+ publish_mqtt(
+ mqtt_client,
+ f"{topic_root}/sensor/{deviceId}/state",
+ {
+ "online": row[0],
+ "down": row[1],
+ "all": row[2],
+ "archived": row[3],
+ "new": row[4],
+ "unknown": row[5]
+ }
+ )
# Generate device-specific MQTT messages if enabled
if get_setting_value('MQTT_SEND_DEVICES'):
@@ -466,11 +467,11 @@ def mqtt_start(db):
# Specific devices processing
# Get all devices
- devices = db.read(get_setting_value('MQTT_DEVICES_SQL').replace('{s-quote}',"'"))
+ devices = db.read(get_setting_value('MQTT_DEVICES_SQL').replace('{s-quote}', "'"))
- sec_delay = len(devices) * int(get_setting_value('MQTT_DELAY_SEC'))*5
+ sec_delay = len(devices) * int(get_setting_value('MQTT_DELAY_SEC')) * 5
- mylog('verbose', [f"[{pluginName}] Estimated delay: ", (sec_delay), 's ', '(', round(sec_delay/60, 1), 'min)'])
+ mylog('verbose', [f"[{pluginName}] Estimated delay: ", (sec_delay), 's ', '(', round(sec_delay / 60, 1), 'min)'])
for device in devices:
@@ -495,27 +496,29 @@ def mqtt_start(db):
# handle device_tracker
# IMPORTANT: shared payload - device_tracker attributes and individual sensors
devJson = {
- "last_ip": device["devLastIP"],
- "is_new": str(device["devIsNew"]),
- "alert_down": str(device["devAlertDown"]),
- "vendor": sanitize_string(device["devVendor"]),
- "mac_address": str(device["devMac"]),
- "model": devDisplayName,
- "last_connection": prepTimeStamp(str(device["devLastConnection"])),
- "first_connection": prepTimeStamp(str(device["devFirstConnection"])),
- "sync_node": device["devSyncHubNode"],
- "group": device["devGroup"],
- "location": device["devLocation"],
- "network_parent_mac": device["devParentMAC"],
- "network_parent_name": next((dev["devName"] for dev in devices if dev["devMAC"] == device["devParentMAC"]), "")
- }
+ "last_ip": device["devLastIP"],
+ "is_new": str(device["devIsNew"]),
+ "alert_down": str(device["devAlertDown"]),
+ "vendor": sanitize_string(device["devVendor"]),
+ "mac_address": str(device["devMac"]),
+ "model": devDisplayName,
+ "last_connection": prepTimeStamp(str(device["devLastConnection"])),
+ "first_connection": prepTimeStamp(str(device["devFirstConnection"])),
+ "sync_node": device["devSyncHubNode"],
+ "group": device["devGroup"],
+ "location": device["devLocation"],
+ "network_parent_mac": device["devParentMAC"],
+ "network_parent_name": next((dev["devName"] for dev in devices if dev["devMAC"] == device["devParentMAC"]), "")
+ }
# bulk update device sensors in home assistant
publish_mqtt(mqtt_client, sensorConfig.state_topic, devJson) # REQUIRED, DON'T DELETE
# create and update is_present sensor
sensorConfig = create_sensor(mqtt_client, deviceId, devDisplayName, 'binary_sensor', 'is_present', 'wifi', device["devMac"])
- publish_mqtt(mqtt_client, sensorConfig.state_topic,
+ publish_mqtt(
+ mqtt_client,
+ sensorConfig.state_topic,
{
"is_present": to_binary_sensor(str(device["devPresentLastScan"]))
}
@@ -547,7 +550,7 @@ def to_binary_sensor(input):
elif isinstance(input, bool) and input:
return "ON"
elif isinstance(input, str) and input == "1":
- return "ON"
+ return "ON"
elif isinstance(input, bytes) and bytes_to_string(input) == "1":
return "ON"
return "OFF"
diff --git a/front/plugins/_publisher_ntfy/ntfy.py b/front/plugins/_publisher_ntfy/ntfy.py
index 46fe7059..71f91811 100755
--- a/front/plugins/_publisher_ntfy/ntfy.py
+++ b/front/plugins/_publisher_ntfy/ntfy.py
@@ -1,5 +1,5 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import json
import os
@@ -11,15 +11,15 @@ from base64 import b64encode
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-import conf
-from const import confFileName, logPath
-from plugin_helper import Plugin_Objects, handleEmpty
-from utils.datetime_utils import timeNowDB
-from logger import mylog, Logger
-from helper import get_setting_value
-from models.notification_instance import NotificationInstance
-from database import DB
-from pytz import timezone
+import conf # noqa: E402 [flake8 lint suppression]
+from const import confFileName, logPath # noqa: E402 [flake8 lint suppression]
+from plugin_helper import Plugin_Objects, handleEmpty # noqa: E402 [flake8 lint suppression]
+from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from models.notification_instance import NotificationInstance # noqa: E402 [flake8 lint suppression]
+from database import DB # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -33,13 +33,12 @@ LOG_PATH = logPath + '/plugins'
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
-
def main():
-
- mylog('verbose', [f'[{pluginName}](publisher) In script'])
-
+
+ mylog('verbose', [f'[{pluginName}](publisher) In script'])
+
# Check if basic config settings supplied
- if check_config() == False:
+ if check_config() is False:
mylog('none', [f'[{pluginName}] ⚠ ERROR: Publisher notification gateway not set up correctly. Check your {confFileName} {pluginName}_* variables.'])
return
@@ -65,9 +64,9 @@ def main():
# Log result
plugin_objects.add_object(
primaryId = pluginName,
- secondaryId = timeNowDB(),
+ secondaryId = timeNowDB(),
watched1 = notification["GUID"],
- watched2 = handleEmpty(response_text),
+ watched2 = handleEmpty(response_text),
watched3 = response_status_code,
watched4 = 'null',
extra = 'null',
@@ -77,15 +76,15 @@ def main():
plugin_objects.write_result_file()
-
-#-------------------------------------------------------------------------------
+# -------------------------------------------------------------------------------
def check_config():
- if get_setting_value('NTFY_HOST') == '' or get_setting_value('NTFY_TOPIC') == '':
+ if get_setting_value('NTFY_HOST') == '' or get_setting_value('NTFY_TOPIC') == '':
return False
else:
return True
-
-#-------------------------------------------------------------------------------
+
+
+# -------------------------------------------------------------------------------
def send(html, text):
response_text = ''
@@ -100,7 +99,7 @@ def send(html, text):
# prepare request headers
headers = {
"Title": "NetAlertX Notification",
- "Actions": "view, Open Dashboard, "+ get_setting_value('REPORT_DASHBOARD_URL'),
+ "Actions": "view, Open Dashboard, " + get_setting_value('REPORT_DASHBOARD_URL'),
"Priority": get_setting_value('NTFY_PRIORITY'),
"Tags": "warning"
}
@@ -109,37 +108,38 @@ def send(html, text):
if token != '':
headers["Authorization"] = "Bearer {}".format(token)
elif user != "" and pwd != "":
- # Generate hash for basic auth
+ # Generate hash for basic auth
basichash = b64encode(bytes(user + ':' + pwd, "utf-8")).decode("ascii")
- # add authorization header with hash
+ # add authorization header with hash
headers["Authorization"] = "Basic {}".format(basichash)
# call NTFY service
try:
- response = requests.post("{}/{}".format( get_setting_value('NTFY_HOST'),
- get_setting_value('NTFY_TOPIC')),
- data = text,
- headers = headers,
- verify = verify_ssl)
+ response = requests.post("{}/{}".format(
+ get_setting_value('NTFY_HOST'),
+ get_setting_value('NTFY_TOPIC')),
+ data = text,
+ headers = headers,
+ verify = verify_ssl
+ )
response_status_code = response.status_code
# Check if the request was successful (status code 200)
if response_status_code == 200:
- response_text = response.text # This captures the response body/message
+ response_text = response.text # This captures the response body/message
else:
- response_text = json.dumps(response.text)
+ response_text = json.dumps(response.text)
- except requests.exceptions.RequestException as e:
+ except requests.exceptions.RequestException as e:
mylog('none', [f'[{pluginName}] ⚠ ERROR: ', e])
response_text = e
return response_text, response_status_code
- return response_text, response_status_code
+ return response_text, response_status_code
if __name__ == '__main__':
sys.exit(main())
-
diff --git a/front/plugins/_publisher_pushover/pushover.py b/front/plugins/_publisher_pushover/pushover.py
index 39140f27..28e87a5a 100755
--- a/front/plugins/_publisher_pushover/pushover.py
+++ b/front/plugins/_publisher_pushover/pushover.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python3
+# !/usr/bin/env python3
import conf
from const import confFileName, logPath
from pytz import timezone
@@ -12,12 +12,12 @@ import requests
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from plugin_helper import Plugin_Objects, handleEmpty # noqa: E402
-from logger import mylog, Logger # noqa: E402
-from helper import get_setting_value, hide_string # noqa: E402
-from utils.datetime_utils import timeNowDB
-from models.notification_instance import NotificationInstance # noqa: E402
-from database import DB # noqa: E402
+from plugin_helper import Plugin_Objects, handleEmpty # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value, hide_string # noqa: E402 [flake8 lint suppression]
+from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
+from models.notification_instance import NotificationInstance # noqa: E402 [flake8 lint suppression]
+from database import DB # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value("TIMEZONE"))
diff --git a/front/plugins/_publisher_pushsafer/pushsafer.py b/front/plugins/_publisher_pushsafer/pushsafer.py
index 95f2159b..dda0b601 100755
--- a/front/plugins/_publisher_pushsafer/pushsafer.py
+++ b/front/plugins/_publisher_pushsafer/pushsafer.py
@@ -1,5 +1,5 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import json
import os
@@ -10,15 +10,15 @@ import requests
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-import conf
-from const import confFileName, logPath
-from plugin_helper import Plugin_Objects, handleEmpty
-from logger import mylog, Logger
-from helper import get_setting_value, hide_string
-from utils.datetime_utils import timeNowDB
-from models.notification_instance import NotificationInstance
-from database import DB
-from pytz import timezone
+import conf # noqa: E402 [flake8 lint suppression]
+from const import confFileName, logPath # noqa: E402 [flake8 lint suppression]
+from plugin_helper import Plugin_Objects, handleEmpty # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value, hide_string # noqa: E402 [flake8 lint suppression]
+from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
+from models.notification_instance import NotificationInstance # noqa: E402 [flake8 lint suppression]
+from database import DB # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -32,13 +32,12 @@ LOG_PATH = logPath + '/plugins'
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
-
def main():
-
- mylog('verbose', [f'[{pluginName}](publisher) In script'])
-
+
+ mylog('verbose', [f'[{pluginName}](publisher) In script'])
+
# Check if basic config settings supplied
- if check_config() == False:
+ if check_config() is False:
mylog('none', [f'[{pluginName}] ⚠ ERROR: Publisher notification gateway not set up correctly. Check your {confFileName} {pluginName}_* variables.'])
return
@@ -59,14 +58,14 @@ def main():
for notification in new_notifications:
# Send notification
- response_text, response_status_code = send(notification["Text"])
+ response_text, response_status_code = send(notification["Text"])
# Log result
plugin_objects.add_object(
primaryId = pluginName,
- secondaryId = timeNowDB(),
+ secondaryId = timeNowDB(),
watched1 = notification["GUID"],
- watched2 = handleEmpty(response_text),
+ watched2 = handleEmpty(response_text),
watched3 = response_status_code,
watched4 = 'null',
extra = 'null',
@@ -76,8 +75,7 @@ def main():
plugin_objects.write_result_file()
-
-#-------------------------------------------------------------------------------
+# -------------------------------------------------------------------------------
def send(text):
response_text = ''
@@ -85,8 +83,7 @@ def send(text):
token = get_setting_value('PUSHSAFER_TOKEN')
- mylog('verbose', [f'[{pluginName}] PUSHSAFER_TOKEN: "{hide_string(token)}"'])
-
+ mylog('verbose', [f'[{pluginName}] PUSHSAFER_TOKEN: "{hide_string(token)}"'])
try:
url = 'https://www.pushsafer.com/api'
@@ -101,40 +98,34 @@ def send(text):
"u" : get_setting_value('REPORT_DASHBOARD_URL'),
"ut" : 'Open NetAlertX',
"k" : token,
- }
+ }
response = requests.post(url, data=post_fields)
-
response_status_code = response.status_code
-
# Check if the request was successful (status code 200)
if response_status_code == 200:
- response_text = response.text # This captures the response body/message
+ response_text = response.text # This captures the response body/message
else:
- response_text = json.dumps(response.text)
+ response_text = json.dumps(response.text)
- except requests.exceptions.RequestException as e:
+ except requests.exceptions.RequestException as e:
mylog('none', [f'[{pluginName}] ⚠ ERROR: ', e])
response_text = e
return response_text, response_status_code
-
- return response_text, response_status_code
+ return response_text, response_status_code
-
-
-
-#-------------------------------------------------------------------------------
+# -------------------------------------------------------------------------------
def check_config():
- if get_setting_value('PUSHSAFER_TOKEN') == 'ApiKey':
- return False
- else:
- return True
+ if get_setting_value('PUSHSAFER_TOKEN') == 'ApiKey':
+ return False
+ else:
+ return True
+
# -------------------------------------------------------
if __name__ == '__main__':
sys.exit(main())
-
diff --git a/front/plugins/_publisher_telegram/tg.py b/front/plugins/_publisher_telegram/tg.py
index 8f63cc88..45bda08e 100755
--- a/front/plugins/_publisher_telegram/tg.py
+++ b/front/plugins/_publisher_telegram/tg.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import subprocess
import os
@@ -8,15 +8,15 @@ import sys
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-import conf
-from const import confFileName, logPath
-from plugin_helper import Plugin_Objects
-from utils.datetime_utils import timeNowDB
-from logger import mylog, Logger
-from helper import get_setting_value
-from models.notification_instance import NotificationInstance
-from database import DB
-from pytz import timezone
+import conf # noqa: E402 [flake8 lint suppression]
+from const import confFileName, logPath # noqa: E402 [flake8 lint suppression]
+from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
+from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from models.notification_instance import NotificationInstance # noqa: E402 [flake8 lint suppression]
+from database import DB # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -30,13 +30,11 @@ LOG_PATH = logPath + '/plugins'
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
-
-
def main():
mylog('verbose', [f'[{pluginName}](publisher) In script'])
# Check if basic config settings supplied
- if check_config() == False:
+ if check_config() is False:
mylog('none', [
f'[{pluginName}] ⚠ ERROR: Publisher notification gateway not set up correctly. Check your {confFileName} {pluginName}_* variables.'])
return
diff --git a/front/plugins/_publisher_webhook/webhook.py b/front/plugins/_publisher_webhook/webhook.py
index 24f34eeb..deace014 100755
--- a/front/plugins/_publisher_webhook/webhook.py
+++ b/front/plugins/_publisher_webhook/webhook.py
@@ -1,5 +1,5 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import json
import subprocess
@@ -13,15 +13,15 @@ INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-import conf
-from const import logPath, confFileName
-from plugin_helper import Plugin_Objects, handleEmpty
-from utils.datetime_utils import timeNowDB
-from logger import mylog, Logger
-from helper import get_setting_value, write_file
-from models.notification_instance import NotificationInstance
-from database import DB
-from pytz import timezone
+import conf # noqa: E402 [flake8 lint suppression]
+from const import logPath, confFileName # noqa: E402 [flake8 lint suppression]
+from plugin_helper import Plugin_Objects, handleEmpty # noqa: E402 [flake8 lint suppression]
+from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value, write_file # noqa: E402 [flake8 lint suppression]
+from models.notification_instance import NotificationInstance # noqa: E402 [flake8 lint suppression]
+from database import DB # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -35,13 +35,12 @@ LOG_PATH = logPath + '/plugins'
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
-
def main():
-
- mylog('verbose', [f'[{pluginName}](publisher) In script'])
-
+
+ mylog('verbose', [f'[{pluginName}](publisher) In script'])
+
# Check if basic config settings supplied
- if check_config() == False:
+ if check_config() is False:
mylog('none', [f'[{pluginName}] ⚠ ERROR: Publisher notification gateway not set up correctly. Check your {confFileName} {pluginName}_* variables.'])
return
@@ -62,15 +61,19 @@ def main():
for notification in new_notifications:
# Send notification
- response_stdout, response_stderr = send(notification["Text"], notification["HTML"], notification["JSON"])
+ response_stdout, response_stderr = send(
+ notification["Text"],
+ notification["HTML"],
+ notification["JSON"]
+ )
# Log result
plugin_objects.add_object(
primaryId = pluginName,
- secondaryId = timeNowDB(),
+ secondaryId = timeNowDB(),
watched1 = notification["GUID"],
- watched2 = handleEmpty(response_stdout),
- watched3 = handleEmpty(response_stderr),
+ watched2 = handleEmpty(response_stdout),
+ watched3 = handleEmpty(response_stderr),
watched4 = 'null',
extra = 'null',
foreignKey = notification["GUID"]
@@ -79,16 +82,16 @@ def main():
plugin_objects.write_result_file()
-#-------------------------------------------------------------------------------
+# -------------------------------------------------------------------------------
def check_config():
- if get_setting_value('WEBHOOK_URL') == '':
- return False
- else:
- return True
-
-#-------------------------------------------------------------------------------
+ if get_setting_value('WEBHOOK_URL') == '':
+ return False
+ else:
+ return True
-def send (text_data, html_data, json_data):
+
+# -------------------------------------------------------------------------------
+def send(text_data, html_data, json_data):
response_stderr = ''
response_stdout = ''
@@ -102,9 +105,9 @@ def send (text_data, html_data, json_data):
# use data type based on specified payload type
if payloadType == 'json':
- # In this code, the truncate_json function is used to recursively traverse the JSON object
- # and remove nodes that exceed the size limit. It checks the size of each node's JSON representation
- # using json.dumps and includes only the nodes that are within the limit.
+ # In this code, the truncate_json function is used to recursively traverse the JSON object
+ # and remove nodes that exceed the size limit. It checks the size of each node's JSON representation
+ # using json.dumps and includes only the nodes that are within the limit.
json_str = json.dumps(json_data)
if len(json_str) <= limit:
@@ -127,45 +130,48 @@ def send (text_data, html_data, json_data):
return obj
payloadData = truncate_json(json_data)
- if payloadType == 'html':
+ if payloadType == 'html':
if len(html_data) > limit:
payloadData = html_data[:limit] + "
(text was truncated)
"
else:
payloadData = html_data
- if payloadType == 'text':
+ if payloadType == 'text':
if len(text_data) > limit:
payloadData = text_data[:limit] + " (text was truncated)"
else:
payloadData = text_data
# Define slack-compatible payload
- _json_payload = { "text": payloadData } if payloadType == 'text' else {
- "username": "NetAlertX",
- "text": "There are new notifications",
- "attachments": [{
- "title": "NetAlertX Notifications",
- "title_link": get_setting_value('REPORT_DASHBOARD_URL'),
- "text": payloadData
- }]
- }
+ if payloadType == 'text':
+ _json_payload = {"text": payloadData}
+ else:
+ _json_payload = {
+ "username": "NetAlertX",
+ "text": "There are new notifications",
+ "attachments": [{
+ "title": "NetAlertX Notifications",
+ "title_link": get_setting_value('REPORT_DASHBOARD_URL'),
+ "text": payloadData
+ }]
+ }
# DEBUG - Write the json payload into a log file for debugging
- write_file (logPath + '/webhook_payload.json', json.dumps(_json_payload))
+ write_file(logPath + '/webhook_payload.json', json.dumps(_json_payload))
# Using the Slack-Compatible Webhook endpoint for Discord so that the same payload can be used for both
# Consider: curl has the ability to load in data to POST from a file + piping
- if(endpointUrl.startswith('https://discord.com/api/webhooks/') and not endpointUrl.endswith("/slack")):
+ if (endpointUrl.startswith('https://discord.com/api/webhooks/') and not endpointUrl.endswith("/slack")):
_WEBHOOK_URL = f"{endpointUrl}/slack"
- curlParams = ["curl","-i","-H", "Content-Type:application/json" ,"-d", json.dumps(_json_payload), _WEBHOOK_URL]
+ curlParams = ["curl", "-i", "-H", "Content-Type:application/json", "-d", json.dumps(_json_payload), _WEBHOOK_URL]
else:
_WEBHOOK_URL = endpointUrl
- curlParams = ["curl","-i","-X", requestMethod , "-H", "Content-Type:application/json", "-d", json.dumps(_json_payload), _WEBHOOK_URL]
+ curlParams = ["curl", "-i", "-X", requestMethod , "-H", "Content-Type:application/json", "-d", json.dumps(_json_payload), _WEBHOOK_URL]
# Add HMAC signature if configured
- if(secret != ''):
+ if (secret != ''):
h = hmac.new(secret.encode("UTF-8"), json.dumps(_json_payload, separators=(',', ':')).encode(), hashlib.sha256).hexdigest()
- curlParams.insert(4,"-H")
- curlParams.insert(5,f"X-Webhook-Signature: sha256={h}")
+ curlParams.insert(4, "-H")
+ curlParams.insert(5, f"X-Webhook-Signature: sha256={h}")
try:
# Execute CURL call
@@ -173,13 +179,11 @@ def send (text_data, html_data, json_data):
result = subprocess.run(curlParams, capture_output=True, text=True)
response_stderr = result.stderr
- response_stdout = result.stdout
+ response_stdout = result.stdout
# Write stdout and stderr into .log files for debugging if needed
mylog('debug', [f'[{pluginName}] stdout: ', response_stdout])
- mylog('debug', [f'[{pluginName}] stderr: ', response_stderr])
-
-
+ mylog('debug', [f'[{pluginName}] stderr: ', response_stderr])
except subprocess.CalledProcessError as e:
# An error occurred, handle it
@@ -187,10 +191,9 @@ def send (text_data, html_data, json_data):
response_stderr = e.output
+ return response_stdout, response_stderr
- return response_stdout, response_stderr
-# -------------------------------------------------------
+# -------------------------------------------------------
if __name__ == '__main__':
sys.exit(main())
-
diff --git a/front/plugins/arp_scan/script.py b/front/plugins/arp_scan/script.py
index f1b14c8d..eda50387 100755
--- a/front/plugins/arp_scan/script.py
+++ b/front/plugins/arp_scan/script.py
@@ -1,7 +1,6 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import time
-import pathlib
import argparse
import sys
import re
@@ -9,16 +8,16 @@ import base64
import subprocess
# Register NetAlertX directories
-INSTALL_PATH="/app"
+INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from database import DB
-from plugin_helper import Plugin_Objects, handleEmpty
-from logger import mylog, Logger, append_line_to_file
-from helper import get_setting_value
-from const import logPath, applicationPath
-import conf
-from pytz import timezone
+from database import DB # noqa: E402 [flake8 lint suppression]
+from plugin_helper import Plugin_Objects, handleEmpty # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value("TIMEZONE"))
diff --git a/front/plugins/asuswrt_import/script.py b/front/plugins/asuswrt_import/script.py
index 463cb270..bb4c1a4b 100755
--- a/front/plugins/asuswrt_import/script.py
+++ b/front/plugins/asuswrt_import/script.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import sys
@@ -6,17 +6,16 @@ INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
pluginName = "ASUSWRT"
-import asyncio
-
-import aiohttp
-import conf
-from asusrouter import AsusData, AsusRouter
-from asusrouter.modules.connection import ConnectionState
-from const import logPath
-from helper import get_setting_value
-from logger import Logger, mylog
-from plugin_helper import (Plugin_Objects, handleEmpty)
-from pytz import timezone
+import asyncio # noqa: E402 [flake8 lint suppression]
+import aiohttp # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from asusrouter import AsusData, AsusRouter # noqa: E402 [flake8 lint suppression]
+from asusrouter.modules.connection import ConnectionState # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from logger import Logger, mylog # noqa: E402 [flake8 lint suppression]
+from plugin_helper import (Plugin_Objects, handleEmpty) # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
conf.tz = timezone(get_setting_value("TIMEZONE"))
diff --git a/front/plugins/avahi_scan/avahi_scan.py b/front/plugins/avahi_scan/avahi_scan.py
index 56daaa0d..119dff22 100755
--- a/front/plugins/avahi_scan/avahi_scan.py
+++ b/front/plugins/avahi_scan/avahi_scan.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python3
+# !/usr/bin/env python3
import os
import sys
import socket
@@ -8,14 +8,14 @@ from zeroconf import Zeroconf
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from plugin_helper import Plugin_Objects
-from logger import mylog, Logger
-from const import logPath
-from helper import get_setting_value
-from database import DB
-from models.device_instance import DeviceInstance
-import conf
-from pytz import timezone
+from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from database import DB # noqa: E402 [flake8 lint suppression]
+from models.device_instance import DeviceInstance # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
# Configure timezone and logging
conf.tz = timezone(get_setting_value("TIMEZONE"))
@@ -67,7 +67,7 @@ def resolve_mdns_name(ip: str, timeout: int = 5) -> str:
hostname = socket.getnameinfo((ip, 0), socket.NI_NAMEREQD)[0]
zeroconf.close()
if hostname and hostname != ip:
- mylog("debug", [f"[{pluginName}] Found mDNS name: {hostname}"])
+ mylog("debug", [f"[{pluginName}] Found mDNS name (rev_name): {hostname} ({rev_name})"])
return hostname
except Exception as e:
mylog("debug", [f"[{pluginName}] Zeroconf lookup failed for {ip}: {e}"])
@@ -89,7 +89,7 @@ def main():
timeout = get_setting_value("AVAHISCAN_RUN_TIMEOUT")
use_mock = "--mockdata" in sys.argv
-
+
if use_mock:
mylog("verbose", [f"[{pluginName}] Running in MOCK mode"])
devices = [
@@ -137,4 +137,4 @@ def main():
# Entrypoint
# =============================================================================
if __name__ == "__main__":
- main()
\ No newline at end of file
+ main()
diff --git a/front/plugins/csv_backup/script.py b/front/plugins/csv_backup/script.py
index 672bc099..124843ee 100755
--- a/front/plugins/csv_backup/script.py
+++ b/front/plugins/csv_backup/script.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import argparse
@@ -11,11 +11,11 @@ from datetime import datetime
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from logger import mylog, Logger
-from helper import get_setting_value
-from const import logPath, fullDbPath
-import conf
-from pytz import timezone
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from const import logPath, fullDbPath # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -29,6 +29,7 @@ LOG_PATH = logPath + '/plugins'
LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
+
def main():
# the script expects a parameter in the format of devices=device1,device2,...
@@ -44,7 +45,7 @@ def main():
else:
overwrite = False
- mylog('verbose', ['[CSVBCKP] In script'])
+ mylog('verbose', ['[CSVBCKP] In script'])
# Connect to the App database
conn = sqlite3.connect(fullDbPath)
@@ -64,7 +65,7 @@ def main():
fullPath = os.path.join(values.location.split('=')[1], filename)
- mylog('verbose', ['[CSVBCKP] Writing file ', fullPath])
+ mylog('verbose', ['[CSVBCKP] Writing file ', fullPath])
# Create a CSV file in the specified location
with open(fullPath, 'w', newline='') as csvfile:
@@ -72,7 +73,7 @@ def main():
csv_writer = csv.writer(csvfile, delimiter=',', quoting=csv.QUOTE_MINIMAL)
# Wrap the header values in double quotes and write the header row
- csv_writer.writerow([ '"' + col + '"' for col in columns])
+ csv_writer.writerow(['"' + col + '"' for col in columns])
# Fetch and write data rows
for row in cursor.fetchall():
@@ -96,8 +97,8 @@ def main():
return 0
-#===============================================================================
+# ===============================================================================
# BEGIN
-#===============================================================================
+# ===============================================================================
if __name__ == '__main__':
- main()
\ No newline at end of file
+ main()
diff --git a/front/plugins/db_cleanup/script.py b/front/plugins/db_cleanup/script.py
index ee538075..e657b75a 100755
--- a/front/plugins/db_cleanup/script.py
+++ b/front/plugins/db_cleanup/script.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import sys
@@ -8,11 +8,11 @@ import sqlite3
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from logger import mylog, Logger
-from helper import get_setting_value
-from const import logPath, fullDbPath
-import conf
-from pytz import timezone
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from const import logPath, fullDbPath # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value("TIMEZONE"))
@@ -81,7 +81,7 @@ def cleanup_database(
)
cursor.execute(
"""DELETE from Online_History where "Index" not in (
- SELECT "Index" from Online_History
+ SELECT "Index" from Online_History
order by Scan_Date desc limit 150)"""
)
@@ -94,7 +94,7 @@ def cleanup_database(
],
)
cursor.execute(
- f"""DELETE FROM Events
+ f"""DELETE FROM Events
WHERE eve_DateTime <= date('now', '-{str(DAYS_TO_KEEP_EVENTS)} day')"""
)
# -----------------------------------------------------
@@ -107,11 +107,11 @@ def cleanup_database(
)
# Build the SQL query to delete entries that exceed the limit per unique "Plugin" column entry
- delete_query = f"""DELETE FROM Plugins_History
+ delete_query = f"""DELETE FROM Plugins_History
WHERE "Index" NOT IN (
SELECT "Index"
FROM (
- SELECT "Index",
+ SELECT "Index",
ROW_NUMBER() OVER(PARTITION BY "Plugin" ORDER BY DateTimeChanged DESC) AS row_num
FROM Plugins_History
) AS ranked_objects
@@ -133,11 +133,11 @@ def cleanup_database(
)
# Build the SQL query to delete entries
- delete_query = f"""DELETE FROM Notifications
+ delete_query = f"""DELETE FROM Notifications
WHERE "Index" NOT IN (
SELECT "Index"
FROM (
- SELECT "Index",
+ SELECT "Index",
ROW_NUMBER() OVER(PARTITION BY "Notifications" ORDER BY DateTimeCreated DESC) AS row_num
FROM Notifications
) AS ranked_objects
@@ -153,11 +153,11 @@ def cleanup_database(
mylog("verbose", [f"[{pluginName}] Trim AppEvents to less than {histCount}"])
# Build the SQL query to delete entries
- delete_query = f"""DELETE FROM AppEvents
+ delete_query = f"""DELETE FROM AppEvents
WHERE "Index" NOT IN (
SELECT "Index"
FROM (
- SELECT "Index",
+ SELECT "Index",
ROW_NUMBER() OVER(PARTITION BY "AppEvents" ORDER BY DateTimeCreated DESC) AS row_num
FROM AppEvents
) AS ranked_objects
diff --git a/front/plugins/ddns_update/script.py b/front/plugins/ddns_update/script.py
index 4ae97fce..f38d231d 100755
--- a/front/plugins/ddns_update/script.py
+++ b/front/plugins/ddns_update/script.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import argparse
@@ -9,11 +9,11 @@ import subprocess
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from logger import mylog, Logger
-from helper import get_setting_value, check_IP_format
-from const import logPath
-import conf
-from pytz import timezone
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value, check_IP_format # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -28,91 +28,88 @@ LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
-
def main():
- mylog('verbose', [f'[{pluginName}] In script'])
-
+ mylog('verbose', [f'[{pluginName}] In script'])
+
parser = argparse.ArgumentParser(description='Check internet connectivity and IP')
-
- parser.add_argument('prev_ip', action="store", help="Previous IP address to compare against the current IP")
+
+ parser.add_argument('prev_ip', action="store", help="Previous IP address to compare against the current IP")
parser.add_argument('DDNS_UPDATE_URL', action="store", help="URL for updating Dynamic DNS (DDNS)")
parser.add_argument('DDNS_USER', action="store", help="Username for Dynamic DNS (DDNS) authentication")
parser.add_argument('DDNS_PASSWORD', action="store", help="Password for Dynamic DNS (DDNS) authentication")
parser.add_argument('DDNS_DOMAIN', action="store", help="Dynamic DNS (DDNS) domain name")
-
values = parser.parse_args()
- PREV_IP = values.prev_ip.split('=')[1]
+ PREV_IP = values.prev_ip.split('=')[1]
DDNS_UPDATE_URL = values.DDNS_UPDATE_URL.split('=')[1]
DDNS_USER = values.DDNS_USER.split('=')[1]
DDNS_PASSWORD = values.DDNS_PASSWORD.split('=')[1]
- DDNS_DOMAIN = values.DDNS_DOMAIN.split('=')[1]
+ DDNS_DOMAIN = values.DDNS_DOMAIN.split('=')[1]
# perform the new IP lookup and DDNS tasks if enabled
- ddns_update( DDNS_UPDATE_URL, DDNS_USER, DDNS_PASSWORD, DDNS_DOMAIN, PREV_IP)
+ ddns_update(DDNS_UPDATE_URL, DDNS_USER, DDNS_PASSWORD, DDNS_DOMAIN, PREV_IP)
+
+ mylog('verbose', [f'[{pluginName}] Finished '])
- mylog('verbose', [f'[{pluginName}] Finished '])
-
return 0
-
-
-#===============================================================================
+
+
+# ===============================================================================
# INTERNET IP CHANGE
-#===============================================================================
-def ddns_update ( DDNS_UPDATE_URL, DDNS_USER, DDNS_PASSWORD, DDNS_DOMAIN, PREV_IP ):
-
+# ===============================================================================
+def ddns_update(DDNS_UPDATE_URL, DDNS_USER, DDNS_PASSWORD, DDNS_DOMAIN, PREV_IP):
+
# Update DDNS record if enabled and IP is different
# Get Dynamic DNS IP
-
+
mylog('verbose', [f'[{pluginName}] Retrieving Dynamic DNS IP'])
dns_IP = get_dynamic_DNS_IP(DDNS_DOMAIN)
# Check Dynamic DNS IP
if dns_IP == "" or dns_IP == "0.0.0.0" :
- mylog('none', [f'[{pluginName}] Error retrieving Dynamic DNS IP'])
+ mylog('none', [f'[{pluginName}] Error retrieving Dynamic DNS IP'])
mylog('none', [f'[{pluginName}] ', dns_IP])
# Check DNS Change
if dns_IP != PREV_IP :
mylog('none', [f'[{pluginName}] Updating Dynamic DNS IP'])
- message = set_dynamic_DNS_IP (DDNS_UPDATE_URL, DDNS_USER, DDNS_PASSWORD, DDNS_DOMAIN)
- mylog('none', [f'[{pluginName}] ', message])
+ message = set_dynamic_DNS_IP(DDNS_UPDATE_URL, DDNS_USER, DDNS_PASSWORD, DDNS_DOMAIN)
+ mylog('none', [f'[{pluginName}] ', message])
- # plugin_objects = Plugin_Objects(RESULT_FILE)
-
+ # plugin_objects = Plugin_Objects(RESULT_FILE)
# plugin_objects.add_object(
# primaryId = 'Internet', # MAC (Device Name)
- # secondaryId = new_internet_IP, # IP Address
+ # secondaryId = new_internet_IP, # IP Address
# watched1 = f'Previous IP: {PREV_IP}',
# watched2 = '',
- # watched3 = '',
+ # watched3 = '',
# watched4 = '',
- # extra = f'Previous IP: {PREV_IP}',
+ # extra = f'Previous IP: {PREV_IP}',
# foreignKey = 'Internet')
- # plugin_objects.write_result_file()
-
+ # plugin_objects.write_result_file()
-#-------------------------------------------------------------------------------
-def get_dynamic_DNS_IP (DDNS_DOMAIN):
+
+# -------------------------------------------------------------------------------
+def get_dynamic_DNS_IP(DDNS_DOMAIN):
# Using supplied DNS server
dig_args = ['dig', '+short', DDNS_DOMAIN]
try:
# try runnning a subprocess
- dig_output = subprocess.check_output (dig_args, universal_newlines=True)
+ dig_output = subprocess.check_output(dig_args, universal_newlines=True)
mylog('none', [f'[{pluginName}] DIG output :', dig_output])
except subprocess.CalledProcessError as e:
# An error occured, handle it
mylog('none', [f'[{pluginName}] ⚠ ERROR - ', e.output])
- dig_output = '' # probably no internet
+ dig_output = '' # probably no internet
# Check result is an IP
- IP = check_IP_format (dig_output)
+ IP = check_IP_format(dig_output)
# Handle invalid response
if IP == '':
@@ -120,28 +117,27 @@ def get_dynamic_DNS_IP (DDNS_DOMAIN):
return IP
-#-------------------------------------------------------------------------------
-def set_dynamic_DNS_IP (DDNS_UPDATE_URL, DDNS_USER, DDNS_PASSWORD, DDNS_DOMAIN):
+
+# -------------------------------------------------------------------------------
+def set_dynamic_DNS_IP(DDNS_UPDATE_URL, DDNS_USER, DDNS_PASSWORD, DDNS_DOMAIN):
try:
# try runnning a subprocess
# Update Dynamic IP
- curl_output = subprocess.check_output (['curl',
- '-s',
- DDNS_UPDATE_URL +
- 'username=' + DDNS_USER +
- '&password=' + DDNS_PASSWORD +
- '&hostname=' + DDNS_DOMAIN],
- universal_newlines=True)
+ curl_output = subprocess.check_output([
+ 'curl',
+ '-s',
+ DDNS_UPDATE_URL + 'username=' + DDNS_USER + '&password=' + DDNS_PASSWORD + '&hostname=' + DDNS_DOMAIN],
+ universal_newlines=True)
except subprocess.CalledProcessError as e:
# An error occured, handle it
- mylog('none', [f'[{pluginName}] ⚠ ERROR - ',e.output])
- curl_output = ""
-
+ mylog('none', [f'[{pluginName}] ⚠ ERROR - ', e.output])
+ curl_output = ""
+
return curl_output
-#===============================================================================
+# ===============================================================================
# BEGIN
-#===============================================================================
+# ===============================================================================
if __name__ == '__main__':
- main()
\ No newline at end of file
+ main()
diff --git a/front/plugins/dhcp_leases/script.py b/front/plugins/dhcp_leases/script.py
index 49be19f5..2366bc93 100755
--- a/front/plugins/dhcp_leases/script.py
+++ b/front/plugins/dhcp_leases/script.py
@@ -1,22 +1,22 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
from __future__ import unicode_literals
import argparse
import os
import sys
-import chardet
+import chardet
# Register NetAlertX directories
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from plugin_helper import Plugin_Objects, handleEmpty, is_mac
-from logger import mylog, Logger
-from dhcp_leases import DhcpLeases
-from helper import get_setting_value
-import conf
-from const import logPath
-from pytz import timezone
+from plugin_helper import Plugin_Objects, handleEmpty, is_mac # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from dhcp_leases import DhcpLeases # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -24,34 +24,38 @@ conf.tz = timezone(get_setting_value('TIMEZONE'))
# Make sure log level is initialized correctly
Logger(get_setting_value('LOG_LEVEL'))
-pluginName= 'DHCPLSS'
+pluginName = 'DHCPLSS'
LOG_PATH = logPath + '/plugins'
LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
-
-
# -------------------------------------------------------------
-def main():
+def main():
mylog('verbose', [f'[{pluginName}] In script'])
- last_run_logfile = open(RESULT_FILE, 'a')
+ last_run_logfile = open(RESULT_FILE, 'a')
last_run_logfile.write("")
parser = argparse.ArgumentParser(description='Import devices from dhcp.leases files')
- parser.add_argument('paths', action="store", help="absolute dhcp.leases file paths to check separated by ','")
+ parser.add_argument(
+ 'paths',
+ action="store",
+ help="absolute dhcp.leases file paths to check separated by ','"
+ )
+
values = parser.parse_args()
plugin_objects = Plugin_Objects(RESULT_FILE)
if values.paths:
- for path in values.paths.split('=')[1].split(','):
+ for path in values.paths.split('=')[1].split(','):
plugin_objects = get_entries(path, plugin_objects)
- mylog('verbose', [f'[{pluginName}] {len(plugin_objects)} Entries found in "{path}"'])
-
+ mylog('verbose', [f'[{pluginName}] {len(plugin_objects)} Entries found in "{path}"'])
+
plugin_objects.write_result_file()
+
# -------------------------------------------------------------
def get_entries(path, plugin_objects):
@@ -66,7 +70,7 @@ def get_entries(path, plugin_objects):
# Use the detected encoding
encoding = result['encoding']
- # Order: MAC, IP, IsActive, NAME, Hardware
+ # Order: MAC, IP, IsActive, NAME, Hardware
# Handle pihole-specific dhcp.leases files
if 'pihole' in path:
with open(path, 'r', encoding=encoding, errors='replace') as f:
@@ -111,9 +115,9 @@ def get_entries(path, plugin_objects):
if is_mac(lease.ethernet):
plugin_objects.add_object(
- primaryId = handleEmpty(lease.ethernet),
- secondaryId = handleEmpty(lease.ip),
- watched1 = handleEmpty(lease.active),
+ primaryId = handleEmpty(lease.ethernet),
+ secondaryId = handleEmpty(lease.ip),
+ watched1 = handleEmpty(lease.active),
watched2 = handleEmpty(lease.hostname),
watched3 = handleEmpty(lease.hardware),
watched4 = handleEmpty(lease.binding_state),
@@ -122,5 +126,6 @@ def get_entries(path, plugin_objects):
)
return plugin_objects
-if __name__ == '__main__':
- main()
+
+if __name__ == '__main__':
+ main()
diff --git a/front/plugins/dhcp_servers/script.py b/front/plugins/dhcp_servers/script.py
index 5948153c..665ae155 100755
--- a/front/plugins/dhcp_servers/script.py
+++ b/front/plugins/dhcp_servers/script.py
@@ -1,9 +1,8 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
# Based on the work of https://github.com/leiweibau/Pi.Alert
import subprocess
import os
-from datetime import datetime
import sys
@@ -11,12 +10,12 @@ import sys
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from plugin_helper import Plugin_Objects, Plugin_Object
-from logger import mylog, Logger
-from helper import get_setting_value
-import conf
-from pytz import timezone
-from const import logPath
+from plugin_helper import Plugin_Objects, Plugin_Object # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
@@ -31,13 +30,14 @@ LOG_PATH = logPath + '/plugins'
LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
+
def main():
mylog('verbose', ['[DHCPSRVS] In script'])
-
- last_run_logfile = open(RESULT_FILE, 'a')
+
+ last_run_logfile = open(RESULT_FILE, 'a')
last_run_logfile.write("")
-
+
plugin_objects = Plugin_Objects(RESULT_FILE)
timeoutSec = get_setting_value('DHCPSRVS_RUN_TIMEOUT')
@@ -46,10 +46,10 @@ def main():
try:
# Number of DHCP discovery probes to send
dhcp_probes = 1
-
+
# Initialize a list to store output lines from the scan
newLines = []
-
+
for _ in range(dhcp_probes):
output = subprocess.check_output(nmapArgs, universal_newlines=True, stderr=subprocess.STDOUT, timeout=timeoutSec)
newLines += output.split("\n")
@@ -57,9 +57,9 @@ def main():
newEntries = []
for line in newLines:
-
+
mylog('verbose', [f'[DHCPSRVS] Processing line: {line} '])
-
+
if 'Response ' in line and ' of ' in line:
newEntries.append(Plugin_Object())
elif 'Server Identifier' in line:
@@ -85,7 +85,7 @@ def main():
newEntries[-1].extra += ',' + newVal
for e in newEntries:
-
+
plugin_objects.add_object(
primaryId=e.primaryId,
secondaryId=e.secondaryId,
@@ -101,5 +101,6 @@ def main():
except Exception as e:
mylog('verbose', ['[DHCPSRVS] Error in main:', str(e)])
+
if __name__ == '__main__':
main()
diff --git a/front/plugins/dig_scan/digscan.py b/front/plugins/dig_scan/digscan.py
index c35fb9b9..90fa17ad 100755
--- a/front/plugins/dig_scan/digscan.py
+++ b/front/plugins/dig_scan/digscan.py
@@ -1,5 +1,4 @@
-#!/usr/bin/env python
-
+# !/usr/bin/env python
import os
import sys
import subprocess
@@ -8,14 +7,14 @@ import subprocess
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from plugin_helper import Plugin_Objects
-from logger import mylog, Logger
-from const import logPath
-from helper import get_setting_value
-from database import DB
-from models.device_instance import DeviceInstance
-import conf
-from pytz import timezone
+from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from database import DB # noqa: E402 [flake8 lint suppression]
+from models.device_instance import DeviceInstance # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -35,7 +34,7 @@ plugin_objects = Plugin_Objects(RESULT_FILE)
def main():
- mylog('verbose', [f'[{pluginName}] In script'])
+ mylog('verbose', [f'[{pluginName}] In script'])
timeout = get_setting_value('DIGSCAN_RUN_TIMEOUT')
@@ -50,13 +49,13 @@ def main():
device_handler = DeviceInstance(db)
# Retrieve devices
- if get_setting_value("REFRESH_FQDN"):
+ if get_setting_value("REFRESH_FQDN"):
devices = device_handler.getAll()
- else:
+ else:
devices = device_handler.getUnknown()
- mylog('verbose', [f'[{pluginName}] Devices count: {len(devices)}'])
-
+ mylog('verbose', [f'[{pluginName}] Devices count: {len(devices)}'])
+
# TEST - below is a WINDOWS host IP
# execute_name_lookup('192.168.1.121', timeout)
@@ -65,27 +64,27 @@ def main():
if domain_name != '':
plugin_objects.add_object(
- # "MAC", "IP", "Server", "Name"
- primaryId = device['devMac'],
- secondaryId = device['devLastIP'],
- watched1 = dns_server,
- watched2 = domain_name,
- watched3 = '',
- watched4 = '',
- extra = '',
- foreignKey = device['devMac'])
+ primaryId = device['devMac'],
+ secondaryId = device['devLastIP'],
+ watched1 = dns_server,
+ watched2 = domain_name,
+ watched3 = '',
+ watched4 = '',
+ extra = '',
+ foreignKey = device['devMac']
+ )
plugin_objects.write_result_file()
-
-
- mylog('verbose', [f'[{pluginName}] Script finished'])
-
+
+ mylog('verbose', [f'[{pluginName}] Script finished'])
+
return 0
-#===============================================================================
+
+# ===============================================================================
# Execute scan
-#===============================================================================
-def execute_name_lookup (ip, timeout):
+# ===============================================================================
+def execute_name_lookup(ip, timeout):
"""
Execute the DIG command on IP.
"""
@@ -97,32 +96,38 @@ def execute_name_lookup (ip, timeout):
try:
mylog('verbose', [f'[{pluginName}] DEBUG CMD :', args])
-
+
# try runnning a subprocess with a forced (timeout) in case the subprocess hangs
- output = subprocess.check_output (args, universal_newlines=True, stderr=subprocess.STDOUT, timeout=(timeout), text=True).strip()
+ output = subprocess.check_output(
+ args,
+ universal_newlines=True,
+ stderr=subprocess.STDOUT,
+ timeout=(timeout),
+ text=True
+ ).strip()
mylog('verbose', [f'[{pluginName}] DEBUG OUTPUT : {output}'])
domain_name = output
- dns_server = ''
+ dns_server = ''
mylog('verbose', [f'[{pluginName}] Domain Name: {domain_name}'])
return domain_name, dns_server
except subprocess.CalledProcessError as e:
- mylog('verbose', [f'[{pluginName}] ⚠ ERROR - {e.output}'])
-
- except subprocess.TimeoutExpired:
- mylog('verbose', [f'[{pluginName}] TIMEOUT - the process forcefully terminated as timeout reached'])
+ mylog('verbose', [f'[{pluginName}] ⚠ ERROR - {e.output}'])
- if output == "": # check if the subprocess failed
- mylog('verbose', [f'[{pluginName}] Scan: FAIL - check logs'])
- else:
+ except subprocess.TimeoutExpired:
+ mylog('verbose', [f'[{pluginName}] TIMEOUT - the process forcefully terminated as timeout reached'])
+
+ if output == "": # check if the subprocess failed
+ mylog('verbose', [f'[{pluginName}] Scan: FAIL - check logs'])
+ else:
mylog('verbose', [f'[{pluginName}] Scan: SUCCESS'])
- return '', ''
+ return '', ''
+
if __name__ == '__main__':
main()
-
diff --git a/front/plugins/freebox/freebox.py b/front/plugins/freebox/freebox.py
index 5dff8717..f3088cb6 100755
--- a/front/plugins/freebox/freebox.py
+++ b/front/plugins/freebox/freebox.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import sys
@@ -17,11 +17,11 @@ from aiofreepybox.exceptions import NotOpenError, AuthorizationError
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from plugin_helper import Plugin_Objects
-from logger import mylog, Logger
-from const import logPath
-from helper import get_setting_value
-import conf
+from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value("TIMEZONE"))
@@ -79,6 +79,7 @@ def map_device_type(type: str):
mylog("minimal", [f"[{pluginName}] Unknown device type: {type}"])
return device_type_map["other"]
+
async def get_device_data(api_version: int, api_address: str, api_port: int):
# ensure existence of db path
config_base = Path(os.getenv("NETALERTX_CONFIG", "/data/config"))
diff --git a/front/plugins/icmp_scan/icmp.py b/front/plugins/icmp_scan/icmp.py
index e86848f9..461a7e32 100755
--- a/front/plugins/icmp_scan/icmp.py
+++ b/front/plugins/icmp_scan/icmp.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
# test script by running:
# tbc
@@ -11,14 +11,14 @@ import re
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from plugin_helper import Plugin_Objects
-from logger import mylog, Logger
-from helper import get_setting_value
-from const import logPath
-from database import DB
-from models.device_instance import DeviceInstance
-import conf
-from pytz import timezone
+from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
+from database import DB # noqa: E402 [flake8 lint suppression]
+from models.device_instance import DeviceInstance # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -33,16 +33,14 @@ LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
-
def main():
- mylog('verbose', [f'[{pluginName}] In script'])
-
+ mylog('verbose', [f'[{pluginName}] In script'])
timeout = get_setting_value('ICMP_RUN_TIMEOUT')
args = get_setting_value('ICMP_ARGS')
in_regex = get_setting_value('ICMP_IN_REGEX')
-
+
# Create a database connection
db = DB() # instance of class DB
db.open()
@@ -61,46 +59,45 @@ def main():
# Filter devices based on the regex match
filtered_devices = [
- device for device in all_devices
+ device for device in all_devices
if regex_pattern.match(device['devLastIP'])
]
-
- mylog('verbose', [f'[{pluginName}] Devices to PING: {len(filtered_devices)}'])
+ mylog('verbose', [f'[{pluginName}] Devices to PING: {len(filtered_devices)}'])
for device in filtered_devices:
is_online, output = execute_scan(device['devLastIP'], timeout, args)
mylog('verbose', [f"[{pluginName}] ip: {device['devLastIP']} is_online: {is_online}"])
-
if is_online:
plugin_objects.add_object(
- # "MAC", "IP", "Name", "Output"
- primaryId = device['devMac'],
- secondaryId = device['devLastIP'],
- watched1 = device['devName'],
- watched2 = output.replace('\n',''),
- watched3 = '',
- watched4 = '',
- extra = '',
- foreignKey = device['devMac'])
+ # "MAC", "IP", "Name", "Output"
+ primaryId = device['devMac'],
+ secondaryId = device['devLastIP'],
+ watched1 = device['devName'],
+ watched2 = output.replace('\n', ''),
+ watched3 = '',
+ watched4 = '',
+ extra = '',
+ foreignKey = device['devMac']
+ )
plugin_objects.write_result_file()
-
-
- mylog('verbose', [f'[{pluginName}] Script finished'])
-
+
+ mylog('verbose', [f'[{pluginName}] Script finished'])
+
return 0
-#===============================================================================
+
+# ===============================================================================
# Execute scan
-#===============================================================================
-def execute_scan (ip, timeout, args):
+# ===============================================================================
+def execute_scan(ip, timeout, args):
"""
Execute the ICMP command on IP.
"""
-
+
icmp_args = ['ping'] + args.split() + [ip]
# Execute command
@@ -108,12 +105,18 @@ def execute_scan (ip, timeout, args):
try:
# try runnning a subprocess with a forced (timeout) in case the subprocess hangs
- output = subprocess.check_output (icmp_args, universal_newlines=True, stderr=subprocess.STDOUT, timeout=(timeout), text=True)
+ output = subprocess.check_output(
+ icmp_args,
+ universal_newlines=True,
+ stderr=subprocess.STDOUT,
+ timeout=(timeout),
+ text=True
+ )
mylog('verbose', [f'[{pluginName}] DEBUG OUTPUT : {output}'])
# Parse output using case-insensitive regular expressions
- #Synology-NAS:/# ping -i 0.5 -c 3 -W 8 -w 9 192.168.1.82
+ # Synology-NAS:/# ping -i 0.5 -c 3 -W 8 -w 9 192.168.1.82
# PING 192.168.1.82 (192.168.1.82): 56 data bytes
# 64 bytes from 192.168.1.82: seq=0 ttl=64 time=0.080 ms
# 64 bytes from 192.168.1.82: seq=1 ttl=64 time=0.081 ms
@@ -130,7 +133,7 @@ def execute_scan (ip, timeout, args):
# --- 192.168.1.92 ping statistics ---
# 3 packets transmitted, 0 packets received, 100% packet loss
- # TODO: parse output and return True if online, False if Offline (100% packet loss, bad address)
+ # TODO: parse output and return True if online, False if Offline (100% packet loss, bad address)
is_online = True
# Check for 0% packet loss in the output
@@ -145,22 +148,20 @@ def execute_scan (ip, timeout, args):
except subprocess.CalledProcessError as e:
# An error occurred, handle it
- mylog('verbose', [f'[{pluginName}] ⚠ ERROR - check logs'])
+ mylog('verbose', [f'[{pluginName}] ⚠ ERROR - check logs'])
mylog('verbose', [f'[{pluginName}]', e.output])
- return False, output
-
+ return False, output
+
except subprocess.TimeoutExpired:
- mylog('verbose', [f'[{pluginName}] TIMEOUT - the process forcefully terminated as timeout reached'])
- return False, output
+ mylog('verbose', [f'[{pluginName}] TIMEOUT - the process forcefully terminated as timeout reached'])
+ return False, output
- return False, output
-
-
-
+ return False, output
-#===============================================================================
+
+# ===============================================================================
# BEGIN
-#===============================================================================
+# ===============================================================================
if __name__ == '__main__':
- main()
\ No newline at end of file
+ main()
diff --git a/front/plugins/internet_ip/script.py b/front/plugins/internet_ip/script.py
index d793e441..5cb98e11 100755
--- a/front/plugins/internet_ip/script.py
+++ b/front/plugins/internet_ip/script.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import time
@@ -11,13 +11,13 @@ import re
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from plugin_helper import Plugin_Objects
-from utils.datetime_utils import timeNowDB
-from logger import mylog, Logger, append_line_to_file
-from helper import check_IP_format, get_setting_value
-from const import logPath
-import conf
-from pytz import timezone
+from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
+from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger, append_line_to_file # noqa: E402 [flake8 lint suppression]
+from helper import check_IP_format, get_setting_value # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -31,39 +31,39 @@ LOG_PATH = logPath + '/plugins'
LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
-
no_internet_ip = '0.0.0.0'
+
def main():
- mylog('verbose', [f'[{pluginName}] In script'])
-
+ mylog('verbose', [f'[{pluginName}] In script'])
+
parser = argparse.ArgumentParser(description='Check internet connectivity and IP')
-
+
parser.add_argument('prev_ip', action="store", help="Previous IP address to compare against the current IP")
- parser.add_argument('DIG_GET_IP_ARG', action="store", help="Arguments for the 'dig' command to retrieve the IP address") # unused
+ parser.add_argument('DIG_GET_IP_ARG', action="store", help="Arguments for the 'dig' command to retrieve the IP address") # unused
values = parser.parse_args()
- PREV_IP = values.prev_ip.split('=')[1]
+ PREV_IP = values.prev_ip.split('=')[1]
DIG_GET_IP_ARG = get_setting_value("INTRNT_DIG_GET_IP_ARG")
new_internet_IP = no_internet_ip
- mylog('verbose', [f'[{pluginName}] INTRNT_DIG_GET_IP_ARG: ', DIG_GET_IP_ARG])
+ mylog('verbose', [f'[{pluginName}] INTRNT_DIG_GET_IP_ARG: ', DIG_GET_IP_ARG])
# METHOD 1: dig
# perform the new IP lookup N times specified by the INTRNT_TRIES setting
-
+
INTRNT_RETRIES = get_setting_value("INTRNT_RETRIES")
retries_needed = 0
for i in range(INTRNT_RETRIES + 1):
- new_internet_IP, cmd_output = check_internet_IP( PREV_IP, DIG_GET_IP_ARG)
+ new_internet_IP, cmd_output = check_internet_IP(PREV_IP, DIG_GET_IP_ARG)
if new_internet_IP == no_internet_ip:
- time.sleep(1*i) # Exponential backoff strategy
+ time.sleep(1 * i) # Exponential backoff strategy
else:
retries_needed = i
break
@@ -71,68 +71,69 @@ def main():
# METHOD 2: curl
if new_internet_IP == no_internet_ip:
new_internet_IP, cmd_output = fallback_check_ip()
- mylog('verbose', [f'[{pluginName}] Curl Fallback (new_internet_IP|cmd_output): {new_internet_IP} | {cmd_output}'])
+ mylog('verbose', [f'[{pluginName}] Curl Fallback (new_internet_IP|cmd_output): {new_internet_IP} | {cmd_output}'])
# logging
- append_line_to_file (logPath + '/IP_changes.log', '['+str(timeNowDB()) +']\t'+ new_internet_IP +'\n')
+ append_line_to_file(logPath + '/IP_changes.log', '[' + str(timeNowDB()) + ']\t' + new_internet_IP + '\n')
+
+ plugin_objects = Plugin_Objects(RESULT_FILE)
- plugin_objects = Plugin_Objects(RESULT_FILE)
-
plugin_objects.add_object(
primaryId = 'Internet', # MAC (Device Name)
- secondaryId = new_internet_IP, # IP Address
+ secondaryId = new_internet_IP, # IP Address
watched1 = f'Previous IP: {PREV_IP}',
- watched2 = cmd_output.replace('\n',''),
- watched3 = retries_needed,
+ watched2 = cmd_output.replace('\n', ''),
+ watched3 = retries_needed,
watched4 = 'Gateway',
- extra = f'Previous IP: {PREV_IP}',
- foreignKey = 'Internet')
+ extra = f'Previous IP: {PREV_IP}',
+ foreignKey = 'Internet'
+ )
- plugin_objects.write_result_file()
+ plugin_objects.write_result_file()
+
+ mylog('verbose', [f'[{pluginName}] Finished '])
- mylog('verbose', [f'[{pluginName}] Finished '])
-
return 0
-
-
-#===============================================================================
+
+
+# ===============================================================================
# INTERNET IP CHANGE
-#===============================================================================
-def check_internet_IP ( PREV_IP, DIG_GET_IP_ARG ):
-
+# ===============================================================================
+def check_internet_IP(PREV_IP, DIG_GET_IP_ARG):
+
# Get Internet IP
mylog('verbose', [f'[{pluginName}] - Retrieving Internet IP'])
internet_IP, cmd_output = get_internet_IP(DIG_GET_IP_ARG)
- mylog('verbose', [f'[{pluginName}] Current internet_IP : {internet_IP}'])
-
- # Check previously stored IP
+ mylog('verbose', [f'[{pluginName}] Current internet_IP : {internet_IP}'])
+
+ # Check previously stored IP
previous_IP = no_internet_ip
- if PREV_IP is not None and len(PREV_IP) > 0 :
+ if PREV_IP is not None and len(PREV_IP) > 0 :
previous_IP = PREV_IP
- mylog('verbose', [f'[{pluginName}] previous_IP : {previous_IP}'])
+ mylog('verbose', [f'[{pluginName}] previous_IP : {previous_IP}'])
return internet_IP, cmd_output
-
-#-------------------------------------------------------------------------------
-def get_internet_IP (DIG_GET_IP_ARG):
+
+# -------------------------------------------------------------------------------
+def get_internet_IP(DIG_GET_IP_ARG):
cmd_output = ''
-
+
# Using 'dig'
dig_args = ['dig', '+short'] + DIG_GET_IP_ARG.strip().split()
try:
- cmd_output = subprocess.check_output (dig_args, universal_newlines=True)
- mylog('verbose', [f'[{pluginName}] DIG result : {cmd_output}'])
+ cmd_output = subprocess.check_output(dig_args, universal_newlines=True)
+ mylog('verbose', [f'[{pluginName}] DIG result : {cmd_output}'])
except subprocess.CalledProcessError as e:
mylog('verbose', [e.output])
- cmd_output = '' # no internet
+ cmd_output = '' # no internet
# Check result is an IP
- IP = check_IP_format (cmd_output)
+ IP = check_IP_format(cmd_output)
# Handle invalid response
if IP == '':
@@ -140,7 +141,8 @@ def get_internet_IP (DIG_GET_IP_ARG):
return IP, cmd_output
-#-------------------------------------------------------------------------------
+
+# -------------------------------------------------------------------------------
def fallback_check_ip():
"""Fallback mechanism using `curl ifconfig.me/ip`."""
try:
@@ -155,8 +157,9 @@ def fallback_check_ip():
mylog('none', [f'[{pluginName}] Fallback curl exception: {e}'])
return no_internet_ip, f'Fallback via curl exception: "{e}"'
-#===============================================================================
+
+# ===============================================================================
# BEGIN
-#===============================================================================
+# ===============================================================================
if __name__ == '__main__':
- main()
\ No newline at end of file
+ main()
diff --git a/front/plugins/internet_speedtest/script.py b/front/plugins/internet_speedtest/script.py
index 6314941e..c0f1a083 100755
--- a/front/plugins/internet_speedtest/script.py
+++ b/front/plugins/internet_speedtest/script.py
@@ -1,6 +1,5 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
-import argparse
import os
import sys
import speedtest
@@ -9,13 +8,13 @@ import speedtest
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from plugin_helper import Plugin_Objects
-from utils.datetime_utils import timeNowDB
-from logger import mylog, Logger
-from helper import get_setting_value
-import conf
-from pytz import timezone
-from const import logPath
+from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
+from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -28,18 +27,16 @@ pluginName = 'INTRSPD'
LOG_PATH = logPath + '/plugins'
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
-def main():
-
- mylog('verbose', ['[INTRSPD] In script'])
- parser = argparse.ArgumentParser(description='Speedtest Plugin for NetAlertX')
- values = parser.parse_args()
+def main():
+
+ mylog('verbose', ['[INTRSPD] In script'])
plugin_objects = Plugin_Objects(RESULT_FILE)
speedtest_result = run_speedtest()
plugin_objects.add_object(
primaryId = 'Speedtest',
- secondaryId = timeNowDB(),
+ secondaryId = timeNowDB(),
watched1 = speedtest_result['download_speed'],
watched2 = speedtest_result['upload_speed'],
watched3 = 'null',
@@ -49,25 +46,27 @@ def main():
)
plugin_objects.write_result_file()
+
def run_speedtest():
try:
st = speedtest.Speedtest(secure=True)
st.get_best_server()
download_speed = round(st.download() / 10**6, 2) # Convert to Mbps
upload_speed = round(st.upload() / 10**6, 2) # Convert to Mbps
-
- mylog('verbose', [f"[INTRSPD] Result (down|up): {str(download_speed)} Mbps|{upload_speed} Mbps"])
+
+ mylog('verbose', [f"[INTRSPD] Result (down|up): {str(download_speed)} Mbps|{upload_speed} Mbps"])
return {
'download_speed': download_speed,
'upload_speed': upload_speed,
}
except Exception as e:
- mylog('verbose', [f"[INTRSPD] Error running speedtest: {str(e)}"])
+ mylog('verbose', [f"[INTRSPD] Error running speedtest: {str(e)}"])
return {
'download_speed': -1,
'upload_speed': -1,
}
+
if __name__ == '__main__':
sys.exit(main())
diff --git a/front/plugins/ipneigh/ipneigh.py b/front/plugins/ipneigh/ipneigh.py
index f805347e..4ca9ee9f 100755
--- a/front/plugins/ipneigh/ipneigh.py
+++ b/front/plugins/ipneigh/ipneigh.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import sys
@@ -11,11 +11,11 @@ from functools import reduce
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from plugin_helper import Plugin_Objects
-from logger import mylog, Logger
-from const import logPath
-from helper import get_setting_value
-import conf
+from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -34,9 +34,8 @@ RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
plugin_objects = Plugin_Objects(RESULT_FILE)
-
def main():
- mylog('verbose', [f'[{pluginName}] In script'])
+ mylog('verbose', [f'[{pluginName}] In script'])
# Retrieve configuration settings
SCAN_SUBNETS = get_setting_value('SCAN_SUBNETS')
@@ -48,33 +47,33 @@ def main():
entry.split('--interface=')[-1].strip() for entry in SCAN_SUBNETS if '--interface=' in entry
)
- mylog('verbose', [f'[{pluginName}] Interfaces value: "{interfaces}"'])
+ mylog('verbose', [f'[{pluginName}] Interfaces value: "{interfaces}"'])
# retrieve data
raw_neighbors = get_neighbors(interfaces)
-
+
neighbors = parse_neighbors(raw_neighbors)
# Process the data into native application tables
if len(neighbors) > 0:
for device in neighbors:
- plugin_objects.add_object(
- primaryId = device['mac'],
- secondaryId = device['ip'],
- watched4 = device['last_seen'],
+ plugin_objects.add_object(
+ primaryId = device['mac'],
+ secondaryId = device['ip'],
+ watched4 = device['last_seen'],
- # The following are always unknown
- watched1 = device['hostname'], # don't use these --> handleEmpty(device['hostname']),
- watched2 = device['vendor'], # handleEmpty(device['vendor']),
- watched3 = device['device_type'], # handleEmpty(device['device_type']),
- extra = '',
- foreignKey = "" #device['mac']
- # helpVal1 = "Something1", # Optional Helper values to be passed for mapping into the app
- # helpVal2 = "Something1", # If you need to use even only 1, add the remaining ones too
- # helpVal3 = "Something1", # and set them to 'null'. Check the the docs for details:
- # helpVal4 = "Something1", # https://github.com/jokob-sk/NetAlertX/blob/main/docs/PLUGINS_DEV.md
- )
+ # The following are always unknown
+ watched1 = device['hostname'], # don't use these --> handleEmpty(device['hostname']),
+ watched2 = device['vendor'], # don't use these --> handleEmpty(device['vendor']),
+ watched3 = device['device_type'], # don't use these --> handleEmpty(device['device_type']),
+ extra = '',
+ foreignKey = "" # device['mac']
+ # helpVal1 = "Something1", # Optional Helper values to be passed for mapping into the app
+ # helpVal2 = "Something1", # If you need to use even only 1, add the remaining ones too
+ # helpVal3 = "Something1", # and set them to 'null'. Check the the docs for details:
+ # helpVal4 = "Something1", # https://github.com/jokob-sk/NetAlertX/blob/main/docs/PLUGINS_DEV.md
+ )
mylog('verbose', [f'[{pluginName}] New entries: "{len(neighbors)}"'])
@@ -83,13 +82,14 @@ def main():
return 0
+
def parse_neighbors(raw_neighbors: list[str]):
neighbors = []
for line in raw_neighbors:
if "lladdr" in line and "REACHABLE" in line:
# Known data
fields = line.split()
-
+
if not is_multicast(fields[0]):
# mylog('verbose', [f'[{pluginName}] adding ip {fields[0]}"'])
neighbor = {}
@@ -101,9 +101,9 @@ def parse_neighbors(raw_neighbors: list[str]):
neighbor['hostname'] = '(unknown)'
neighbor['vendor'] = '(unknown)'
neighbor['device_type'] = '(unknown)'
-
+
neighbors.append(neighbor)
-
+
return neighbors
@@ -111,6 +111,7 @@ def is_multicast(ip):
prefixes = ['ff', '224', '231', '232', '233', '234', '238', '239']
return reduce(lambda acc, prefix: acc or ip.startswith(prefix), prefixes, False)
+
# retrieve data
def get_neighbors(interfaces):
@@ -119,7 +120,7 @@ def get_neighbors(interfaces):
for interface in interfaces.split(","):
try:
- # Ping all IPv6 devices in multicast to trigger NDP
+ # Ping all IPv6 devices in multicast to trigger NDP
mylog('verbose', [f'[{pluginName}] Pinging on interface: "{interface}"'])
command = f"ping ff02::1%{interface} -c 2".split()
@@ -136,11 +137,11 @@ def get_neighbors(interfaces):
mylog('verbose', [f'[{pluginName}] Scanning interface succeded: "{interface}"'])
except subprocess.CalledProcessError as e:
# An error occurred, handle it
-
- mylog('verbose', [f'[{pluginName}] Scanning interface failed: "{interface}"'])
error_type = type(e).__name__ # Capture the error type
+ mylog('verbose', [f'[{pluginName}] Scanning interface failed: "{interface}" ({error_type})'])
return results
+
if __name__ == '__main__':
main()
diff --git a/front/plugins/luci_import/script.py b/front/plugins/luci_import/script.py
index a1adb837..692fa55b 100755
--- a/front/plugins/luci_import/script.py
+++ b/front/plugins/luci_import/script.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import sys
@@ -7,18 +7,18 @@ INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
pluginName = 'LUCIRPC'
-from plugin_helper import Plugin_Objects
-from logger import mylog, Logger
-from helper import get_setting_value
-from const import logPath
-import conf
-from pytz import timezone
+from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
try:
from openwrt_luci_rpc import OpenWrtRpc
-except:
- mylog('error', [f'[{pluginName}] Failed import openwrt_luci_rpc'])
- exit()
+except ImportError as e:
+ mylog('error', [f'[{pluginName}] Failed import openwrt_luci_rpc: {e}'])
+ exit(1)
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -30,13 +30,14 @@ RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
plugin_objects = Plugin_Objects(RESULT_FILE)
-def main():
- mylog('verbose', [f'[{pluginName}] start script.'])
+
+def main():
+ mylog('verbose', [f'[{pluginName}] start script.'])
device_data = get_device_data()
for entry in device_data:
- mylog('verbose', [f'[{pluginName}] found: ', str(entry.mac).lower()])
+ mylog('verbose', [f'[{pluginName}] found: ', str(entry.mac).lower()])
name = str(entry.hostname)
@@ -45,36 +46,38 @@ def main():
plugin_objects.add_object(
primaryId = str(entry.mac).lower(),
- secondaryId = entry.ip,
+ secondaryId = entry.ip,
watched1 = entry.host,
watched2 = name,
- watched3 = "",
+ watched3 = "",
watched4 = "",
- extra = pluginName,
+ extra = pluginName,
foreignKey = str(entry.mac).lower())
plugin_objects.write_result_file()
- mylog('verbose', [f'[{pluginName}] Script finished'])
-
+ mylog('verbose', [f'[{pluginName}] Script finished'])
+
return 0
+
def get_device_data():
router = OpenWrtRpc(
get_setting_value("LUCIRPC_host"),
- get_setting_value("LUCIRPC_user"),
- get_setting_value("LUCIRPC_password"),
- get_setting_value("LUCIRPC_ssl"),
+ get_setting_value("LUCIRPC_user"),
+ get_setting_value("LUCIRPC_password"),
+ get_setting_value("LUCIRPC_ssl"),
get_setting_value("LUCIRPC_verify_ssl")
- )
+ )
if router.is_logged_in():
- mylog('verbose', [f'[{pluginName}] login successfully.'])
+ mylog('verbose', [f'[{pluginName}] login successfully.'])
else:
- mylog('error', [f'[{pluginName}] login fail.'])
-
+ mylog('error', [f'[{pluginName}] login fail.'])
+
device_data = router.get_all_connected_devices(only_reachable=get_setting_value("LUCIRPC_only_reachable"))
return device_data
+
if __name__ == '__main__':
- main()
\ No newline at end of file
+ main()
diff --git a/front/plugins/maintenance/maintenance.py b/front/plugins/maintenance/maintenance.py
index 1785bb00..379f88a1 100755
--- a/front/plugins/maintenance/maintenance.py
+++ b/front/plugins/maintenance/maintenance.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import sys
@@ -8,12 +8,12 @@ from collections import deque
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from logger import mylog, Logger
-from helper import get_setting_value
-from const import logPath
-from messaging.in_app import remove_old
-import conf
-from pytz import timezone
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
+from messaging.in_app import remove_old # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -28,10 +28,9 @@ LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
-
def main():
- mylog('verbose', [f'[{pluginName}] In script'])
+ mylog('verbose', [f'[{pluginName}] In script'])
MAINT_LOG_LENGTH = int(get_setting_value('MAINT_LOG_LENGTH'))
MAINT_NOTI_LENGTH = int(get_setting_value('MAINT_NOTI_LENGTH'))
@@ -39,7 +38,7 @@ def main():
# Check if set
if MAINT_LOG_LENGTH != 0:
- mylog('verbose', [f'[{pluginName}] Cleaning file'])
+ mylog('verbose', [f'[{pluginName}] Cleaning file'])
logFile = logPath + "/app.log"
@@ -54,19 +53,19 @@ def main():
with open(logFile, 'w') as file:
# Write the last N lines back to the file
file.writelines(lines_to_keep)
-
- mylog('verbose', [f'[{pluginName}] Cleanup finished'])
+
+ mylog('verbose', [f'[{pluginName}] Cleanup finished'])
# Check if set
if MAINT_NOTI_LENGTH != 0:
- mylog('verbose', [f'[{pluginName}] Cleaning in-app notification history'])
+ mylog('verbose', [f'[{pluginName}] Cleaning in-app notification history'])
remove_old(MAINT_NOTI_LENGTH)
return 0
-#===============================================================================
+# ===============================================================================
# BEGIN
-#===============================================================================
+# ===============================================================================
if __name__ == '__main__':
- main()
\ No newline at end of file
+ main()
diff --git a/front/plugins/mikrotik_scan/mikrotik.py b/front/plugins/mikrotik_scan/mikrotik.py
index 5b446c9e..e25631aa 100755
--- a/front/plugins/mikrotik_scan/mikrotik.py
+++ b/front/plugins/mikrotik_scan/mikrotik.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import sys
@@ -7,14 +7,14 @@ import sys
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from plugin_helper import Plugin_Objects
-from logger import mylog, Logger
-from helper import get_setting_value
-from const import logPath
-import conf
-from pytz import timezone
-from librouteros import connect
-from librouteros.exceptions import TrapError
+from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
+from librouteros import connect # noqa: E402 [flake8 lint suppression]
+from librouteros.exceptions import TrapError # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -29,7 +29,6 @@ LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
-
def main():
mylog('verbose', [f'[{pluginName}] In script'])
@@ -49,7 +48,7 @@ def main():
plugin_objects = get_entries(plugin_objects)
plugin_objects.write_result_file()
-
+
mylog('verbose', [f'[{pluginName}] Scan finished, found {len(plugin_objects)} devices'])
@@ -58,10 +57,10 @@ def get_entries(plugin_objects: Plugin_Objects) -> Plugin_Objects:
try:
# connect router
api = connect(username=MT_USER, password=MT_PASS, host=MT_HOST, port=MT_PORT)
-
+
# get dhcp leases
leases = api('/ip/dhcp-server/lease/print')
-
+
for lease in leases:
lease_id = lease.get('.id')
address = lease.get('address')
@@ -71,8 +70,11 @@ def get_entries(plugin_objects: Plugin_Objects) -> Plugin_Objects:
last_seen = lease.get('last-seen')
status = lease.get('status')
device_name = comment or host_name or "(unknown)"
-
- mylog('verbose', [f"ID: {lease_id}, Address: {address}, MAC Address: {mac_address}, Host Name: {host_name}, Comment: {comment}, Last Seen: {last_seen}, Status: {status}"])
+
+ mylog(
+ 'verbose',
+ [f"ID: {lease_id}, Address: {address}, MAC Address: {mac_address}, Host Name: {host_name}, Comment: {comment}, Last Seen: {last_seen}, Status: {status}"]
+ )
if (status == "bound"):
plugin_objects.add_object(
@@ -83,7 +85,7 @@ def get_entries(plugin_objects: Plugin_Objects) -> Plugin_Objects:
watched3 = host_name,
watched4 = last_seen,
extra = '',
- helpVal1 = comment,
+ helpVal1 = comment,
foreignKey = mac_address)
except TrapError as e:
@@ -91,13 +93,13 @@ def get_entries(plugin_objects: Plugin_Objects) -> Plugin_Objects:
except Exception as e:
mylog('error', [f"Failed to connect to MikroTik API: {e}"])
- mylog('verbose', [f'[{pluginName}] Script finished'])
-
+ mylog('verbose', [f'[{pluginName}] Script finished'])
+
return plugin_objects
-#===============================================================================
+# ===============================================================================
# BEGIN
-#===============================================================================
+# ===============================================================================
if __name__ == '__main__':
main()
diff --git a/front/plugins/nbtscan_scan/nbtscan.py b/front/plugins/nbtscan_scan/nbtscan.py
index 505fbcda..729b4842 100755
--- a/front/plugins/nbtscan_scan/nbtscan.py
+++ b/front/plugins/nbtscan_scan/nbtscan.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import sys
@@ -8,14 +8,14 @@ import subprocess
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from plugin_helper import Plugin_Objects
-from logger import mylog, Logger
-from const import logPath
-from helper import get_setting_value
-from database import DB
-from models.device_instance import DeviceInstance
-import conf
-from pytz import timezone
+from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from database import DB # noqa: E402 [flake8 lint suppression]
+from models.device_instance import DeviceInstance # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -34,9 +34,8 @@ RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
plugin_objects = Plugin_Objects(RESULT_FILE)
-
def main():
- mylog('verbose', [f'[{pluginName}] In script'])
+ mylog('verbose', [f'[{pluginName}] In script'])
# timeout = get_setting_value('NBLOOKUP_RUN_TIMEOUT')
timeout = 20
@@ -52,13 +51,13 @@ def main():
device_handler = DeviceInstance(db)
# Retrieve devices
- if get_setting_value("REFRESH_FQDN"):
+ if get_setting_value("REFRESH_FQDN"):
devices = device_handler.getAll()
- else:
+ else:
devices = device_handler.getUnknown()
- mylog('verbose', [f'[{pluginName}] Devices count: {len(devices)}'])
-
+ mylog('verbose', [f'[{pluginName}] Devices count: {len(devices)}'])
+
# TEST - below is a WINDOWS host IP
# execute_name_lookup('192.168.1.121', timeout)
@@ -67,31 +66,32 @@ def main():
if domain_name != '':
plugin_objects.add_object(
- # "MAC", "IP", "Server", "Name"
- primaryId = device['devMac'],
- secondaryId = device['devLastIP'],
- watched1 = dns_server,
- watched2 = domain_name,
- watched3 = '',
- watched4 = '',
- extra = '',
- foreignKey = device['devMac'])
+ # "MAC", "IP", "Server", "Name"
+ primaryId = device['devMac'],
+ secondaryId = device['devLastIP'],
+ watched1 = dns_server,
+ watched2 = domain_name,
+ watched3 = '',
+ watched4 = '',
+ extra = '',
+ foreignKey = device['devMac']
+ )
plugin_objects.write_result_file()
-
-
- mylog('verbose', [f'[{pluginName}] Script finished'])
-
+
+ mylog('verbose', [f'[{pluginName}] Script finished'])
+
return 0
-#===============================================================================
+
+# ===============================================================================
# Execute scan
-#===============================================================================
-def execute_name_lookup (ip, timeout):
+# ===============================================================================
+def execute_name_lookup(ip, timeout):
"""
Execute the NBTSCAN command on IP.
"""
-
+
args = ['nbtscan', ip]
# Execute command
@@ -99,20 +99,25 @@ def execute_name_lookup (ip, timeout):
try:
mylog('verbose', [f'[{pluginName}] DEBUG CMD :', args])
-
+
# try runnning a subprocess with a forced (timeout) in case the subprocess hangs
- output = subprocess.check_output (args, universal_newlines=True, stderr=subprocess.STDOUT, timeout=(timeout), text=True)
+ output = subprocess.check_output(
+ args,
+ universal_newlines=True,
+ stderr=subprocess.STDOUT,
+ timeout=(timeout),
+ text=True
+ )
mylog('verbose', [f'[{pluginName}] DEBUG OUTPUT : {output}'])
-
+
domain_name = ''
dns_server = ''
-
+
# Split the output into lines
lines = output.splitlines()
# Look for the first line containing a valid NetBIOS name entry
- index = 0
for line in lines:
if 'Doing NBT name scan' not in line and ip in line:
# Split the line and extract the primary NetBIOS name
@@ -121,7 +126,6 @@ def execute_name_lookup (ip, timeout):
domain_name = parts[1]
else:
mylog('verbose', [f'[{pluginName}] ⚠ ERROR - Unexpected output format: {line}'])
-
mylog('verbose', [f'[{pluginName}] Domain Name: {domain_name}'])
@@ -132,18 +136,21 @@ def execute_name_lookup (ip, timeout):
# if "NXDOMAIN" in e.output:
# mylog('verbose', [f'[{pluginName}]', f"No PTR record found for IP: {ip}"])
# else:
- mylog('verbose', [f'[{pluginName}] ⚠ ERROR - {e.output}'])
-
- except subprocess.TimeoutExpired:
- mylog('verbose', [f'[{pluginName}] TIMEOUT - the process forcefully terminated as timeout reached'])
+ mylog('verbose', [f'[{pluginName}] ⚠ ERROR - {e.output}'])
- if output == "": # check if the subprocess failed
- mylog('verbose', [f'[{pluginName}] Scan: FAIL - check logs'])
- else:
+ except subprocess.TimeoutExpired:
+ mylog('verbose', [f'[{pluginName}] TIMEOUT - the process forcefully terminated as timeout reached'])
+
+ if output == "": # check if the subprocess failed
+ mylog('verbose', [f'[{pluginName}] Scan: FAIL - check logs'])
+ else:
mylog('verbose', [f'[{pluginName}] Scan: SUCCESS'])
- return '', ''
+ return '', ''
+
+# ===============================================================================
+# BEGIN
+# ===============================================================================
if __name__ == '__main__':
main()
-
diff --git a/front/plugins/nmap_dev_scan/nmap_dev.py b/front/plugins/nmap_dev_scan/nmap_dev.py
index 54d859ed..70641d5a 100755
--- a/front/plugins/nmap_dev_scan/nmap_dev.py
+++ b/front/plugins/nmap_dev_scan/nmap_dev.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
# test script by running:
# tbc
@@ -7,19 +7,18 @@ import subprocess
import sys
import hashlib
import re
-import nmap
+import nmap
# Register NetAlertX directories
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from plugin_helper import Plugin_Objects
-from logger import mylog, Logger
-from helper import get_setting_value
-from const import logPath
-from database import DB
-import conf
-from pytz import timezone
+from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -37,46 +36,46 @@ RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
def main():
- mylog('verbose', [f'[{pluginName}] In script'])
+ mylog('verbose', [f'[{pluginName}] In script'])
timeout = get_setting_value('NMAPDEV_RUN_TIMEOUT')
fakeMac = get_setting_value('NMAPDEV_FAKE_MAC')
subnets = get_setting_value('SCAN_SUBNETS')
args = get_setting_value('NMAPDEV_ARGS')
- mylog('verbose', [f'[{pluginName}] subnets: ', subnets])
-
+ mylog('verbose', [f'[{pluginName}] subnets: ', subnets])
# Initialize the Plugin obj output file
plugin_objects = Plugin_Objects(RESULT_FILE)
unique_devices = execute_scan(subnets, timeout, fakeMac, args)
- mylog('verbose', [f'[{pluginName}] Devices found: {len(unique_devices)}'])
+ mylog('verbose', [f'[{pluginName}] Devices found: {len(unique_devices)}'])
for device in unique_devices:
plugin_objects.add_object(
- # "MAC", "IP", "Name", "Vendor", "Interface"
- primaryId = device['mac'].lower(),
- secondaryId = device['ip'],
- watched1 = device['name'],
- watched2 = device['vendor'],
- watched3 = device['interface'],
- watched4 = '',
- extra = '',
- foreignKey = device['mac'])
+ # "MAC", "IP", "Name", "Vendor", "Interface"
+ primaryId = device['mac'].lower(),
+ secondaryId = device['ip'],
+ watched1 = device['name'],
+ watched2 = device['vendor'],
+ watched3 = device['interface'],
+ watched4 = '',
+ extra = '',
+ foreignKey = device['mac']
+ )
plugin_objects.write_result_file()
-
-
- mylog('verbose', [f'[{pluginName}] Script finished'])
-
+
+ mylog('verbose', [f'[{pluginName}] Script finished'])
+
return 0
-#===============================================================================
+
+# ===============================================================================
# Execute scan
-#===============================================================================
+# ===============================================================================
def execute_scan(subnets_list, timeout, fakeMac, args):
devices_list = []
@@ -103,22 +102,21 @@ def execute_scan(subnets_list, timeout, fakeMac, args):
return devices_list
-
-def execute_scan_on_interface (interface, timeout, args):
- # Remove unsupported VLAN flags
+def execute_scan_on_interface(interface, timeout, args):
+ # Remove unsupported VLAN flags
interface = re.sub(r'--vlan=\S+', '', interface).strip()
# Prepare command arguments
- scan_args = args.split() + interface.replace('--interface=','-e ').split()
+ scan_args = args.split() + interface.replace('--interface=', '-e ').split()
+
+ mylog('verbose', [f'[{pluginName}] scan_args: ', scan_args])
- mylog('verbose', [f'[{pluginName}] scan_args: ', scan_args])
-
try:
result = subprocess.check_output(scan_args, universal_newlines=True)
except subprocess.CalledProcessError as e:
error_type = type(e).__name__
result = ""
- mylog('verbose', [f'[{pluginName}] ERROR: ', error_type])
+ mylog('verbose', [f'[{pluginName}] ERROR: ', error_type])
return result
@@ -130,28 +128,25 @@ def parse_nmap_xml(xml_output, interface, fakeMac):
nm = nmap.PortScanner()
nm.analyse_nmap_xml_scan(xml_output)
- mylog('verbose', [f'[{pluginName}] Number of hosts: ', len(nm.all_hosts())])
+ mylog('verbose', [f'[{pluginName}] Number of hosts: ', len(nm.all_hosts())])
for host in nm.all_hosts():
- hostname = nm[host].hostname() or '(unknown)'
+ hostname = nm[host].hostname() or '(unknown)'
ip = nm[host]['addresses']['ipv4'] if 'ipv4' in nm[host]['addresses'] else ''
mac = nm[host]['addresses']['mac'] if 'mac' in nm[host]['addresses'] else ''
-
- mylog('verbose', [f'[{pluginName}] nm[host]: ', nm[host]])
+ mylog('verbose', [f'[{pluginName}] nm[host]: ', nm[host]])
vendor = ''
-
+
if nm[host]['vendor']:
- mylog('verbose', [f'[{pluginName}] entry: ', nm[host]['vendor']])
-
+ mylog('verbose', [f'[{pluginName}] entry: ', nm[host]['vendor']])
+
for key, value in nm[host]['vendor'].items():
vendor = value
-
break
-
# Log debug information
mylog('verbose', [f"[{pluginName}] Hostname: {hostname}, IP: {ip}, MAC: {mac}, Vendor: {vendor}"])
@@ -172,24 +167,24 @@ def parse_nmap_xml(xml_output, interface, fakeMac):
# MAC or IP missing
mylog('verbose', [f"[{pluginName}] Skipping: {hostname}, IP or MAC missing, or NMAPDEV_GENERATE_MAC setting not enabled"])
-
except Exception as e:
mylog('verbose', [f"[{pluginName}] Error parsing nmap XML: ", str(e)])
return devices_list
-
-
+
+
def string_to_mac_hash(input_string):
# Calculate a hash using SHA-256
sha256_hash = hashlib.sha256(input_string.encode()).hexdigest()
# Take the first 12 characters of the hash and format as a MAC address
- mac_hash = ':'.join(sha256_hash[i:i+2] for i in range(0, 12, 2))
-
+ mac_hash = ':'.join(sha256_hash[i:i + 2] for i in range(0, 12, 2))
+
return mac_hash
-#===============================================================================
+
+# ===============================================================================
# BEGIN
-#===============================================================================
+# ===============================================================================
if __name__ == '__main__':
- main()
\ No newline at end of file
+ main()
diff --git a/front/plugins/nmap_scan/script.py b/front/plugins/nmap_scan/script.py
index fab672ce..2d149d05 100755
--- a/front/plugins/nmap_scan/script.py
+++ b/front/plugins/nmap_scan/script.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import argparse
@@ -9,13 +9,13 @@ import subprocess
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from plugin_helper import Plugin_Objects
-from logger import mylog, Logger, append_line_to_file
-from utils.datetime_utils import timeNowDB
-from helper import get_setting_value
-from const import logPath
-import conf
-from pytz import timezone
+from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger, append_line_to_file # noqa: E402 [flake8 lint suppression]
+from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -32,7 +32,8 @@ RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
# Initialize the Plugin obj output file
plugin_objects = Plugin_Objects(RESULT_FILE)
-#-------------------------------------------------------------------------------
+
+# -------------------------------------------------------------------------------
def main():
parser = argparse.ArgumentParser(
description='Scan ports of devices specified by IP addresses'
@@ -85,7 +86,7 @@ def main():
mylog('verbose', [f'[{pluginName}] Total number of ports found by NMAP: ', len(entries)])
- for entry in entries:
+ for entry in entries:
plugin_objects.add_object(
primaryId = entry.mac, # MAC (Device Name)
@@ -94,14 +95,14 @@ def main():
watched2 = entry.service,
watched3 = entry.ip + ":" + entry.port,
watched4 = "",
- extra = entry.extra,
- foreignKey = entry.mac
+ extra = entry.extra,
+ foreignKey = entry.mac
)
plugin_objects.write_result_file()
-
-#-------------------------------------------------------------------------------
+
+# -------------------------------------------------------------------------------
class nmap_entry:
def __init__(self, ip, mac, time, port, state, service, name = '', extra = '', index = 0):
self.ip = ip
@@ -109,13 +110,13 @@ class nmap_entry:
self.time = time
self.port = port
self.state = state
- self.service = service
+ self.service = service
self.extra = extra
self.index = index
- self.hash = str(mac) + str(port)+ str(state)+ str(service)
+ self.hash = str(mac) + str(port) + str(state) + str(service)
-#-------------------------------------------------------------------------------
+# -------------------------------------------------------------------------------
def parse_kv_args(raw_args):
"""
Converts ['ips=a,b,c', 'macs=x,y,z', 'timeout=5'] to a dict.
@@ -125,26 +126,28 @@ def parse_kv_args(raw_args):
for item in raw_args:
if '=' not in item:
- mylog('none', [f"[{pluginName}] Scan: Invalid parameter (missing '='): {item}"])
+ mylog('none', [f"[{pluginName}] Scan: Invalid parameter (missing '='): {item}"])
key, value = item.split('=', 1)
if key in parsed:
- mylog('none', [f"[{pluginName}] Scan: Duplicate parameter supplied: {key}"])
+ mylog('none', [f"[{pluginName}] Scan: Duplicate parameter supplied: {key}"])
parsed[key] = value
return parsed
-#-------------------------------------------------------------------------------
+
+# -------------------------------------------------------------------------------
def safe_split_list(value, keyname):
"""Split comma list safely and ensure no empty items."""
items = [x.strip() for x in value.split(',') if x.strip()]
if not items:
- mylog('none', [f"[{pluginName}] Scan: {keyname} list is empty or invalid"])
+ mylog('none', [f"[{pluginName}] Scan: {keyname} list is empty or invalid"])
return items
-#-------------------------------------------------------------------------------
+
+# -------------------------------------------------------------------------------
def performNmapScan(deviceIPs, deviceMACs, timeoutSec, args):
"""
run nmap scan on a list of devices
@@ -154,15 +157,12 @@ def performNmapScan(deviceIPs, deviceMACs, timeoutSec, args):
# collect ports / new Nmap Entries
newEntriesTmp = []
-
- if len(deviceIPs) > 0:
+ if len(deviceIPs) > 0:
devTotal = len(deviceIPs)
-
-
- mylog('verbose', [f'[{pluginName}] Scan: Nmap for max ', str(timeoutSec), 's ('+ str(round(int(timeoutSec) / 60, 1)) +'min) per device'])
- mylog('verbose', ["[NMAP Scan] Estimated max delay: ", (devTotal * int(timeoutSec)), 's ', '(', round((devTotal * int(timeoutSec))/60,1) , 'min)' ])
+ mylog('verbose', [f'[{pluginName}] Scan: Nmap for max ', str(timeoutSec), 's (' + str(round(int(timeoutSec) / 60, 1)) + 'min) per device'])
+ mylog('verbose', ["[NMAP Scan] Estimated max delay: ", (devTotal * int(timeoutSec)), 's ', '(', round((devTotal * int(timeoutSec)) / 60, 1) , 'min)'])
devIndex = 0
for ip in deviceIPs:
@@ -171,67 +171,63 @@ def performNmapScan(deviceIPs, deviceMACs, timeoutSec, args):
# prepare arguments from user supplied ones
nmapArgs = ['nmap'] + args.split() + [ip]
- progress = ' (' + str(devIndex+1) + '/' + str(devTotal) + ')'
+ progress = ' (' + str(devIndex + 1) + '/' + str(devTotal) + ')'
try:
# try runnning a subprocess with a forced (timeout) in case the subprocess hangs
- output = subprocess.check_output (nmapArgs, universal_newlines=True, stderr=subprocess.STDOUT, timeout=(float(timeoutSec)))
+ output = subprocess.check_output(
+ nmapArgs,
+ universal_newlines=True,
+ stderr=subprocess.STDOUT,
+ timeout=(float(timeoutSec))
+ )
except subprocess.CalledProcessError as e:
# An error occured, handle it
- mylog('none', ["[NMAP Scan] " ,e.output])
- mylog('none', ["[NMAP Scan] ⚠ ERROR - Nmap Scan - check logs", progress])
+ mylog('none', ["[NMAP Scan] ", e.output])
+ mylog('none', ["[NMAP Scan] ⚠ ERROR - Nmap Scan - check logs", progress])
except subprocess.TimeoutExpired:
- mylog('verbose', [f'[{pluginName}] Nmap TIMEOUT - the process forcefully terminated as timeout reached for ', ip, progress])
+ mylog('verbose', [f'[{pluginName}] Nmap TIMEOUT - the process forcefully terminated as timeout reached for ', ip, progress])
- if output == "": # check if the subprocess failed
- mylog('minimal', [f'[{pluginName}] Nmap FAIL for ', ip, progress ,' check logs for details'])
- else:
+ if output == "": # check if the subprocess failed
+ mylog('minimal', [f'[{pluginName}] Nmap FAIL for ', ip, progress, ' check logs for details'])
+ else:
mylog('verbose', [f'[{pluginName}] Nmap SUCCESS for ', ip, progress])
-
-
- # check the last run output
+ # check the last run output
newLines = output.split('\n')
# regular logging
for line in newLines:
- append_line_to_file (logPath + '/app_nmap.log', line +'\n')
-
+ append_line_to_file(logPath + '/app_nmap.log', line + '\n')
index = 0
startCollecting = False
- duration = ""
+ duration = ""
newPortsPerDevice = 0
- for line in newLines:
+ for line in newLines:
if 'Starting Nmap' in line:
- if len(newLines) > index+1 and 'Note: Host seems down' in newLines[index+1]:
- break # this entry is empty
+ if len(newLines) > index + 1 and 'Note: Host seems down' in newLines[index + 1]:
+ break # this entry is empty
elif 'PORT' in line and 'STATE' in line and 'SERVICE' in line:
startCollecting = True
- elif 'PORT' in line and 'STATE' in line and 'SERVICE' in line:
- startCollecting = False # end reached
- elif startCollecting and len(line.split()) == 3:
+ elif 'PORT' in line and 'STATE' in line and 'SERVICE' in line:
+ startCollecting = False # end reached
+ elif startCollecting and len(line.split()) == 3:
newEntriesTmp.append(nmap_entry(ip, deviceMACs[devIndex], timeNowDB(), line.split()[0], line.split()[1], line.split()[2]))
newPortsPerDevice += 1
elif 'Nmap done' in line:
- duration = line.split('scanned in ')[1]
-
- mylog('verbose', [f'[{pluginName}] {newPortsPerDevice} ports found on {deviceMACs[devIndex]}'])
+ duration = line.split('scanned in ')[1]
+
+ mylog('verbose', [f'[{pluginName}] {newPortsPerDevice} ports found on {deviceMACs[devIndex]} after {duration}'])
index += 1
devIndex += 1
-
-
- #end for loop
-
return newEntriesTmp
-#===============================================================================
+
+# ===============================================================================
# BEGIN
-#===============================================================================
+# ===============================================================================
if __name__ == '__main__':
main()
-
-
-
diff --git a/front/plugins/nslookup_scan/nslookup.py b/front/plugins/nslookup_scan/nslookup.py
index 5e169bac..5fd3360c 100755
--- a/front/plugins/nslookup_scan/nslookup.py
+++ b/front/plugins/nslookup_scan/nslookup.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
# test script by running:
# tbc
@@ -11,14 +11,14 @@ import re
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from plugin_helper import Plugin_Objects
-from logger import mylog, Logger
-from helper import get_setting_value
-from const import logPath
-from database import DB
-from models.device_instance import DeviceInstance
-import conf
-from pytz import timezone
+from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
+from database import DB # noqa: E402 [flake8 lint suppression]
+from models.device_instance import DeviceInstance # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -33,11 +33,9 @@ LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
-
def main():
- mylog('verbose', [f'[{pluginName}] In script'])
-
+ mylog('verbose', [f'[{pluginName}] In script'])
timeout = get_setting_value('NSLOOKUP_RUN_TIMEOUT')
@@ -52,13 +50,13 @@ def main():
device_handler = DeviceInstance(db)
# Retrieve devices
- if get_setting_value("REFRESH_FQDN"):
+ if get_setting_value("REFRESH_FQDN"):
devices = device_handler.getAll()
- else:
+ else:
devices = device_handler.getUnknown()
- mylog('verbose', [f'[{pluginName}] Devices count: {len(devices)}'])
-
+ mylog('verbose', [f'[{pluginName}] Devices count: {len(devices)}'])
+
# TEST - below is a WINDOWS host IP
# execute_name_lookup('192.168.1.121', timeout)
@@ -67,31 +65,32 @@ def main():
if domain_name != '':
plugin_objects.add_object(
- # "MAC", "IP", "Server", "Name"
- primaryId = device['devMac'],
- secondaryId = device['devLastIP'],
- watched1 = dns_server,
- watched2 = domain_name,
- watched3 = '',
- watched4 = '',
- extra = '',
- foreignKey = device['devMac'])
+ # "MAC", "IP", "Server", "Name"
+ primaryId = device['devMac'],
+ secondaryId = device['devLastIP'],
+ watched1 = dns_server,
+ watched2 = domain_name,
+ watched3 = '',
+ watched4 = '',
+ extra = '',
+ foreignKey = device['devMac']
+ )
plugin_objects.write_result_file()
-
-
- mylog('verbose', [f'[{pluginName}] Script finished'])
-
+
+ mylog('verbose', [f'[{pluginName}] Script finished'])
+
return 0
-#===============================================================================
+
+# ===============================================================================
# Execute scan
-#===============================================================================
-def execute_nslookup (ip, timeout):
+# ===============================================================================
+def execute_nslookup(ip, timeout):
"""
Execute the NSLOOKUP command on IP.
"""
-
+
nslookup_args = ['nslookup', ip]
# Execute command
@@ -99,7 +98,13 @@ def execute_nslookup (ip, timeout):
try:
# try runnning a subprocess with a forced (timeout) in case the subprocess hangs
- output = subprocess.check_output (nslookup_args, universal_newlines=True, stderr=subprocess.STDOUT, timeout=(timeout), text=True)
+ output = subprocess.check_output(
+ nslookup_args,
+ universal_newlines=True,
+ stderr=subprocess.STDOUT,
+ timeout=(timeout),
+ text=True
+ )
domain_name = ''
dns_server = ''
@@ -110,8 +115,7 @@ def execute_nslookup (ip, timeout):
domain_pattern = re.compile(r'name\s*=\s*([^\s]+)', re.IGNORECASE)
server_pattern = re.compile(r'Server:\s+(.+)', re.IGNORECASE)
-
- domain_match = domain_pattern.search(output)
+ domain_match = domain_pattern.search(output)
server_match = server_pattern.search(output)
if domain_match:
@@ -131,24 +135,20 @@ def execute_nslookup (ip, timeout):
else:
mylog('verbose', [f'[{pluginName}]', e.output])
# Handle other errors here
- # mylog('verbose', [f'[{pluginName}] ⚠ ERROR - check logs'])
-
- except subprocess.TimeoutExpired:
- mylog('verbose', [f'[{pluginName}] TIMEOUT - the process forcefully terminated as timeout reached'])
+ # mylog('verbose', [f'[{pluginName}] ⚠ ERROR - check logs'])
+
+ except subprocess.TimeoutExpired:
+ mylog('verbose', [f'[{pluginName}] TIMEOUT - the process forcefully terminated as timeout reached'])
+
+ if output != "": # check if the subprocess failed
- if output == "": # check if the subprocess failed
- tmp = 1 # can't have empty
- # mylog('verbose', [f'[{pluginName}] Scan: FAIL - check logs'])
- else:
mylog('verbose', [f'[{pluginName}] Scan: SUCCESS'])
- return '', ''
-
-
-
+ return '', ''
-#===============================================================================
+
+# ===============================================================================
# BEGIN
-#===============================================================================
+# ===============================================================================
if __name__ == '__main__':
- main()
\ No newline at end of file
+ main()
diff --git a/front/plugins/omada_sdn_imp/omada_sdn.py b/front/plugins/omada_sdn_imp/omada_sdn.py
index b8d80b11..0957f163 100755
--- a/front/plugins/omada_sdn_imp/omada_sdn.py
+++ b/front/plugins/omada_sdn_imp/omada_sdn.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
__author__ = "ffsb"
__version__ = "0.1" # initial
__version__ = "0.2" # added logic to retry omada api call once as it seems to sometimes fail for some reasons, and error handling logic...
@@ -15,10 +15,9 @@ __version__ = "1.3" # fix detection of the default gateway IP address that woul
# try to identify and populate their connections by switch/accesspoints and ports/SSID
# try to differentiate root bridges from accessory
-
-#
# sample code to update unbound on opnsense - for reference...
-# curl -X POST -d '{"host":{"enabled":"1","hostname":"test","domain":"testdomain.com","rr":"A","mxprio":"","mx":"","server":"10.0.1.1","description":""}}' -H "Content-Type: application/json" -k -u $OPNS_KEY:$OPNS_SECRET https://$IPFW/api/unbound/settings/AddHostOverride
+# curl -X POST -d '{"host":{"enabled":"1","hostname":"test","domain":"testdomain.com","rr":"A","mxprio":"","mx":"","server":"10.0.1.1","description":""}}'\
+# -H "Content-Type: application/json" -k -u $OPNS_KEY:$OPNS_SECRET https://$IPFW/api/unbound/settings/AddHostOverride
#
import os
import sys
@@ -35,12 +34,12 @@ import multiprocessing
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from plugin_helper import Plugin_Objects
-from logger import mylog, Logger
-from const import logPath
-from helper import get_setting_value
-from pytz import timezone
-import conf
+from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -87,8 +86,6 @@ cMAC, cIP, cNAME, cSWITCH_AP, cPORT_SSID = range(5)
OMDLOGLEVEL = "debug"
-
-#
# translate MAC address from standard ieee model to ietf draft
# AA-BB-CC-DD-EE-FF to aa:bb:cc:dd:ee:ff
# tplink adheres to ieee, Nax adheres to ietf
@@ -142,7 +139,7 @@ def callomada(myargs):
try:
mf = io.StringIO()
with redirect_stdout(mf):
- bar = omada(myargs)
+ omada(myargs)
omada_output = mf.getvalue()
except Exception:
mylog(
@@ -190,12 +187,12 @@ def add_uplink(
if switch_mac not in device_data_bymac:
mylog("none", [f"[{pluginName}] switch_mac '{switch_mac}' not found in device_data_bymac"])
return
-
+
# Ensure SWITCH_AP key exists in the dictionary
if SWITCH_AP not in device_data_bymac[switch_mac]:
mylog("none", [f"[{pluginName}] Missing key '{SWITCH_AP}' in device_data_bymac[{switch_mac}]"])
return
-
+
# Check if uplink should be added
if device_data_bymac[switch_mac][SWITCH_AP] in [None, "null"]:
device_data_bymac[switch_mac][SWITCH_AP] = uplink_mac
@@ -204,11 +201,10 @@ def add_uplink(
if uplink_mac not in device_data_bymac:
mylog("none", [f"[{pluginName}] uplink_mac '{uplink_mac}' not found in device_data_bymac"])
return
-
+
# Determine port to uplink
if (
- device_data_bymac[switch_mac].get(TYPE) == "Switch"
- and device_data_bymac[uplink_mac].get(TYPE) == "Switch"
+ device_data_bymac[switch_mac].get(TYPE) == "Switch" and device_data_bymac[uplink_mac].get(TYPE) == "Switch"
):
port_to_uplink = port_byswitchmac_byclientmac.get(switch_mac, {}).get(uplink_mac)
if port_to_uplink is None:
@@ -216,16 +212,14 @@ def add_uplink(
return
else:
port_to_uplink = device_data_bymac[uplink_mac].get(PORT_SSID)
-
+
# Assign port to switch_mac
device_data_bymac[switch_mac][PORT_SSID] = port_to_uplink
-
+
# Recursively add uplinks for linked devices
for link in sadevices_linksbymac.get(switch_mac, []):
if (
- link in device_data_bymac
- and device_data_bymac[link].get(SWITCH_AP) in [None, "null"]
- and device_data_bymac[switch_mac].get(TYPE) == "Switch"
+ link in device_data_bymac and device_data_bymac[link].get(SWITCH_AP) in [None, "null"] and device_data_bymac[switch_mac].get(TYPE) == "Switch"
):
add_uplink(
switch_mac,
@@ -236,7 +230,6 @@ def add_uplink(
)
-
# ----------------------------------------------
# Main initialization
def main():
@@ -324,16 +317,16 @@ def main():
)
mymac = ieee2ietf_mac_formater(device[MAC])
plugin_objects.add_object(
- primaryId=mymac, # MAC
- secondaryId=device[IP], # IP
- watched1=device[NAME], # NAME/HOSTNAME
- watched2=ParentNetworkNode, # PARENT NETWORK NODE MAC
- watched3=myport, # PORT
- watched4=myssid, # SSID
+ primaryId=mymac, # MAC
+ secondaryId=device[IP], # IP
+ watched1=device[NAME], # NAME/HOSTNAME
+ watched2=ParentNetworkNode, # PARENT NETWORK NODE MAC
+ watched3=myport, # PORT
+ watched4=myssid, # SSID
extra=device[TYPE],
# omada_site, # SITENAME (cur_NetworkSite) or VENDOR (cur_Vendor) (PICK one and adjust config.json -> "column": "Extra")
foreignKey=device[MAC].lower().replace("-", ":"),
- ) # usually MAC
+ ) # usually MAC
mylog(
"verbose",
@@ -369,7 +362,6 @@ def get_omada_devices_details(msadevice_data):
mswitch_dump = callomada(["-t", "myomada", "switch", "-d", mthisswitch])
else:
mswitch_detail = ""
- nswitch_dump = ""
return mswitch_detail, mswitch_dump
@@ -414,7 +406,6 @@ def get_device_data(omada_clients_output, switches_and_aps, device_handler):
# 17:27:10 [] token: "['1A-2B-3C-4D-5E-6F', '192.168.0.217', '1A-2B-3C-4D-5E-6F', '17', '40-AE-30-A5-A7-50, 'Switch']"
# constants
sadevices_macbyname = {}
- sadevices_macbymac = {}
sadevices_linksbymac = {}
port_byswitchmac_byclientmac = {}
device_data_bymac = {}
@@ -427,7 +418,7 @@ def get_device_data(omada_clients_output, switches_and_aps, device_handler):
def run_command(command, index):
result = subprocess.run(command, capture_output=True, text=True, shell=True)
return str(index), result.stdout.strip()
-
+
myindex, command_output= run_command(command, 2)
mylog('verbose', [f'[{pluginName}] command={command} index={myindex} results={command_output}'])
"""
@@ -556,11 +547,11 @@ def get_device_data(omada_clients_output, switches_and_aps, device_handler):
#
naxname = real_naxname
- if real_naxname != None:
+ if real_naxname is not None:
if "(" in real_naxname:
# removing parenthesis and domains from the name
naxname = real_naxname.split("(")[0]
- if naxname != None and "." in naxname:
+ if naxname is not None and "." in naxname:
naxname = naxname.split(".")[0]
if naxname in (None, "null", ""):
naxname = (
diff --git a/front/plugins/omada_sdn_openapi/script.py b/front/plugins/omada_sdn_openapi/script.py
index 90bd0068..a15af4f5 100755
--- a/front/plugins/omada_sdn_openapi/script.py
+++ b/front/plugins/omada_sdn_openapi/script.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
"""
This plugin imports devices and clients from Omada Controller using their OpenAPI.
@@ -25,7 +25,6 @@ import sys
import urllib3
import requests
import time
-import datetime
import pytz
from datetime import datetime
@@ -35,11 +34,11 @@ from typing import Literal, Any, Dict
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from plugin_helper import Plugin_Objects, is_typical_router_ip, is_mac
-from logger import mylog, Logger
-from const import logPath
-from helper import get_setting_value
-import conf
+from plugin_helper import Plugin_Objects, is_typical_router_ip, is_mac # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = pytz.timezone(get_setting_value('TIMEZONE'))
@@ -176,7 +175,10 @@ class OmadaHelper:
# If it's not a gateway try to assign parent node MAC
if data.get("type", "") != "gateway":
parent_mac = OmadaHelper.normalize_mac(data.get("uplinkDeviceMac"))
- entry["parent_node_mac_address"] = parent_mac.get("response_result") if isinstance(parent_mac, dict) and parent_mac.get("response_type") == "success" else ""
+
+ resp_type = parent_mac.get("response_type")
+
+ entry["parent_node_mac_address"] = parent_mac.get("response_result") if isinstance(parent_mac, dict) and resp_type == "success" else ""
# Applicable only for CLIENT
if input_type == "client":
@@ -185,15 +187,26 @@ class OmadaHelper:
# Try to assign parent node MAC and PORT/SSID to the CLIENT
if data.get("connectDevType", "") == "gateway":
parent_mac = OmadaHelper.normalize_mac(data.get("gatewayMac"))
- entry["parent_node_mac_address"] = parent_mac.get("response_result") if isinstance(parent_mac, dict) and parent_mac.get("response_type") == "success" else ""
+
+ resp_type = parent_mac.get("response_type")
+
+ entry["parent_node_mac_address"] = parent_mac.get("response_result") if isinstance(parent_mac, dict) and resp_type == "success" else ""
entry["parent_node_port"] = data.get("port", "")
+
elif data.get("connectDevType", "") == "switch":
parent_mac = OmadaHelper.normalize_mac(data.get("switchMac"))
- entry["parent_node_mac_address"] = parent_mac.get("response_result") if isinstance(parent_mac, dict) and parent_mac.get("response_type") == "success" else ""
+
+ resp_type = parent_mac.get("response_type")
+
+ entry["parent_node_mac_address"] = parent_mac.get("response_result") if isinstance(parent_mac, dict) and resp_type == "success" else ""
entry["parent_node_port"] = data.get("port", "")
+
elif data.get("connectDevType", "") == "ap":
parent_mac = OmadaHelper.normalize_mac(data.get("apMac"))
- entry["parent_node_mac_address"] = parent_mac.get("response_result") if isinstance(parent_mac, dict) and parent_mac.get("response_type") == "success" else ""
+
+ resp_type = parent_mac.get("response_type")
+
+ entry["parent_node_mac_address"] = parent_mac.get("response_result") if isinstance(parent_mac, dict) and resp_type == "success" else ""
entry["parent_node_ssid"] = data.get("ssid", "")
# Add the entry to the result
@@ -253,7 +266,7 @@ class OmadaAPI:
"""Return request headers."""
headers = {"Content-type": "application/json"}
# Add access token to header if requested and available
- if include_auth == True:
+ if include_auth is True:
if not self.access_token:
OmadaHelper.debug("No access token available for headers")
else:
@@ -368,7 +381,7 @@ class OmadaAPI:
# Failed site population
OmadaHelper.debug(f"Site population response: {response}")
- return OmadaHelper.response("error", f"Site population failed - error: {response.get('response_message', 'Not provided')}")
+ return OmadaHelper.response("error", f"Site population failed - error: {response.get('response_message', 'Not provided')}")
def requested_sites(self) -> list:
"""Returns sites requested by user."""
diff --git a/front/plugins/pihole_api_scan/pihole_api_scan.py b/front/plugins/pihole_api_scan/pihole_api_scan.py
index a6b08baf..37a01d49 100644
--- a/front/plugins/pihole_api_scan/pihole_api_scan.py
+++ b/front/plugins/pihole_api_scan/pihole_api_scan.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
"""
NetAlertX plugin: PIHOLEAPI
Imports devices from Pi-hole v6 API (Network endpoints) into NetAlertX plugin results.
@@ -17,12 +17,12 @@ sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
pluginName = 'PIHOLEAPI'
-from plugin_helper import Plugin_Objects, is_mac
-from logger import mylog, Logger
-from helper import get_setting_value
-from const import logPath
-import conf
-from pytz import timezone
+from plugin_helper import Plugin_Objects, is_mac # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
# Setup timezone & logger using standard NAX helpers
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -179,7 +179,7 @@ def get_pihole_network_devices():
resp = requests.get(PIHOLEAPI_URL + 'api/network/devices', headers=headers, params=params, verify=PIHOLEAPI_VERIFY_SSL, timeout=PIHOLEAPI_RUN_TIMEOUT)
resp.raise_for_status()
data = resp.json()
-
+
mylog('debug', [f'[{pluginName}] Pi-hole API returned data: {json.dumps(data)}'])
except Exception as e:
@@ -267,7 +267,7 @@ def main():
for entry in device_entries:
if is_mac(entry['mac']):
- # Map to Plugin_Objects fields
+ # Map to Plugin_Objects fields
mylog('verbose', [f'[{pluginName}] found: {entry['name']}|{entry['mac']}|{entry['ip']}'])
plugin_objects.add_object(
diff --git a/front/plugins/plugin_helper.py b/front/plugins/plugin_helper.py
index bb501cdd..45c867f5 100755
--- a/front/plugins/plugin_helper.py
+++ b/front/plugins/plugin_helper.py
@@ -5,18 +5,18 @@ import os
import re
import base64
import json
-from datetime import datetime
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.append(f"{INSTALL_PATH}/front/plugins")
-sys.path.append(f'{INSTALL_PATH}/server')
+sys.path.append(f'{INSTALL_PATH}/server')
-from logger import mylog, Logger
-from utils.datetime_utils import timeNowDB
-from const import default_tz, fullConfPath
+from logger import mylog # noqa: E402 [flake8 lint suppression]
+from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
+from const import default_tz, fullConfPath # noqa: E402 [flake8 lint suppression]
-#-------------------------------------------------------------------------------
+
+# -------------------------------------------------------------------------------
def read_config_file():
"""
retuns dict on the config file key:value pairs
@@ -25,15 +25,15 @@ def read_config_file():
filename = fullConfPath
-
print('[plugin_helper] reading config file')
+
# load the variables from .conf
with open(filename, "r") as file:
code = compile(file.read(), filename, "exec")
- confDict = {} # config dictionary
+ confDict = {} # config dictionary
exec(code, {"__builtins__": {}}, confDict)
- return confDict
+ return confDict
configFile = read_config_file()
@@ -42,6 +42,7 @@ if timeZoneSetting not in all_timezones:
timeZoneSetting = default_tz
timeZone = pytz.timezone(timeZoneSetting)
+
# -------------------------------------------------------------------
# Sanitizes plugin output
def handleEmpty(input):
@@ -54,70 +55,72 @@ def handleEmpty(input):
input = re.sub(r'[^\x00-\x7F]+', ' ', input)
input = input.replace('\n', '') # Removing new lines
return input
-
+
+
# -------------------------------------------------------------------
# Sanitizes string
def rmBadChars(input):
-
+
input = handleEmpty(input)
input = input.replace("'", '_') # Removing ' (single quotes)
-
+
return input
+
# -------------------------------------------------------------------
# check if this is a router IP
def is_typical_router_ip(ip_address):
- # List of common default gateway IP addresses
- common_router_ips = [
- "192.168.0.1", "192.168.1.1", "192.168.1.254", "192.168.0.254",
- "10.0.0.1", "10.1.1.1", "192.168.2.1", "192.168.10.1", "192.168.11.1",
- "192.168.100.1", "192.168.101.1", "192.168.123.254", "192.168.223.1",
- "192.168.31.1", "192.168.8.1", "192.168.254.254", "192.168.50.1",
- "192.168.3.1", "192.168.4.1", "192.168.5.1", "192.168.9.1",
- "192.168.15.1", "192.168.16.1", "192.168.20.1", "192.168.30.1",
- "192.168.42.1", "192.168.62.1", "192.168.178.1", "192.168.1.1",
- "192.168.1.254", "192.168.0.1", "192.168.0.10", "192.168.0.100",
- "192.168.0.254"
- ]
-
- return ip_address in common_router_ips
+ # List of common default gateway IP addresses
+ common_router_ips = [
+ "192.168.0.1", "192.168.1.1", "192.168.1.254", "192.168.0.254",
+ "10.0.0.1", "10.1.1.1", "192.168.2.1", "192.168.10.1", "192.168.11.1",
+ "192.168.100.1", "192.168.101.1", "192.168.123.254", "192.168.223.1",
+ "192.168.31.1", "192.168.8.1", "192.168.254.254", "192.168.50.1",
+ "192.168.3.1", "192.168.4.1", "192.168.5.1", "192.168.9.1",
+ "192.168.15.1", "192.168.16.1", "192.168.20.1", "192.168.30.1",
+ "192.168.42.1", "192.168.62.1", "192.168.178.1", "192.168.1.1",
+ "192.168.1.254", "192.168.0.1", "192.168.0.10", "192.168.0.100",
+ "192.168.0.254"
+ ]
+
+ return ip_address in common_router_ips
+
# -------------------------------------------------------------------
# Check if a valid MAC address
def is_mac(input):
input_str = str(input).lower() # Convert to string and lowercase so non-string values won't raise errors
-
+
isMac = bool(re.match("[0-9a-f]{2}([-:]?)[0-9a-f]{2}(\\1[0-9a-f]{2}){4}$", input_str))
-
+
if not isMac: # If it's not a MAC address, log the input
mylog('verbose', [f'[is_mac] not a MAC: {input_str}'])
-
+
return isMac
+
# -------------------------------------------------------------------
def decodeBase64(inputParamBase64):
# Printing the input list to check its content.
- mylog('debug', ['[Plugins] Helper base64 input: ', input])
- print('[Plugins] Helper base64 input: ')
- print(input)
-
-
- # Extract the base64-encoded subnet information from the first element
- # The format of the element is assumed to be like 'param=b'.
+ mylog('debug', ['[Plugins] Helper base64 input: ', input])
+ print('[Plugins] Helper base64 input: ')
+ print(input)
+ # Extract the base64-encoded subnet information from the first element
+ # The format of the element is assumed to be like 'param=b'.
# Printing the extracted base64-encoded information.
- mylog('debug', ['[Plugins] Helper base64 inputParamBase64: ', inputParamBase64])
-
+ mylog('debug', ['[Plugins] Helper base64 inputParamBase64: ', inputParamBase64])
# Decode the base64-encoded subnet information to get the actual subnet information in ASCII format.
result = base64.b64decode(inputParamBase64).decode('ascii')
# Print the decoded subnet information.
- mylog('debug', ['[Plugins] Helper base64 result: ', result])
+ mylog('debug', ['[Plugins] Helper base64 result: ', result])
return result
+
# -------------------------------------------------------------------
def decode_settings_base64(encoded_str, convert_types=True):
"""
@@ -167,7 +170,7 @@ def decode_settings_base64(encoded_str, convert_types=True):
def normalize_mac(mac):
# Split the MAC address by colon (:) or hyphen (-) and convert each part to uppercase
parts = mac.upper().split(':')
-
+
# If the MAC address is split by hyphen instead of colon
if len(parts) == 1:
parts = mac.upper().split('-')
@@ -177,14 +180,15 @@ def normalize_mac(mac):
# Join the parts with colon (:)
normalized_mac = ':'.join(normalized_parts)
-
+
return normalized_mac
+
# -------------------------------------------------------------------
class Plugin_Object:
- """
+ """
Plugin_Object class to manage one object introduced by the plugin.
- An object typically is a device but could also be a website or something
+ An object typically is a device but could also be a website or something
else that is monitored by the plugin.
"""
@@ -222,8 +226,8 @@ class Plugin_Object:
self.helpVal4 = helpVal4 or ""
def write(self):
- """
- Write the object details as a string in the
+ """
+ Write the object details as a string in the
format required to write the result file.
"""
line = "{}|{}|{}|{}|{}|{}|{}|{}|{}|{}|{}|{}|{}\n".format(
@@ -243,6 +247,7 @@ class Plugin_Object:
)
return line
+
class Plugin_Objects:
"""
Plugin_Objects is the class that manages and holds all the objects created by the plugin.
@@ -303,7 +308,3 @@ class Plugin_Objects:
def __len__(self):
return len(self.objects)
-
-
-
-
diff --git a/front/plugins/snmp_discovery/script.py b/front/plugins/snmp_discovery/script.py
index c85e5fe4..a0583e8f 100755
--- a/front/plugins/snmp_discovery/script.py
+++ b/front/plugins/snmp_discovery/script.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
from __future__ import unicode_literals
import subprocess
@@ -10,12 +10,12 @@ import sys
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from plugin_helper import Plugin_Objects, handleEmpty, normalize_mac
-from logger import mylog, Logger
-from helper import get_setting_value
-from const import logPath
-import conf
-from pytz import timezone
+from plugin_helper import Plugin_Objects, handleEmpty, normalize_mac # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -28,53 +28,60 @@ pluginName = "SNMPDSC"
LOG_PATH = logPath + '/plugins'
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
-# Workflow
-def main():
- mylog('verbose', ['[SNMPDSC] In script '])
+def main():
+ mylog('verbose', ['[SNMPDSC] In script '])
# init global variables
global snmpWalkCmds
-
- parser = argparse.ArgumentParser(description='This plugin is used to discover devices via the arp table(s) of a RFC1213 compliant router or switch.')
- parser.add_argument('routers', action="store", help="IP(s) of routers, separated by comma (,) if passing multiple")
+ parser = argparse.ArgumentParser(description='This plugin is used to discover devices via the arp table(s) of a RFC1213 compliant router or switch.')
+ parser.add_argument(
+ 'routers',
+ action="store",
+ help="IP(s) of routers, separated by comma (,) if passing multiple"
+ )
+
values = parser.parse_args()
timeoutSetting = get_setting_value("SNMPDSC_RUN_TIMEOUT")
plugin_objects = Plugin_Objects(RESULT_FILE)
- if values.routers:
- snmpWalkCmds = values.routers.split('=')[1].replace('\'','')
-
+ if values.routers:
+ snmpWalkCmds = values.routers.split('=')[1].replace('\'', '')
if ',' in snmpWalkCmds:
commands = snmpWalkCmds.split(',')
else:
commands = [snmpWalkCmds]
-
+
for cmd in commands:
- mylog('verbose', ['[SNMPDSC] Router snmpwalk command: ', cmd])
+ mylog('verbose', ['[SNMPDSC] Router snmpwalk command: ', cmd])
# split the string, remove white spaces around each item, and exclude any empty strings
snmpwalkArgs = [arg.strip() for arg in cmd.split(' ') if arg.strip()]
# Execute N probes and insert in list
probes = 1 # N probes
-
- for _ in range(probes):
- output = subprocess.check_output (snmpwalkArgs, universal_newlines=True, stderr=subprocess.STDOUT, timeout=(timeoutSetting))
- mylog('verbose', ['[SNMPDSC] output: ', output])
+ for _ in range(probes):
+ output = subprocess.check_output(
+ snmpwalkArgs,
+ universal_newlines=True,
+ stderr=subprocess.STDOUT,
+ timeout=(timeoutSetting)
+ )
+
+ mylog('verbose', ['[SNMPDSC] output: ', output])
lines = output.split('\n')
- for line in lines:
+ for line in lines:
- tmpSplt = line.split('"')
+ tmpSplt = line.split('"')
if len(tmpSplt) == 3:
-
+
ipStr = tmpSplt[0].split('.')[-4:] # Get the last 4 elements to extract the IP
macStr = tmpSplt[1].strip().split(' ') # Remove leading/trailing spaces from MAC
@@ -82,19 +89,18 @@ def main():
macAddress = ':'.join(macStr)
ipAddress = '.'.join(ipStr)
- mylog('verbose', [f'[SNMPDSC] IP: {ipAddress} MAC: {macAddress}'])
-
+ mylog('verbose', [f'[SNMPDSC] IP: {ipAddress} MAC: {macAddress}'])
+
plugin_objects.add_object(
primaryId = handleEmpty(macAddress),
- secondaryId = handleEmpty(ipAddress.strip()), # Remove leading/trailing spaces from IP
+ secondaryId = handleEmpty(ipAddress.strip()), # Remove leading/trailing spaces from IP
watched1 = '(unknown)',
watched2 = handleEmpty(snmpwalkArgs[6]), # router IP
extra = handleEmpty(line),
foreignKey = handleEmpty(macAddress) # Use the primary ID as the foreign key
)
else:
- mylog('verbose', ['[SNMPDSC] ipStr does not seem to contain a valid IP:', ipStr])
-
+ mylog('verbose', ['[SNMPDSC] ipStr does not seem to contain a valid IP:', ipStr])
elif line.startswith('ipNetToMediaPhysAddress'):
# Format: snmpwalk -OXsq output
@@ -115,12 +121,11 @@ def main():
foreignKey = handleEmpty(macAddress)
)
- mylog('verbose', ['[SNMPDSC] Entries found: ', len(plugin_objects)])
+ mylog('verbose', ['[SNMPDSC] Entries found: ', len(plugin_objects)])
plugin_objects.write_result_file()
-
# BEGIN
-if __name__ == '__main__':
+if __name__ == '__main__':
main()
diff --git a/front/plugins/sync/sync.py b/front/plugins/sync/sync.py
index dea6b8f2..f17d169b 100755
--- a/front/plugins/sync/sync.py
+++ b/front/plugins/sync/sync.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import sys
@@ -12,16 +12,16 @@ import base64
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from plugin_helper import Plugin_Objects
-from utils.plugin_utils import get_plugins_configs, decode_and_rename_files
-from logger import mylog, Logger
-from const import fullDbPath, logPath
-from helper import get_setting_value
-from utils.datetime_utils import timeNowDB
-from utils.crypto_utils import encrypt_data
-from messaging.in_app import write_notification
-import conf
-from pytz import timezone
+from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
+from utils.plugin_utils import get_plugins_configs, decode_and_rename_files # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from const import fullDbPath, logPath # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
+from utils.crypto_utils import encrypt_data # noqa: E402 [flake8 lint suppression]
+from messaging.in_app import write_notification # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -41,21 +41,21 @@ plugin_objects = Plugin_Objects(RESULT_FILE)
def main():
- mylog('verbose', [f'[{pluginName}] In script'])
+ mylog('verbose', [f'[{pluginName}] In script'])
# Retrieve configuration settings
plugins_to_sync = get_setting_value('SYNC_plugins')
- api_token = get_setting_value('API_TOKEN')
+ api_token = get_setting_value('API_TOKEN')
encryption_key = get_setting_value('SYNC_encryption_key')
hub_url = get_setting_value('SYNC_hub_url')
node_name = get_setting_value('SYNC_node_name')
send_devices = get_setting_value('SYNC_devices')
pull_nodes = get_setting_value('SYNC_nodes')
-
+
# variables to determine operation mode
is_hub = False
is_node = False
-
+
# Check if api_token set
if not api_token:
mylog('verbose', [f'[{pluginName}] ⚠ ERROR api_token not defined - quitting.'])
@@ -63,23 +63,23 @@ def main():
# check if this is a hub or a node
if len(hub_url) > 0 and (send_devices or plugins_to_sync):
- is_node = True
- mylog('verbose', [f'[{pluginName}] Mode 1: PUSH (NODE) - This is a NODE as SYNC_hub_url, SYNC_devices or SYNC_plugins are set'])
- if len(pull_nodes) > 0:
+ is_node = True
+ mylog('verbose', [f'[{pluginName}] Mode 1: PUSH (NODE) - This is a NODE as SYNC_hub_url, SYNC_devices or SYNC_plugins are set'])
+ if len(pull_nodes) > 0:
is_hub = True
- mylog('verbose', [f'[{pluginName}] Mode 2: PULL (HUB) - This is a HUB as SYNC_nodes is set'])
+ mylog('verbose', [f'[{pluginName}] Mode 2: PULL (HUB) - This is a HUB as SYNC_nodes is set'])
- # Mode 1: PUSH/SEND (NODE)
+ # Mode 1: PUSH/SEND (NODE)
if is_node:
- # PUSHING/SENDING Plugins
-
+ # PUSHING/SENDING Plugins
+
# Get all plugin configurations
all_plugins = get_plugins_configs(False)
mylog('verbose', [f'[{pluginName}] plugins_to_sync {plugins_to_sync}'])
-
+
for plugin in all_plugins:
- pref = plugin["unique_prefix"]
+ pref = plugin["unique_prefix"]
index = 0
if pref in plugins_to_sync:
@@ -100,9 +100,8 @@ def main():
send_data(api_token, file_content, encryption_key, file_path, node_name, pref, hub_url)
else:
- mylog('verbose', [f'[{pluginName}] {file_path} not found'])
-
-
+ mylog('verbose', [f'[{pluginName}] {file_path} not found'])
+
# PUSHING/SENDING devices
if send_devices:
@@ -117,27 +116,27 @@ def main():
mylog('verbose', [f'[{pluginName}] Sending file_content: "{file_content}"'])
send_data(api_token, file_content, encryption_key, file_path, node_name, pref, hub_url)
else:
- mylog('verbose', [f'[{pluginName}] SYNC_hub_url not defined, skipping posting "Devices" data'])
+ mylog('verbose', [f'[{pluginName}] SYNC_hub_url not defined, skipping posting "Devices" data'])
else:
- mylog('verbose', [f'[{pluginName}] SYNC_hub_url not defined, skipping posting "Plugins" and "Devices" data'])
+ mylog('verbose', [f'[{pluginName}] SYNC_hub_url not defined, skipping posting "Plugins" and "Devices" data'])
# Mode 2: PULL/GET (HUB)
-
- # PULLING DEVICES
+
+ # PULLING DEVICES
file_prefix = 'last_result'
-
+
# pull data from nodes if specified
if is_hub:
for node_url in pull_nodes:
response_json = get_data(api_token, node_url)
-
+
# Extract node_name and base64 data
node_name = response_json.get('node_name', 'unknown_node')
data_base64 = response_json.get('data_base64', '')
# Decode base64 data
decoded_data = base64.b64decode(data_base64)
-
+
# Create log file name using node name
log_file_name = f'{file_prefix}.{node_name}.log'
@@ -148,18 +147,17 @@ def main():
message = f'[{pluginName}] Device data from node "{node_name}" written to {log_file_name}'
mylog('verbose', [message])
if lggr.isAbove('verbose'):
- write_notification(message, 'info', timeNowDB())
-
+ write_notification(message, 'info', timeNowDB())
# Process any received data for the Device DB table (ONLY JSON)
# Create the file path
# Get all "last_result" files from the sync folder, decode, rename them, and get the list of files
files_to_process = decode_and_rename_files(LOG_PATH, file_prefix)
-
+
if len(files_to_process) > 0:
-
- mylog('verbose', [f'[{pluginName}] Mode 3: RECEIVE (HUB) - This is a HUB as received data found'])
+
+ mylog('verbose', [f'[{pluginName}] Mode 3: RECEIVE (HUB) - This is a HUB as received data found'])
# Connect to the App database
conn = sqlite3.connect(fullDbPath)
@@ -176,24 +174,24 @@ def main():
# only process received .log files, skipping the one logging the progress of this plugin
if file_name != 'last_result.log':
mylog('verbose', [f'[{pluginName}] Processing: "{file_name}"'])
-
+
# make sure the file has the correct name (e.g last_result.encoded.Node_1.1.log) to skip any otehr plugin files
if len(file_name.split('.')) > 2:
# Extract node name from either last_result.decoded.Node_1.1.log or last_result.Node_1.log
parts = file_name.split('.')
# If decoded/encoded file, node name is at index 2; otherwise at index 1
- syncHubNodeName = parts[2] if 'decoded' in file_name or 'encoded' in file_name else parts[1]
+ syncHubNodeName = parts[2] if 'decoded' in file_name or 'encoded' in file_name else parts[1]
file_path = f"{LOG_PATH}/{file_name}"
-
+
with open(file_path, 'r') as f:
data = json.load(f)
for device in data['data']:
if device['devMac'] not in unique_mac_addresses:
device['devSyncHubNode'] = syncHubNodeName
unique_mac_addresses.add(device['devMac'])
- device_data.append(device)
-
+ device_data.append(device)
+
# Rename the file to "processed_" + current name
new_file_name = f"processed_{file_name}"
new_file_path = os.path.join(LOG_PATH, new_file_name)
@@ -209,7 +207,6 @@ def main():
placeholders = ', '.join('?' for _ in unique_mac_addresses)
cursor.execute(f'SELECT devMac FROM Devices WHERE devMac IN ({placeholders})', tuple(unique_mac_addresses))
existing_mac_addresses = set(row[0] for row in cursor.fetchall())
-
# insert devices into the last_result.log and thus CurrentScan table to manage state
for device in device_data:
@@ -228,7 +225,7 @@ def main():
# Filter out existing devices
new_devices = [device for device in device_data if device['devMac'] not in existing_mac_addresses]
- # Remove 'rowid' key if it exists
+ # Remove 'rowid' key if it exists
for device in new_devices:
device.pop('rowid', None)
device.pop('devStatus', None)
@@ -257,7 +254,6 @@ def main():
mylog('verbose', [message])
write_notification(message, 'info', timeNowDB())
-
# Commit and close the connection
conn.commit()
@@ -268,6 +264,7 @@ def main():
return 0
+
# ------------------------------------------------------------------
# Data retrieval methods
api_endpoints = [
@@ -275,6 +272,7 @@ api_endpoints = [
"/plugins/sync/hub.php" # Legacy PHP endpoint
]
+
# send data to the HUB
def send_data(api_token, file_content, encryption_key, file_path, node_name, pref, hub_url):
"""Send encrypted data to HUB, preferring /sync endpoint and falling back to PHP version."""
@@ -345,6 +343,5 @@ def get_data(api_token, node_url):
return ""
-
if __name__ == '__main__':
main()
diff --git a/front/plugins/unifi_api_import/unifi_api_import.py b/front/plugins/unifi_api_import/unifi_api_import.py
index 77abe899..2d2e3e30 100755
--- a/front/plugins/unifi_api_import/unifi_api_import.py
+++ b/front/plugins/unifi_api_import/unifi_api_import.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import sys
@@ -10,12 +10,11 @@ from unifi_sm_api.api import SiteManagerAPI
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from plugin_helper import Plugin_Objects, decode_settings_base64
-from logger import mylog, Logger
-from const import logPath
-from helper import get_setting_value
-
-import conf
+from plugin_helper import Plugin_Objects, decode_settings_base64 # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -35,13 +34,13 @@ plugin_objects = Plugin_Objects(RESULT_FILE)
def main():
- mylog('verbose', [f'[{pluginName}] In script'])
+ mylog('verbose', [f'[{pluginName}] In script'])
# Retrieve configuration settings
unifi_sites_configs = get_setting_value('UNIFIAPI_sites')
mylog('verbose', [f'[{pluginName}] number of unifi_sites_configs: {len(unifi_sites_configs)}'])
-
+
for site_config in unifi_sites_configs:
siteDict = decode_settings_base64(site_config)
@@ -50,11 +49,11 @@ def main():
mylog('none', [f'[{pluginName}] Connecting to: {siteDict["UNIFIAPI_site_name"]}'])
api = SiteManagerAPI(
- api_key=siteDict["UNIFIAPI_api_key"],
- version=siteDict["UNIFIAPI_api_version"],
- base_url=siteDict["UNIFIAPI_base_url"],
- verify_ssl=siteDict["UNIFIAPI_verify_ssl"]
- )
+ api_key=siteDict["UNIFIAPI_api_key"],
+ version=siteDict["UNIFIAPI_api_version"],
+ base_url=siteDict["UNIFIAPI_base_url"],
+ verify_ssl=siteDict["UNIFIAPI_verify_ssl"]
+ )
sites_resp = api.get_sites()
sites = sites_resp.get("data", [])
@@ -67,18 +66,18 @@ def main():
# Process the data into native application tables
if len(device_data) > 0:
- # insert devices into the lats_result.log
+ # insert devices into the lats_result.log
for device in device_data:
- plugin_objects.add_object(
- primaryId = device['dev_mac'], # mac
- secondaryId = device['dev_ip'], # IP
- watched1 = device['dev_name'], # name
- watched2 = device['dev_type'], # device_type (AP/Switch etc)
- watched3 = device['dev_connected'], # connectedAt or empty
- watched4 = device['dev_parent_mac'],# parent_mac or "Internet"
- extra = '',
- foreignKey = device['dev_mac']
- )
+ plugin_objects.add_object(
+ primaryId = device['dev_mac'], # mac
+ secondaryId = device['dev_ip'], # IP
+ watched1 = device['dev_name'], # name
+ watched2 = device['dev_type'], # device_type (AP/Switch etc)
+ watched3 = device['dev_connected'], # connectedAt or empty
+ watched4 = device['dev_parent_mac'], # parent_mac or "Internet"
+ extra = '',
+ foreignKey = device['dev_mac']
+ )
mylog('verbose', [f'[{pluginName}] New entries: "{len(device_data)}"'])
@@ -87,6 +86,7 @@ def main():
return 0
+
# retrieve data
def get_device_data(site, api):
device_data = []
@@ -146,8 +146,8 @@ def get_device_data(site, api):
dev_parent_mac = resolve_parent_mac(uplinkDeviceId)
device_data.append({
- "dev_mac": dev_mac,
- "dev_ip": dev_ip,
+ "dev_mac": dev_mac,
+ "dev_ip": dev_ip,
"dev_name": dev_name,
"dev_type": dev_type,
"dev_connected": dev_connected,
diff --git a/front/plugins/unifi_import/script.py b/front/plugins/unifi_import/script.py
index 7970860d..57775c68 100755
--- a/front/plugins/unifi_import/script.py
+++ b/front/plugins/unifi_import/script.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
# Inspired by https://github.com/stevehoek/Pi.Alert
from __future__ import unicode_literals
@@ -14,12 +14,12 @@ from pyunifi.controller import Controller
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from plugin_helper import Plugin_Objects, rmBadChars, is_typical_router_ip, is_mac
-from logger import mylog, Logger
-from helper import get_setting_value, normalize_string
-import conf
-from pytz import timezone
-from const import logPath
+from plugin_helper import Plugin_Objects, rmBadChars, is_typical_router_ip, is_mac # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value, normalize_string # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -37,21 +37,16 @@ LOCK_FILE = os.path.join(LOG_PATH, f'full_run.{pluginName}.lock')
urllib3.disable_warnings(InsecureRequestWarning)
-
-
-# Workflow
-
def main():
-
+
mylog('verbose', [f'[{pluginName}] In script'])
-
# init global variables
global UNIFI_USERNAME, UNIFI_PASSWORD, UNIFI_HOST, UNIFI_SITES, PORT, VERIFYSSL, VERSION, FULL_IMPORT
# parse output
- plugin_objects = Plugin_Objects(RESULT_FILE)
-
+ plugin_objects = Plugin_Objects(RESULT_FILE)
+
UNIFI_USERNAME = get_setting_value("UNFIMP_username")
UNIFI_PASSWORD = get_setting_value("UNFIMP_password")
UNIFI_HOST = get_setting_value("UNFIMP_host")
@@ -64,12 +59,11 @@ def main():
plugin_objects = get_entries(plugin_objects)
plugin_objects.write_result_file()
-
mylog('verbose', [f'[{pluginName}] Scan finished, found {len(plugin_objects)} devices'])
-# .............................................
+# .............................................
def get_entries(plugin_objects: Plugin_Objects) -> Plugin_Objects:
global VERIFYSSL
@@ -79,27 +73,26 @@ def get_entries(plugin_objects: Plugin_Objects) -> Plugin_Objects:
mylog('verbose', [f'[{pluginName}] sites: {UNIFI_SITES}'])
-
if (VERIFYSSL.upper() == "TRUE"):
VERIFYSSL = True
else:
VERIFYSSL = False
-
+
# mylog('verbose', [f'[{pluginName}] sites: {sites}'])
-
+
for site in UNIFI_SITES:
-
+
mylog('verbose', [f'[{pluginName}] site: {site}'])
c = Controller(
- UNIFI_HOST,
- UNIFI_USERNAME,
- UNIFI_PASSWORD,
- port=PORT,
- version=VERSION,
- ssl_verify=VERIFYSSL,
+ UNIFI_HOST,
+ UNIFI_USERNAME,
+ UNIFI_PASSWORD,
+ port=PORT,
+ version=VERSION,
+ ssl_verify=VERIFYSSL,
site_id=site)
-
+
online_macs = set()
processed_macs = []
@@ -114,7 +107,7 @@ def get_entries(plugin_objects: Plugin_Objects) -> Plugin_Objects:
plugin_objects=plugin_objects,
device_label='client',
device_vendor="",
- force_import=True # These are online clients, force import
+ force_import=True # These are online clients, force import
)
mylog('verbose', [f'[{pluginName}] Found {len(plugin_objects)} Online Devices'])
@@ -154,11 +147,9 @@ def get_entries(plugin_objects: Plugin_Objects) -> Plugin_Objects:
mylog('verbose', [f'[{pluginName}] Found {len(plugin_objects)} Users'])
-
mylog('verbose', [f'[{pluginName}] check if Lock file needs to be modified'])
set_lock_file_value(FULL_IMPORT, lock_file_value)
-
mylog('verbose', [f'[{pluginName}] Found {len(plugin_objects)} Clients overall'])
return plugin_objects
@@ -173,19 +164,19 @@ def collect_details(device_type, devices, online_macs, processed_macs, plugin_ob
name = get_name(get_unifi_val(device, 'name'), get_unifi_val(device, 'hostname'))
ipTmp = get_ip(get_unifi_val(device, 'lan_ip'), get_unifi_val(device, 'last_ip'), get_unifi_val(device, 'fixed_ip'), get_unifi_val(device, 'ip'))
macTmp = device['mac']
-
+
# continue only if valid MAC address
if is_mac(macTmp):
status = 1 if macTmp in online_macs else device.get('state', 0)
deviceType = device_type.get(device.get('type'), '')
parentMac = get_parent_mac(get_unifi_val(device, 'uplink_mac'), get_unifi_val(device, 'ap_mac'), get_unifi_val(device, 'sw_mac'))
-
+
# override parent MAC if this is a router
if parentMac == 'null' and is_typical_router_ip(ipTmp):
- parentMac = 'Internet'
+ parentMac = 'Internet'
# Add object only if not processed
- if macTmp not in processed_macs and ( status == 1 or force_import is True ):
+ if macTmp not in processed_macs and (status == 1 or force_import is True):
plugin_objects.add_object(
primaryId=macTmp,
secondaryId=ipTmp,
@@ -203,7 +194,8 @@ def collect_details(device_type, devices, online_macs, processed_macs, plugin_ob
processed_macs.append(macTmp)
else:
mylog('verbose', [f'[{pluginName}] Skipping, not a valid MAC address: {macTmp}'])
-
+
+
# -----------------------------------------------------------------------------
def get_unifi_val(obj, key, default='null'):
if isinstance(obj, dict):
@@ -212,9 +204,9 @@ def get_unifi_val(obj, key, default='null'):
for k, v in obj.items():
if isinstance(v, dict):
result = get_unifi_val(v, key, default)
- if result not in ['','None', None, 'null']:
+ if result not in ['', 'None', None, 'null']:
return result
-
+
mylog('trace', [f'[{pluginName}] Value not found for key "{key}" in obj "{json.dumps(obj)}"'])
return default
@@ -226,6 +218,7 @@ def get_name(*names: str) -> str:
return rmBadChars(name)
return 'null'
+
# -----------------------------------------------------------------------------
def get_parent_mac(*macs: str) -> str:
for mac in macs:
@@ -233,6 +226,7 @@ def get_parent_mac(*macs: str) -> str:
return mac
return 'null'
+
# -----------------------------------------------------------------------------
def get_port(*ports: str) -> str:
for port in ports:
@@ -240,12 +234,6 @@ def get_port(*ports: str) -> str:
return port
return 'null'
-# -----------------------------------------------------------------------------
-def get_port(*macs: str) -> str:
- for mac in macs:
- if mac and mac != 'null':
- return mac
- return 'null'
# -----------------------------------------------------------------------------
def get_ip(*ips: str) -> str:
@@ -271,7 +259,7 @@ def set_lock_file_value(config_value: str, lock_file_value: bool) -> None:
mylog('verbose', [f'[{pluginName}] Setting lock value for "full import" to {out}'])
with open(LOCK_FILE, 'w') as lock_file:
- lock_file.write(str(out))
+ lock_file.write(str(out))
# -----------------------------------------------------------------------------
@@ -286,15 +274,16 @@ def read_lock_file() -> bool:
# -----------------------------------------------------------------------------
def check_full_run_state(config_value: str, lock_file_value: bool) -> bool:
- if config_value == 'always' or (config_value == 'once' and lock_file_value == False):
+ if config_value == 'always' or (config_value == 'once' and lock_file_value is False):
mylog('verbose', [f'[{pluginName}] Full import needs to be done: config_value: {config_value} and lock_file_value: {lock_file_value}'])
return True
else:
mylog('verbose', [f'[{pluginName}] Full import NOT needed: config_value: {config_value} and lock_file_value: {lock_file_value}'])
return False
-#===============================================================================
+
+# ===============================================================================
# BEGIN
-#===============================================================================
+# ===============================================================================
if __name__ == '__main__':
main()
diff --git a/front/plugins/vendor_update/script.py b/front/plugins/vendor_update/script.py
index cc3a5b45..7b0c3661 100755
--- a/front/plugins/vendor_update/script.py
+++ b/front/plugins/vendor_update/script.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import sys
@@ -9,13 +9,13 @@ import sqlite3
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from plugin_helper import Plugin_Objects, handleEmpty
-from logger import mylog, Logger
-from helper import get_setting_value
-from const import logPath, applicationPath, fullDbPath
-from scan.device_handling import query_MAC_vendor
-import conf
-from pytz import timezone
+from plugin_helper import Plugin_Objects, handleEmpty # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from const import logPath, applicationPath, fullDbPath # noqa: E402 [flake8 lint suppression]
+from scan.device_handling import query_MAC_vendor # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -25,17 +25,17 @@ Logger(get_setting_value('LOG_LEVEL'))
pluginName = 'VNDRPDT'
-
LOG_PATH = logPath + '/plugins'
LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
-def main():
- mylog('verbose', ['[VNDRPDT] In script'])
+def main():
- # Get newest DB
- update_vendor_database()
+ mylog('verbose', ['[VNDRPDT] In script'])
+
+ # Get newest DB
+ update_vendor_database()
# Resolve missing vendors
plugin_objects = Plugin_Objects(RESULT_FILE)
@@ -43,67 +43,67 @@ def main():
plugin_objects = update_vendors(fullDbPath, plugin_objects)
plugin_objects.write_result_file()
-
- mylog('verbose', ['[VNDRPDT] Update complete'])
-
+
+ mylog('verbose', ['[VNDRPDT] Update complete'])
+
return 0
-#===============================================================================
+
+# ===============================================================================
# Update device vendors database
-#===============================================================================
+# ===============================================================================
def update_vendor_database():
# Update vendors DB (iab oui)
- mylog('verbose', [' Updating vendors DB (iab & oui)'])
+ mylog('verbose', [' Updating vendors DB (iab & oui)'])
update_args = ['sh', applicationPath + '/services/update_vendors.sh']
- # Execute command
+ # Execute command
try:
# try runnning a subprocess safely
- update_output = subprocess.check_output (update_args)
+ subprocess.check_output(update_args)
except subprocess.CalledProcessError as e:
# An error occured, handle it
- mylog('verbose', [' FAILED: Updating vendors DB, set LOG_LEVEL=debug for more info'])
- mylog('verbose', [e.output])
+ mylog('verbose', [' FAILED: Updating vendors DB, set LOG_LEVEL=debug for more info'])
+ mylog('verbose', [e.output])
+
# ------------------------------------------------------------------------------
# resolve missing vendors
-def update_vendors (dbPath, plugin_objects):
-
+def update_vendors(dbPath, plugin_objects):
+
# Connect to the App SQLite database
conn = sqlite3.connect(dbPath)
sql = conn.cursor()
# Initialize variables
- recordsToUpdate = []
ignored = 0
notFound = 0
-
- mylog('verbose', [' Searching devices vendor'])
+ mylog('verbose', [' Searching devices vendor'])
# Get devices without a vendor
- sql.execute ("""SELECT
- devMac,
- devLastIP,
- devName,
- devVendor
+ sql.execute("""SELECT
+ devMac,
+ devLastIP,
+ devName,
+ devVendor
FROM Devices
- WHERE devVendor = '(unknown)'
- OR devVendor = '(Unknown)'
+ WHERE devVendor = '(unknown)'
+ OR devVendor = '(Unknown)'
OR devVendor = ''
OR devVendor IS NULL
""")
- devices = sql.fetchall()
- conn.commit()
+ devices = sql.fetchall()
+ conn.commit()
# Close the database connection
- conn.close()
+ conn.close()
# All devices loop
for device in devices:
# Search vendor in HW Vendors DB
- vendor = query_MAC_vendor (device[0])
+ vendor = query_MAC_vendor(device[0])
if vendor == -1 :
notFound += 1
elif vendor == -2 :
@@ -112,27 +112,25 @@ def update_vendors (dbPath, plugin_objects):
plugin_objects.add_object(
primaryId = handleEmpty(device[0]), # MAC (Device Name)
secondaryId = handleEmpty(device[1]), # IP Address (always 0.0.0.0)
- watched1 = handleEmpty(vendor),
+ watched1 = handleEmpty(vendor),
watched2 = handleEmpty(device[2]), # Device name
watched3 = "",
watched4 = "",
- extra = "",
- foreignKey = handleEmpty(device[0])
- )
-
- # Print log
+ extra = "",
+ foreignKey = handleEmpty(device[0])
+ )
+
+ # Print log
mylog('verbose', [" Devices Ignored : ", ignored])
mylog('verbose', [" Devices with missing vendor : ", len(devices)])
mylog('verbose', [" Vendors Not Found : ", notFound])
- mylog('verbose', [" Vendors updated : ", len(plugin_objects) ])
-
+ mylog('verbose', [" Vendors updated : ", len(plugin_objects)])
return plugin_objects
-
-#===============================================================================
+# ===============================================================================
# BEGIN
-#===============================================================================
+# ===============================================================================
if __name__ == '__main__':
- main()
\ No newline at end of file
+ main()
diff --git a/front/plugins/wake_on_lan/wake_on_lan.py b/front/plugins/wake_on_lan/wake_on_lan.py
index b5d44d99..02008184 100755
--- a/front/plugins/wake_on_lan/wake_on_lan.py
+++ b/front/plugins/wake_on_lan/wake_on_lan.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import sys
@@ -9,13 +9,13 @@ from wakeonlan import send_magic_packet
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from plugin_helper import Plugin_Objects
-from logger import mylog, Logger
-from const import logPath
-from helper import get_setting_value
-from database import DB
-from models.device_instance import DeviceInstance
-import conf
+from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from database import DB # noqa: E402 [flake8 lint suppression]
+from models.device_instance import DeviceInstance # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -34,9 +34,8 @@ RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
plugin_objects = Plugin_Objects(RESULT_FILE)
-
def main():
- mylog('none', [f'[{pluginName}] In script'])
+ mylog('none', [f'[{pluginName}] In script'])
# Retrieve configuration settings
broadcast_ips = get_setting_value('WOL_broadcast_ips')
@@ -58,7 +57,7 @@ def main():
devices_to_wake = device_handler.getOffline()
elif 'down' in devices_to_wake:
-
+
devices_to_wake = device_handler.getDown()
else:
@@ -89,15 +88,16 @@ def main():
# log result
plugin_objects.write_result_file()
else:
- mylog('none', [f'[{pluginName}] No devices to wake'])
+ mylog('none', [f'[{pluginName}] No devices to wake'])
- mylog('none', [f'[{pluginName}] Script finished'])
+ mylog('none', [f'[{pluginName}] Script finished'])
return 0
+
# wake
def execute(port, ip, mac, name):
-
+
result = 'null'
try:
# Send the magic packet to wake up the device
@@ -105,7 +105,7 @@ def execute(port, ip, mac, name):
mylog('verbose', [f'[{pluginName}] Magic packet sent to {mac} ({name})'])
result = 'success'
-
+
except Exception as e:
result = str(e)
mylog('verbose', [f'[{pluginName}] Failed to send magic packet to {mac} ({name}): {e}'])
@@ -113,5 +113,6 @@ def execute(port, ip, mac, name):
# Return the data result
return result
+
if __name__ == '__main__':
main()
diff --git a/front/plugins/website_monitor/script.py b/front/plugins/website_monitor/script.py
index 14443619..79e108ec 100755
--- a/front/plugins/website_monitor/script.py
+++ b/front/plugins/website_monitor/script.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
# Based on the work of https://github.com/leiweibau/Pi.Alert
import requests
@@ -12,12 +12,12 @@ from urllib3.exceptions import InsecureRequestWarning
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from plugin_helper import Plugin_Objects
-from const import logPath
-from helper import get_setting_value
-import conf
-from pytz import timezone
-from logger import mylog, Logger
+from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
+from const import logPath # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+import conf # noqa: E402 [flake8 lint suppression]
+from pytz import timezone # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
# Make sure the TIMEZONE for logging is correct
conf.tz = timezone(get_setting_value('TIMEZONE'))
@@ -30,15 +30,14 @@ pluginName = 'WEBMON'
LOG_PATH = logPath + '/plugins'
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
-
mylog('verbose', [f'[{pluginName}] In script'])
+
def main():
values = get_setting_value('WEBMON_urls_to_check')
mylog('verbose', [f'[{pluginName}] Checking URLs: {values}'])
-
if len(values) > 0:
plugin_objects = Plugin_Objects(RESULT_FILE)
@@ -48,12 +47,13 @@ def main():
else:
return
+
def check_services_health(site):
mylog('verbose', [f'[{pluginName}] Checking {site}'])
urllib3.disable_warnings(InsecureRequestWarning)
-
+
try:
resp = requests.get(site, verify=False, timeout=get_setting_value('WEBMON_RUN_TIMEOUT'), headers={"User-Agent": "NetAlertX"})
latency = resp.elapsed.total_seconds()
@@ -79,12 +79,13 @@ def check_services_health(site):
return status, latency
+
def service_monitoring(urls, plugin_objects):
for site in urls:
status, latency = check_services_health(site)
plugin_objects.add_object(
primaryId=site,
- secondaryId='null',
+ secondaryId='null',
watched1=status,
watched2=latency,
watched3='null',
@@ -94,7 +95,6 @@ def service_monitoring(urls, plugin_objects):
)
return plugin_objects
+
if __name__ == '__main__':
sys.exit(main())
-
-
diff --git a/install/production-filesystem/entrypoint.d/10-mounts.py b/install/production-filesystem/entrypoint.d/10-mounts.py
index e10033c9..b021bb84 100755
--- a/install/production-filesystem/entrypoint.d/10-mounts.py
+++ b/install/production-filesystem/entrypoint.d/10-mounts.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python3
+# !/usr/bin/env python3
import os
import sys
diff --git a/scripts/checkmk/script.py b/scripts/checkmk/script.py
index d0fb1009..d1f5b6f2 100755
--- a/scripts/checkmk/script.py
+++ b/scripts/checkmk/script.py
@@ -1,5 +1,5 @@
-#!/usr/bin/env python3
+# !/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
NetAlertX-New-Devices-Checkmk-Script
@@ -19,11 +19,12 @@ import subprocess
import json
import os
+
def check_new_devices():
# Get API path from environment variable, fallback to /tmp/api
api_path = os.environ.get('NETALERTX_API', '/tmp/api')
table_devices_path = f'{api_path}/table_devices.json'
-
+
try:
# Rufe die JSON-Datei aus dem Docker-Container ab
result = subprocess.run(
@@ -73,6 +74,6 @@ def check_new_devices():
)
print(f"1 NetAlertX_New_Devices - WARNING - Found {len(new_devices)} new device(s): {device_list_str}")
+
if __name__ == "__main__":
check_new_devices()
-
diff --git a/scripts/db_cleanup/db_cleanup.py b/scripts/db_cleanup/db_cleanup.py
index 7a34df4f..e1e7dc75 100755
--- a/scripts/db_cleanup/db_cleanup.py
+++ b/scripts/db_cleanup/db_cleanup.py
@@ -1,8 +1,8 @@
-#!/usr/bin/env python3
+# !/usr/bin/env python3
import subprocess
-import sys
import os
+
def run_sqlite_command(command):
# Use environment variable with fallback
db_path = os.path.join(
@@ -19,18 +19,19 @@ def run_sqlite_command(command):
print(f"Error executing command: {e}")
return None
+
def check_and_clean_device():
while True:
print("\nDevice Cleanup Tool")
print("1. Check/Clean by MAC address")
print("2. Check/Clean by IP address")
print("3. Exit")
-
+
choice = input("\nSelect option (1-3): ")
-
+
if choice == "1":
mac = input("Enter MAC address (format: xx:xx:xx:xx:xx:xx): ").lower()
-
+
# Check all tables for MAC
tables_checks = [
f"SELECT 'Events' as source, * FROM Events WHERE eve_MAC='{mac}'",
@@ -40,14 +41,14 @@ def check_and_clean_device():
f"SELECT 'AppEvents' as source, * FROM AppEvents WHERE ObjectPrimaryID LIKE '%{mac}%' OR ObjectSecondaryID LIKE '%{mac}%'",
f"SELECT 'Plugins_Objects' as source, * FROM Plugins_Objects WHERE Object_PrimaryID LIKE '%{mac}%'"
]
-
+
found = False
for check in tables_checks:
result = run_sqlite_command(check)
if result and result.strip():
found = True
print(f"\nFound entries:\n{result}")
-
+
if found:
confirm = input("\nWould you like to clean these entries? (y/n): ")
if confirm.lower() == 'y':
@@ -60,16 +61,16 @@ def check_and_clean_device():
f"DELETE FROM AppEvents WHERE ObjectPrimaryID LIKE '%{mac}%' OR ObjectSecondaryID LIKE '%{mac}%'",
f"DELETE FROM Plugins_Objects WHERE Object_PrimaryID LIKE '%{mac}%'"
]
-
+
for delete in deletes:
run_sqlite_command(delete)
print("Cleanup completed!")
else:
print("\nNo entries found for this MAC address")
-
+
elif choice == "2":
ip = input("Enter IP address (format: xxx.xxx.xxx.xxx): ")
-
+
# Check all tables for IP
tables_checks = [
f"SELECT 'Events' as source, * FROM Events WHERE eve_IP='{ip}'",
@@ -79,14 +80,14 @@ def check_and_clean_device():
f"SELECT 'AppEvents' as source, * FROM AppEvents WHERE ObjectSecondaryID LIKE '%{ip}%'",
f"SELECT 'Plugins_Objects' as source, * FROM Plugins_Objects WHERE Object_SecondaryID LIKE '%{ip}%'"
]
-
+
found = False
for check in tables_checks:
result = run_sqlite_command(check)
if result and result.strip():
found = True
print(f"\nFound entries:\n{result}")
-
+
if found:
confirm = input("\nWould you like to clean these entries? (y/n): ")
if confirm.lower() == 'y':
@@ -99,19 +100,20 @@ def check_and_clean_device():
f"DELETE FROM AppEvents WHERE ObjectSecondaryID LIKE '%{ip}%'",
f"DELETE FROM Plugins_Objects WHERE Object_SecondaryID LIKE '%{ip}%'"
]
-
+
for delete in deletes:
run_sqlite_command(delete)
print("Cleanup completed!")
else:
print("\nNo entries found for this IP address")
-
+
elif choice == "3":
print("\nExiting...")
break
-
+
else:
print("\nInvalid option, please try again")
+
if __name__ == "__main__":
check_and_clean_device()
diff --git a/scripts/opnsense_leases/opnsense_leases.py b/scripts/opnsense_leases/opnsense_leases.py
index ab8f7f00..1715d0bb 100755
--- a/scripts/opnsense_leases/opnsense_leases.py
+++ b/scripts/opnsense_leases/opnsense_leases.py
@@ -1,6 +1,5 @@
#!/usr/bin/env python3
import paramiko
-import re
from datetime import datetime
import argparse
import sys
@@ -8,6 +7,9 @@ from pathlib import Path
import time
import logging
+logger = None
+
+
def setup_logging(debug=False):
"""Configure logging based on debug flag."""
level = logging.DEBUG if debug else logging.INFO
@@ -18,6 +20,7 @@ def setup_logging(debug=False):
)
return logging.getLogger(__name__)
+
def parse_timestamp(date_str):
"""Convert OPNsense timestamp to Unix epoch time."""
try:
@@ -27,7 +30,7 @@ def parse_timestamp(date_str):
dt = datetime.strptime(clean_date, '%Y/%m/%d %H:%M:%S')
return int(dt.timestamp())
except Exception as e:
- logger.error(f"Failed to parse timestamp: {date_str}")
+ logger.error(f"Failed to parse timestamp: {date_str} ({e})")
return None
@@ -39,8 +42,14 @@ def get_lease_file(hostname, username, password=None, key_filename=None, port=22
try:
logger.debug(f"Attempting to connect to {hostname}:{port} as {username}")
- ssh.connect(hostname, port=port, username=username,
- password=password, key_filename=key_filename)
+
+ ssh.connect(
+ hostname,
+ port=port,
+ username=username,
+ password=password,
+ key_filename=key_filename
+ )
# Get an interactive shell session
logger.debug("Opening interactive SSH channel")
@@ -75,10 +84,23 @@ def get_lease_file(hostname, username, password=None, key_filename=None, port=22
# Clean up the output by removing the command echo and shell prompts
lines = output.split('\n')
# Remove first line (command echo) and any lines containing shell prompts
- cleaned_lines = [line for line in lines
- if not line.strip().startswith(command.strip())
- and not line.strip().endswith('> ')
- and not line.strip().endswith('# ')]
+ # cleaned_lines = [line for line in lines
+ # if not line.strip().startswith(command.strip()) and not line.strip().endswith('> ') and not line.strip().endswith('# ')]
+ cmd = command.strip()
+
+ cleaned_lines = []
+ for line in lines:
+ stripped = line.strip()
+
+ if stripped.startswith(cmd):
+ continue
+ if stripped.endswith('> '):
+ continue
+ if stripped.endswith('# '):
+ continue
+
+ cleaned_lines.append(line)
+
cleaned_output = '\n'.join(cleaned_lines)
logger.debug(f"Final cleaned output length: {len(cleaned_output)} characters")
@@ -156,9 +178,7 @@ def parse_lease_file(lease_content):
# Filter only active leases
active_leases = [lease for lease in leases
- if lease.get('state') == 'active'
- and 'mac' in lease
- and 'ip' in lease]
+ if lease.get('state') == 'active' and 'mac' in lease and 'ip' in lease]
logger.debug(f"Found {len(active_leases)} active leases out of {len(leases)} total leases")
logger.debug("Active leases:")
@@ -206,6 +226,7 @@ def convert_to_dnsmasq(leases):
return dnsmasq_lines
+
def main():
parser = argparse.ArgumentParser(description='Convert OPNsense DHCP leases to dnsmasq format')
parser.add_argument('--host', required=True, help='OPNsense hostname or IP')
@@ -219,6 +240,7 @@ def main():
args = parser.parse_args()
# Setup logging
+ global logger
logger = setup_logging(args.debug)
try:
@@ -255,5 +277,6 @@ def main():
logger.error(f"Error: {str(e)}")
sys.exit(1)
+
if __name__ == '__main__':
main()
diff --git a/server/__main__.py b/server/__main__.py
index bb149e32..86ccd6bb 100755
--- a/server/__main__.py
+++ b/server/__main__.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
#
# -------------------------------------------------------------------------------
# NetAlertX v2.70 / 2021-02-01
@@ -22,9 +22,9 @@ from pathlib import Path
# Register NetAlertX modules
import conf
-from const import *
-from logger import mylog
-from helper import filePermissions
+from const import fullConfPath, sql_new_devices
+from logger import mylog
+from helper import filePermissions
from utils.datetime_utils import timeNowTZ
from app_state import updateState
from api import update_api
@@ -48,12 +48,12 @@ main structure of NetAlertX
Initialise All
Rename old settings
start Loop forever
- initialise loop
+ initialise loop
(re)import config
(re)import plugin config
run plugins (once)
run frontend events
- update API
+ update API
run plugins (scheduled)
processing scan results
run plugins (after Scan)
@@ -111,7 +111,7 @@ def main():
loop_start_time = conf.loop_start_time # TODO fix
# Handle plugins executed ONCE
- if conf.plugins_once_run == False:
+ if conf.plugins_once_run is False:
pm.run_plugin_scripts("once")
conf.plugins_once_run = True
@@ -146,7 +146,7 @@ def main():
processScan = updateState("Check scan").processScan
mylog("debug", [f"[MAIN] processScan: {processScan}"])
- if processScan == True:
+ if processScan is True:
mylog("debug", "[MAIN] start processing scan results")
process_scan(db)
updateState("Scan processed", None, None, None, None, False)
diff --git a/server/api.py b/server/api.py
index 50092f23..0876bbae 100755
--- a/server/api.py
+++ b/server/api.py
@@ -1,3 +1,4 @@
+# !/usr/bin/env python
import json
import time
import threading
@@ -145,8 +146,7 @@ class api_endpoint_class:
self.needsUpdate = True
# Only update changeDetectedWhen if it hasn't been set recently
if not self.changeDetectedWhen or current_time > (
- self.changeDetectedWhen
- + datetime.timedelta(seconds=self.debounce_interval)
+ self.changeDetectedWhen + datetime.timedelta(seconds=self.debounce_interval)
):
self.changeDetectedWhen = (
current_time # Set timestamp for change detection
@@ -164,8 +164,7 @@ class api_endpoint_class:
self.needsUpdate = True
# Only update changeDetectedWhen if it hasn't been set recently
if not self.changeDetectedWhen or current_time > (
- self.changeDetectedWhen
- + datetime.timedelta(seconds=self.debounce_interval)
+ self.changeDetectedWhen + datetime.timedelta(seconds=self.debounce_interval)
):
self.changeDetectedWhen = (
current_time # Initialize timestamp for new endpoint
@@ -180,17 +179,15 @@ class api_endpoint_class:
current_time = timeNowTZ()
# Debugging info to understand the issue
- # mylog('debug', [f'[API] api_endpoint_class: {self.fileName} is_ad_hoc_user_event {self.is_ad_hoc_user_event} last_update_time={self.last_update_time}, debounce time={self.last_update_time + datetime.timedelta(seconds=self.debounce_interval)}.'])
+ # mylog('debug', [f'[API] api_endpoint_class: {self.fileName} is_ad_hoc_user_event
+ # {self.is_ad_hoc_user_event} last_update_time={self.last_update_time},
+ # debounce time={self.last_update_time + datetime.timedelta(seconds=self.debounce_interval)}.'])
# Only attempt to write if the debounce time has passed
- if forceUpdate == True or (
- self.needsUpdate
- and (
- self.changeDetectedWhen is None
- or current_time
- > (
- self.changeDetectedWhen
- + datetime.timedelta(seconds=self.debounce_interval)
+ if forceUpdate is True or (
+ self.needsUpdate and (
+ self.changeDetectedWhen is None or current_time > (
+ self.changeDetectedWhen + datetime.timedelta(seconds=self.debounce_interval)
)
)
):
diff --git a/server/api_server/api_server_start.py b/server/api_server/api_server_start.py
index 3bc9f7db..980dcbd0 100755
--- a/server/api_server/api_server_start.py
+++ b/server/api_server/api_server_start.py
@@ -9,25 +9,68 @@ from flask_cors import CORS
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
sys.path.extend([f"{INSTALL_PATH}/server"])
-from logger import mylog
-from helper import get_setting_value
-from db.db_helper import get_date_from_period
-from app_state import updateState
+from logger import mylog # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from db.db_helper import get_date_from_period # noqa: E402 [flake8 lint suppression]
+from app_state import updateState # noqa: E402 [flake8 lint suppression]
-
-from .graphql_endpoint import devicesSchema
-from .device_endpoint import get_device_data, set_device_data, delete_device, delete_device_events, reset_device_props, copy_device, update_device_column
-from .devices_endpoint import get_all_devices, delete_unknown_devices, delete_all_with_empty_macs, delete_devices, export_devices, import_csv, devices_totals, devices_by_status
-from .events_endpoint import delete_events, delete_events_older_than, get_events, create_event, get_events_totals
-from .history_endpoint import delete_online_history
-from .prometheus_endpoint import get_metric_stats
-from .sessions_endpoint import get_sessions, delete_session, create_session, get_sessions_calendar, get_device_sessions, get_session_events
-from .nettools_endpoint import wakeonlan, traceroute, speedtest, nslookup, nmap_scan, internet_info
-from .dbquery_endpoint import read_query, write_query, update_query, delete_query
-from .sync_endpoint import handle_sync_post, handle_sync_get
-from .logs_endpoint import clean_log
-from models.user_events_queue_instance import UserEventsQueueInstance
-from messaging.in_app import write_notification, mark_all_notifications_read, delete_notifications, get_unread_notifications, delete_notification, mark_notification_as_read
+from .graphql_endpoint import devicesSchema # noqa: E402 [flake8 lint suppression]
+from .device_endpoint import ( # noqa: E402 [flake8 lint suppression]
+ get_device_data,
+ set_device_data,
+ delete_device,
+ delete_device_events,
+ reset_device_props,
+ copy_device,
+ update_device_column
+)
+from .devices_endpoint import ( # noqa: E402 [flake8 lint suppression]
+ get_all_devices,
+ delete_unknown_devices,
+ delete_all_with_empty_macs,
+ delete_devices,
+ export_devices,
+ import_csv,
+ devices_totals,
+ devices_by_status
+)
+from .events_endpoint import ( # noqa: E402 [flake8 lint suppression]
+ delete_events,
+ delete_events_older_than,
+ get_events,
+ create_event,
+ get_events_totals
+)
+from .history_endpoint import delete_online_history # noqa: E402 [flake8 lint suppression]
+from .prometheus_endpoint import get_metric_stats # noqa: E402 [flake8 lint suppression]
+from .sessions_endpoint import ( # noqa: E402 [flake8 lint suppression]
+ get_sessions,
+ delete_session,
+ create_session,
+ get_sessions_calendar,
+ get_device_sessions,
+ get_session_events
+)
+from .nettools_endpoint import ( # noqa: E402 [flake8 lint suppression]
+ wakeonlan,
+ traceroute,
+ speedtest,
+ nslookup,
+ nmap_scan,
+ internet_info
+)
+from .dbquery_endpoint import read_query, write_query, update_query, delete_query # noqa: E402 [flake8 lint suppression]
+from .sync_endpoint import handle_sync_post, handle_sync_get # noqa: E402 [flake8 lint suppression]
+from .logs_endpoint import clean_log # noqa: E402 [flake8 lint suppression]
+from models.user_events_queue_instance import UserEventsQueueInstance # noqa: E402 [flake8 lint suppression]
+from messaging.in_app import ( # noqa: E402 [flake8 lint suppression]
+ write_notification,
+ mark_all_notifications_read,
+ delete_notifications,
+ get_unread_notifications,
+ delete_notification,
+ mark_notification_as_read
+)
# Flask application
app = Flask(__name__)
@@ -50,13 +93,14 @@ CORS(
allow_headers=["Authorization", "Content-Type"],
)
+
# -------------------------------------------------------------------
# Custom handler for 404 - Route not found
# -------------------------------------------------------------------
@app.errorhandler(404)
def not_found(error):
response = {
- "success": False,
+ "success": False,
"error": "API route not found",
"message": f"The requested URL {error.description if hasattr(error, 'description') else ''} was not found on the server.",
}
@@ -200,7 +244,7 @@ def api_get_devices():
def api_delete_devices():
if not is_authorized():
return jsonify({"success": False, "message": "ERROR: Not authorized", "error": "Forbidden"}), 403
-
+
macs = request.json.get("macs") if request.is_json else None
return delete_devices(macs)
@@ -338,7 +382,7 @@ def dbquery_read():
if not raw_sql_b64:
return jsonify({"success": False, "message": "ERROR: Missing parameters", "error": "rawSql is required"}), 400
-
+
return read_query(raw_sql_b64)
@@ -350,7 +394,7 @@ def dbquery_write():
data = request.get_json() or {}
raw_sql_b64 = data.get("rawSql")
if not raw_sql_b64:
- return jsonify({"success": False, "message": "ERROR: Missing parameters", "error": "rawSql is required"}), 400
+ return jsonify({"success": False, "message": "ERROR: Missing parameters", "error": "rawSql is required"}), 400
return write_query(raw_sql_b64)
@@ -363,7 +407,13 @@ def dbquery_update():
data = request.get_json() or {}
required = ["columnName", "id", "dbtable", "columns", "values"]
if not all(data.get(k) for k in required):
- return jsonify({"success": False, "message": "ERROR: Missing parameters", "error": "Missing required 'columnName', 'id', 'dbtable', 'columns', or 'values' query parameter"}), 400
+ return jsonify(
+ {
+ "success": False,
+ "message": "ERROR: Missing parameters",
+ "error": "Missing required 'columnName', 'id', 'dbtable', 'columns', or 'values' query parameter"
+ }
+ ), 400
return update_query(
column_name=data["columnName"],
@@ -418,12 +468,13 @@ def api_clean_log():
return clean_log(file)
+
@app.route("/logs/add-to-execution-queue", methods=["POST"])
def api_add_to_execution_queue():
if not is_authorized():
return jsonify({"success": False, "message": "ERROR: Not authorized", "error": "Forbidden"}), 403
-
+
queue = UserEventsQueueInstance()
# Get JSON payload safely
@@ -499,7 +550,7 @@ def api_delete_old_events(days: int):
"""
if not is_authorized():
return jsonify({"success": False, "message": "ERROR: Not authorized", "error": "Forbidden"}), 403
-
+
return delete_events_older_than(days)
@@ -619,7 +670,7 @@ def api_write_notification():
if not content:
return jsonify({"success": False, "message": "ERROR: Missing parameters", "error": "Missing content"}), 400
-
+
write_notification(content, level)
return jsonify({"success": True})
@@ -672,7 +723,8 @@ def api_mark_notification_read(guid):
return jsonify({"success": True})
else:
return jsonify({"success": False, "message": "ERROR", "error": result.get("error")}), 500
-
+
+
# --------------------------
# SYNC endpoint
# --------------------------
diff --git a/server/api_server/dbquery_endpoint.py b/server/api_server/dbquery_endpoint.py
index b2bbb8b0..98db3991 100755
--- a/server/api_server/dbquery_endpoint.py
+++ b/server/api_server/dbquery_endpoint.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import base64
@@ -9,7 +9,7 @@ from flask import jsonify
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from database import get_temp_db_connection
+from database import get_temp_db_connection # noqa: E402 [flake8 lint suppression]
def read_query(raw_sql_b64):
diff --git a/server/api_server/device_endpoint.py b/server/api_server/device_endpoint.py
index e1dda839..bec8ff73 100755
--- a/server/api_server/device_endpoint.py
+++ b/server/api_server/device_endpoint.py
@@ -1,18 +1,17 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import sys
-from datetime import datetime
from flask import jsonify, request
# Register NetAlertX directories
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from database import get_temp_db_connection
-from helper import is_random_mac, get_setting_value
-from utils.datetime_utils import timeNowDB, format_date
-from db.db_helper import row_to_json, get_date_from_period
+from database import get_temp_db_connection # noqa: E402 [flake8 lint suppression]
+from helper import is_random_mac, get_setting_value # noqa: E402 [flake8 lint suppression]
+from utils.datetime_utils import timeNowDB, format_date # noqa: E402 [flake8 lint suppression]
+from db.db_helper import row_to_json, get_date_from_period # noqa: E402 [flake8 lint suppression]
# --------------------------
# Device Endpoints Functions
@@ -27,10 +26,10 @@ def get_device_data(mac):
cur = conn.cursor()
now = timeNowDB()
-
+
# Special case for new device
if mac.lower() == "new":
-
+
device_data = {
"devMac": "",
"devName": "",
@@ -89,10 +88,10 @@ def get_device_data(mac):
ELSE 'Off-line'
END AS devStatus,
- (SELECT COUNT(*) FROM Sessions
+ (SELECT COUNT(*) FROM Sessions
WHERE ses_MAC = d.devMac AND (
- ses_DateTimeConnection >= {period_date_sql} OR
- ses_DateTimeDisconnection >= {period_date_sql} OR
+ ses_DateTimeConnection >= {period_date_sql} OR
+ ses_DateTimeDisconnection >= {period_date_sql} OR
ses_StillConnected = 1
)) AS devSessions,
diff --git a/server/api_server/devices_endpoint.py b/server/api_server/devices_endpoint.py
index 0ae2415a..92d5baeb 100755
--- a/server/api_server/devices_endpoint.py
+++ b/server/api_server/devices_endpoint.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import base64
@@ -14,16 +14,13 @@ from logger import mylog
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from database import get_temp_db_connection
-from db.db_helper import get_table_json, get_device_condition_by_status
-from utils.datetime_utils import format_date
+from database import get_temp_db_connection # noqa: E402 [flake8 lint suppression]
+from db.db_helper import get_table_json, get_device_condition_by_status # noqa: E402 [flake8 lint suppression]
# --------------------------
# Device Endpoints Functions
# --------------------------
-
-
def get_all_devices():
"""Retrieve all devices from the database."""
conn = get_temp_db_connection()
@@ -139,7 +136,6 @@ def export_devices(export_format):
def import_csv(file_storage=None):
data = ""
skipped = []
- error = None
# 1. Try JSON `content` (base64-encoded CSV)
if request.is_json and request.json.get("content"):
diff --git a/server/api_server/events_endpoint.py b/server/api_server/events_endpoint.py
index 3689fed8..799b2263 100755
--- a/server/api_server/events_endpoint.py
+++ b/server/api_server/events_endpoint.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import sys
@@ -9,10 +9,10 @@ from flask import jsonify
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from database import get_temp_db_connection
-from helper import is_random_mac, mylog
-from db.db_helper import row_to_json, get_date_from_period
-from utils.datetime_utils import format_date, format_date_iso, format_event_date, ensure_datetime
+from database import get_temp_db_connection # noqa: E402 [flake8 lint suppression]
+from helper import mylog # noqa: E402 [flake8 lint suppression]
+from db.db_helper import row_to_json, get_date_from_period # noqa: E402 [flake8 lint suppression]
+from utils.datetime_utils import ensure_datetime # noqa: E402 [flake8 lint suppression]
# --------------------------
@@ -120,14 +120,14 @@ def get_events_totals(period: str = "7 days"):
cur = conn.cursor()
sql = f"""
- SELECT
+ SELECT
(SELECT COUNT(*) FROM Events WHERE eve_DateTime >= {period_date_sql}) AS all_events,
- (SELECT COUNT(*) FROM Sessions WHERE
+ (SELECT COUNT(*) FROM Sessions WHERE
ses_DateTimeConnection >= {period_date_sql}
OR ses_DateTimeDisconnection >= {period_date_sql}
OR ses_StillConnected = 1
) AS sessions,
- (SELECT COUNT(*) FROM Sessions WHERE
+ (SELECT COUNT(*) FROM Sessions WHERE
(ses_DateTimeConnection IS NULL AND ses_DateTimeDisconnection >= {period_date_sql})
OR (ses_DateTimeDisconnection IS NULL AND ses_StillConnected = 0 AND ses_DateTimeConnection >= {period_date_sql})
) AS missing,
diff --git a/server/api_server/graphql_endpoint.py b/server/api_server/graphql_endpoint.py
index 7e12931c..9ea995bf 100755
--- a/server/api_server/graphql_endpoint.py
+++ b/server/api_server/graphql_endpoint.py
@@ -1,5 +1,7 @@
import graphene
-from graphene import ObjectType, String, Int, Boolean, List, Field, InputObjectType, Argument
+from graphene import (
+ ObjectType, String, Int, Boolean, List, Field, InputObjectType, Argument
+)
import json
import sys
import os
@@ -8,9 +10,9 @@ import os
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
sys.path.extend([f"{INSTALL_PATH}/server"])
-from logger import mylog
-from const import apiPath
-from helper import (
+from logger import mylog # noqa: E402 [flake8 lint suppression]
+from const import apiPath # noqa: E402 [flake8 lint suppression]
+from helper import ( # noqa: E402 [flake8 lint suppression]
is_random_mac,
get_number_of_children,
format_ip_long,
@@ -111,12 +113,14 @@ class SettingResult(ObjectType):
settings = List(Setting)
count = Int()
-# --- LANGSTRINGS ---
+# --- LANGSTRINGS ---
+
# In-memory cache for lang strings
_langstrings_cache = {} # caches lists per file (core JSON or plugin)
_langstrings_cache_mtime = {} # tracks last modified times
+
# LangString ObjectType
class LangString(ObjectType):
langCode = String()
@@ -128,6 +132,7 @@ class LangStringResult(ObjectType):
langStrings = List(LangString)
count = Int()
+
# Define Query Type with Pagination Support
class Query(ObjectType):
# --- DEVICES ---
@@ -184,31 +189,39 @@ class Query(ObjectType):
if (device.get("devParentRelType") not in hidden_relationships)
]
- devices_data = [
- device
- for device in devices_data
- if (
- (
- device["devPresentLastScan"] == 1
- and "online" in allowed_statuses
- )
- or (device["devIsNew"] == 1 and "new" in allowed_statuses)
- or (
- device["devPresentLastScan"] == 0
- and device["devAlertDown"]
- and "down" in allowed_statuses
- )
- or (
- device["devPresentLastScan"] == 0
- and "offline" in allowed_statuses
- )
- and device["devIsArchived"] == 0
- or (
- device["devIsArchived"] == 1
- and "archived" in allowed_statuses
- )
+ filtered = []
+
+ for device in devices_data:
+ is_online = (
+ device["devPresentLastScan"] == 1 and "online" in allowed_statuses
)
- ]
+
+ is_new = (
+ device["devIsNew"] == 1 and "new" in allowed_statuses
+ )
+
+ is_down = (
+ device["devPresentLastScan"] == 0 and device["devAlertDown"] and "down" in allowed_statuses
+ )
+
+ is_offline = (
+ device["devPresentLastScan"] == 0 and "offline" in allowed_statuses
+ )
+
+ is_archived = (
+ device["devIsArchived"] == 1 and "archived" in allowed_statuses
+ )
+
+ # Matches if not archived and status matches OR it is archived and allowed
+ matches = (
+ (is_online or is_new or is_down or is_offline) and device["devIsArchived"] == 0
+ ) or is_archived
+
+ if matches:
+ filtered.append(device)
+
+ devices_data = filtered
+
elif status == "connected":
devices_data = [
device
@@ -257,8 +270,7 @@ class Query(ObjectType):
devices_data = [
device
for device in devices_data
- if str(device.get(filter.filterColumn, "")).lower()
- == str(filter.filterValue).lower()
+ if str(device.get(filter.filterColumn, "")).lower() == str(filter.filterValue).lower()
]
# Search data if a search term is provided
@@ -340,7 +352,7 @@ class Query(ObjectType):
return SettingResult(settings=settings, count=len(settings))
- # --- LANGSTRINGS ---
+ # --- LANGSTRINGS ---
langStrings = Field(
LangStringResult,
langCode=Argument(String, required=False),
@@ -437,11 +449,11 @@ class Query(ObjectType):
if en_fallback:
langStrings[i] = en_fallback[0]
- mylog('trace', f'[graphql_schema] Collected {len(langStrings)} language strings '
- f'(langCode={langCode}, key={langStringKey}, fallback_to_en={fallback_to_en})')
+ mylog('trace', f'[graphql_schema] Collected {len(langStrings)} language strings (langCode={langCode}, key={langStringKey}, fallback_to_en={fallback_to_en})')
return LangStringResult(langStrings=langStrings, count=len(langStrings))
+
# helps sorting inconsistent dataset mixed integers and strings
def mixed_type_sort_key(value):
if value is None or value == "":
diff --git a/server/api_server/history_endpoint.py b/server/api_server/history_endpoint.py
index fd766167..8a28ca4a 100755
--- a/server/api_server/history_endpoint.py
+++ b/server/api_server/history_endpoint.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import sys
@@ -8,7 +8,7 @@ from flask import jsonify
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from database import get_temp_db_connection
+from database import get_temp_db_connection # noqa: E402 [flake8 lint suppression]
# --------------------------------------------------
diff --git a/server/api_server/logs_endpoint.py b/server/api_server/logs_endpoint.py
index 120644b7..d3a8fd50 100644
--- a/server/api_server/logs_endpoint.py
+++ b/server/api_server/logs_endpoint.py
@@ -3,18 +3,18 @@ import sys
from flask import jsonify
# Register NetAlertX directories
-INSTALL_PATH="/app"
+INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from const import logPath
-from logger import mylog, Logger
-from helper import get_setting_value
-from utils.datetime_utils import timeNowDB
-from messaging.in_app import write_notification
+from const import logPath # noqa: E402 [flake8 lint suppression]
+from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from messaging.in_app import write_notification # noqa: E402 [flake8 lint suppression]
# Make sure log level is initialized correctly
Logger(get_setting_value('LOG_LEVEL'))
+
def clean_log(log_file):
"""
Purge the content of an allowed log file within the /app/log/ directory.
@@ -55,4 +55,3 @@ def clean_log(log_file):
mylog('none', [msg])
write_notification(msg, 'interrupt')
return jsonify({"success": False, "message": msg}), 500
-
diff --git a/server/api_server/prometheus_endpoint.py b/server/api_server/prometheus_endpoint.py
index 7a32937f..593e3f4e 100755
--- a/server/api_server/prometheus_endpoint.py
+++ b/server/api_server/prometheus_endpoint.py
@@ -6,8 +6,8 @@ import os
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
sys.path.extend([f"{INSTALL_PATH}/server"])
-from logger import mylog
-from const import apiPath
+from logger import mylog # noqa: E402 [flake8 lint suppression]
+from const import apiPath # noqa: E402 [flake8 lint suppression]
def escape_label_value(val):
diff --git a/server/api_server/sessions_endpoint.py b/server/api_server/sessions_endpoint.py
index 1e0643f1..703ad307 100755
--- a/server/api_server/sessions_endpoint.py
+++ b/server/api_server/sessions_endpoint.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
import os
import sqlite3
@@ -9,10 +9,10 @@ from flask import jsonify
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from database import get_temp_db_connection
-from helper import is_random_mac, get_setting_value, mylog, format_ip_long
-from db.db_helper import row_to_json, get_date_from_period
-from utils.datetime_utils import format_date_iso, format_event_date, format_date_diff, parse_datetime, format_date
+from database import get_temp_db_connection # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value, format_ip_long # noqa: E402 [flake8 lint suppression]
+from db.db_helper import get_date_from_period # noqa: E402 [flake8 lint suppression]
+from utils.datetime_utils import format_date_iso, format_event_date, format_date_diff, format_date # noqa: E402 [flake8 lint suppression]
# --------------------------
@@ -33,7 +33,7 @@ def create_session(
cur.execute(
"""
- INSERT INTO Sessions (ses_MAC, ses_IP, ses_DateTimeConnection, ses_DateTimeDisconnection,
+ INSERT INTO Sessions (ses_MAC, ses_IP, ses_DateTimeConnection, ses_DateTimeDisconnection,
ses_EventTypeConnection, ses_EventTypeDisconnection)
VALUES (?, ?, ?, ?, ?, ?)
""",
@@ -105,7 +105,7 @@ def get_sessions_calendar(start_date, end_date):
-- If ses_EventTypeConnection is missing, backfill from last disconnection
-- If ses_EventTypeDisconnection is missing, forward-fill from next connection
- SELECT
+ SELECT
SES1.ses_MAC, SES1.ses_EventTypeConnection, SES1.ses_DateTimeConnection,
SES1.ses_EventTypeDisconnection, SES1.ses_DateTimeDisconnection, SES1.ses_IP,
SES1.ses_AdditionalInfo, SES1.ses_StillConnected,
@@ -113,9 +113,9 @@ def get_sessions_calendar(start_date, end_date):
CASE
WHEN SES1.ses_EventTypeConnection = '' THEN
IFNULL(
- (SELECT MAX(SES2.ses_DateTimeDisconnection)
- FROM Sessions AS SES2
- WHERE SES2.ses_MAC = SES1.ses_MAC
+ (SELECT MAX(SES2.ses_DateTimeDisconnection)
+ FROM Sessions AS SES2
+ WHERE SES2.ses_MAC = SES1.ses_MAC
AND SES2.ses_DateTimeDisconnection < SES1.ses_DateTimeDisconnection
AND SES2.ses_DateTimeDisconnection BETWEEN Date(?) AND Date(?)
),
@@ -126,9 +126,9 @@ def get_sessions_calendar(start_date, end_date):
CASE
WHEN SES1.ses_EventTypeDisconnection = '' THEN
- (SELECT MIN(SES2.ses_DateTimeConnection)
- FROM Sessions AS SES2
- WHERE SES2.ses_MAC = SES1.ses_MAC
+ (SELECT MIN(SES2.ses_DateTimeConnection)
+ FROM Sessions AS SES2
+ WHERE SES2.ses_MAC = SES1.ses_MAC
AND SES2.ses_DateTimeConnection > SES1.ses_DateTimeConnection
AND SES2.ses_DateTimeConnection BETWEEN Date(?) AND Date(?)
)
@@ -162,8 +162,7 @@ def get_sessions_calendar(start_date, end_date):
# Determine color
if (
- row["ses_EventTypeConnection"] == ""
- or row["ses_EventTypeDisconnection"] == ""
+ row["ses_EventTypeConnection"] == "" or row["ses_EventTypeDisconnection"] == ""
):
color = "#f39c12"
elif row["ses_StillConnected"] == 1:
@@ -205,7 +204,7 @@ def get_device_sessions(mac, period):
cur = conn.cursor()
sql = f"""
- SELECT
+ SELECT
IFNULL(ses_DateTimeConnection, ses_DateTimeDisconnection) AS ses_DateTimeOrder,
ses_EventTypeConnection,
ses_DateTimeConnection,
@@ -293,7 +292,7 @@ def get_session_events(event_type, period_date):
# Base SQLs
sql_events = f"""
- SELECT
+ SELECT
eve_DateTime AS eve_DateTimeOrder,
devName,
devOwner,
@@ -314,7 +313,7 @@ def get_session_events(event_type, period_date):
"""
sql_sessions = """
- SELECT
+ SELECT
IFNULL(ses_DateTimeConnection, ses_DateTimeDisconnection) AS ses_DateTimeOrder,
devName,
devOwner,
@@ -337,8 +336,7 @@ def get_session_events(event_type, period_date):
sql = sql_events
elif event_type == "sessions":
sql = (
- sql_sessions
- + f"""
+ sql_sessions + f"""
WHERE (
ses_DateTimeConnection >= {period_date}
OR ses_DateTimeDisconnection >= {period_date}
@@ -348,8 +346,7 @@ def get_session_events(event_type, period_date):
)
elif event_type == "missing":
sql = (
- sql_sessions
- + f"""
+ sql_sessions + f"""
WHERE (
(ses_DateTimeConnection IS NULL AND ses_DateTimeDisconnection >= {period_date})
OR (ses_DateTimeDisconnection IS NULL AND ses_StillConnected = 0 AND ses_DateTimeConnection >= {period_date})
diff --git a/server/app_state.py b/server/app_state.py
index 8120a8b5..28f469db 100755
--- a/server/app_state.py
+++ b/server/app_state.py
@@ -1,7 +1,7 @@
import os
import json
-from const import *
+from const import applicationPath, apiPath
from logger import mylog
from helper import checkNewVersion
from utils.datetime_utils import timeNowDB, timeNow
@@ -32,14 +32,17 @@ class app_state_class:
isNewVersionChecked (int): Timestamp of last version check.
"""
- def __init__(self, currentState=None,
- settingsSaved=None,
- settingsImported=None,
- showSpinner=None,
- graphQLServerStarted=0,
- processScan=False,
- pluginsStates=None,
- appVersion=None):
+ def __init__(
+ self,
+ currentState=None,
+ settingsSaved=None,
+ settingsImported=None,
+ showSpinner=None,
+ graphQLServerStarted=0,
+ processScan=False,
+ pluginsStates=None,
+ appVersion=None
+ ):
"""
Initialize the application state, optionally overwriting previous values.
@@ -62,7 +65,7 @@ class app_state_class:
# Update self
self.lastUpdated = str(timeNowDB())
-
+
if os.path.exists(stateFile):
try:
with open(stateFile, "r") as json_file:
@@ -73,7 +76,7 @@ class app_state_class:
)
# Check if the file exists and recover previous values
- if previousState != "":
+ if previousState != "":
self.settingsSaved = previousState.get("settingsSaved", 0)
self.settingsImported = previousState.get("settingsImported", 0)
self.processScan = previousState.get("processScan", False)
@@ -82,9 +85,9 @@ class app_state_class:
self.isNewVersionChecked = previousState.get("isNewVersionChecked", 0)
self.graphQLServerStarted = previousState.get("graphQLServerStarted", 0)
self.currentState = previousState.get("currentState", "Init")
- self.pluginsStates = previousState.get("pluginsStates", {})
- self.appVersion = previousState.get("appVersion", "")
- else: # init first time values
+ self.pluginsStates = previousState.get("pluginsStates", {})
+ self.appVersion = previousState.get("appVersion", "")
+ else: # init first time values
self.settingsSaved = 0
self.settingsImported = 0
self.showSpinner = False
@@ -158,12 +161,12 @@ class app_state_class:
# -------------------------------------------------------------------------------
# method to update the state
-def updateState(newState = None,
- settingsSaved = None,
- settingsImported = None,
- showSpinner = None,
- graphQLServerStarted = None,
- processScan = None,
+def updateState(newState = None,
+ settingsSaved = None,
+ settingsImported = None,
+ showSpinner = None,
+ graphQLServerStarted = None,
+ processScan = None,
pluginsStates=None,
appVersion=None):
"""
@@ -182,14 +185,16 @@ def updateState(newState = None,
Returns:
app_state_class: Updated state object.
"""
- return app_state_class( newState,
- settingsSaved,
- settingsImported,
- showSpinner,
- graphQLServerStarted,
- processScan,
- pluginsStates,
- appVersion)
+ return app_state_class(
+ newState,
+ settingsSaved,
+ settingsImported,
+ showSpinner,
+ graphQLServerStarted,
+ processScan,
+ pluginsStates,
+ appVersion
+ )
# -------------------------------------------------------------------------------
diff --git a/server/const.py b/server/const.py
index 2714bcd3..fe2c2317 100755
--- a/server/const.py
+++ b/server/const.py
@@ -52,7 +52,7 @@ default_tz = "Europe/Berlin"
# SQL queries
# ===============================================================================
sql_devices_all = """
- SELECT
+ SELECT
rowid,
IFNULL(devMac, '') AS devMac,
IFNULL(devName, '') AS devName,
@@ -88,7 +88,7 @@ sql_devices_all = """
IFNULL(devFQDN, '') AS devFQDN,
IFNULL(devParentRelType, '') AS devParentRelType,
IFNULL(devReqNicsOnline, '') AS devReqNicsOnline,
- CASE
+ CASE
WHEN devIsNew = 1 THEN 'New'
WHEN devPresentLastScan = 1 THEN 'On-line'
WHEN devPresentLastScan = 0 AND devAlertDown != 0 THEN 'Down'
@@ -133,7 +133,7 @@ sql_devices_tiles = """
(SELECT COUNT(*) FROM Devices) AS "all_devices",
-- My Devices count
(SELECT COUNT(*) FROM MyDevicesFilter) AS my_devices
- FROM Statuses;
+ FROM Statuses;
"""
sql_devices_filters = """
SELECT DISTINCT 'devSite' AS columnName, devSite AS columnValue
@@ -164,9 +164,9 @@ sql_devices_filters = """
FROM Devices WHERE devSSID NOT IN ('', 'null') AND devSSID IS NOT NULL
ORDER BY columnName;
"""
-sql_devices_stats = """SELECT Online_Devices as online, Down_Devices as down, All_Devices as 'all', Archived_Devices as archived,
- (select count(*) from Devices a where devIsNew = 1 ) as new,
- (select count(*) from Devices a where devName = '(unknown)' or devName = '(name not found)' ) as unknown
+sql_devices_stats = """SELECT Online_Devices as online, Down_Devices as down, All_Devices as 'all', Archived_Devices as archived,
+ (select count(*) from Devices a where devIsNew = 1 ) as new,
+ (select count(*) from Devices a where devName = '(unknown)' or devName = '(name not found)' ) as unknown
from Online_History order by Scan_Date desc limit 1"""
sql_events_pending_alert = "SELECT * FROM Events where eve_PendingAlertEmail is not 0"
sql_settings = "SELECT * FROM Settings"
@@ -176,23 +176,23 @@ sql_notifications_all = "SELECT * FROM Notifications"
sql_online_history = "SELECT * FROM Online_History"
sql_plugins_events = "SELECT * FROM Plugins_Events"
sql_plugins_history = "SELECT * FROM Plugins_History ORDER BY DateTimeChanged DESC"
-sql_new_devices = """SELECT * FROM (
- SELECT eve_IP as devLastIP, eve_MAC as devMac
+sql_new_devices = """SELECT * FROM (
+ SELECT eve_IP as devLastIP, eve_MAC as devMac
FROM Events_Devices
WHERE eve_PendingAlertEmail = 1
AND eve_EventType = 'New Device'
ORDER BY eve_DateTime ) t1
- LEFT JOIN
- ( SELECT devName, devMac as devMac_t2 FROM Devices) t2
+ LEFT JOIN
+ ( SELECT devName, devMac as devMac_t2 FROM Devices) t2
ON t1.devMac = t2.devMac_t2"""
sql_generateGuid = """
lower(
- hex(randomblob(4)) || '-' || hex(randomblob(2)) || '-' || '4' ||
- substr(hex( randomblob(2)), 2) || '-' ||
+ hex(randomblob(4)) || '-' || hex(randomblob(2)) || '-' || '4' ||
+ substr(hex( randomblob(2)), 2) || '-' ||
substr('AB89', 1 + (abs(random()) % 4) , 1) ||
- substr(hex(randomblob(2)), 2) || '-' ||
+ substr(hex(randomblob(2)), 2) || '-' ||
hex(randomblob(6))
)
"""
diff --git a/server/database.py b/server/database.py
index 3c19fdb7..8f7845bf 100755
--- a/server/database.py
+++ b/server/database.py
@@ -180,7 +180,7 @@ class DB:
# Init the AppEvent database table
AppEvent_obj(self)
- # #-------------------------------------------------------------------------------
+ # # -------------------------------------------------------------------------------
# def get_table_as_json(self, sqlQuery):
# # mylog('debug',[ '[Database] - get_table_as_json - Query: ', sqlQuery])
diff --git a/server/db/db_helper.py b/server/db/db_helper.py
index 01a5ccd8..3d394d7f 100755
--- a/server/db/db_helper.py
+++ b/server/db/db_helper.py
@@ -6,8 +6,8 @@ import os
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
sys.path.extend([f"{INSTALL_PATH}/server"])
-from helper import if_byte_then_to_str
-from logger import mylog
+from helper import if_byte_then_to_str # noqa: E402 [flake8 lint suppression]
+from logger import mylog # noqa: E402 [flake8 lint suppression]
# -------------------------------------------------------------------------------
diff --git a/server/db/db_upgrade.py b/server/db/db_upgrade.py
index f634f5b7..d3b5f4d8 100755
--- a/server/db/db_upgrade.py
+++ b/server/db/db_upgrade.py
@@ -5,8 +5,8 @@ import os
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
sys.path.extend([f"{INSTALL_PATH}/server"])
-from logger import mylog
-from messaging.in_app import write_notification
+from logger import mylog # noqa: E402 [flake8 lint suppression]
+from messaging.in_app import write_notification # noqa: E402 [flake8 lint suppression]
def ensure_column(sql, table: str, column_name: str, column_type: str) -> bool:
@@ -108,23 +108,23 @@ def ensure_views(sql) -> bool:
- sql: database cursor or connection wrapper (must support execute() and fetchall()).
"""
sql.execute(""" DROP VIEW IF EXISTS Events_Devices;""")
- sql.execute(""" CREATE VIEW Events_Devices AS
- SELECT *
- FROM Events
+ sql.execute(""" CREATE VIEW Events_Devices AS
+ SELECT *
+ FROM Events
LEFT JOIN Devices ON eve_MAC = devMac;
""")
sql.execute(""" DROP VIEW IF EXISTS LatestEventsPerMAC;""")
sql.execute("""CREATE VIEW LatestEventsPerMAC AS
WITH RankedEvents AS (
- SELECT
+ SELECT
e.*,
ROW_NUMBER() OVER (PARTITION BY e.eve_MAC ORDER BY e.eve_DateTime DESC) AS row_num
FROM Events AS e
)
- SELECT
- e.*,
- d.*,
+ SELECT
+ e.*,
+ d.*,
c.*
FROM RankedEvents AS e
LEFT JOIN Devices AS d ON e.eve_MAC = d.devMac
@@ -138,14 +138,14 @@ def ensure_views(sql) -> bool:
sql.execute(""" CREATE VIEW IF NOT EXISTS LatestEventsPerMAC AS
WITH RankedEvents AS (
- SELECT
+ SELECT
e.*,
ROW_NUMBER() OVER (PARTITION BY e.eve_MAC ORDER BY e.eve_DateTime DESC) AS row_num
FROM Events AS e
)
- SELECT
- e.*,
- d.*,
+ SELECT
+ e.*,
+ d.*,
c.*
FROM RankedEvents AS e
LEFT JOIN Devices AS d ON e.eve_MAC = d.devMac
@@ -272,7 +272,7 @@ def ensure_CurrentScan(sql) -> bool:
"""
# 🐛 CurrentScan DEBUG: comment out below when debugging to keep the CurrentScan table after restarts/scan finishes
sql.execute("DROP TABLE IF EXISTS CurrentScan;")
- sql.execute(""" CREATE TABLE IF NOT EXISTS CurrentScan (
+ sql.execute(""" CREATE TABLE IF NOT EXISTS CurrentScan (
cur_MAC STRING(50) NOT NULL COLLATE NOCASE,
cur_IP STRING(50) NOT NULL COLLATE NOCASE,
cur_Vendor STRING(250),
@@ -354,7 +354,7 @@ def ensure_plugins_tables(sql) -> bool:
# Plugin state
sql_Plugins_Objects = """ CREATE TABLE IF NOT EXISTS Plugins_Objects(
"Index" INTEGER,
- Plugin TEXT NOT NULL,
+ Plugin TEXT NOT NULL,
Object_PrimaryID TEXT NOT NULL,
Object_SecondaryID TEXT NOT NULL,
DateTimeCreated TEXT NOT NULL,
diff --git a/server/db/sql_safe_builder.py b/server/db/sql_safe_builder.py
index fc5ec003..42c504dc 100755
--- a/server/db/sql_safe_builder.py
+++ b/server/db/sql_safe_builder.py
@@ -18,7 +18,7 @@ from typing import Dict, List, Tuple, Any, Optional
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
sys.path.extend([f"{INSTALL_PATH}/server"])
-from logger import mylog
+from logger import mylog # noqa: E402 [flake8 lint suppression]
class SafeConditionBuilder:
@@ -494,8 +494,6 @@ class SafeConditionBuilder:
if logical_op and not self._validate_logical_operator(logical_op):
raise ValueError(f"Invalid logical operator: {logical_op}")
- # Parse values from the IN clause
- values = []
# Simple regex to extract quoted values
value_pattern = r"'([^']*)'"
matches = re.findall(value_pattern, values_str)
diff --git a/server/helper.py b/server/helper.py
index a10dd9f8..dbb9588e 100755
--- a/server/helper.py
+++ b/server/helper.py
@@ -7,25 +7,22 @@ import os
import re
import unicodedata
import subprocess
-import pytz
import json
import requests
import base64
import hashlib
import random
-import email
import string
import ipaddress
import conf
-from const import *
+from const import applicationPath, fullConfPath, fullDbPath, dbPath, confPath, apiPath
from logger import mylog, logResult
# Register NetAlertX directories using runtime configuration
INSTALL_PATH = applicationPath
-
# -------------------------------------------------------------------------------
# File system permission handling
# -------------------------------------------------------------------------------
@@ -58,12 +55,6 @@ def checkPermissionsOK():
return (confR_access, dbR_access)
-# -------------------------------------------------------------------------------
-def fixPermissions():
- # Try fixing access rights if needed
- chmodCommands = []
-
-
# -------------------------------------------------------------------------------
def initialiseFile(pathToCheck, defaultFile):
# if file not readable (missing?) try to copy over the backed-up (default) one
@@ -71,9 +62,7 @@ def initialiseFile(pathToCheck, defaultFile):
mylog(
"none",
[
- "[Setup] ("
- + pathToCheck
- + ") file is not readable or missing. Trying to copy over the default one."
+ "[Setup] (" + pathToCheck + ") file is not readable or missing. Trying to copy over the default one."
],
)
try:
@@ -89,22 +78,14 @@ def initialiseFile(pathToCheck, defaultFile):
mylog(
"none",
[
- "[Setup] ⚠ ERROR copying ("
- + defaultFile
- + ") to ("
- + pathToCheck
- + "). Make sure the app has Read & Write access to the parent directory."
+ "[Setup] ⚠ ERROR copying (" + defaultFile + ") to (" + pathToCheck + "). Make sure the app has Read & Write access to the parent directory."
],
)
else:
mylog(
"none",
[
- "[Setup] ("
- + defaultFile
- + ") copied over successfully to ("
- + pathToCheck
- + ")."
+ "[Setup] (" + defaultFile + ") copied over successfully to (" + pathToCheck + ")."
],
)
@@ -116,10 +97,7 @@ def initialiseFile(pathToCheck, defaultFile):
mylog(
"none",
[
- "[Setup] ⚠ ERROR copying ("
- + defaultFile
- + "). Make sure the app has Read & Write access to "
- + pathToCheck
+ "[Setup] ⚠ ERROR copying (" + defaultFile + "). Make sure the app has Read & Write access to " + pathToCheck
],
)
mylog("none", [e.output])
@@ -130,16 +108,13 @@ def filePermissions():
# check and initialize .conf
(confR_access, dbR_access) = checkPermissionsOK() # Initial check
- if confR_access == False:
+ if confR_access is False:
initialiseFile(fullConfPath, f"{INSTALL_PATH}/back/app.conf")
# check and initialize .db
- if dbR_access == False:
+ if dbR_access is False:
initialiseFile(fullDbPath, f"{INSTALL_PATH}/back/app.db")
- # last attempt
- fixPermissions()
-
# -------------------------------------------------------------------------------
# File manipulation methods
@@ -292,7 +267,7 @@ def get_setting_value(key):
value = setting_value_to_python_type(set_type, set_value)
else:
value = setting_value_to_python_type(set_type, str(set_value))
-
+
SETTINGS_SECONDARYCACHE[key] = value
return value
@@ -382,7 +357,7 @@ def setting_value_to_python_type(set_type, set_value):
if isinstance(set_value, str):
try:
value = json.loads(set_value.replace("'", "\""))
-
+
except json.JSONDecodeError as e:
mylog(
"none",
@@ -413,17 +388,12 @@ def setting_value_to_python_type(set_type, set_value):
value = set_value
elif (
- dataType == "string"
- and elementType == "input"
- and any(opt.get("readonly") == "true" for opt in elementOptions)
+ dataType == "string" and elementType == "input" and any(opt.get("readonly") == "true" for opt in elementOptions)
):
value = reverseTransformers(str(set_value), transformers)
elif (
- dataType == "string"
- and elementType == "input"
- and any(opt.get("type") == "password" for opt in elementOptions)
- and "sha256" in transformers
+ dataType == "string" and elementType == "input" and any(opt.get("type") == "password" for opt in elementOptions) and "sha256" in transformers
):
value = hashlib.sha256(set_value.encode()).hexdigest()
@@ -602,23 +572,23 @@ def normalize_string(text):
# -------------------------------------------------------------------------------
-# -------------------------------------------------------------------------------------------
-def is_random_mac(mac: str) -> bool:
- """Determine if a MAC address is random, respecting user-defined prefixes not to mark as random."""
+# # -------------------------------------------------------------------------------------------
+# def is_random_mac(mac: str) -> bool:
+# """Determine if a MAC address is random, respecting user-defined prefixes not to mark as random."""
- is_random = mac[1].upper() in ["2", "6", "A", "E"]
+# is_random = mac[1].upper() in ["2", "6", "A", "E"]
- # Get prefixes from settings
- prefixes = get_setting_value("UI_NOT_RANDOM_MAC")
+# # Get prefixes from settings
+# prefixes = get_setting_value("UI_NOT_RANDOM_MAC")
- # If detected as random, make sure it doesn't start with a prefix the user wants to exclude
- if is_random:
- for prefix in prefixes:
- if mac.upper().startswith(prefix.upper()):
- is_random = False
- break
+# # If detected as random, make sure it doesn't start with a prefix the user wants to exclude
+# if is_random:
+# for prefix in prefixes:
+# if mac.upper().startswith(prefix.upper()):
+# is_random = False
+# break
- return is_random
+# return is_random
# -------------------------------------------------------------------------------------------
@@ -653,6 +623,7 @@ def extract_ip_addresses(text):
# -------------------------------------------------------------------------------
# Helper function to determine if a MAC address is random
def is_random_mac(mac):
+ """Determine if a MAC address is random, respecting user-defined prefixes not to mark as random."""
# Check if second character matches "2", "6", "A", "E" (case insensitive)
is_random = mac[1].upper() in ["2", "6", "A", "E"]
@@ -773,7 +744,6 @@ def getBuildTimeStampAndVersion():
return tuple(results)
-
# -------------------------------------------------------------------------------
def checkNewVersion():
mylog("debug", ["[Version check] Checking if new version available"])
diff --git a/server/initialise.py b/server/initialise.py
index 75de99d4..1552d976 100755
--- a/server/initialise.py
+++ b/server/initialise.py
@@ -8,9 +8,9 @@ import shutil
import re
# Register NetAlertX libraries
-import conf
-from const import fullConfPath, applicationPath, fullConfFolder, default_tz
-from helper import getBuildTimeStampAndVersion, fixPermissions, collect_lang_strings, updateSubnets, isJsonObject, setting_value_to_python_type, get_setting_value, generate_random_string
+import conf
+from const import fullConfPath, fullConfFolder, default_tz
+from helper import getBuildTimeStampAndVersion, fixPermissions, collect_lang_strings, updateSubnets, generate_random_string
from utils.datetime_utils import timeNowDB
from app_state import updateState
from logger import mylog
@@ -19,7 +19,6 @@ from scheduler import schedule_class
from plugin import plugin_manager, print_plugin_info
from utils.plugin_utils import get_plugins_configs, get_set_value_for_init
from messaging.in_app import write_notification
-from utils.crypto_utils import get_random_bytes
# ===============================================================================
# Initialise user defined values
@@ -59,7 +58,7 @@ def ccd(
result = default
# Use existing value if already supplied, otherwise default value is used
- if forceDefault == False and key in config_dir:
+ if forceDefault is False and key in config_dir:
result = config_dir[key]
# Single quotes might break SQL queries, replacing them
@@ -216,7 +215,7 @@ def importConfigs(pm, db, all_plugins):
[],
c_d,
"Loaded plugins",
- '{"dataType":"array","elements":[{"elementType":"select","elementHasInputValue":1,"elementOptions":[{"multiple":"true","ordeable":"true"}],"transformers":[]},{"elementType":"button","elementOptions":[{"sourceSuffixes":[]},{"separator":""},{"cssClasses":"col-xs-12"},{"onClick":"selectChange(this)"},{"getStringKey":"Gen_Change"}],"transformers":[]}]}',
+ '{"dataType":"array","elements":[{"elementType":"select","elementHasInputValue":1,"elementOptions":[{"multiple":"true","ordeable":"true"}],"transformers":[]},{"elementType":"button","elementOptions":[{"sourceSuffixes":[]},{"separator":""},{"cssClasses":"col-xs-12"},{"onClick":"selectChange(this)"},{"getStringKey":"Gen_Change"}],"transformers":[]}]}', # noqa: E501
"[]",
"General",
)
@@ -234,7 +233,7 @@ def importConfigs(pm, db, all_plugins):
["192.168.1.0/24 --interface=eth1", "192.168.1.0/24 --interface=eth0"],
c_d,
"Subnets to scan",
- """{"dataType": "array","elements": [{"elementType": "input","elementOptions": [{"placeholder": "192.168.1.0/24 --interface=eth1"},{"suffix": "_in"},{"cssClasses": "col-sm-10"},{"prefillValue": "null"}],"transformers": []},{"elementType": "button","elementOptions": [{"sourceSuffixes": ["_in"]},{"separator": ""},{"cssClasses": "col-xs-12"},{"onClick": "addList(this, false)"},{"getStringKey": "Gen_Add"}],"transformers": []},{"elementType": "select","elementHasInputValue": 1,"elementOptions": [{"multiple": "true"},{"readonly": "true"},{"editable": "true"}],"transformers": []},{"elementType": "button","elementOptions": [{"sourceSuffixes": []},{"separator": ""},{"cssClasses": "col-xs-6"},{"onClick": "removeAllOptions(this)"},{"getStringKey": "Gen_Remove_All"}],"transformers": []},{"elementType": "button","elementOptions": [{"sourceSuffixes": []},{"separator": ""},{"cssClasses": "col-xs-6"},{"onClick": "removeFromList(this)"},{"getStringKey": "Gen_Remove_Last"}],"transformers": []}]}""",
+ """{"dataType": "array","elements": [{"elementType": "input","elementOptions": [{"placeholder": "192.168.1.0/24 --interface=eth1"},{"suffix": "_in"},{"cssClasses": "col-sm-10"},{"prefillValue": "null"}],"transformers": []},{"elementType": "button","elementOptions": [{"sourceSuffixes": ["_in"]},{"separator": ""},{"cssClasses": "col-xs-12"},{"onClick": "addList(this, false)"},{"getStringKey": "Gen_Add"}],"transformers": []},{"elementType": "select","elementHasInputValue": 1,"elementOptions": [{"multiple": "true"},{"readonly": "true"},{"editable": "true"}],"transformers": []},{"elementType": "button","elementOptions": [{"sourceSuffixes": []},{"separator": ""},{"cssClasses": "col-xs-6"},{"onClick": "removeAllOptions(this)"},{"getStringKey": "Gen_Remove_All"}],"transformers": []},{"elementType": "button","elementOptions": [{"sourceSuffixes": []},{"separator": ""},{"cssClasses": "col-xs-6"},{"onClick": "removeFromList(this)"},{"getStringKey": "Gen_Remove_Last"}],"transformers": []}]}""", # noqa: E501 - inline JSON
"[]",
"General",
)
@@ -356,7 +355,7 @@ def importConfigs(pm, db, all_plugins):
],
c_d,
"Network device types",
- '{"dataType":"array","elements":[{"elementType":"input","elementOptions":[{"placeholder":"Enter value"},{"suffix":"_in"},{"cssClasses":"col-sm-10"},{"prefillValue":"null"}],"transformers":[]},{"elementType":"button","elementOptions":[{"sourceSuffixes":["_in"]},{"separator":""},{"cssClasses":"col-xs-12"},{"onClick":"addList(this,false)"},{"getStringKey":"Gen_Add"}],"transformers":[]},{"elementType":"select", "elementHasInputValue":1,"elementOptions":[{"multiple":"true"},{"readonly":"true"},{"editable":"true"}],"transformers":[]},{"elementType":"button","elementOptions":[{"sourceSuffixes":[]},{"separator":""},{"cssClasses":"col-xs-6"},{"onClick":"removeAllOptions(this)"},{"getStringKey":"Gen_Remove_All"}],"transformers":[]},{"elementType":"button","elementOptions":[{"sourceSuffixes":[]},{"separator":""},{"cssClasses":"col-xs-6"},{"onClick":"removeFromList(this)"},{"getStringKey":"Gen_Remove_Last"}],"transformers":[]}]}',
+ '{"dataType":"array","elements":[{"elementType":"input","elementOptions":[{"placeholder":"Enter value"},{"suffix":"_in"},{"cssClasses":"col-sm-10"},{"prefillValue":"null"}],"transformers":[]},{"elementType":"button","elementOptions":[{"sourceSuffixes":["_in"]},{"separator":""},{"cssClasses":"col-xs-12"},{"onClick":"addList(this,false)"},{"getStringKey":"Gen_Add"}],"transformers":[]},{"elementType":"select", "elementHasInputValue":1,"elementOptions":[{"multiple":"true"},{"readonly":"true"},{"editable":"true"}],"transformers":[]},{"elementType":"button","elementOptions":[{"sourceSuffixes":[]},{"separator":""},{"cssClasses":"col-xs-6"},{"onClick":"removeAllOptions(this)"},{"getStringKey":"Gen_Remove_All"}],"transformers":[]},{"elementType":"button","elementOptions":[{"sourceSuffixes":[]},{"separator":""},{"cssClasses":"col-xs-6"},{"onClick":"removeFromList(this)"},{"getStringKey":"Gen_Remove_Last"}],"transformers":[]}]}', # noqa: E501 - inline JSON
"[]",
"General",
)
@@ -374,7 +373,7 @@ def importConfigs(pm, db, all_plugins):
"t_" + generate_random_string(20),
c_d,
"API token",
- '{"dataType": "string","elements": [{"elementType": "input","elementHasInputValue": 1,"elementOptions": [{ "cssClasses": "col-xs-12" }],"transformers": []},{"elementType": "button","elementOptions": [{ "getStringKey": "Gen_Generate" },{ "customParams": "API_TOKEN" },{ "onClick": "generateApiToken(this, 20)" },{ "cssClasses": "col-xs-12" }],"transformers": []}]}',
+ '{"dataType": "string","elements": [{"elementType": "input","elementHasInputValue": 1,"elementOptions": [{ "cssClasses": "col-xs-12" }],"transformers": []},{"elementType": "button","elementOptions": [{ "getStringKey": "Gen_Generate" },{ "customParams": "API_TOKEN" },{ "onClick": "generateApiToken(this, 20)" },{ "cssClasses": "col-xs-12" }],"transformers": []}]}', # noqa: E501 - inline JSON
"[]",
"General",
)
@@ -386,7 +385,7 @@ def importConfigs(pm, db, all_plugins):
c_d,
"Language Interface",
'{"dataType":"string", "elements": [{"elementType" : "select", "elementOptions" : [] ,"transformers": []}]}',
- "['English (en_us)', 'Arabic (ar_ar)', 'Catalan (ca_ca)', 'Czech (cs_cz)', 'German (de_de)', 'Spanish (es_es)', 'Farsi (fa_fa)', 'French (fr_fr)', 'Italian (it_it)', 'Japanese (ja_jp)', 'Norwegian (nb_no)', 'Polish (pl_pl)', 'Portuguese (pt_br)', 'Portuguese (pt_pt)', 'Russian (ru_ru)', 'Swedish (sv_sv)', 'Turkish (tr_tr)', 'Ukrainian (uk_ua)', 'Chinese (zh_cn)']",
+ "['English (en_us)', 'Arabic (ar_ar)', 'Catalan (ca_ca)', 'Czech (cs_cz)', 'German (de_de)', 'Spanish (es_es)', 'Farsi (fa_fa)', 'French (fr_fr)', 'Italian (it_it)', 'Japanese (ja_jp)', 'Norwegian (nb_no)', 'Polish (pl_pl)', 'Portuguese (pt_br)', 'Portuguese (pt_pt)', 'Russian (ru_ru)', 'Swedish (sv_sv)', 'Turkish (tr_tr)', 'Ukrainian (uk_ua)', 'Chinese (zh_cn)']", # noqa: E501 - inline JSON
"UI",
)
@@ -483,9 +482,7 @@ def importConfigs(pm, db, all_plugins):
# only include loaded plugins, and the ones that are enabled
if (
- pref in conf.LOADED_PLUGINS
- or plugin_run != "disabled"
- or plugin_run is None
+ pref in conf.LOADED_PLUGINS or plugin_run != "disabled" or plugin_run is None
):
print_plugin_info(plugin, ["display_name", "description"])
@@ -524,9 +521,7 @@ def importConfigs(pm, db, all_plugins):
if "popupForm" in option:
for popup_entry in option["popupForm"]:
popup_pref = (
- key
- + "_popupform_"
- + popup_entry.get("function", "")
+ key + "_popupform_" + popup_entry.get("function", "")
)
stringSqlParams = collect_lang_strings(
popup_entry, popup_pref, stringSqlParams
@@ -606,7 +601,7 @@ def importConfigs(pm, db, all_plugins):
# Loop through settings_override dictionary
for setting_name, value in settings_override.items():
# Ensure the value is treated as a string and passed directly
- if isinstance(value, str) == False:
+ if isinstance(value, str) is False:
value = str(value)
# Log the value being passed
@@ -669,23 +664,31 @@ def importConfigs(pm, db, all_plugins):
# -----------------
# HANDLE APP was upgraded message - clear cache
-
+
# Check if app was upgraded
-
+
buildTimestamp, new_version = getBuildTimeStampAndVersion()
prev_version = conf.VERSION if conf.VERSION != '' else "unknown"
-
+
mylog('debug', [f"[Config] buildTimestamp | prev_version | .VERSION file: '{buildTimestamp}|{prev_version}|{new_version}'"])
-
+
if str(prev_version) != str(new_version):
-
- mylog('none', ['[Config] App upgraded 🚀'])
-
+
+ mylog('none', ['[Config] App upgraded 🚀'])
+
# ccd(key, default, config_dir, name, inputtype, options, group, events=None, desc="", setJsonMetadata=None, overrideTemplate=None, forceDefault=False)
ccd('VERSION', new_version , c_d, '_KEEP_', '_KEEP_', '_KEEP_', '_KEEP_', None, "_KEEP_", None, None, True)
-
- write_notification(f'[Upgrade] : App upgraded from {prev_version} to {new_version} 🚀 Please clear the cache: - Click OK below
- Clear the browser cache (shift + browser refresh button)
- Clear app cache with the (reload) button in the header
- Go to Settings and click Save
Check out new features and what has changed in the 📓 release notes.', 'interrupt', timeNowDB())
-
+
+ write_notification(f'[Upgrade]: App upgraded from {prev_version} to \
+ {new_version} code> 🚀 Please clear the cache: \
+ - Click OK below
- Clear the browser cache (shift + \
+ browser refresh button)
- Clear app cache with the \
+ (reload) button in the header
- Go to Settings and click Save
\
+ Check out new features and what has changed in the \
+ 📓 release notes.',
+ 'interrupt',
+ timeNowDB()
+ )
# -----------------
# Initialization finished, update DB and API endpoints
@@ -717,13 +720,13 @@ def importConfigs(pm, db, all_plugins):
# settingsImported = None (timestamp),
# showSpinner = False (1/0),
# graphQLServerStarted = 1 (1/0))
- updateState("Config imported", conf.lastImportedConfFile, conf.lastImportedConfFile, False, 1, None, None, new_version)
-
+ updateState("Config imported", conf.lastImportedConfFile, conf.lastImportedConfFile, False, 1, None, None, new_version)
+
msg = '[Config] Imported new settings config'
mylog('minimal', msg)
-
+
# front end app log loggging
- write_notification(msg, 'info', timeNowDB())
+ write_notification(msg, 'info', timeNowDB())
return pm, all_plugins, True
diff --git a/server/logger.py b/server/logger.py
index b49b047f..079edda6 100755
--- a/server/logger.py
+++ b/server/logger.py
@@ -1,19 +1,14 @@
import sys
import io
-import datetime
+# import datetime
import threading
import queue
import logging
-from zoneinfo import ZoneInfo
-
-# Register NetAlertX directories
-INSTALL_PATH="/app"
-
-sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
+# from zoneinfo import ZoneInfo
# NetAlertX imports
import conf
-from const import *
+from const import logPath
from utils.datetime_utils import timeNowTZ
diff --git a/server/messaging/in_app.py b/server/messaging/in_app.py
index f988687b..39b9bd17 100755
--- a/server/messaging/in_app.py
+++ b/server/messaging/in_app.py
@@ -11,13 +11,9 @@ from flask import jsonify
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
sys.path.extend([f"{INSTALL_PATH}/server"])
-from const import apiPath
-from logger import mylog
-
-import conf
-from const import applicationPath, logPath, apiPath, confFileName, reportTemplatesPath
-from logger import mylog
-from utils.datetime_utils import timeNowDB
+from const import apiPath # noqa: E402 [flake8 lint suppression]
+from logger import mylog # noqa: E402 [flake8 lint suppression]
+from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
NOTIFICATION_API_FILE = apiPath + 'user_notifications.json'
@@ -38,7 +34,7 @@ def write_notification(content, level="alert", timestamp=None):
None
"""
if timestamp is None:
- timestamp = timeNowDB()
+ timestamp = timeNowDB()
# Generate GUID
guid = str(uuid.uuid4())
diff --git a/server/messaging/reporting.py b/server/messaging/reporting.py
index 6dcfeed0..b102e4f6 100755
--- a/server/messaging/reporting.py
+++ b/server/messaging/reporting.py
@@ -18,12 +18,12 @@ import sys
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
sys.path.extend([f"{INSTALL_PATH}/server"])
-from helper import (
+from helper import ( # noqa: E402 [flake8 lint suppression]
get_setting_value,
)
-from logger import mylog
-from db.sql_safe_builder import create_safe_condition_builder
-from utils.datetime_utils import get_timezone_offset
+from logger import mylog # noqa: E402 [flake8 lint suppression]
+from db.sql_safe_builder import create_safe_condition_builder # noqa: E402 [flake8 lint suppression]
+from utils.datetime_utils import get_timezone_offset # noqa: E402 [flake8 lint suppression]
# ===============================================================================
# REPORTING
@@ -56,14 +56,14 @@ def get_notifications(db):
WHERE eve_PendingAlertEmail = 1 AND eve_EventType not in ('Device Down', 'Down Reconnected', 'New Device' ) AND eve_MAC IN
(
SELECT devMac FROM Devices WHERE devAlertEvents = 0
- )""")
+ )""")
# Disable down/down reconnected notifications on devices where devAlertDown is disabled
sql.execute("""UPDATE Events SET eve_PendingAlertEmail = 0
WHERE eve_PendingAlertEmail = 1 AND eve_EventType in ('Device Down', 'Down Reconnected') AND eve_MAC IN
(
SELECT devMac FROM Devices WHERE devAlertDown = 0
- )""")
+ )""")
sections = get_setting_value("NTFPRCS_INCLUDED_SECTIONS")
@@ -79,20 +79,32 @@ def get_notifications(db):
safe_condition, parameters = condition_builder.get_safe_condition_legacy(
new_dev_condition_setting
)
- sqlQuery = """SELECT eve_MAC as MAC, eve_DateTime as Datetime, devLastIP as IP, eve_EventType as "Event Type", devName as "Device name", devComments as Comments FROM Events_Devices
- WHERE eve_PendingAlertEmail = 1
+ sqlQuery = """SELECT
+ eve_MAC as MAC,
+ eve_DateTime as Datetime,
+ devLastIP as IP,
+ eve_EventType as "Event Type",
+ devName as "Device name",
+ devComments as Comments FROM Events_Devices
+ WHERE eve_PendingAlertEmail = 1
AND eve_EventType = 'New Device' {}
- ORDER BY eve_DateTime""".format(safe_condition)
+ ORDER BY eve_DateTime""".format(safe_condition)
except Exception as e:
mylog(
"verbose",
["[Notification] Error building safe condition for new devices: ", e],
)
# Fall back to safe default (no additional conditions)
- sqlQuery = """SELECT eve_MAC as MAC, eve_DateTime as Datetime, devLastIP as IP, eve_EventType as "Event Type", devName as "Device name", devComments as Comments FROM Events_Devices
- WHERE eve_PendingAlertEmail = 1
+ sqlQuery = """SELECT
+ eve_MAC as MAC,
+ eve_DateTime as Datetime,
+ devLastIP as IP,
+ eve_EventType as "Event Type",
+ devName as "Device name",
+ devComments as Comments FROM Events_Devices
+ WHERE eve_PendingAlertEmail = 1
AND eve_EventType = 'New Device'
- ORDER BY eve_DateTime"""
+ ORDER BY eve_DateTime"""
parameters = {}
mylog("debug", ["[Notification] new_devices SQL query: ", sqlQuery])
@@ -114,17 +126,17 @@ def get_notifications(db):
minutes = int(get_setting_value("NTFPRCS_alert_down_time") or 0)
tz_offset = get_timezone_offset()
sqlQuery = f"""
- SELECT devName, eve_MAC, devVendor, eve_IP, eve_DateTime, eve_EventType
+ SELECT devName, eve_MAC, devVendor, eve_IP, eve_DateTime, eve_EventType
FROM Events_Devices AS down_events
- WHERE eve_PendingAlertEmail = 1
- AND down_events.eve_EventType = 'Device Down'
+ WHERE eve_PendingAlertEmail = 1
+ AND down_events.eve_EventType = 'Device Down'
AND eve_DateTime < datetime('now', '-{minutes} minutes', '{tz_offset}')
AND NOT EXISTS (
SELECT 1
FROM Events AS connected_events
WHERE connected_events.eve_MAC = down_events.eve_MAC
AND connected_events.eve_EventType = 'Connected'
- AND connected_events.eve_DateTime > down_events.eve_DateTime
+ AND connected_events.eve_DateTime > down_events.eve_DateTime
)
ORDER BY down_events.eve_DateTime;
"""
@@ -181,20 +193,32 @@ def get_notifications(db):
safe_condition, parameters = condition_builder.get_safe_condition_legacy(
event_condition_setting
)
- sqlQuery = """SELECT eve_MAC as MAC, eve_DateTime as Datetime, devLastIP as IP, eve_EventType as "Event Type", devName as "Device name", devComments as Comments FROM Events_Devices
- WHERE eve_PendingAlertEmail = 1
+ sqlQuery = """SELECT
+ eve_MAC as MAC,
+ eve_DateTime as Datetime,
+ devLastIP as IP,
+ eve_EventType as "Event Type",
+ devName as "Device name",
+ devComments as Comments FROM Events_Devices
+ WHERE eve_PendingAlertEmail = 1
AND eve_EventType IN ('Connected', 'Down Reconnected', 'Disconnected','IP Changed') {}
- ORDER BY eve_DateTime""".format(safe_condition)
+ ORDER BY eve_DateTime""".format(safe_condition)
except Exception as e:
mylog(
"verbose",
["[Notification] Error building safe condition for events: ", e],
)
# Fall back to safe default (no additional conditions)
- sqlQuery = """SELECT eve_MAC as MAC, eve_DateTime as Datetime, devLastIP as IP, eve_EventType as "Event Type", devName as "Device name", devComments as Comments FROM Events_Devices
- WHERE eve_PendingAlertEmail = 1
+ sqlQuery = """SELECT
+ eve_MAC as MAC,
+ eve_DateTime as Datetime,
+ devLastIP as IP,
+ eve_EventType as "Event Type",
+ devName as "Device name",
+ devComments as Comments FROM Events_Devices
+ WHERE eve_PendingAlertEmail = 1
AND eve_EventType IN ('Connected', 'Down Reconnected', 'Disconnected','IP Changed')
- ORDER BY eve_DateTime"""
+ ORDER BY eve_DateTime"""
parameters = {}
mylog("debug", ["[Notification] events SQL query: ", sqlQuery])
@@ -208,7 +232,17 @@ def get_notifications(db):
if "plugins" in sections:
# Compose Plugins Section
- sqlQuery = """SELECT Plugin, Object_PrimaryId, Object_SecondaryId, DateTimeChanged, Watched_Value1, Watched_Value2, Watched_Value3, Watched_Value4, Status from Plugins_Events"""
+ sqlQuery = """SELECT
+ Plugin,
+ Object_PrimaryId,
+ Object_SecondaryId,
+ DateTimeChanged,
+ Watched_Value1,
+ Watched_Value2,
+ Watched_Value3,
+ Watched_Value4,
+ Status
+ from Plugins_Events"""
# Get the events as JSON
json_obj = db.get_table_as_json(sqlQuery)
diff --git a/server/models/notification_instance.py b/server/models/notification_instance.py
index c4367c67..b2f5526f 100755
--- a/server/models/notification_instance.py
+++ b/server/models/notification_instance.py
@@ -1,13 +1,12 @@
import json
import uuid
import socket
-import subprocess
from yattag import indent
from json2table import convert
# Register NetAlertX modules
import conf
-from const import applicationPath, logPath, apiPath, reportTemplatesPath
+from const import logPath, apiPath, reportTemplatesPath
from logger import mylog, Logger
from helper import (
generate_mac_links,
@@ -62,11 +61,7 @@ class NotificationInstance:
# Check if nothing to report, end
if (
- JSON["new_devices"] == []
- and JSON["down_devices"] == []
- and JSON["events"] == []
- and JSON["plugins"] == []
- and JSON["down_reconnected"] == []
+ JSON["new_devices"] == [] and JSON["down_devices"] == [] and JSON["events"] == [] and JSON["plugins"] == [] and JSON["down_reconnected"] == []
):
self.HasNotifications = False
else:
@@ -88,8 +83,6 @@ class NotificationInstance:
# else:
# mylog('debug', ['[Notification] notiStruc:', json.dumps(notiStruc.__dict__, indent=4)])
- Text = ""
- HTML = ""
template_file_path = reportTemplatesPath + "report_template.html"
# Open text Template
@@ -274,7 +267,7 @@ class NotificationInstance:
# Clear the Pending Email flag from all events and devices
def clearPendingEmailFlag(self):
- # Clean Pending Alert Events
+ # Clean Pending Alert Events
self.db.sql.execute("""
UPDATE Devices SET devLastNotification = ?
WHERE devMac IN (
diff --git a/server/models/user_events_queue_instance.py b/server/models/user_events_queue_instance.py
index 204c5b6b..94fb0e31 100755
--- a/server/models/user_events_queue_instance.py
+++ b/server/models/user_events_queue_instance.py
@@ -100,7 +100,7 @@ class UserEventsQueueInstance:
if not action or not isinstance(action, str):
msg = "[UserEventsQueueInstance] Invalid or missing action"
mylog('none', [msg])
-
+
return False, msg
try:
@@ -109,15 +109,11 @@ class UserEventsQueueInstance:
msg = f'[UserEventsQueueInstance] Action "{action}" added to the execution queue.'
mylog('minimal', [msg])
-
+
return True, msg
except Exception as e:
msg = f"[UserEventsQueueInstance] ERROR Failed to write to {self.log_file}: {e}"
mylog('none', [msg])
-
+
return False, msg
-
-
-
-
diff --git a/server/plugin.py b/server/plugin.py
index 45b9b9a4..ee64290b 100755
--- a/server/plugin.py
+++ b/server/plugin.py
@@ -9,12 +9,21 @@ from concurrent.futures import ThreadPoolExecutor, as_completed
# Register NetAlertX modules
import conf
from const import pluginsPath, logPath, applicationPath, reportTemplatesPath
-from logger import mylog, Logger
-from helper import get_file_content, write_file, get_setting, get_setting_value
+from logger import mylog, Logger
+from helper import get_file_content, get_setting, get_setting_value
from utils.datetime_utils import timeNowTZ, timeNowDB
from app_state import updateState
from api import update_api
-from utils.plugin_utils import logEventStatusCounts, get_plugin_string, get_plugin_setting_obj, print_plugin_info, list_to_csv, combine_plugin_objects, resolve_wildcards_arr, handle_empty, custom_plugin_decoder, decode_and_rename_files
+from utils.plugin_utils import (
+ logEventStatusCounts,
+ get_plugin_setting_obj,
+ print_plugin_info,
+ list_to_csv,
+ combine_plugin_objects,
+ resolve_wildcards_arr,
+ handle_empty,
+ decode_and_rename_files
+)
from models.notification_instance import NotificationInstance
from messaging.in_app import write_notification
from models.user_events_queue_instance import UserEventsQueueInstance
@@ -57,13 +66,7 @@ class plugin_manager:
# Header
updateState("Run: Plugins")
- mylog(
- "debug",
- [
- "[Plugins] Check if any plugins need to be executed on run type: ",
- runType,
- ],
- )
+ mylog("debug", f"[Plugins] Check if any plugins need to be executed on run type: {runType}")
for plugin in self.all_plugins:
shouldRun = False
@@ -72,7 +75,7 @@ class plugin_manager:
# 🔹 Lookup RUN setting from cache instead of calling get_plugin_setting_obj each time
run_setting = self._cache["settings"].get(prefix, {}).get("RUN")
- if run_setting != None and run_setting["value"] == runType:
+ if run_setting is not None and run_setting["value"] == runType:
if runType != "schedule":
shouldRun = True
elif runType == "schedule":
@@ -91,10 +94,7 @@ class plugin_manager:
# 🔹 CMD also retrieved from cache
cmd_setting = self._cache["settings"].get(prefix, {}).get("CMD")
- mylog(
- "debug",
- ["[Plugins] CMD: ", cmd_setting["value"] if cmd_setting else None],
- )
+ mylog("debug", f"[Plugins] CMD: {cmd_setting["value"] if cmd_setting else None}")
execute_plugin(self.db, self.all_plugins, plugin)
@@ -130,13 +130,7 @@ class plugin_manager:
mylog("debug", ["[check_and_run_user_event] User Execution Queue is empty"])
return # Exit early if the log file is empty
else:
- mylog(
- "debug",
- [
- "[check_and_run_user_event] Process User Execution Queue:"
- + ", ".join(map(str, lines))
- ],
- )
+ mylog("debug", "[check_and_run_user_event] Process User Execution Queue:" + ", ".join(map(str, lines)))
for line in lines:
# Extract event name and parameters from the log line
@@ -160,15 +154,7 @@ class plugin_manager:
update_api(self.db, self.all_plugins, False, param.split(","), True)
else:
- mylog(
- "minimal",
- [
- "[check_and_run_user_event] WARNING: Unhandled event in execution queue: ",
- event,
- " | ",
- param,
- ],
- )
+ mylog("minimal", f"[check_and_run_user_event] WARNING: Unhandled event in execution queue: {event} | {param}")
execution_log.finalize_event(
event
) # Finalize unknown events to remove them
@@ -183,9 +169,9 @@ class plugin_manager:
# -------------------------------------------------------------------------------
def handle_run(self, runType):
-
+
mylog('minimal', ['[', timeNowDB(), '] START Run: ', runType])
-
+
# run the plugin
for plugin in self.all_plugins:
if plugin["unique_prefix"] == runType:
@@ -201,7 +187,7 @@ class plugin_manager:
pluginsStates={pluginName: current_plugin_state.get(pluginName, {})}
)
- mylog('minimal', ['[', timeNowDB(), '] END Run: ', runType])
+ mylog('minimal', ['[', timeNowDB(), '] END Run: ', runType])
return
@@ -210,7 +196,7 @@ class plugin_manager:
mylog("minimal", ["[", timeNowTZ(), "] [Test] START Test: ", runType])
mylog('minimal', ['[', timeNowDB(), '] [Test] START Test: ', runType])
-
+
# Prepare test samples
sample_json = json.loads(
get_file_content(reportTemplatesPath + "webhook_json_sample.json")
@@ -312,7 +298,7 @@ class plugin_param:
if param["type"] == "setting":
inputValue = get_setting(param["value"])
- if inputValue != None:
+ if inputValue is not None:
setVal = inputValue["setValue"] # setting value
setTyp = inputValue["setType"] # setting type
@@ -337,9 +323,7 @@ class plugin_param:
resolved = list_to_csv(setVal)
else:
- mylog(
- "none", ["[Plugins] ⚠ ERROR: Parameter probably not converted."]
- )
+ mylog("none", "[Plugins] ⚠ ERROR: Parameter probably not converted.")
return json.dumps(setVal)
# Get SQL result
@@ -390,15 +374,10 @@ def run_plugin(command, set_RUN_TIMEOUT, plugin):
)
except subprocess.CalledProcessError as e:
mylog("none", [e.output])
- mylog("none", ["[Plugins] ⚠ ERROR - enable LOG_LEVEL=debug and check logs"])
+ mylog("none", "[Plugins] ⚠ ERROR - enable LOG_LEVEL=debug and check logs")
return None
except subprocess.TimeoutExpired:
- mylog(
- "none",
- [
- f"[Plugins] ⚠ ERROR - TIMEOUT - the plugin {plugin['unique_prefix']} forcefully terminated as timeout reached. Increase TIMEOUT setting and scan interval."
- ],
- )
+ mylog("none", f"[Plugins] ⚠ ERROR - TIMEOUT - the plugin {plugin['unique_prefix']} forcefully terminated as timeout reached. Increase TIMEOUT setting and scan interval.")
return None
@@ -411,11 +390,11 @@ def execute_plugin(db, all_plugins, plugin):
set = get_plugin_setting_obj(plugin, "CMD")
# handle missing "function":"CMD" setting
- if set == None:
+ if set is None:
return
set_CMD = set["value"]
-
+
# Replace hardcoded /app paths with environment-aware path
if "/app/front/plugins" in set_CMD:
set_CMD = set_CMD.replace("/app/front/plugins", str(pluginsPath))
@@ -441,13 +420,8 @@ def execute_plugin(db, all_plugins, plugin):
for param in plugin["params"]:
tempParam = plugin_param(param, plugin, db)
- if tempParam.resolved == None:
- mylog(
- "none",
- [
- f'[Plugins] The parameter "name":"{tempParam.name}" for "value": {tempParam.value} was resolved as None'
- ],
- )
+ if tempParam.resolved is None:
+ mylog("none", f'[Plugins] The parameter "name":"{tempParam.name}" for "value": {tempParam.value} was resolved as None')
else:
# params.append( [param["name"], resolved] )
@@ -456,14 +430,9 @@ def execute_plugin(db, all_plugins, plugin):
if tempParam.multiplyTimeout:
set_RUN_TIMEOUT = set_RUN_TIMEOUT * tempParam.paramValuesCount
- mylog(
- "debug",
- [
- f'[Plugins] The parameter "name":"{param["name"]}" will multiply the timeout {tempParam.paramValuesCount} times. Total timeout: {set_RUN_TIMEOUT}s'
- ],
- )
+ mylog("debug", f'[Plugins] The parameter "name":"{param["name"]}" will multiply timeout {tempParam.paramValuesCount}x. Total timeout: {set_RUN_TIMEOUT}s')
- mylog("debug", ["[Plugins] Timeout: ", set_RUN_TIMEOUT])
+ mylog("debug", f"[Plugins] Timeout: {set_RUN_TIMEOUT}")
# build SQL query parameters to insert into the DB
sqlParams = []
@@ -475,8 +444,8 @@ def execute_plugin(db, all_plugins, plugin):
command = resolve_wildcards_arr(set_CMD.split(), params)
# Execute command
- mylog("verbose", ["[Plugins] Executing: ", set_CMD])
- mylog("debug", ["[Plugins] Resolved : ", command])
+ mylog("verbose", f"[Plugins] Executing: {set_CMD}")
+ mylog("debug", f"[Plugins] Resolved : {command}")
# Using ThreadPoolExecutor to handle concurrent subprocesses
with ThreadPoolExecutor(max_workers=5) as executor:
@@ -521,12 +490,7 @@ def execute_plugin(db, all_plugins, plugin):
columns = line.split("|")
# There have to be 9 or 13 columns
if len(columns) not in [9, 13]:
- mylog(
- "none",
- [
- f"[Plugins] Wrong number of input values, must be 9 or 13, got {len(columns)} from: {line}"
- ],
- )
+ mylog("none", f"[Plugins] Wrong number of input values, must be 9 or 13, got {len(columns)} from: {line}")
continue # Skip lines with incorrect number of columns
# Common part of the SQL parameters
@@ -581,9 +545,7 @@ def execute_plugin(db, all_plugins, plugin):
# keep current instance log file, delete all from other nodes
if filename != "last_result.log" and os.path.exists(full_path):
os.remove(full_path) # DEBUG:TODO uncomment 🐛
- mylog(
- "verbose", [f"[Plugins] Processed and deleted file: {full_path} "]
- )
+ mylog("verbose", f"[Plugins] Processed and deleted file: {full_path} ")
# app-db-query
if plugin["data_source"] == "app-db-query":
@@ -591,7 +553,7 @@ def execute_plugin(db, all_plugins, plugin):
q = set_CMD.replace("{s-quote}", "'")
# Execute command
- mylog("verbose", ["[Plugins] Executing: ", q])
+ mylog("verbose", f"[Plugins] Executing: {q}")
# set_CMD should contain a SQL query
arr = db.get_sql_array(q)
@@ -650,7 +612,7 @@ def execute_plugin(db, all_plugins, plugin):
# Append the final parameters to sqlParams
sqlParams.append(tuple(base_params))
else:
- mylog("none", ["[Plugins] Skipped invalid sql result"])
+ mylog("none", "[Plugins] Skipped invalid sql result")
# app-db-query
if plugin["data_source"] == "sqlite-db-query":
@@ -659,19 +621,14 @@ def execute_plugin(db, all_plugins, plugin):
q = set_CMD.replace("{s-quote}", "'")
# Execute command
- mylog("verbose", ["[Plugins] Executing: ", q])
+ mylog("verbose", f"[Plugins] Executing: {q}")
# ------- necessary settings check --------
set = get_plugin_setting_obj(plugin, "DB_PATH")
# handle missing "function":"DB_PATH" setting
- if set == None:
- mylog(
- "none",
- [
- "[Plugins] ⚠ ERROR: DB_PATH setting for plugin type sqlite-db-query missing."
- ],
- )
+ if set is None:
+ mylog("none", "[Plugins] ⚠ ERROR: DB_PATH setting for plugin type sqlite-db-query missing.")
return
fullSqlitePath = set["value"]
@@ -679,25 +636,14 @@ def execute_plugin(db, all_plugins, plugin):
# try attaching the sqlite DB
try:
sql.execute(
- "ATTACH DATABASE '"
- + fullSqlitePath
- + "' AS EXTERNAL_"
- + plugin["unique_prefix"]
+ "ATTACH DATABASE '" + fullSqlitePath + "' AS EXTERNAL_" + plugin["unique_prefix"]
)
arr = db.get_sql_array(q)
sql.execute("DETACH DATABASE EXTERNAL_" + plugin["unique_prefix"])
except sqlite3.Error as e:
- mylog(
- "none",
- [
- f"[Plugins] ⚠ ERROR: DB_PATH setting ({fullSqlitePath}) for plugin {plugin['unique_prefix']}. Did you mount it correctly?"
- ],
- )
- mylog(
- "none",
- ["[Plugins] ⚠ ERROR: ATTACH DATABASE failed with SQL ERROR: ", e],
- )
+ mylog("none", f"[Plugins] ⚠ ERROR: DB_PATH setting ({fullSqlitePath}) for plugin {plugin['unique_prefix']}. Did you mount it correctly?")
+ mylog("none", f"[Plugins] ⚠ ERROR: ATTACH DATABASE failed with SQL ERROR: {e}")
return
for row in arr:
@@ -748,24 +694,14 @@ def execute_plugin(db, all_plugins, plugin):
# Append the final parameters to sqlParams
sqlParams.append(tuple(base_params))
else:
- mylog("none", ["[Plugins] Skipped invalid sql result"])
+ mylog("none", "[Plugins] Skipped invalid sql result")
# check if the subprocess / SQL query failed / there was no valid output
if len(sqlParams) == 0:
- mylog(
- "none",
- [
- f'[Plugins] No output received from the plugin "{plugin["unique_prefix"]}"'
- ],
- )
+ mylog("none", f'[Plugins] No output received from the plugin "{plugin["unique_prefix"]}"')
else:
- mylog(
- "verbose",
- [
- f"[Plugins] SUCCESS for {plugin['unique_prefix']} received {len(sqlParams)} entries"
- ],
- )
+ mylog("verbose", f"[Plugins] SUCCESS for {plugin['unique_prefix']} received {len(sqlParams)} entries")
# mylog('debug', ['[Plugins] sqlParam entries: ', sqlParams])
# create objects
@@ -782,12 +718,7 @@ def execute_plugin(db, all_plugins, plugin):
# check if we need to update devices api endpoint as well to prevent long user waits on Loading...
userUpdatedDevices = UserEventsQueueInstance().has_update_devices()
- mylog(
- "verbose",
- [
- f"[Plugins] Should I update API (userUpdatedDevices): {userUpdatedDevices}"
- ],
- )
+ mylog("verbose", f"[Plugins] Should I update API (userUpdatedDevices): {userUpdatedDevices}")
if userUpdatedDevices:
endpoints += ["devices"]
@@ -807,7 +738,7 @@ def process_plugin_events(db, plugin, plugEventsArr):
pluginPref = plugin["unique_prefix"]
- mylog("verbose", ["[Plugins] Processing : ", pluginPref])
+ mylog("verbose", f"[Plugins] Processing : {pluginPref}")
try:
# Begin a transaction
@@ -827,20 +758,8 @@ def process_plugin_events(db, plugin, plugEventsArr):
for eve in plugEventsArr:
pluginEvents.append(plugin_object_class(plugin, eve))
- mylog(
- "debug",
- [
- "[Plugins] Existing objects from Plugins_Objects: ",
- len(pluginObjects),
- ],
- )
- mylog(
- "debug",
- [
- "[Plugins] Logged events from the plugin run : ",
- len(pluginEvents),
- ],
- )
+ mylog("debug", f"[Plugins] Existing objects from Plugins_Objects: {len(pluginObjects)}")
+ mylog("debug", f"[Plugins] Logged events from the plugin run : {len(pluginEvents)}")
# Loop thru all current events and update the status to "exists" if the event matches an existing object
index = 0
@@ -857,8 +776,7 @@ def process_plugin_events(db, plugin, plugEventsArr):
if tmpObjFromEvent.status == "exists":
# compare hash of the changed watched columns for uniqueness - make sure you compare the values with the same idsHash before checking watchedHash
if any(
- x.idsHash == tmpObjFromEvent.idsHash
- and x.watchedHash != tmpObjFromEvent.watchedHash
+ x.idsHash == tmpObjFromEvent.idsHash and x.watchedHash != tmpObjFromEvent.watchedHash
for x in pluginObjects
):
pluginEvents[index].status = "watched-changed"
@@ -879,7 +797,7 @@ def process_plugin_events(db, plugin, plugEventsArr):
# if wasn't missing before, mark as changed
if tmpObj.status != "missing-in-last-scan":
tmpObj.changed = timeNowDB()
- tmpObj.status = "missing-in-last-scan"
+ tmpObj.status = "missing-in-last-scan"
# mylog('debug', [f'[Plugins] Missing from last scan (PrimaryID | SecondaryID): {tmpObj.primaryId} | {tmpObj.secondaryId}'])
# Merge existing plugin objects with newly discovered ones and update existing ones with new values
@@ -955,25 +873,17 @@ def process_plugin_events(db, plugin, plugEventsArr):
# combine all DB insert and update events into one for history
history_to_insert.append(values)
- mylog("debug", ["[Plugins] pluginEvents count: ", len(pluginEvents)])
- mylog("debug", ["[Plugins] pluginObjects count: ", len(pluginObjects)])
+ mylog("debug", f"[Plugins] pluginEvents count: {len(pluginEvents)}")
+ mylog("debug", f"[Plugins] pluginObjects count: {len(pluginObjects)}")
- mylog(
- "debug", ["[Plugins] events_to_insert count: ", len(events_to_insert)]
- )
- mylog(
- "debug", ["[Plugins] history_to_insert count: ", len(history_to_insert)]
- )
- mylog(
- "debug", ["[Plugins] objects_to_insert count: ", len(objects_to_insert)]
- )
- mylog(
- "debug", ["[Plugins] objects_to_update count: ", len(objects_to_update)]
- )
+ mylog("debug", f"[Plugins] events_to_insert count: {len(events_to_insert)}")
+ mylog("debug", f"[Plugins] history_to_insert count: {len(history_to_insert)}")
+ mylog("debug", f"[Plugins] objects_to_insert count: {len(objects_to_insert)}")
+ mylog("debug", f"[Plugins] objects_to_update count: {len(objects_to_update)}")
- mylog("trace", ["[Plugins] objects_to_update: ", objects_to_update])
- mylog("trace", ["[Plugins] events_to_insert: ", events_to_insert])
- mylog("trace", ["[Plugins] history_to_insert: ", history_to_insert])
+ mylog("trace", f"[Plugins] objects_to_update: {objects_to_update}")
+ mylog("trace", f"[Plugins] events_to_insert: {events_to_insert}")
+ mylog("trace", f"[Plugins] history_to_insert: {history_to_insert}")
logEventStatusCounts("pluginEvents", pluginEvents)
logEventStatusCounts("pluginObjects", pluginObjects)
@@ -982,12 +892,12 @@ def process_plugin_events(db, plugin, plugEventsArr):
if objects_to_insert:
sql.executemany(
"""
- INSERT INTO Plugins_Objects
- ("Plugin", "Object_PrimaryID", "Object_SecondaryID", "DateTimeCreated",
- "DateTimeChanged", "Watched_Value1", "Watched_Value2", "Watched_Value3",
+ INSERT INTO Plugins_Objects
+ ("Plugin", "Object_PrimaryID", "Object_SecondaryID", "DateTimeCreated",
+ "DateTimeChanged", "Watched_Value1", "Watched_Value2", "Watched_Value3",
"Watched_Value4", "Status", "Extra", "UserData", "ForeignKey", "SyncHubNodeName",
- "HelpVal1", "HelpVal2", "HelpVal3", "HelpVal4",
- "ObjectGUID")
+ "HelpVal1", "HelpVal2", "HelpVal3", "HelpVal4",
+ "ObjectGUID")
VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
""",
objects_to_insert,
@@ -998,10 +908,10 @@ def process_plugin_events(db, plugin, plugEventsArr):
sql.executemany(
"""
UPDATE Plugins_Objects
- SET "Plugin" = ?, "Object_PrimaryID" = ?, "Object_SecondaryID" = ?, "DateTimeCreated" = ?,
- "DateTimeChanged" = ?, "Watched_Value1" = ?, "Watched_Value2" = ?, "Watched_Value3" = ?,
- "Watched_Value4" = ?, "Status" = ?, "Extra" = ?, "UserData" = ?, "ForeignKey" = ?, "SyncHubNodeName" = ?,
- "HelpVal1" = ?, "HelpVal2" = ?, "HelpVal3" = ?, "HelpVal4" = ?,
+ SET "Plugin" = ?, "Object_PrimaryID" = ?, "Object_SecondaryID" = ?, "DateTimeCreated" = ?,
+ "DateTimeChanged" = ?, "Watched_Value1" = ?, "Watched_Value2" = ?, "Watched_Value3" = ?,
+ "Watched_Value4" = ?, "Status" = ?, "Extra" = ?, "UserData" = ?, "ForeignKey" = ?, "SyncHubNodeName" = ?,
+ "HelpVal1" = ?, "HelpVal2" = ?, "HelpVal3" = ?, "HelpVal4" = ?,
"ObjectGUID" = ?
WHERE "Index" = ?
""",
@@ -1012,12 +922,12 @@ def process_plugin_events(db, plugin, plugEventsArr):
if events_to_insert:
sql.executemany(
"""
- INSERT INTO Plugins_Events
- ("Plugin", "Object_PrimaryID", "Object_SecondaryID", "DateTimeCreated",
- "DateTimeChanged", "Watched_Value1", "Watched_Value2", "Watched_Value3",
+ INSERT INTO Plugins_Events
+ ("Plugin", "Object_PrimaryID", "Object_SecondaryID", "DateTimeCreated",
+ "DateTimeChanged", "Watched_Value1", "Watched_Value2", "Watched_Value3",
"Watched_Value4", "Status", "Extra", "UserData", "ForeignKey", "SyncHubNodeName",
"HelpVal1", "HelpVal2", "HelpVal3", "HelpVal4",
- "ObjectGUID")
+ "ObjectGUID")
VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
""",
events_to_insert,
@@ -1027,12 +937,12 @@ def process_plugin_events(db, plugin, plugEventsArr):
if history_to_insert:
sql.executemany(
"""
- INSERT INTO Plugins_History
- ("Plugin", "Object_PrimaryID", "Object_SecondaryID", "DateTimeCreated",
- "DateTimeChanged", "Watched_Value1", "Watched_Value2", "Watched_Value3",
+ INSERT INTO Plugins_History
+ ("Plugin", "Object_PrimaryID", "Object_SecondaryID", "DateTimeCreated",
+ "DateTimeChanged", "Watched_Value1", "Watched_Value2", "Watched_Value3",
"Watched_Value4", "Status", "Extra", "UserData", "ForeignKey", "SyncHubNodeName",
"HelpVal1", "HelpVal2", "HelpVal3", "HelpVal4",
- "ObjectGUID")
+ "ObjectGUID")
VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)
""",
history_to_insert,
@@ -1044,7 +954,7 @@ def process_plugin_events(db, plugin, plugEventsArr):
except Exception as e:
# Rollback the transaction in case of an error
conn.rollback()
- mylog("none", ["[Plugins] ⚠ ERROR: ", e])
+ mylog("none", f"[Plugins] ⚠ ERROR: {e}")
raise e
# Perform database table mapping if enabled for the plugin
@@ -1056,7 +966,7 @@ def process_plugin_events(db, plugin, plugEventsArr):
dbTable = plugin["mapped_to_table"]
# Log a debug message indicating the mapping of objects to the database table.
- mylog("debug", ["[Plugins] Mapping objects to database table: ", dbTable])
+ mylog("debug", f"[Plugins] Mapping objects to database table: {dbTable}")
# Initialize lists to hold mapped column names, columnsStr, and valuesStr for SQL query.
mappedCols = []
@@ -1121,8 +1031,7 @@ def process_plugin_events(db, plugin, plugEventsArr):
# Check if there's a default value specified for this column in the JSON.
if (
- "mapped_to_column_data" in col
- and "value" in col["mapped_to_column_data"]
+ "mapped_to_column_data" in col and "value" in col["mapped_to_column_data"]
):
tmpList.append(col["mapped_to_column_data"]["value"])
@@ -1133,8 +1042,8 @@ def process_plugin_events(db, plugin, plugEventsArr):
q = f"INSERT OR IGNORE INTO {dbTable} ({columnsStr}) VALUES ({valuesStr})"
# Log a debug message showing the generated SQL query for mapping.
- mylog("debug", ["[Plugins] SQL query for mapping: ", q])
- mylog("debug", ["[Plugins] SQL sqlParams for mapping: ", sqlParams])
+ mylog("debug", f"[Plugins] SQL query for mapping: {q}")
+ mylog("debug", f"[Plugins] SQL sqlParams for mapping: {sqlParams}")
# Execute the SQL query using 'sql.executemany()' and the 'sqlParams' list of tuples.
# This will insert multiple rows into the database in one go.
diff --git a/server/scan/device_handling.py b/server/scan/device_handling.py
index 41ad707d..cefe4ebb 100755
--- a/server/scan/device_handling.py
+++ b/server/scan/device_handling.py
@@ -1,14 +1,6 @@
-import sys
import subprocess
import os
import re
-import datetime
-from dateutil import parser
-
-# Register NetAlertX directories
-INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
-sys.path.extend([f"{INSTALL_PATH}/server"])
-
from helper import get_setting_value, check_IP_format
from utils.datetime_utils import timeNowDB, normalizeTimeStamp
from logger import mylog, Logger
@@ -44,7 +36,7 @@ def exclude_ignored_devices(db):
# Join conditions and prepare the query
conditions_str = " OR ".join(conditions)
if conditions_str:
- query = f"""DELETE FROM CurrentScan WHERE
+ query = f"""DELETE FROM CurrentScan WHERE
1=1
AND (
{conditions_str}
@@ -57,22 +49,23 @@ def exclude_ignored_devices(db):
sql.execute(query)
-#-------------------------------------------------------------------------------
-def update_devices_data_from_scan (db):
- sql = db.sql #TO-DO
+
+# -------------------------------------------------------------------------------
+def update_devices_data_from_scan(db):
+ sql = db.sql # TO-DO
startTime = timeNowDB()
# Update Last Connection
mylog("debug", "[Update Devices] 1 Last Connection")
sql.execute(f"""UPDATE Devices SET devLastConnection = '{startTime}',
devPresentLastScan = 1
- WHERE EXISTS (SELECT 1 FROM CurrentScan
+ WHERE EXISTS (SELECT 1 FROM CurrentScan
WHERE devMac = cur_MAC) """)
# Clean no active devices
mylog("debug", "[Update Devices] 2 Clean no active devices")
sql.execute("""UPDATE Devices SET devPresentLastScan = 0
- WHERE NOT EXISTS (SELECT 1 FROM CurrentScan
+ WHERE NOT EXISTS (SELECT 1 FROM CurrentScan
WHERE devMac = cur_MAC) """)
# Update IP
@@ -103,7 +96,7 @@ def update_devices_data_from_scan (db):
FROM CurrentScan
WHERE Devices.devMac = CurrentScan.cur_MAC
)
- WHERE
+ WHERE
(devVendor IS NULL OR devVendor IN ("", "null", "(unknown)", "(Unknown)"))
AND EXISTS (
SELECT 1
@@ -116,12 +109,12 @@ def update_devices_data_from_scan (db):
sql.execute("""UPDATE Devices
SET devParentPort = (
SELECT cur_Port
- FROM CurrentScan
- WHERE Devices.devMac = CurrentScan.cur_MAC
+ FROM CurrentScan
+ WHERE Devices.devMac = CurrentScan.cur_MAC
)
- WHERE
+ WHERE
(devParentPort IS NULL OR devParentPort IN ("", "null", "(unknown)", "(Unknown)"))
- AND
+ AND
EXISTS (
SELECT 1
FROM CurrentScan
@@ -139,9 +132,9 @@ def update_devices_data_from_scan (db):
FROM CurrentScan
WHERE Devices.devMac = CurrentScan.cur_MAC
)
- WHERE
+ WHERE
(devParentMAC IS NULL OR devParentMAC IN ("", "null", "(unknown)", "(Unknown)"))
- AND
+ AND
EXISTS (
SELECT 1
FROM CurrentScan
@@ -161,7 +154,7 @@ def update_devices_data_from_scan (db):
FROM CurrentScan
WHERE Devices.devMac = CurrentScan.cur_MAC
)
- WHERE
+ WHERE
(devSite IS NULL OR devSite IN ("", "null"))
AND EXISTS (
SELECT 1
@@ -178,7 +171,7 @@ def update_devices_data_from_scan (db):
FROM CurrentScan
WHERE Devices.devMac = CurrentScan.cur_MAC
)
- WHERE
+ WHERE
(devSSID IS NULL OR devSSID IN ("", "null"))
AND EXISTS (
SELECT 1
@@ -195,7 +188,7 @@ def update_devices_data_from_scan (db):
FROM CurrentScan
WHERE Devices.devMac = CurrentScan.cur_MAC
)
- WHERE
+ WHERE
(devType IS NULL OR devType IN ("", "null"))
AND EXISTS (
SELECT 1
@@ -208,17 +201,17 @@ def update_devices_data_from_scan (db):
mylog("debug", "[Update Devices] - (if not empty) cur_Name -> (if empty) devName")
sql.execute(""" UPDATE Devices
SET devName = COALESCE((
- SELECT cur_Name
+ SELECT cur_Name
FROM CurrentScan
WHERE cur_MAC = devMac
AND cur_Name IS NOT NULL
AND cur_Name <> 'null'
AND cur_Name <> ''
), devName)
- WHERE (devName IN ('(unknown)', '(name not found)', '')
+ WHERE (devName IN ('(unknown)', '(name not found)', '')
OR devName IS NULL)
AND EXISTS (
- SELECT 1
+ SELECT 1
FROM CurrentScan
WHERE cur_MAC = devMac
AND cur_Name IS NOT NULL
@@ -425,9 +418,9 @@ def print_scan_stats(db):
mylog("verbose", f" {row['cur_ScanMethod']}: {row['scan_method_count']}")
-#-------------------------------------------------------------------------------
-def create_new_devices (db):
- sql = db.sql # TO-DO
+# -------------------------------------------------------------------------------
+def create_new_devices(db):
+ sql = db.sql # TO-DO
startTime = timeNowDB()
# Insert events for new devices from CurrentScan (not yet in Devices)
@@ -474,36 +467,36 @@ def create_new_devices (db):
mylog("debug", "[New Devices] 2 Create devices")
# default New Device values preparation
- newDevColumns = """devAlertEvents,
- devAlertDown,
- devPresentLastScan,
- devIsArchived,
- devIsNew,
- devSkipRepeated,
- devScan,
- devOwner,
- devFavorite,
- devGroup,
- devComments,
- devLogEvents,
+ newDevColumns = """devAlertEvents,
+ devAlertDown,
+ devPresentLastScan,
+ devIsArchived,
+ devIsNew,
+ devSkipRepeated,
+ devScan,
+ devOwner,
+ devFavorite,
+ devGroup,
+ devComments,
+ devLogEvents,
devLocation,
devCustomProps,
devParentRelType,
devReqNicsOnline
"""
- newDevDefaults = f"""{get_setting_value("NEWDEV_devAlertEvents")},
- {get_setting_value("NEWDEV_devAlertDown")},
- {get_setting_value("NEWDEV_devPresentLastScan")},
- {get_setting_value("NEWDEV_devIsArchived")},
- {get_setting_value("NEWDEV_devIsNew")},
- {get_setting_value("NEWDEV_devSkipRepeated")},
- {get_setting_value("NEWDEV_devScan")},
- '{sanitize_SQL_input(get_setting_value("NEWDEV_devOwner"))}',
- {get_setting_value("NEWDEV_devFavorite")},
- '{sanitize_SQL_input(get_setting_value("NEWDEV_devGroup"))}',
- '{sanitize_SQL_input(get_setting_value("NEWDEV_devComments"))}',
- {get_setting_value("NEWDEV_devLogEvents")},
+ newDevDefaults = f"""{get_setting_value("NEWDEV_devAlertEvents")},
+ {get_setting_value("NEWDEV_devAlertDown")},
+ {get_setting_value("NEWDEV_devPresentLastScan")},
+ {get_setting_value("NEWDEV_devIsArchived")},
+ {get_setting_value("NEWDEV_devIsNew")},
+ {get_setting_value("NEWDEV_devSkipRepeated")},
+ {get_setting_value("NEWDEV_devScan")},
+ '{sanitize_SQL_input(get_setting_value("NEWDEV_devOwner"))}',
+ {get_setting_value("NEWDEV_devFavorite")},
+ '{sanitize_SQL_input(get_setting_value("NEWDEV_devGroup"))}',
+ '{sanitize_SQL_input(get_setting_value("NEWDEV_devComments"))}',
+ {get_setting_value("NEWDEV_devLogEvents")},
'{sanitize_SQL_input(get_setting_value("NEWDEV_devLocation"))}',
'{sanitize_SQL_input(get_setting_value("NEWDEV_devCustomProps"))}',
'{sanitize_SQL_input(get_setting_value("NEWDEV_devParentRelType"))}',
@@ -511,7 +504,7 @@ def create_new_devices (db):
"""
# Fetch data from CurrentScan skipping ignored devices by IP and MAC
- query = """SELECT cur_MAC, cur_Name, cur_Vendor, cur_ScanMethod, cur_IP, cur_SyncHubNodeName, cur_NetworkNodeMAC, cur_PORT, cur_NetworkSite, cur_SSID, cur_Type
+ query = """SELECT cur_MAC, cur_Name, cur_Vendor, cur_ScanMethod, cur_IP, cur_SyncHubNodeName, cur_NetworkNodeMAC, cur_PORT, cur_NetworkSite, cur_SSID, cur_Type
FROM CurrentScan """
mylog("debug", f"[New Devices] Collecting New Devices Query: {query}")
@@ -554,40 +547,40 @@ def create_new_devices (db):
)
# Preparing the individual insert statement
- sqlQuery = f"""INSERT OR IGNORE INTO Devices
+ sqlQuery = f"""INSERT OR IGNORE INTO Devices
(
- devMac,
- devName,
+ devMac,
+ devName,
devVendor,
- devLastIP,
- devFirstConnection,
- devLastConnection,
- devSyncHubNode,
+ devLastIP,
+ devFirstConnection,
+ devLastConnection,
+ devSyncHubNode,
devGUID,
- devParentMAC,
+ devParentMAC,
devParentPort,
- devSite,
+ devSite,
devSSID,
- devType,
- devSourcePlugin,
+ devType,
+ devSourcePlugin,
{newDevColumns}
)
- VALUES
+ VALUES
(
- '{sanitize_SQL_input(cur_MAC)}',
+ '{sanitize_SQL_input(cur_MAC)}',
'{sanitize_SQL_input(cur_Name)}',
- '{sanitize_SQL_input(cur_Vendor)}',
- '{sanitize_SQL_input(cur_IP)}',
- ?,
- ?,
- '{sanitize_SQL_input(cur_SyncHubNodeName)}',
+ '{sanitize_SQL_input(cur_Vendor)}',
+ '{sanitize_SQL_input(cur_IP)}',
+ ?,
+ ?,
+ '{sanitize_SQL_input(cur_SyncHubNodeName)}',
{sql_generateGuid},
'{sanitize_SQL_input(cur_NetworkNodeMAC)}',
'{sanitize_SQL_input(cur_PORT)}',
- '{sanitize_SQL_input(cur_NetworkSite)}',
+ '{sanitize_SQL_input(cur_NetworkSite)}',
'{sanitize_SQL_input(cur_SSID)}',
- '{sanitize_SQL_input(cur_Type)}',
- '{sanitize_SQL_input(cur_ScanMethod)}',
+ '{sanitize_SQL_input(cur_Type)}',
+ '{sanitize_SQL_input(cur_ScanMethod)}',
{newDevDefaults}
)"""
@@ -598,7 +591,8 @@ def create_new_devices (db):
mylog("debug", "[New Devices] New Devices end")
db.commitDB()
-#-------------------------------------------------------------------------------
+
+# -------------------------------------------------------------------------------
# Check if plugins data changed
def check_plugin_data_changed(pm, plugins_to_check):
"""
@@ -630,7 +624,7 @@ def check_plugin_data_changed(pm, plugins_to_check):
for plugin_name in plugins_to_check:
- last_data_change = pm.plugin_states.get(plugin_name, {}).get("lastDataChange")
+ last_data_change = pm.plugin_states.get(plugin_name, {}).get("lastDataChange")
last_data_check = pm.plugin_checks.get(plugin_name, "")
if not last_data_change:
@@ -639,13 +633,13 @@ def check_plugin_data_changed(pm, plugins_to_check):
# Normalize and validate last_changed timestamp
last_changed_ts = normalizeTimeStamp(last_data_change)
- if last_changed_ts == None:
+ if last_changed_ts is None:
mylog('none', f'[check_plugin_data_changed] Unexpected last_data_change timestamp for {plugin_name} (input|output): ({last_data_change}|{last_changed_ts})')
# Normalize and validate last_data_check timestamp
last_data_check_ts = normalizeTimeStamp(last_data_check)
- if last_data_check_ts == None:
+ if last_data_check_ts is None:
mylog('none', f'[check_plugin_data_changed] Unexpected last_data_check timestamp for {plugin_name} (input|output): ({last_data_check}|{last_data_check_ts})')
# Track which plugins have newer state than last_checked
@@ -660,15 +654,19 @@ def check_plugin_data_changed(pm, plugins_to_check):
# Continue if changes detected
for p in plugins_changed:
- mylog('debug', f'[check_plugin_data_changed] {p} changed (last_data_change|last_data_check): ({pm.plugin_states.get(p, {}).get("lastDataChange")}|{pm.plugin_checks.get(p)})')
+ mylog(
+ 'debug',
+ f'[check_plugin_data_changed] {p} changed (last_data_change|last_data_check): ({pm.plugin_states.get(p, {}).get("lastDataChange")}|{pm.plugin_checks.get(p)})'
+ )
return True
-#-------------------------------------------------------------------------------
+
+# -------------------------------------------------------------------------------
def update_devices_names(pm):
# --- Short-circuit if no name-resolution plugin has changed ---
- if check_plugin_data_changed(pm, ["DIGSCAN", "NSLOOKUP", "NBTSCAN", "AVAHISCAN"]) == False:
+ if check_plugin_data_changed(pm, ["DIGSCAN", "NSLOOKUP", "NBTSCAN", "AVAHISCAN"]) is False:
mylog('debug', '[Update Device Name] No relevant plugin changes since last check.')
return
@@ -676,8 +674,8 @@ def update_devices_names(pm):
sql = pm.db.sql
resolver = NameResolver(pm.db)
- device_handler = DeviceInstance(pm.db)
-
+ device_handler = DeviceInstance(pm.db)
+
nameNotFound = "(name not found)"
# Define resolution strategies in priority order
@@ -722,8 +720,7 @@ def update_devices_names(pm):
# If a valid result is found, record it and stop further attempts
if (
- newFQDN not in [nameNotFound, "", "localhost."]
- and " communications error to " not in newFQDN
+ newFQDN not in [nameNotFound, "", "localhost."] and " communications error to " not in newFQDN
):
foundStats[label] += 1
@@ -750,14 +747,14 @@ def update_devices_names(pm):
)
# Try resolving both name and FQDN
- recordsToUpdate, recordsNotFound, foundStats, notFound = resolve_devices(
+ recordsToUpdate, recordsNotFound, fs, notFound = resolve_devices(
unknownDevices
)
# Log summary
mylog(
"verbose",
- f"[Update Device Name] Names Found (DIGSCAN/AVAHISCAN/NSLOOKUP/NBTSCAN): {len(recordsToUpdate)} ({foundStats['DIGSCAN']}/{foundStats['AVAHISCAN']}/{foundStats['NSLOOKUP']}/{foundStats['NBTSCAN']})",
+ f"[Update Device Name] Names Found (DIGSCAN/AVAHISCAN/NSLOOKUP/NBTSCAN): {len(recordsToUpdate)} ({fs['DIGSCAN']}/{fs['AVAHISCAN']}/{fs['NSLOOKUP']}/{fs['NBTSCAN']})",
)
mylog("verbose", f"[Update Device Name] Names Not Found : {notFound}")
@@ -780,16 +777,14 @@ def update_devices_names(pm):
)
# Try resolving only FQDN
- recordsToUpdate, _, foundStats, notFound = resolve_devices(
+ recordsToUpdate, _, fs, notFound = resolve_devices(
allDevices, resolve_both_name_and_fqdn=False
)
# Log summary
mylog(
"verbose",
- f"[Update FQDN] Names Found (DIGSCAN/AVAHISCAN/NSLOOKUP/NBTSCAN): {len(recordsToUpdate)}"+
- f"({foundStats['DIGSCAN']}/{foundStats['AVAHISCAN']}/{foundStats['NSLOOKUP']}"+
- f"/{foundStats['NBTSCAN']})",
+ f"[Update FQDN] Names Found (DIGSCAN/AVAHISCAN/NSLOOKUP/NBTSCAN): {len(recordsToUpdate)}({fs['DIGSCAN']}/{fs['AVAHISCAN']}/{fs['NSLOOKUP']}/{fs['NBTSCAN']})",
)
mylog("verbose", f"[Update FQDN] Names Not Found : {notFound}")
@@ -803,7 +798,7 @@ def update_devices_names(pm):
# --- Step 3: Log last checked time ---
# After resolving names, update last checked
- pm.plugin_checks = {"DIGSCAN": timeNowDB(), "AVAHISCAN": timeNowDB(), "NSLOOKUP": timeNowDB(), "NBTSCAN": timeNowDB() }
+ pm.plugin_checks = {"DIGSCAN": timeNowDB(), "AVAHISCAN": timeNowDB(), "NSLOOKUP": timeNowDB(), "NBTSCAN": timeNowDB()}
# -------------------------------------------------------------------------------
@@ -901,7 +896,6 @@ def query_MAC_vendor(pMAC):
# Search vendor in HW Vendors DB
mac_start_string6 = mac[0:6]
- mac_start_string9 = mac[0:9]
try:
with open(filePath, "r") as f:
diff --git a/server/scan/device_heuristics.py b/server/scan/device_heuristics.py
index 6ff975b3..24d06558 100755
--- a/server/scan/device_heuristics.py
+++ b/server/scan/device_heuristics.py
@@ -1,16 +1,13 @@
-import sys
import os
import re
import json
import base64
from pathlib import Path
from typing import Optional, Tuple
+from logger import mylog
# Register NetAlertX directories
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
-sys.path.extend([f"{INSTALL_PATH}/server"])
-
-from logger import mylog
# Load MAC/device-type/icon rules from external file
MAC_TYPE_ICON_PATH = Path(f"{INSTALL_PATH}/back/device_heuristics_rules.json")
@@ -83,7 +80,7 @@ def match_vendor(vendor: str, default_type: str, default_icon: str) -> Tuple[str
for pattern in patterns:
# Only apply fallback when no MAC prefix is specified
- mac_prefix = pattern.get("mac_prefix", "")
+ # mac_prefix = pattern.get("mac_prefix", "")
vendor_pattern = pattern.get("vendor", "").lower()
if vendor_pattern and vendor_pattern in vendor_lc:
diff --git a/server/scan/name_resolution.py b/server/scan/name_resolution.py
index e331b786..8a9e226c 100755
--- a/server/scan/name_resolution.py
+++ b/server/scan/name_resolution.py
@@ -1,11 +1,4 @@
-import sys
-import os
import re
-
-# Register NetAlertX directories
-INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
-sys.path.extend([f"{INSTALL_PATH}/server"])
-
from logger import mylog
from helper import get_setting_value
@@ -31,7 +24,7 @@ class NameResolver:
# Check by MAC
sql.execute(f"""
- SELECT Watched_Value2 FROM Plugins_Objects
+ SELECT Watched_Value2 FROM Plugins_Objects
WHERE Plugin = '{plugin}' AND Object_PrimaryID = '{pMAC}'
""")
result = sql.fetchall()
@@ -42,9 +35,9 @@ class NameResolver:
# Check name by IP if enabled
if get_setting_value('NEWDEV_IP_MATCH_NAME'):
-
+
sql.execute(f"""
- SELECT Watched_Value2 FROM Plugins_Objects
+ SELECT Watched_Value2 FROM Plugins_Objects
WHERE Plugin = '{plugin}' AND Object_SecondaryID = '{pIP}'
""")
result = sql.fetchall()
diff --git a/server/scan/session_events.py b/server/scan/session_events.py
index cd549e80..e04961f0 100755
--- a/server/scan/session_events.py
+++ b/server/scan/session_events.py
@@ -1,10 +1,3 @@
-import sys
-import os
-
-# Register NetAlertX directories
-INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
-sys.path.extend([f"{INSTALL_PATH}/server"])
-
from scan.device_handling import (
create_new_devices,
print_scan_stats,
@@ -14,7 +7,7 @@ from scan.device_handling import (
)
from helper import get_setting_value
from db.db_helper import print_table_schema
-from utils.datetime_utils import timeNowDB, timeNowTZ
+from utils.datetime_utils import timeNowDB
from logger import mylog, Logger
from messaging.reporting import skip_repeated_notifications
@@ -133,20 +126,20 @@ def create_sessions_snapshot(db):
db.commitDB()
-#-------------------------------------------------------------------------------
-def insert_events (db):
- sql = db.sql #TO-DO
- startTime = timeNowDB()
-
+# -------------------------------------------------------------------------------
+def insert_events(db):
+ sql = db.sql # TO-DO
+ startTime = timeNowDB()
+
# Check device down
mylog("debug", "[Events] - 1 - Devices down")
sql.execute(f"""INSERT INTO Events (eve_MAC, eve_IP, eve_DateTime,
eve_EventType, eve_AdditionalInfo,
eve_PendingAlertEmail)
SELECT devMac, devLastIP, '{startTime}', 'Device Down', '', 1
- FROM Devices
+ FROM Devices
WHERE devAlertDown != 0
- AND devPresentLastScan = 1
+ AND devPresentLastScan = 1
AND NOT EXISTS (SELECT 1 FROM CurrentScan
WHERE devMac = cur_MAC
) """)
@@ -156,15 +149,15 @@ def insert_events (db):
sql.execute(f""" INSERT INTO Events (eve_MAC, eve_IP, eve_DateTime,
eve_EventType, eve_AdditionalInfo,
eve_PendingAlertEmail)
- SELECT DISTINCT c.cur_MAC, c.cur_IP, '{startTime}',
- CASE
- WHEN last_event.eve_EventType = 'Device Down' and last_event.eve_PendingAlertEmail = 0 THEN 'Down Reconnected'
- ELSE 'Connected'
+ SELECT DISTINCT c.cur_MAC, c.cur_IP, '{startTime}',
+ CASE
+ WHEN last_event.eve_EventType = 'Device Down' and last_event.eve_PendingAlertEmail = 0 THEN 'Down Reconnected'
+ ELSE 'Connected'
END,
'',
1
- FROM CurrentScan AS c
- LEFT JOIN LatestEventsPerMAC AS last_event ON c.cur_MAC = last_event.eve_MAC
+ FROM CurrentScan AS c
+ LEFT JOIN LatestEventsPerMAC AS last_event ON c.cur_MAC = last_event.eve_MAC
WHERE last_event.devPresentLastScan = 0 OR last_event.eve_MAC IS NULL
""")
@@ -190,7 +183,7 @@ def insert_events (db):
SELECT cur_MAC, cur_IP, '{startTime}', 'IP Changed',
'Previous IP: '|| devLastIP, devAlertEvents
FROM Devices, CurrentScan
- WHERE devMac = cur_MAC
+ WHERE devMac = cur_MAC
AND devLastIP <> cur_IP """)
mylog("debug", "[Events] - Events end")
diff --git a/server/utils/datetime_utils.py b/server/utils/datetime_utils.py
index 21fab835..cb51a16d 100644
--- a/server/utils/datetime_utils.py
+++ b/server/utils/datetime_utils.py
@@ -1,49 +1,43 @@
-#!/usr/bin/env python
+# !/usr/bin/env python
-import os
-import pathlib
-import sys
-from datetime import datetime
+# from datetime import datetime
from dateutil import parser
import datetime
import re
import pytz
-from pytz import timezone
from typing import Union
from zoneinfo import ZoneInfo
import email.utils
-
-# Register NetAlertX directories
-INSTALL_PATH="/app"
-sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-
import conf
-from const import *
+# from const import *
-
-#-------------------------------------------------------------------------------
+# -------------------------------------------------------------------------------
# DateTime
-#-------------------------------------------------------------------------------
+# -------------------------------------------------------------------------------
DATETIME_PATTERN = "%Y-%m-%d %H:%M:%S"
DATETIME_REGEX = re.compile(r'^\d{4}-\d{2}-\d{2} \d{2}:\d{2}:\d{2}$')
+
def timeNowTZ():
if conf.tz:
return datetime.datetime.now(conf.tz).replace(microsecond=0)
else:
return datetime.datetime.now().replace(microsecond=0)
+
def timeNow():
return datetime.datetime.now().replace(microsecond=0)
-def get_timezone_offset():
+
+def get_timezone_offset():
now = datetime.datetime.now(conf.tz)
- offset_hours = now.utcoffset().total_seconds() / 3600
+ offset_hours = now.utcoffset().total_seconds() / 3600
offset_formatted = "{:+03d}:{:02d}".format(int(offset_hours), int((offset_hours % 1) * 60))
return offset_formatted
+
def timeNowDB(local=True):
"""
Return the current time (local or UTC) as ISO 8601 for DB storage.
@@ -67,9 +61,9 @@ def timeNowDB(local=True):
return datetime.datetime.now(datetime.UTC).strftime(DATETIME_PATTERN)
-#-------------------------------------------------------------------------------
+# -------------------------------------------------------------------------------
# Date and time methods
-#-------------------------------------------------------------------------------
+# -------------------------------------------------------------------------------
def normalizeTimeStamp(inputTimeStamp):
"""
@@ -91,7 +85,7 @@ def normalizeTimeStamp(inputTimeStamp):
# Epoch timestamp (integer or float)
if isinstance(inputTimeStamp, (int, float)):
- try:
+ try:
return datetime.datetime.fromtimestamp(inputTimeStamp)
except (OSError, OverflowError, ValueError):
return None
@@ -125,6 +119,7 @@ def format_date_iso(date1: str) -> str:
dt = datetime.datetime.fromisoformat(date1) if isinstance(date1, str) else date1
return dt.isoformat()
+
# -------------------------------------------------------------------------------------------
def format_event_date(date_str: str, event_type: str) -> str:
"""Format event date with fallback rules."""
@@ -135,6 +130,7 @@ def format_event_date(date_str: str, event_type: str) -> str:
else:
return ""
+
# -------------------------------------------------------------------------------------------
def ensure_datetime(dt: Union[str, datetime.datetime, None]) -> datetime.datetime:
if dt is None:
@@ -157,6 +153,7 @@ def parse_datetime(dt_str):
except ValueError:
return None
+
def format_date(date_str: str) -> str:
try:
dt = parse_datetime(date_str)
@@ -168,13 +165,14 @@ def format_date(date_str: str) -> str:
except (ValueError, AttributeError, TypeError):
return "invalid"
+
def format_date_diff(date1, date2, tz_name):
"""
Return difference between two datetimes as 'Xd HH:MM'.
Uses app timezone if datetime is naive.
date2 can be None (uses now).
"""
- # Get timezone from settings
+ # Get timezone from settings
tz = pytz.timezone(tz_name)
def parse_dt(dt):
@@ -184,8 +182,8 @@ def format_date_diff(date1, date2, tz_name):
try:
dt_parsed = email.utils.parsedate_to_datetime(dt)
except (ValueError, TypeError):
- # fallback: parse ISO string
- dt_parsed = datetime.datetime.fromisoformat(dt)
+ # fallback: parse ISO string
+ dt_parsed = datetime.datetime.fromisoformat(dt)
# convert naive GMT/UTC to app timezone
if dt_parsed.tzinfo is None:
dt_parsed = tz.localize(dt_parsed)
@@ -208,4 +206,4 @@ def format_date_diff(date1, date2, tz_name):
"hours": hours,
"minutes": minutes,
"total_minutes": total_minutes
- }
\ No newline at end of file
+ }
diff --git a/server/utils/plugin_utils.py b/server/utils/plugin_utils.py
index edc820cf..4b4b3bd4 100755
--- a/server/utils/plugin_utils.py
+++ b/server/utils/plugin_utils.py
@@ -1,6 +1,6 @@
import os
import json
-
+from collections import namedtuple
import conf
from logger import mylog
from utils.crypto_utils import decrypt_data
@@ -220,9 +220,7 @@ def get_plugins_configs(loadAll):
# Load all plugins if `loadAll` is True, the plugin is in the enabled list,
# or no specific plugins are enabled (enabledPlugins is empty)
if (
- loadAll
- or plugJson["unique_prefix"] in enabledPlugins
- or enabledPlugins == []
+ loadAll or plugJson["unique_prefix"] in enabledPlugins or enabledPlugins == []
):
# Load the contents of the config.json file as a JSON object and append it to pluginsList
pluginsList.append(plugJson)
diff --git a/server/workflows/actions.py b/server/workflows/actions.py
index f8d01de1..8ef30bdd 100755
--- a/server/workflows/actions.py
+++ b/server/workflows/actions.py
@@ -1,11 +1,4 @@
import sqlite3
-import os
-import sys
-
-# Register NetAlertX directories
-INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
-sys.path.extend([f"{INSTALL_PATH}/server"])
-
from logger import mylog, Logger
from helper import get_setting_value
from models.device_instance import DeviceInstance
@@ -15,7 +8,6 @@ from models.plugin_object_instance import PluginObjectInstance
Logger(get_setting_value("LOG_LEVEL"))
-
class Action:
"""Base class for all actions."""
diff --git a/server/workflows/app_events.py b/server/workflows/app_events.py
index adbd5b8b..6396e30a 100755
--- a/server/workflows/app_events.py
+++ b/server/workflows/app_events.py
@@ -1,10 +1,3 @@
-import os
-import sys
-
-# Register NetAlertX directories
-INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
-sys.path.extend([f"{INSTALL_PATH}/server"])
-
from helper import get_setting_value
from logger import Logger
from const import sql_generateGuid
@@ -96,11 +89,11 @@ class AppEvent_obj:
"ObjectPrimaryID" TEXT,
"ObjectSecondaryID" TEXT,
"ObjectForeignKey" TEXT,
- "ObjectIndex" TEXT,
- "ObjectIsNew" BOOLEAN,
- "ObjectIsArchived" BOOLEAN,
+ "ObjectIndex" TEXT,
+ "ObjectIsNew" BOOLEAN,
+ "ObjectIsArchived" BOOLEAN,
"ObjectStatusColumn" TEXT,
- "ObjectStatus" TEXT,
+ "ObjectStatus" TEXT,
"AppEventType" TEXT,
"Helper1" TEXT,
"Helper2" TEXT,
@@ -117,11 +110,11 @@ class AppEvent_obj:
CREATE TRIGGER IF NOT EXISTS "{trigger_name}"
AFTER {event.upper()} ON "{table_name}"
WHEN NOT EXISTS (
- SELECT 1 FROM AppEvents
- WHERE AppEventProcessed = 0
+ SELECT 1 FROM AppEvents
+ WHERE AppEventProcessed = 0
AND ObjectType = '{table_name}'
AND ObjectGUID = {manage_prefix(config["fields"]["ObjectGUID"], event)}
- AND ObjectStatus = {manage_prefix(config["fields"]["ObjectStatus"], event)}
+ AND ObjectStatus = {manage_prefix(config["fields"]["ObjectStatus"], event)}
AND AppEventType = '{event.lower()}'
)
BEGIN
@@ -142,10 +135,10 @@ class AppEvent_obj:
"AppEventType"
)
VALUES (
- {sql_generateGuid},
- DATETIME('now'),
- FALSE,
- '{table_name}',
+ {sql_generateGuid},
+ DATETIME('now'),
+ FALSE,
+ '{table_name}',
{manage_prefix(config["fields"]["ObjectGUID"], event)}, -- ObjectGUID
{manage_prefix(config["fields"]["ObjectPrimaryID"], event)}, -- ObjectPrimaryID
{manage_prefix(config["fields"]["ObjectSecondaryID"], event)}, -- ObjectSecondaryID
diff --git a/server/workflows/conditions.py b/server/workflows/conditions.py
index bac0a8a0..cff3b44a 100755
--- a/server/workflows/conditions.py
+++ b/server/workflows/conditions.py
@@ -1,12 +1,5 @@
import re
import json
-import os
-import sys
-
-# Register NetAlertX directories
-INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
-sys.path.extend([f"{INSTALL_PATH}/server"])
-
from logger import mylog, Logger
from helper import get_setting_value
diff --git a/server/workflows/manager.py b/server/workflows/manager.py
index 97546fa9..787426cb 100755
--- a/server/workflows/manager.py
+++ b/server/workflows/manager.py
@@ -1,22 +1,17 @@
import json
-import os
-import sys
-
-# Register NetAlertX directories
-INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
-sys.path.extend([f"{INSTALL_PATH}/server"])
-
from const import fullConfFolder
from logger import mylog, Logger
from helper import get_setting_value
-# Make sure log level is initialized correctly
-Logger(get_setting_value("LOG_LEVEL"))
-
from workflows.triggers import Trigger
from workflows.conditions import ConditionGroup
from workflows.actions import DeleteObjectAction, RunPluginAction, UpdateFieldAction
+
+# Make sure log level is initialized correctly
+Logger(get_setting_value("LOG_LEVEL"))
+
+
class WorkflowManager:
def __init__(self, db):
self.db = db
diff --git a/server/workflows/triggers.py b/server/workflows/triggers.py
index 6080624e..81cd947a 100755
--- a/server/workflows/triggers.py
+++ b/server/workflows/triggers.py
@@ -1,11 +1,4 @@
import json
-import os
-import sys
-
-# Register NetAlertX directories
-INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
-sys.path.extend([f"{INSTALL_PATH}/server"])
-
from logger import mylog, Logger
from helper import get_setting_value
from database import get_array_from_sql_rows
@@ -28,8 +21,7 @@ class Trigger:
self.event_type = triggerJson["event_type"]
self.event = event # Store the triggered event context, if provided
self.triggered = (
- self.object_type == event["ObjectType"]
- and self.event_type == event["AppEventType"]
+ self.object_type == event["ObjectType"] and self.event_type == event["AppEventType"]
)
mylog(
@@ -53,9 +45,9 @@ class Trigger:
raise ValueError(m)
query = f"""
- SELECT * FROM
+ SELECT * FROM
{db_table}
- WHERE {refField} = '{event["ObjectGUID"]}'
+ WHERE {refField} = '{event["ObjectGUID"]}'
"""
mylog("debug", [query])
diff --git a/test/__init__.py b/test/__init__.py
index af5e39a5..c9e082a6 100755
--- a/test/__init__.py
+++ b/test/__init__.py
@@ -1 +1 @@
-""" tests for NetAlertX """
\ No newline at end of file
+""" tests for NetAlertX """
diff --git a/test/api_endpoints/test_dbquery_endpoints.py b/test/api_endpoints/test_dbquery_endpoints.py
index b7fffb6a..74202136 100644
--- a/test/api_endpoints/test_dbquery_endpoints.py
+++ b/test/api_endpoints/test_dbquery_endpoints.py
@@ -7,9 +7,9 @@ import pytest
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from helper import get_setting_value
-from utils.datetime_utils import timeNowDB
-from api_server.api_server_start import app
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
+from api_server.api_server_start import app # noqa: E402 [flake8 lint suppression]
@pytest.fixture(scope="session")
@@ -26,7 +26,7 @@ def client():
@pytest.fixture(scope="session")
def test_mac():
# Generate a unique MAC for each test run
- return "AA:BB:CC:" + ":".join(f"{random.randint(0,255):02X}" for _ in range(3))
+ return "AA:BB:CC:" + ":".join(f"{random.randint(0, 255):02X}" for _ in range(3))
def auth_headers(token):
diff --git a/test/api_endpoints/test_device_endpoints.py b/test/api_endpoints/test_device_endpoints.py
index 5d9d1fb7..f0e4c1c3 100644
--- a/test/api_endpoints/test_device_endpoints.py
+++ b/test/api_endpoints/test_device_endpoints.py
@@ -1,17 +1,17 @@
import sys
-import pathlib
-import sqlite3
+# import pathlib
+# import sqlite3
import random
-import string
-import uuid
+# import string
+# import uuid
import os
import pytest
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from helper import get_setting_value
-from api_server.api_server_start import app
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from api_server.api_server_start import app # noqa: E402 [flake8 lint suppression]
@pytest.fixture(scope="session")
@@ -28,7 +28,7 @@ def client():
@pytest.fixture
def test_mac():
# Generate a unique MAC for each test run
- return "AA:BB:CC:" + ":".join(f"{random.randint(0,255):02X}" for _ in range(3))
+ return "AA:BB:CC:" + ":".join(f"{random.randint(0, 255):02X}" for _ in range(3))
def auth_headers(token):
@@ -38,7 +38,6 @@ def auth_headers(token):
def test_create_device(client, api_token, test_mac):
payload = {
"createNew": True,
- "devType": "Test Device",
"devOwner": "Unit Test",
"devType": "Router",
"devVendor": "TestVendor",
@@ -103,7 +102,7 @@ def test_copy_device(client, api_token, test_mac):
# Step 2: Generate a target MAC
target_mac = "AA:BB:CC:" + ":".join(
- f"{random.randint(0,255):02X}" for _ in range(3)
+ f"{random.randint(0, 255):02X}" for _ in range(3)
)
# Step 3: Copy device
diff --git a/test/api_endpoints/test_devices_endpoints.py b/test/api_endpoints/test_devices_endpoints.py
index dadc99d6..7d8fbb8c 100644
--- a/test/api_endpoints/test_devices_endpoints.py
+++ b/test/api_endpoints/test_devices_endpoints.py
@@ -1,32 +1,36 @@
import sys
-import pathlib
-import sqlite3
+# import pathlib
+# import sqlite3
import base64
import random
-import string
-import uuid
+# import string
+# import uuid
import os
import pytest
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from helper import get_setting_value
-from api_server.api_server_start import app
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from api_server.api_server_start import app # noqa: E402 [flake8 lint suppression]
+
@pytest.fixture(scope="session")
def api_token():
return get_setting_value("API_TOKEN")
+
@pytest.fixture
def client():
with app.test_client() as client:
yield client
+
@pytest.fixture
def test_mac():
# Generate a unique MAC for each test run
- return "AA:BB:CC:" + ":".join(f"{random.randint(0,255):02X}" for _ in range(3))
+ return "AA:BB:CC:" + ":".join(f"{random.randint(0, 255):02X}" for _ in range(3))
+
def auth_headers(token):
return {"Authorization": f"Bearer {token}"}
@@ -40,12 +44,13 @@ def create_dummy(client, api_token, test_mac):
"devType": "Router",
"devVendor": "TestVendor",
}
- resp = client.post(f"/device/{test_mac}", json=payload, headers=auth_headers(api_token))
+ client.post(f"/device/{test_mac}", json=payload, headers=auth_headers(api_token))
+
def test_get_all_devices(client, api_token, test_mac):
# Ensure there is at least one device
create_dummy(client, api_token, test_mac)
-
+
# Fetch all devices
resp = client.get("/devices", headers=auth_headers(api_token))
assert resp.status_code == 200
@@ -59,7 +64,7 @@ def test_get_all_devices(client, api_token, test_mac):
def test_delete_devices_with_macs(client, api_token, test_mac):
# First create device so it exists
create_dummy(client, api_token, test_mac)
-
+
client.post(f"/device/{test_mac}", json={"createNew": True}, headers=auth_headers(api_token))
# Delete by MAC
@@ -67,6 +72,7 @@ def test_delete_devices_with_macs(client, api_token, test_mac):
assert resp.status_code == 200
assert resp.json.get("success") is True
+
def test_delete_all_empty_macs(client, api_token):
resp = client.delete("/devices/empty-macs", headers=auth_headers(api_token))
assert resp.status_code == 200
@@ -79,6 +85,7 @@ def test_delete_unknown_devices(client, api_token):
assert resp.status_code == 200
assert resp.json.get("success") is True
+
def test_export_devices_csv(client, api_token, test_mac):
# Create a device first
create_dummy(client, api_token, test_mac)
@@ -92,6 +99,7 @@ def test_export_devices_csv(client, api_token, test_mac):
# CSV should contain test_mac
assert test_mac in resp.data.decode()
+
def test_export_devices_json(client, api_token, test_mac):
# Create a device first
create_dummy(client, api_token, test_mac)
@@ -101,7 +109,7 @@ def test_export_devices_json(client, api_token, test_mac):
assert resp.status_code == 200
assert resp.is_json
data = resp.get_json()
- assert any(dev.get("devMac") == test_mac for dev in data["data"])
+ assert any(dev.get("devMac") == test_mac for dev in data["data"])
def test_export_devices_invalid_format(client, api_token):
@@ -143,6 +151,7 @@ def test_export_import_cycle_base64(client, api_token, test_mac):
assert resp.json.get("inserted") >= 1
assert resp.json.get("skipped_lines") == []
+
def test_devices_totals(client, api_token, test_mac):
# 1. Create a dummy device
create_dummy(client, api_token, test_mac)
@@ -189,9 +198,10 @@ def test_devices_by_status(client, api_token, test_mac):
assert fav_data is not None
assert "★" in fav_data["title"]
+
def test_delete_test_devices(client, api_token, test_mac):
# Delete by MAC
resp = client.delete("/devices", json={"macs": ["AA:BB:CC:*"]}, headers=auth_headers(api_token))
assert resp.status_code == 200
- assert resp.json.get("success") is True
\ No newline at end of file
+ assert resp.json.get("success") is True
diff --git a/test/api_endpoints/test_events_endpoints.py b/test/api_endpoints/test_events_endpoints.py
index 512cb62f..c5ba46fd 100644
--- a/test/api_endpoints/test_events_endpoints.py
+++ b/test/api_endpoints/test_events_endpoints.py
@@ -1,37 +1,38 @@
import sys
-import pathlib
-import sqlite3
-import random
-import string
-import uuid
import os
import pytest
-from datetime import datetime, timedelta
+import random
+from datetime import timedelta
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from helper import get_setting_value
-from utils.datetime_utils import timeNowTZ
-from api_server.api_server_start import app
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from utils.datetime_utils import timeNowTZ # noqa: E402 [flake8 lint suppression]
+from api_server.api_server_start import app # noqa: E402 [flake8 lint suppression]
+
@pytest.fixture(scope="session")
def api_token():
return get_setting_value("API_TOKEN")
+
@pytest.fixture
def client():
with app.test_client() as client:
yield client
+
@pytest.fixture
def test_mac():
# Generate a unique MAC for each test run
- return "AA:BB:CC:" + ":".join(f"{random.randint(0,255):02X}" for _ in range(3))
+ return "AA:BB:CC:" + ":".join(f"{random.randint(0, 255):02X}" for _ in range(3))
+
def auth_headers(token):
return {"Authorization": f"Bearer {token}"}
+
def create_event(client, api_token, mac, event="UnitTest Event", days_old=None):
payload = {"ip": "0.0.0.0", "event_type": event}
@@ -43,10 +44,12 @@ def create_event(client, api_token, mac, event="UnitTest Event", days_old=None):
return client.post(f"/events/create/{mac}", json=payload, headers=auth_headers(api_token))
+
def list_events(client, api_token, mac=None):
url = "/events" if mac is None else f"/events?mac={mac}"
return client.get(url, headers=auth_headers(api_token))
+
def test_create_event(client, api_token, test_mac):
# create event
resp = create_event(client, api_token, test_mac)
@@ -82,6 +85,7 @@ def test_delete_events_for_mac(client, api_token, test_mac):
assert resp.status_code == 200
assert len(resp.json.get("events", [])) == 0
+
def test_get_events_totals(client, api_token):
# 1. Request totals with default period
resp = client.get(
@@ -108,7 +112,6 @@ def test_get_events_totals(client, api_token):
assert len(data_month) == 6
-
def test_delete_all_events(client, api_token, test_mac):
# create two events
create_event(client, api_token, test_mac)
@@ -146,5 +149,3 @@ def test_delete_events_dynamic_days(client, api_token, test_mac):
events = resp.get_json().get("events", [])
mac_events = [ev for ev in events if ev.get("eve_MAC") == test_mac]
assert len(mac_events) == 1
-
-
diff --git a/test/api_endpoints/test_graphq_endpoints.py b/test/api_endpoints/test_graphq_endpoints.py
index e7b7d4ee..26255ffb 100644
--- a/test/api_endpoints/test_graphq_endpoints.py
+++ b/test/api_endpoints/test_graphq_endpoints.py
@@ -1,31 +1,30 @@
import sys
-import pathlib
-import sqlite3
import random
-import string
-import uuid
import pytest
-from datetime import datetime, timedelta
INSTALL_PATH = "/app"
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from helper import get_setting_value
-from api_server.api_server_start import app
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from api_server.api_server_start import app # noqa: E402 [flake8 lint suppression]
+
@pytest.fixture(scope="session")
def api_token():
return get_setting_value("API_TOKEN")
+
@pytest.fixture
def client():
with app.test_client() as client:
yield client
+
@pytest.fixture
def test_mac():
# Generate a unique MAC for each test run
- return "AA:BB:CC:" + ":".join(f"{random.randint(0,255):02X}" for _ in range(3))
+ return "AA:BB:CC:" + ":".join(f"{random.randint(0, 255):02X}" for _ in range(3))
+
def auth_headers(token):
return {"Authorization": f"Bearer {token}"}
@@ -37,6 +36,7 @@ def test_graphql_debug_get(client):
assert resp.status_code == 200
assert resp.data.decode() == "NetAlertX GraphQL server running."
+
def test_graphql_post_unauthorized(client):
"""POST /graphql without token should return 401"""
query = {"query": "{ devices { devName devMac } }"}
@@ -47,13 +47,14 @@ def test_graphql_post_unauthorized(client):
# --- DEVICES TESTS ---
+
def test_graphql_post_devices(client, api_token):
"""POST /graphql with a valid token should return device data"""
query = {
"query": """
{
devices {
- devices {
+ devices {
devGUID
devGroup
devIsRandomMac
@@ -77,8 +78,8 @@ def test_graphql_post_devices(client, api_token):
assert isinstance(data["devices"]["devices"], list)
assert isinstance(data["devices"]["count"], int)
-# --- SETTINGS TESTS ---
+# --- SETTINGS TESTS ---
def test_graphql_post_settings(client, api_token):
"""POST /graphql should return settings data"""
query = {
@@ -97,8 +98,8 @@ def test_graphql_post_settings(client, api_token):
assert "settings" in data
assert isinstance(data["settings"]["settings"], list)
-# --- LANGSTRINGS TESTS ---
+# --- LANGSTRINGS TESTS ---
def test_graphql_post_langstrings_specific(client, api_token):
"""Retrieve a specific langString in a given language"""
query = {
@@ -167,4 +168,4 @@ def test_graphql_post_langstrings_all_languages(client, api_token):
assert data["enStrings"]["count"] >= 1
assert data["deStrings"]["count"] >= 1
# Ensure langCode matches
- assert all(e["langCode"] == "en_us" for e in data["enStrings"]["langStrings"])
\ No newline at end of file
+ assert all(e["langCode"] == "en_us" for e in data["enStrings"]["langStrings"])
diff --git a/test/api_endpoints/test_history_endpoints.py b/test/api_endpoints/test_history_endpoints.py
index cb53cafd..e1ec74ce 100644
--- a/test/api_endpoints/test_history_endpoints.py
+++ b/test/api_endpoints/test_history_endpoints.py
@@ -1,17 +1,13 @@
import sys
-import pathlib
-import sqlite3
import random
-import string
-import uuid
import os
import pytest
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from helper import get_setting_value
-from api_server.api_server_start import app
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from api_server.api_server_start import app # noqa: E402 [flake8 lint suppression]
@pytest.fixture(scope="session")
@@ -28,7 +24,7 @@ def client():
@pytest.fixture
def test_mac():
# Generate a unique MAC for each test run
- return "AA:BB:CC:" + ":".join(f"{random.randint(0,255):02X}" for _ in range(3))
+ return "AA:BB:CC:" + ":".join(f"{random.randint(0, 255):02X}" for _ in range(3))
def auth_headers(token):
@@ -36,6 +32,6 @@ def auth_headers(token):
def test_delete_history(client, api_token):
- resp = client.delete(f"/history", headers=auth_headers(api_token))
+ resp = client.delete("/history", headers=auth_headers(api_token))
assert resp.status_code == 200
assert resp.json.get("success") is True
diff --git a/test/api_endpoints/test_logs_endpoints.py b/test/api_endpoints/test_logs_endpoints.py
index cd62fd17..7bf81b3a 100644
--- a/test/api_endpoints/test_logs_endpoints.py
+++ b/test/api_endpoints/test_logs_endpoints.py
@@ -5,8 +5,9 @@ import pytest
INSTALL_PATH = "/app"
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from helper import get_setting_value
-from api_server.api_server_start import app
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from api_server.api_server_start import app # noqa: E402 [flake8 lint suppression]
+
# ----------------------------
# Fixtures
@@ -15,14 +16,17 @@ from api_server.api_server_start import app
def api_token():
return get_setting_value("API_TOKEN")
+
@pytest.fixture
def client():
with app.test_client() as client:
yield client
+
def auth_headers(token):
return {"Authorization": f"Bearer {token}"}
+
# ----------------------------
# Logs Endpoint Tests
# ----------------------------
@@ -31,16 +35,18 @@ def test_clean_log(client, api_token):
assert resp.status_code == 200
assert resp.json.get("success") is True
+
def test_clean_log_not_allowed(client, api_token):
resp = client.delete("/logs?file=not_allowed.log", headers=auth_headers(api_token))
assert resp.status_code == 400
assert resp.json.get("success") is False
+
# ----------------------------
# Execution Queue Endpoint Tests
# ----------------------------
def test_add_to_execution_queue(client, api_token):
- action_name = f"test_action_{random.randint(0,9999)}"
+ action_name = f"test_action_{random.randint(0, 9999)}"
resp = client.post(
"/logs/add-to-execution-queue",
json={"action": action_name},
@@ -50,6 +56,7 @@ def test_add_to_execution_queue(client, api_token):
assert resp.json.get("success") is True
assert action_name in resp.json.get("message", "")
+
def test_add_to_execution_queue_missing_action(client, api_token):
resp = client.post(
"/logs/add-to-execution-queue",
diff --git a/test/api_endpoints/test_messaging_in_app_endpoints.py b/test/api_endpoints/test_messaging_in_app_endpoints.py
index 5bd115a5..8d7271bd 100644
--- a/test/api_endpoints/test_messaging_in_app_endpoints.py
+++ b/test/api_endpoints/test_messaging_in_app_endpoints.py
@@ -1,11 +1,8 @@
# -----------------------------
# In-app notifications tests with cleanup
# -----------------------------
-
-import json
import random
import string
-import uuid
import pytest
import os
import sys
@@ -14,26 +11,31 @@ import sys
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from api_server.api_server_start import app
-from messaging.in_app import NOTIFICATION_API_FILE # Import the path to notifications file
-from helper import get_setting_value
+from api_server.api_server_start import app # noqa: E402 [flake8 lint suppression]
+from messaging.in_app import NOTIFICATION_API_FILE # noqa: E402 [flake8 lint suppression]
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+
@pytest.fixture(scope="session")
def api_token():
return get_setting_value("API_TOKEN")
+
@pytest.fixture
def client():
with app.test_client() as client:
yield client
+
def auth_headers(token):
return {"Authorization": f"Bearer {token}"}
+
@pytest.fixture
def random_content():
return "Test Notification " + "".join(random.choices(string.ascii_letters + string.digits, k=6))
+
@pytest.fixture
def notification_guid(client, api_token, random_content):
# Write a notification and return its GUID
@@ -50,6 +52,7 @@ def notification_guid(client, api_token, random_content):
assert guid is not None
return guid
+
@pytest.fixture(autouse=True)
def cleanup_notifications():
# Runs before and after each test
@@ -70,6 +73,7 @@ def cleanup_notifications():
with open(NOTIFICATION_API_FILE, "w") as f:
f.write(backup)
+
# -----------------------------
def test_write_notification(client, api_token, random_content):
resp = client.post(
@@ -80,6 +84,7 @@ def test_write_notification(client, api_token, random_content):
assert resp.status_code == 200
assert resp.json.get("success") is True
+
def test_get_unread_notifications(client, api_token, random_content):
client.post("/messaging/in-app/write", json={"content": random_content}, headers=auth_headers(api_token))
resp = client.get("/messaging/in-app/unread", headers=auth_headers(api_token))
@@ -87,22 +92,26 @@ def test_get_unread_notifications(client, api_token, random_content):
notifications = resp.json
assert any(n["content"] == random_content for n in notifications)
+
def test_mark_all_notifications_read(client, api_token, random_content):
client.post("/messaging/in-app/write", json={"content": random_content}, headers=auth_headers(api_token))
resp = client.post("/messaging/in-app/read/all", headers=auth_headers(api_token))
assert resp.status_code == 200
assert resp.json.get("success") is True
+
def test_mark_single_notification_read(client, api_token, notification_guid):
resp = client.post(f"/messaging/in-app/read/{notification_guid}", headers=auth_headers(api_token))
assert resp.status_code == 200
assert resp.json.get("success") is True
+
def test_delete_single_notification(client, api_token, notification_guid):
resp = client.delete(f"/messaging/in-app/delete/{notification_guid}", headers=auth_headers(api_token))
assert resp.status_code == 200
assert resp.json.get("success") is True
+
def test_delete_all_notifications(client, api_token, random_content):
# Add a notification first
client.post("/messaging/in-app/write", json={"content": random_content}, headers=auth_headers(api_token))
diff --git a/test/api_endpoints/test_nettools_endpoints.py b/test/api_endpoints/test_nettools_endpoints.py
index 19b35eba..20d2825d 100644
--- a/test/api_endpoints/test_nettools_endpoints.py
+++ b/test/api_endpoints/test_nettools_endpoints.py
@@ -1,32 +1,31 @@
import sys
-import pathlib
-import sqlite3
-import base64
import random
-import string
-import uuid
import os
import pytest
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from helper import get_setting_value
-from api_server.api_server_start import app
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from api_server.api_server_start import app # noqa: E402 [flake8 lint suppression]
+
@pytest.fixture(scope="session")
def api_token():
return get_setting_value("API_TOKEN")
+
@pytest.fixture
def client():
with app.test_client() as client:
yield client
+
@pytest.fixture
def test_mac():
# Generate a unique MAC for each test run
- return "AA:BB:CC:" + ":".join(f"{random.randint(0,255):02X}" for _ in range(3))
+ return "AA:BB:CC:" + ":".join(f"{random.randint(0, 255):02X}" for _ in range(3))
+
def auth_headers(token):
return {"Authorization": f"Bearer {token}"}
@@ -40,7 +39,8 @@ def create_dummy(client, api_token, test_mac):
"devType": "Router",
"devVendor": "TestVendor",
}
- resp = client.post(f"/device/{test_mac}", json=payload, headers=auth_headers(api_token))
+ client.post(f"/device/{test_mac}", json=payload, headers=auth_headers(api_token))
+
def test_wakeonlan_device(client, api_token, test_mac):
# 1. Ensure at least one device exists
@@ -73,6 +73,7 @@ def test_wakeonlan_device(client, api_token, test_mac):
assert data.get("success") is True
assert "WOL packet sent" in data.get("message", "")
+
def test_speedtest_endpoint(client, api_token):
# 1. Call the speedtest endpoint
resp = client.get("/nettools/speedtest", headers=auth_headers(api_token))
@@ -92,7 +93,8 @@ def test_speedtest_endpoint(client, api_token):
assert isinstance(data["output"], list)
# Optionally check that output lines are strings
assert all(isinstance(line, str) for line in data["output"])
-
+
+
def test_traceroute_device(client, api_token, test_mac):
# 1. Ensure at least one device exists
create_dummy(client, api_token, test_mac)
@@ -127,6 +129,7 @@ def test_traceroute_device(client, api_token, test_mac):
assert "output" in data
assert isinstance(data["output"], str)
+
@pytest.mark.parametrize("ip,expected_status", [
("8.8.8.8", 200),
("256.256.256.256", 400), # Invalid IP
@@ -147,6 +150,7 @@ def test_nslookup_endpoint(client, api_token, ip, expected_status):
assert data.get("success") is False
assert "error" in data
+
@pytest.mark.parametrize("ip,mode,expected_status", [
("127.0.0.1", "fast", 200),
pytest.param("127.0.0.1", "normal", 200, marks=pytest.mark.feature_complete),
@@ -172,6 +176,7 @@ def test_nmap_endpoint(client, api_token, ip, mode, expected_status):
assert data.get("success") is False
assert "error" in data
+
def test_nslookup_unauthorized(client):
# No auth headers
resp = client.post("/nettools/nslookup", json={"devLastIP": "8.8.8.8"})
@@ -180,6 +185,7 @@ def test_nslookup_unauthorized(client):
assert data.get("success") is False
assert data.get("error") == "Forbidden"
+
def test_nmap_unauthorized(client):
# No auth headers
resp = client.post("/nettools/nmap", json={"scan": "127.0.0.1", "mode": "fast"})
@@ -201,4 +207,4 @@ def test_internet_info_endpoint(client, api_token):
# Handle errors, e.g., curl failure
assert data.get("success") is False
assert "error" in data
- assert "details" in data
\ No newline at end of file
+ assert "details" in data
diff --git a/test/api_endpoints/test_sessions_endpoints.py b/test/api_endpoints/test_sessions_endpoints.py
index f222c532..55e7fa66 100644
--- a/test/api_endpoints/test_sessions_endpoints.py
+++ b/test/api_endpoints/test_sessions_endpoints.py
@@ -1,9 +1,5 @@
import sys
-import pathlib
-import sqlite3
import random
-import string
-import uuid
import os
import pytest
from datetime import datetime, timedelta
@@ -11,31 +7,35 @@ from datetime import datetime, timedelta
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from helper import get_setting_value
-from utils.datetime_utils import timeNowTZ, timeNowDB
-from api_server.api_server_start import app
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from utils.datetime_utils import timeNowTZ, timeNowDB # noqa: E402 [flake8 lint suppression]
+from api_server.api_server_start import app # noqa: E402 [flake8 lint suppression]
+
@pytest.fixture(scope="session")
def api_token():
return get_setting_value("API_TOKEN")
+
@pytest.fixture
def client():
with app.test_client() as client:
yield client
+
@pytest.fixture
def test_mac():
# Generate a unique MAC for each test run
- return "AA:BB:CC:" + ":".join(f"{random.randint(0,255):02X}" for _ in range(3))
+ return "AA:BB:CC:" + ":".join(f"{random.randint(0, 255):02X}" for _ in range(3))
+
def auth_headers(token):
return {"Authorization": f"Bearer {token}"}
+
def test_create_device(client, api_token, test_mac):
payload = {
"createNew": True,
- "devType": "Test Device",
"devOwner": "Unit Test",
"devType": "Router",
"devVendor": "TestVendor",
@@ -129,7 +129,7 @@ def test_device_session_events(client, api_token, test_mac):
# 2. Fetch session events with default type ('all') and period ('7 days')
resp = client.get(
- f"/sessions/session-events?type=all&period=7 days",
+ "/sessions/session-events?type=all&period=7 days",
headers=auth_headers(api_token)
)
assert resp.status_code == 200
@@ -159,6 +159,7 @@ def test_device_session_events(client, api_token, test_mac):
sessions = resp_sessions.json["data"]
assert isinstance(sessions, list)
+
# -----------------------------
def test_delete_session(client, api_token, test_mac):
# First create session
@@ -180,15 +181,12 @@ def test_delete_session(client, api_token, test_mac):
assert not any(ses["ses_MAC"] == test_mac for ses in sessions)
-
def test_get_sessions_calendar(client, api_token, test_mac):
"""
Test the /sessions/calendar endpoint.
Creates session and ensures the calendar output is correct.
Cleans up test sessions after test.
"""
-
-
# --- Setup: create two sessions for the test MAC ---
now = timeNowTZ()
start1 = (now - timedelta(days=2)).isoformat(timespec="seconds")
@@ -256,4 +254,4 @@ def test_get_sessions_calendar(client, api_token, test_mac):
assert "" in ses["tooltip"], f"End is None but session not marked as still connected: {ses}"
# --- Cleanup: delete all test sessions for this MAC ---
- client.delete(f"/sessions/delete?mac={test_mac}", headers=auth_headers(api_token))
\ No newline at end of file
+ client.delete(f"/sessions/delete?mac={test_mac}", headers=auth_headers(api_token))
diff --git a/test/api_endpoints/test_settings_endpoints.py b/test/api_endpoints/test_settings_endpoints.py
index a6b29e6f..95c1732b 100644
--- a/test/api_endpoints/test_settings_endpoints.py
+++ b/test/api_endpoints/test_settings_endpoints.py
@@ -1,36 +1,36 @@
import sys
-import pathlib
-import sqlite3
import random
-import string
-import uuid
import os
import pytest
-from datetime import datetime, timedelta
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from helper import get_setting_value
-from api_server.api_server_start import app
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from api_server.api_server_start import app # noqa: E402 [flake8 lint suppression]
+
@pytest.fixture(scope="session")
def api_token():
return get_setting_value("API_TOKEN")
+
@pytest.fixture
def client():
with app.test_client() as client:
yield client
+
@pytest.fixture
def test_mac():
# Generate a unique MAC for each test run
- return "AA:BB:CC:" + ":".join(f"{random.randint(0,255):02X}" for _ in range(3))
+ return "AA:BB:CC:" + ":".join(f"{random.randint(0, 255):02X}" for _ in range(3))
+
def auth_headers(token):
return {"Authorization": f"Bearer {token}"}
+
def test_get_setting_unauthorized(client):
resp = client.get("/settings/API_TOKEN") # no auth header
assert resp.status_code == 403
diff --git a/test/backend/test_compound_conditions.py b/test/backend/test_compound_conditions.py
index 5790dc4c..06367b3a 100644
--- a/test/backend/test_compound_conditions.py
+++ b/test/backend/test_compound_conditions.py
@@ -6,16 +6,17 @@ Tests the fix for Issue #1210 - compound conditions with multiple AND/OR clauses
import sys
import pytest
+import os
from unittest.mock import MagicMock
# Mock the logger module before importing SafeConditionBuilder
sys.modules['logger'] = MagicMock()
# Add parent directory to path for imports
-import os
+
sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..'))
-from server.db.sql_safe_builder import SafeConditionBuilder
+from server.db.sql_safe_builder import SafeConditionBuilder # noqa: E402 [flake8 lint suppression]
@pytest.fixture
@@ -100,6 +101,7 @@ def test_multiple_or_clauses(builder):
assert 'Device2' in param_values
assert 'Device3' in param_values
+
def test_mixed_and_or_clauses(builder):
"""Test mixed AND/OR logical operators."""
condition = "AND devName = 'Device1' OR devName = 'Device2' AND devFavorite = '1'"
diff --git a/test/backend/test_safe_builder_unit.py b/test/backend/test_safe_builder_unit.py
index 39ed08b1..38b7c2e2 100644
--- a/test/backend/test_safe_builder_unit.py
+++ b/test/backend/test_safe_builder_unit.py
@@ -137,7 +137,7 @@ def test_unicode_support(builder, unicode_str):
@pytest.mark.parametrize("case", [
- "", " ", "AND devName = ''", "AND devName = 'a'", "AND devName = '" + "x"*500 + "'"
+ "", " ", "AND devName = ''", "AND devName = 'a'", "AND devName = '" + "x" * 500 + "'"
])
def test_edge_cases(builder, case):
try:
diff --git a/test/backend/test_sql_injection_prevention.py b/test/backend/test_sql_injection_prevention.py
index 5a43534f..33550064 100644
--- a/test/backend/test_sql_injection_prevention.py
+++ b/test/backend/test_sql_injection_prevention.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python3
+# !/usr/bin/env python3
"""
Comprehensive SQL Injection Prevention Tests for NetAlertX
@@ -15,7 +15,7 @@ sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..', 'server'))
sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..', 'server', 'db'))
# Now import our module
-from sql_safe_builder import SafeConditionBuilder
+from sql_safe_builder import SafeConditionBuilder # noqa: E402 [flake8 lint suppression]
@pytest.fixture
@@ -28,7 +28,7 @@ def test_sql_injection_attempt_single_quote(builder):
"""Test that single quote injection attempts are blocked."""
malicious_input = "'; DROP TABLE users; --"
condition, params = builder.get_safe_condition_legacy(malicious_input)
-
+
# Should return empty condition when invalid
assert condition == ""
assert params == {}
@@ -38,7 +38,7 @@ def test_sql_injection_attempt_union(builder):
"""Test that UNION injection attempts are blocked."""
malicious_input = "1' UNION SELECT * FROM passwords --"
condition, params = builder.get_safe_condition_legacy(malicious_input)
-
+
# Should return empty condition when invalid
assert condition == ""
assert params == {}
@@ -48,7 +48,7 @@ def test_sql_injection_attempt_or_true(builder):
"""Test that OR 1=1 injection attempts are blocked."""
malicious_input = "' OR '1'='1"
condition, params = builder.get_safe_condition_legacy(malicious_input)
-
+
# Should return empty condition when invalid
assert condition == ""
assert params == {}
@@ -58,7 +58,7 @@ def test_valid_simple_condition(builder):
"""Test that valid simple conditions are handled correctly."""
valid_input = "AND devName = 'Test Device'"
condition, params = builder.get_safe_condition_legacy(valid_input)
-
+
# Should create parameterized query
assert "AND devName = :" in condition
assert len(params) == 1
@@ -69,7 +69,7 @@ def test_empty_condition(builder):
"""Test that empty conditions are handled safely."""
empty_input = ""
condition, params = builder.get_safe_condition_legacy(empty_input)
-
+
# Should return empty condition
assert condition == ""
assert params == {}
@@ -79,7 +79,7 @@ def test_whitespace_only_condition(builder):
"""Test that whitespace-only conditions are handled safely."""
whitespace_input = " \n\t "
condition, params = builder.get_safe_condition_legacy(whitespace_input)
-
+
# Should return empty condition
assert condition == ""
assert params == {}
@@ -90,7 +90,7 @@ def test_multiple_conditions_valid(builder):
# Test with a single condition first (our current parser handles single conditions well)
valid_input = "AND devName = 'Device1'"
condition, params = builder.get_safe_condition_legacy(valid_input)
-
+
# Should create parameterized query
assert "devName = :" in condition
assert len(params) == 1
@@ -101,7 +101,7 @@ def test_disallowed_column_name(builder):
"""Test that non-whitelisted column names are rejected."""
invalid_input = "AND malicious_column = 'value'"
condition, params = builder.get_safe_condition_legacy(invalid_input)
-
+
# Should return empty condition when column not in whitelist
assert condition == ""
assert params == {}
@@ -111,7 +111,7 @@ def test_disallowed_operator(builder):
"""Test that non-whitelisted operators are rejected."""
invalid_input = "AND devName SOUNDS LIKE 'test'"
condition, params = builder.get_safe_condition_legacy(invalid_input)
-
+
# Should return empty condition when operator not allowed
assert condition == ""
assert params == {}
@@ -121,7 +121,7 @@ def test_nested_select_attempt(builder):
"""Test that nested SELECT attempts are blocked."""
malicious_input = "AND devName IN (SELECT password FROM users)"
condition, params = builder.get_safe_condition_legacy(malicious_input)
-
+
# Should return empty condition when nested SELECT detected
assert condition == ""
assert params == {}
@@ -131,7 +131,7 @@ def test_hex_encoding_attempt(builder):
"""Test that hex-encoded injection attempts are blocked."""
malicious_input = "AND 0x44524f50205441424c45"
condition, params = builder.get_safe_condition_legacy(malicious_input)
-
+
# Should return empty condition when hex encoding detected
assert condition == ""
assert params == {}
@@ -141,7 +141,7 @@ def test_comment_injection_attempt(builder):
"""Test that comment injection attempts are handled."""
malicious_input = "AND devName = 'test' /* comment */ --"
condition, params = builder.get_safe_condition_legacy(malicious_input)
-
+
# Comments should be stripped and condition validated
if condition:
assert "/*" not in condition
@@ -152,7 +152,7 @@ def test_special_placeholder_replacement(builder):
"""Test that {s-quote} placeholder is safely replaced."""
input_with_placeholder = "AND devName = {s-quote}Test{s-quote}"
condition, params = builder.get_safe_condition_legacy(input_with_placeholder)
-
+
# Should handle placeholder safely
if condition:
assert "{s-quote}" not in condition
@@ -163,7 +163,7 @@ def test_null_byte_injection(builder):
"""Test that null byte injection attempts are blocked."""
malicious_input = "AND devName = 'test\x00' DROP TABLE --"
condition, params = builder.get_safe_condition_legacy(malicious_input)
-
+
# Null bytes should be sanitized
if condition:
assert "\x00" not in condition
@@ -178,7 +178,7 @@ def test_build_condition_with_allowed_values(builder):
{"column": "devName", "operator": "LIKE", "value": "%test%"}
]
condition, params = builder.build_condition(conditions, "AND")
-
+
# Should create valid parameterized condition
assert "eve_EventType = :" in condition
assert "devName LIKE :" in condition
@@ -191,7 +191,7 @@ def test_build_condition_with_invalid_column(builder):
{"column": "invalid_column", "operator": "=", "value": "test"}
]
condition, params = builder.build_condition(conditions)
-
+
# Should return empty when invalid column
assert condition == ""
assert params == {}
@@ -204,7 +204,7 @@ def test_case_variations_injection(builder):
"oR 1=1",
"UnIoN SeLeCt * FrOm users"
]
-
+
for malicious_input in malicious_inputs:
condition, params = builder.get_safe_condition_legacy(malicious_input)
# Should handle case variations safely
@@ -217,7 +217,7 @@ def test_time_based_injection_attempt(builder):
"""Test that time-based injection attempts are blocked."""
malicious_input = "AND IF(1=1, SLEEP(5), 0)"
condition, params = builder.get_safe_condition_legacy(malicious_input)
-
+
# Should return empty condition when SQL functions detected
assert condition == ""
assert params == {}
@@ -227,7 +227,7 @@ def test_stacked_queries_attempt(builder):
"""Test that stacked query attempts are blocked."""
malicious_input = "'; INSERT INTO admin VALUES ('hacker', 'password'); --"
condition, params = builder.get_safe_condition_legacy(malicious_input)
-
+
# Should return empty condition when semicolon detected
assert condition == ""
assert params == {}
diff --git a/test/backend/test_sql_security.py b/test/backend/test_sql_security.py
index cbec10b4..3434344e 100644
--- a/test/backend/test_sql_security.py
+++ b/test/backend/test_sql_security.py
@@ -13,16 +13,15 @@ import unittest
import sqlite3
import tempfile
import os
-from unittest.mock import Mock, patch, MagicMock
+from unittest.mock import Mock, patch
# Add the server directory to the path for imports
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/server"])
sys.path.append('/home/dell/coding/bash/10x-agentic-setup/netalertx-sql-fix/server')
-from db.sql_safe_builder import SafeConditionBuilder, create_safe_condition_builder
-from database import DB
-from messaging.reporting import get_notifications
+from db.sql_safe_builder import SafeConditionBuilder # noqa: E402 [flake8 lint suppression]
+from messaging.reporting import get_notifications # noqa: E402 [flake8 lint suppression]
class TestSafeConditionBuilder(unittest.TestCase):
@@ -83,7 +82,7 @@ class TestSafeConditionBuilder(unittest.TestCase):
def test_build_simple_condition_valid(self):
"""Test building valid simple conditions."""
sql, params = self.builder._build_simple_condition('AND', 'devName', '=', 'TestDevice')
-
+
self.assertIn('AND devName = :param_', sql)
self.assertEqual(len(params), 1)
self.assertIn('TestDevice', params.values())
@@ -92,20 +91,20 @@ class TestSafeConditionBuilder(unittest.TestCase):
"""Test that invalid column names are rejected."""
with self.assertRaises(ValueError) as context:
self.builder._build_simple_condition('AND', 'invalid_column', '=', 'value')
-
+
self.assertIn('Invalid column name', str(context.exception))
def test_build_simple_condition_invalid_operator(self):
"""Test that invalid operators are rejected."""
with self.assertRaises(ValueError) as context:
self.builder._build_simple_condition('AND', 'devName', 'UNION', 'value')
-
+
self.assertIn('Invalid operator', str(context.exception))
def test_build_in_condition_valid(self):
"""Test building valid IN conditions."""
sql, params = self.builder._build_in_condition('AND', 'eve_EventType', 'IN', "'Connected', 'Disconnected'")
-
+
self.assertIn('AND eve_EventType IN', sql)
self.assertEqual(len(params), 2)
self.assertIn('Connected', params.values())
@@ -114,7 +113,7 @@ class TestSafeConditionBuilder(unittest.TestCase):
def test_build_null_condition(self):
"""Test building NULL check conditions."""
sql, params = self.builder._build_null_condition('AND', 'devComments', 'IS NULL')
-
+
self.assertEqual(sql, 'AND devComments IS NULL')
self.assertEqual(len(params), 0)
@@ -154,7 +153,7 @@ class TestSafeConditionBuilder(unittest.TestCase):
def test_device_name_filter(self):
"""Test the device name filter helper method."""
sql, params = self.builder.build_device_name_filter("TestDevice")
-
+
self.assertIn('AND devName = :device_name_', sql)
self.assertIn('TestDevice', params.values())
@@ -162,14 +161,13 @@ class TestSafeConditionBuilder(unittest.TestCase):
"""Test the event type filter helper method."""
event_types = ['Connected', 'Disconnected']
sql, params = self.builder.build_event_type_filter(event_types)
-
+
self.assertIn('AND eve_EventType IN', sql)
self.assertEqual(len(params), 2)
self.assertIn('Connected', params.values())
self.assertIn('Disconnected', params.values())
-
class TestDatabaseParameterSupport(unittest.TestCase):
"""Test that database layer supports parameterized queries."""
@@ -177,7 +175,7 @@ class TestDatabaseParameterSupport(unittest.TestCase):
"""Set up test database."""
self.temp_db = tempfile.NamedTemporaryFile(delete=False, suffix='.db')
self.temp_db.close()
-
+
# Create test database
self.conn = sqlite3.connect(self.temp_db.name)
self.conn.execute('''CREATE TABLE test_table (
@@ -197,23 +195,23 @@ class TestDatabaseParameterSupport(unittest.TestCase):
def test_parameterized_query_execution(self):
"""Test that parameterized queries work correctly."""
cursor = self.conn.cursor()
-
+
# Test named parameters
cursor.execute("SELECT * FROM test_table WHERE name = :name", {'name': 'test1'})
results = cursor.fetchall()
-
+
self.assertEqual(len(results), 1)
self.assertEqual(results[0][1], 'test1')
def test_parameterized_query_prevents_injection(self):
"""Test that parameterized queries prevent SQL injection."""
cursor = self.conn.cursor()
-
+
# This should not cause SQL injection
malicious_input = "'; DROP TABLE test_table; --"
cursor.execute("SELECT * FROM test_table WHERE name = :name", {'name': malicious_input})
- results = cursor.fetchall()
-
+ # results = cursor.fetchall()
+
# The table should still exist and be queryable
cursor.execute("SELECT COUNT(*) FROM test_table")
count = cursor.fetchone()[0]
@@ -228,7 +226,7 @@ class TestReportingSecurityIntegration(unittest.TestCase):
self.mock_db = Mock()
self.mock_db.sql = Mock()
self.mock_db.get_table_as_json = Mock()
-
+
# Mock successful JSON response
mock_json_obj = Mock()
mock_json_obj.columnNames = ['MAC', 'Datetime', 'IP', 'Event Type', 'Device name', 'Comments']
@@ -245,7 +243,7 @@ class TestReportingSecurityIntegration(unittest.TestCase):
}.get(key, '')
# Call the function
- result = get_notifications(self.mock_db)
+ get_notifications(self.mock_db)
# Verify that get_table_as_json was called with parameters
self.mock_db.get_table_as_json.assert_called()
@@ -265,7 +263,6 @@ class TestReportingSecurityIntegration(unittest.TestCase):
# Ensure the parameter dict has the correct value (using actual param name)
self.assertEqual(list(params.values())[0], "TestDevice")
-
@patch('messaging.reporting.get_setting_value')
def test_events_section_security(self, mock_get_setting):
"""Test that events section uses safe SQL building."""
@@ -276,7 +273,7 @@ class TestReportingSecurityIntegration(unittest.TestCase):
}.get(key, '')
# Call the function
- result = get_notifications(self.mock_db)
+ get_notifications(self.mock_db)
# Verify that get_table_as_json was called with parameters
self.mock_db.get_table_as_json.assert_called()
@@ -291,7 +288,7 @@ class TestReportingSecurityIntegration(unittest.TestCase):
}.get(key, '')
# Call the function - should not raise an exception
- result = get_notifications(self.mock_db)
+ get_notifications(self.mock_db)
# Should still call get_table_as_json (with safe fallback query)
self.mock_db.get_table_as_json.assert_called()
@@ -306,7 +303,7 @@ class TestReportingSecurityIntegration(unittest.TestCase):
}.get(key, '')
# Call the function
- result = get_notifications(self.mock_db)
+ get_notifications(self.mock_db)
# Should call get_table_as_json
self.mock_db.get_table_as_json.assert_called()
@@ -322,12 +319,12 @@ class TestSecurityBenchmarks(unittest.TestCase):
def test_performance_simple_condition(self):
"""Test performance of simple condition building."""
import time
-
+
start_time = time.time()
for _ in range(1000):
sql, params = self.builder.build_safe_condition("AND devName = 'TestDevice'")
end_time = time.time()
-
+
execution_time = end_time - start_time
self.assertLess(execution_time, 1.0, "Simple condition building should be fast")
@@ -339,7 +336,7 @@ class TestSecurityBenchmarks(unittest.TestCase):
self.skipTest("psutil not available")
return
import os
-
+
process = psutil.Process(os.getpid())
initial_memory = process.memory_info().rss
@@ -350,7 +347,7 @@ class TestSecurityBenchmarks(unittest.TestCase):
final_memory = process.memory_info().rss
memory_increase = final_memory - initial_memory
-
+
# Memory increase should be reasonable (less than 10MB)
self.assertLess(memory_increase, 10 * 1024 * 1024, "Memory usage should be reasonable")
@@ -376,4 +373,4 @@ class TestSecurityBenchmarks(unittest.TestCase):
if __name__ == '__main__':
# Run the test suite
- unittest.main(verbosity=2)
\ No newline at end of file
+ unittest.main(verbosity=2)
diff --git a/test/docker_tests/test_mount_diagnostics_pytest.py b/test/docker_tests/test_mount_diagnostics_pytest.py
index c186d1a3..2a1bf40b 100644
--- a/test/docker_tests/test_mount_diagnostics_pytest.py
+++ b/test/docker_tests/test_mount_diagnostics_pytest.py
@@ -1,4 +1,4 @@
-#!/usr/bin/env python3
+# !/usr/bin/env python3
"""
Pytest-based Mount Diagnostic Tests for NetAlertX
diff --git a/test/docker_tests/test_ports_available.py b/test/docker_tests/test_ports_available.py
index 48876c5b..5b9ef3c0 100644
--- a/test/docker_tests/test_ports_available.py
+++ b/test/docker_tests/test_ports_available.py
@@ -38,7 +38,7 @@ def dummy_container(tmp_path):
f.write(" network_mode: host\n")
f.write(" userns_mode: host\n")
f.write(" command: sh -c \"while true; do nc -l -p 20211 < /dev/null > /dev/null; done & while true; do nc -l -p 20212 < /dev/null > /dev/null; done & sleep 30\"\n")
-
+
# Start the dummy container
import subprocess
result = subprocess.run(
@@ -47,12 +47,12 @@ def dummy_container(tmp_path):
)
if result.returncode != 0:
pytest.fail(f"Failed to start dummy container: {result.stderr}")
-
+
# Wait a bit for the container to start listening
time.sleep(3)
-
+
yield "dummy"
-
+
# Cleanup
subprocess.run(["docker-compose", "-f", str(compose_file), "down"], capture_output=True)
@@ -139,10 +139,10 @@ def _run_container(
# Copy the script content and run it
script_path = pathlib.Path("install/production-filesystem/entrypoint.d/99-ports-available.sh")
with script_path.open('r', encoding='utf-8') as f:
- script_content = f.read()
+ script_cont = f.read()
# Use printf to avoid shell interpretation issues
- script = f"printf '%s\\n' '{script_content.replace(chr(39), chr(39)+chr(92)+chr(39)+chr(39))}' > /tmp/ports-check.sh && chmod +x /tmp/ports-check.sh && sh /tmp/ports-check.sh"
+ script = f"printf '%s\\n' '{script_cont.replace(chr(39), chr(39) + chr(92) + chr(39) + chr(39))}' > /tmp/ports-check.sh && chmod +x /tmp/ports-check.sh && sh /tmp/ports-check.sh" # noqa: E501 - inline script
cmd.extend(["--entrypoint", "/bin/sh", IMAGE, "-c", script])
print(f"\n--- DOCKER CMD ---\n{' '.join(cmd)}\n--- END CMD ---\n")
@@ -157,8 +157,7 @@ def _run_container(
# Combine and clean stdout and stderr
stdouterr = (
- re.sub(r'\x1b\[[0-9;]*m', '', result.stdout or '') +
- re.sub(r'\x1b\[[0-9;]*m', '', result.stderr or '')
+ re.sub(r'\x1b\[[0-9;]*m', '', result.stdout or '') + re.sub(r'\x1b\[[0-9;]*m', '', result.stderr or '')
)
result.output = stdouterr
print(f"\n--- CONTAINER stdout ---\n{result.stdout}")
@@ -255,4 +254,4 @@ def test_ports_in_use_warning(dummy_container, tmp_path: pathlib.Path) -> None:
_assert_contains(result, "Port Warning: Application port 20211 is already in use")
_assert_contains(result, "Port Warning: GraphQL API port 20212 is already in use")
- assert result.returncode == 0
\ No newline at end of file
+ assert result.returncode == 0
diff --git a/test/integration/integration_test.py b/test/integration/integration_test.py
index 43eec190..65acfc9f 100755
--- a/test/integration/integration_test.py
+++ b/test/integration/integration_test.py
@@ -9,12 +9,14 @@ import sys
sys.path.insert(0, os.path.join(os.path.dirname(__file__), 'server'))
sys.path.insert(0, os.path.join(os.path.dirname(__file__), 'server', 'db'))
-from db.sql_safe_builder import SafeConditionBuilder, create_safe_condition_builder
-from messaging.reporting import get_notifications
+from db.sql_safe_builder import create_safe_condition_builder # noqa: E402 [flake8 lint suppression]
+from messaging.reporting import get_notifications # noqa: E402 [flake8 lint suppression]
# -----------------------------
# Fixtures
# -----------------------------
+
+
@pytest.fixture
def test_db_path():
path = tempfile.mktemp(suffix=".db")
@@ -22,10 +24,12 @@ def test_db_path():
if os.path.exists(path):
os.remove(path)
+
@pytest.fixture
def builder():
return create_safe_condition_builder()
+
@pytest.fixture
def test_db(test_db_path):
conn = sqlite3.connect(test_db_path)
@@ -96,6 +100,7 @@ def test_db(test_db_path):
# Tests
# -----------------------------
+
def test_fresh_install_compatibility(builder):
condition, params = builder.get_safe_condition_legacy("")
assert condition == ""
@@ -105,6 +110,7 @@ def test_fresh_install_compatibility(builder):
assert "devName = :" in condition
assert 'TestDevice' in params.values()
+
def test_existing_db_compatibility():
mock_db = Mock()
mock_result = Mock()
@@ -129,6 +135,7 @@ def test_existing_db_compatibility():
assert 'events_meta' in result
assert mock_db.get_table_as_json.called
+
def test_notification_system_integration(builder):
email_condition = "AND devName = 'EmailTestDevice'"
condition, params = builder.get_safe_condition_legacy(email_condition)
@@ -150,6 +157,7 @@ def test_notification_system_integration(builder):
assert "eve_MAC = :" in condition
assert 'aa:bb:cc:dd:ee:ff' in params.values()
+
def test_settings_persistence(builder):
test_settings = [
"AND devName = 'Persistent Device'",
@@ -163,6 +171,7 @@ def test_settings_persistence(builder):
assert isinstance(condition, str)
assert isinstance(params, dict)
+
def test_device_operations(builder):
device_conditions = [
"AND devName = 'Updated Device'",
@@ -175,6 +184,7 @@ def test_device_operations(builder):
assert len(params) > 0 or safe_condition == ""
assert "'" not in safe_condition
+
def test_plugin_functionality(builder):
plugin_conditions = [
"AND Plugin = 'TestPlugin'",
@@ -187,6 +197,7 @@ def test_plugin_functionality(builder):
assert ":" in safe_condition
assert len(params) > 0
+
def test_sql_injection_prevention(builder):
malicious_inputs = [
"'; DROP TABLE Events_Devices; --",
@@ -200,6 +211,7 @@ def test_sql_injection_prevention(builder):
assert condition == ""
assert params == {}
+
def test_error_handling(builder):
invalid_condition = "INVALID SQL SYNTAX HERE"
condition, params = builder.get_safe_condition_legacy(invalid_condition)
@@ -213,6 +225,7 @@ def test_error_handling(builder):
assert isinstance(condition, str)
assert isinstance(params, dict)
+
def test_backward_compatibility(builder):
legacy_conditions = [
"AND devName = {s-quote}Legacy Device{s-quote}",
@@ -226,6 +239,7 @@ def test_backward_compatibility(builder):
assert ":" in condition
assert len(params) > 0
+
def test_performance_impact(builder):
import time
test_condition = "AND devName = 'Performance Test Device'"
diff --git a/test/test_graphq_endpoints.py b/test/test_graphq_endpoints.py
index 58a185af..38788f36 100755
--- a/test/test_graphq_endpoints.py
+++ b/test/test_graphq_endpoints.py
@@ -1,32 +1,31 @@
import sys
-import pathlib
-import sqlite3
import random
-import string
-import uuid
import os
import pytest
-from datetime import datetime, timedelta
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
-from helper import get_setting_value
-from api_server.api_server_start import app
+from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
+from api_server.api_server_start import app # noqa: E402 [flake8 lint suppression]
+
@pytest.fixture(scope="session")
def api_token():
return get_setting_value("API_TOKEN")
+
@pytest.fixture
def client():
with app.test_client() as client:
yield client
+
@pytest.fixture
def test_mac():
# Generate a unique MAC for each test run
- return "AA:BB:CC:" + ":".join(f"{random.randint(0,255):02X}" for _ in range(3))
+ return "AA:BB:CC:" + ":".join(f"{random.randint(0, 255):02X}" for _ in range(3))
+
def auth_headers(token):
return {"Authorization": f"Bearer {token}"}
@@ -38,6 +37,7 @@ def test_graphql_debug_get(client):
assert resp.status_code == 200
assert resp.data.decode() == "NetAlertX GraphQL server running."
+
def test_graphql_post_unauthorized(client):
"""POST /graphql without token should return 401"""
query = {"query": "{ devices { devName devMac } }"}
@@ -47,13 +47,14 @@ def test_graphql_post_unauthorized(client):
error_text = resp.json.get("error", "") or resp.json.get("message", "")
assert "Unauthorized" in error_text or "Forbidden" in error_text
+
def test_graphql_post_devices(client, api_token):
"""POST /graphql with a valid token should return device data"""
query = {
"query": """
{
devices {
- devices {
+ devices {
devGUID
devGroup
devIsRandomMac
@@ -77,6 +78,7 @@ def test_graphql_post_devices(client, api_token):
assert isinstance(data["devices"]["devices"], list)
assert isinstance(data["devices"]["count"], int)
+
def test_graphql_post_settings(client, api_token):
"""POST /graphql should return settings data"""
query = {