mirror of
https://github.com/jokob-sk/NetAlertX.git
synced 2025-12-07 01:26:11 -08:00
@@ -1,6 +1,6 @@
|
|||||||
import json
|
import json
|
||||||
import os
|
import os
|
||||||
import sys
|
|
||||||
|
|
||||||
def merge_translations(main_file, other_files):
|
def merge_translations(main_file, other_files):
|
||||||
# Load main file
|
# Load main file
|
||||||
@@ -30,10 +30,14 @@ def merge_translations(main_file, other_files):
|
|||||||
json.dump(data, f, indent=4, ensure_ascii=False)
|
json.dump(data, f, indent=4, ensure_ascii=False)
|
||||||
f.truncate()
|
f.truncate()
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
current_path = os.path.dirname(os.path.abspath(__file__))
|
current_path = os.path.dirname(os.path.abspath(__file__))
|
||||||
# language codes can be found here: http://www.lingoes.net/en/translator/langcode.htm
|
# language codes can be found here: http://www.lingoes.net/en/translator/langcode.htm
|
||||||
# "en_us.json" has to be first!
|
# ⚠ "en_us.json" has to be first!
|
||||||
json_files = [ "en_us.json", "ar_ar.json", "ca_ca.json", "cs_cz.json", "de_de.json", "es_es.json", "fa_fa.json", "fr_fr.json", "it_it.json", "ja_jp.json", "nb_no.json", "pl_pl.json", "pt_br.json", "pt_pt.json", "ru_ru.json", "sv_sv.json", "tr_tr.json", "uk_ua.json", "zh_cn.json"]
|
json_files = ["en_us.json", "ar_ar.json", "ca_ca.json", "cs_cz.json", "de_de.json",
|
||||||
|
"es_es.json", "fa_fa.json", "fr_fr.json", "it_it.json", "ja_jp.json",
|
||||||
|
"nb_no.json", "pl_pl.json", "pt_br.json", "pt_pt.json", "ru_ru.json",
|
||||||
|
"sv_sv.json", "tr_tr.json", "uk_ua.json", "zh_cn.json"]
|
||||||
file_paths = [os.path.join(current_path, file) for file in json_files]
|
file_paths = [os.path.join(current_path, file) for file in json_files]
|
||||||
merge_translations(file_paths[0], file_paths[1:])
|
merge_translations(file_paths[0], file_paths[1:])
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
@@ -8,12 +8,12 @@ from pytz import timezone
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from const import logPath
|
from const import logPath # noqa: E402, E261 [flake8 lint suppression]
|
||||||
from plugin_helper import Plugin_Objects
|
from plugin_helper import Plugin_Objects # noqa: E402, E261 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402, E261 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402, E261 [flake8 lint suppression]
|
||||||
|
|
||||||
import conf
|
import conf # noqa: E402, E261 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -32,7 +32,6 @@ RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
|||||||
plugin_objects = Plugin_Objects(RESULT_FILE)
|
plugin_objects = Plugin_Objects(RESULT_FILE)
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
mylog('verbose', [f'[{pluginName}] In script'])
|
mylog('verbose', [f'[{pluginName}] In script'])
|
||||||
|
|
||||||
@@ -69,7 +68,7 @@ def main():
|
|||||||
# helpVal2 = "Something1", # If you need to use even only 1, add the remaining ones too
|
# helpVal2 = "Something1", # If you need to use even only 1, add the remaining ones too
|
||||||
# helpVal3 = "Something1", # and set them to 'null'. Check the the docs for details:
|
# helpVal3 = "Something1", # and set them to 'null'. Check the the docs for details:
|
||||||
# helpVal4 = "Something1", # https://github.com/jokob-sk/NetAlertX/blob/main/docs/PLUGINS_DEV.md
|
# helpVal4 = "Something1", # https://github.com/jokob-sk/NetAlertX/blob/main/docs/PLUGINS_DEV.md
|
||||||
)
|
)
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] New entries: "{len(device_data)}"'])
|
mylog('verbose', [f'[{pluginName}] New entries: "{len(device_data)}"'])
|
||||||
|
|
||||||
@@ -78,6 +77,7 @@ def main():
|
|||||||
|
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
|
|
||||||
# retrieve data
|
# retrieve data
|
||||||
def get_device_data(some_setting):
|
def get_device_data(some_setting):
|
||||||
|
|
||||||
@@ -116,5 +116,6 @@ def get_device_data(some_setting):
|
|||||||
# Return the data to be detected by the main application
|
# Return the data to be detected by the main application
|
||||||
return device_data
|
return device_data
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
main()
|
main()
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
# Just a testing library plugin for development purposes
|
# Just a testing library plugin for development purposes
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
@@ -11,10 +11,10 @@ INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
|||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
# NetAlertX modules
|
# NetAlertX modules
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from plugin_helper import Plugin_Objects
|
from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog
|
from logger import mylog # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
pluginName = 'TESTONLY'
|
pluginName = 'TESTONLY'
|
||||||
|
|
||||||
@@ -28,10 +28,7 @@ plugin_objects = Plugin_Objects(RESULT_FILE)
|
|||||||
md5_hash = hashlib.md5()
|
md5_hash = hashlib.md5()
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
# globals
|
# globals
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
# START
|
# START
|
||||||
mylog('verbose', [f'[{pluginName}] In script'])
|
mylog('verbose', [f'[{pluginName}] In script'])
|
||||||
@@ -43,7 +40,6 @@ def main():
|
|||||||
# result = cleanDeviceName(str, True)
|
# result = cleanDeviceName(str, True)
|
||||||
|
|
||||||
regexes = get_setting_value('NEWDEV_NAME_CLEANUP_REGEX')
|
regexes = get_setting_value('NEWDEV_NAME_CLEANUP_REGEX')
|
||||||
|
|
||||||
print(regexes)
|
print(regexes)
|
||||||
subnets = get_setting_value('SCAN_SUBNETS')
|
subnets = get_setting_value('SCAN_SUBNETS')
|
||||||
|
|
||||||
@@ -57,16 +53,12 @@ def main():
|
|||||||
mylog('trace', ["[cleanDeviceName] name after regex : " + str])
|
mylog('trace', ["[cleanDeviceName] name after regex : " + str])
|
||||||
|
|
||||||
mylog('debug', ["[cleanDeviceName] output: " + str])
|
mylog('debug', ["[cleanDeviceName] output: " + str])
|
||||||
|
|
||||||
|
|
||||||
# SPACE FOR TESTING 🔼
|
# SPACE FOR TESTING 🔼
|
||||||
|
|
||||||
# END
|
# END
|
||||||
mylog('verbose', [f'[{pluginName}] result "{str}"'])
|
mylog('verbose', [f'[{pluginName}] result "{str}"'])
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
# -------------INIT---------------------
|
# -------------INIT---------------------
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
sys.exit(main())
|
sys.exit(main())
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import json
|
import json
|
||||||
import subprocess
|
import subprocess
|
||||||
@@ -9,15 +9,15 @@ import sys
|
|||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from const import confFileName, logPath
|
from const import confFileName, logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.datetime_utils import timeNowDB
|
from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
|
||||||
from plugin_helper import Plugin_Objects
|
from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from models.notification_instance import NotificationInstance
|
from models.notification_instance import NotificationInstance # noqa: E402 [flake8 lint suppression]
|
||||||
from database import DB
|
from database import DB # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value("TIMEZONE"))
|
conf.tz = timezone(get_setting_value("TIMEZONE"))
|
||||||
@@ -35,7 +35,7 @@ def main():
|
|||||||
mylog("verbose", [f"[{pluginName}](publisher) In script"])
|
mylog("verbose", [f"[{pluginName}](publisher) In script"])
|
||||||
|
|
||||||
# Check if basic config settings supplied
|
# Check if basic config settings supplied
|
||||||
if check_config() == False:
|
if check_config() is False:
|
||||||
mylog(
|
mylog(
|
||||||
"none",
|
"none",
|
||||||
[
|
[
|
||||||
@@ -80,8 +80,7 @@ def main():
|
|||||||
# -------------------------------------------------------------------------------
|
# -------------------------------------------------------------------------------
|
||||||
def check_config():
|
def check_config():
|
||||||
if get_setting_value("APPRISE_HOST") == "" or (
|
if get_setting_value("APPRISE_HOST") == "" or (
|
||||||
get_setting_value("APPRISE_URL") == ""
|
get_setting_value("APPRISE_URL") == "" and get_setting_value("APPRISE_TAG") == ""
|
||||||
and get_setting_value("APPRISE_TAG") == ""
|
|
||||||
):
|
):
|
||||||
return False
|
return False
|
||||||
else:
|
else:
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
import re
|
import re
|
||||||
@@ -16,15 +16,15 @@ INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
|||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
# NetAlertX modules
|
# NetAlertX modules
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from const import confFileName, logPath
|
from const import confFileName, logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from plugin_helper import Plugin_Objects
|
from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.datetime_utils import timeNowDB
|
from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value, hide_email
|
from helper import get_setting_value, hide_email # noqa: E402 [flake8 lint suppression]
|
||||||
from models.notification_instance import NotificationInstance
|
from models.notification_instance import NotificationInstance # noqa: E402 [flake8 lint suppression]
|
||||||
from database import DB
|
from database import DB # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -38,13 +38,12 @@ LOG_PATH = logPath + '/plugins'
|
|||||||
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}](publisher) In script'])
|
mylog('verbose', [f'[{pluginName}](publisher) In script'])
|
||||||
|
|
||||||
# Check if basic config settings supplied
|
# Check if basic config settings supplied
|
||||||
if check_config() == False:
|
if check_config() is False:
|
||||||
mylog('none', [f'[{pluginName}] ⚠ ERROR: Publisher notification gateway not set up correctly. Check your {confFileName} {pluginName}_* variables.'])
|
mylog('none', [f'[{pluginName}] ⚠ ERROR: Publisher notification gateway not set up correctly. Check your {confFileName} {pluginName}_* variables.'])
|
||||||
return
|
return
|
||||||
|
|
||||||
@@ -72,7 +71,6 @@ def main():
|
|||||||
# mylog('verbose', [f'[{pluginName}] SMTP_REPORT_TO: ', get_setting_value("SMTP_REPORT_TO")])
|
# mylog('verbose', [f'[{pluginName}] SMTP_REPORT_TO: ', get_setting_value("SMTP_REPORT_TO")])
|
||||||
# mylog('verbose', [f'[{pluginName}] SMTP_REPORT_FROM: ', get_setting_value("SMTP_REPORT_FROM")])
|
# mylog('verbose', [f'[{pluginName}] SMTP_REPORT_FROM: ', get_setting_value("SMTP_REPORT_FROM")])
|
||||||
|
|
||||||
|
|
||||||
# Process the new notifications (see the Notifications DB table for structure or check the /php/server/query_json.php?file=table_notifications.json endpoint)
|
# Process the new notifications (see the Notifications DB table for structure or check the /php/server/query_json.php?file=table_notifications.json endpoint)
|
||||||
for notification in new_notifications:
|
for notification in new_notifications:
|
||||||
|
|
||||||
@@ -93,8 +91,9 @@ def main():
|
|||||||
|
|
||||||
plugin_objects.write_result_file()
|
plugin_objects.write_result_file()
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
def check_config ():
|
# -------------------------------------------------------------------------------
|
||||||
|
def check_config():
|
||||||
|
|
||||||
server = get_setting_value('SMTP_SERVER')
|
server = get_setting_value('SMTP_SERVER')
|
||||||
report_to = get_setting_value("SMTP_REPORT_TO")
|
report_to = get_setting_value("SMTP_REPORT_TO")
|
||||||
@@ -106,12 +105,19 @@ def check_config ():
|
|||||||
else:
|
else:
|
||||||
return True
|
return True
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
|
# -------------------------------------------------------------------------------
|
||||||
def send(pHTML, pText):
|
def send(pHTML, pText):
|
||||||
|
|
||||||
mylog('debug', [f'[{pluginName}] SMTP_REPORT_TO: {hide_email(str(get_setting_value("SMTP_REPORT_TO")))} SMTP_USER: {hide_email(str(get_setting_value("SMTP_USER")))}'])
|
mylog('debug', [f'[{pluginName}] SMTP_REPORT_TO: {hide_email(str(get_setting_value("SMTP_REPORT_TO")))} SMTP_USER: {hide_email(str(get_setting_value("SMTP_USER")))}'])
|
||||||
|
|
||||||
subject, from_email, to_email, message_html, message_text = sanitize_email_content(str(get_setting_value("SMTP_SUBJECT")), get_setting_value("SMTP_REPORT_FROM"), get_setting_value("SMTP_REPORT_TO"), pHTML, pText)
|
subject, from_email, to_email, message_html, message_text = sanitize_email_content(
|
||||||
|
str(get_setting_value("SMTP_SUBJECT")),
|
||||||
|
get_setting_value("SMTP_REPORT_FROM"),
|
||||||
|
get_setting_value("SMTP_REPORT_TO"),
|
||||||
|
pHTML,
|
||||||
|
pText
|
||||||
|
)
|
||||||
|
|
||||||
emails = []
|
emails = []
|
||||||
|
|
||||||
@@ -134,8 +140,8 @@ def send(pHTML, pText):
|
|||||||
msg['To'] = mail_addr
|
msg['To'] = mail_addr
|
||||||
msg['Date'] = formatdate(localtime=True)
|
msg['Date'] = formatdate(localtime=True)
|
||||||
|
|
||||||
msg.attach (MIMEText (message_text, 'plain'))
|
msg.attach(MIMEText(message_text, 'plain'))
|
||||||
msg.attach (MIMEText (message_html, 'html'))
|
msg.attach(MIMEText(message_html, 'html'))
|
||||||
|
|
||||||
# Set a timeout for the SMTP connection (in seconds)
|
# Set a timeout for the SMTP connection (in seconds)
|
||||||
smtp_timeout = 30
|
smtp_timeout = 30
|
||||||
@@ -144,12 +150,12 @@ def send(pHTML, pText):
|
|||||||
|
|
||||||
if get_setting_value("LOG_LEVEL") == 'debug':
|
if get_setting_value("LOG_LEVEL") == 'debug':
|
||||||
|
|
||||||
send_email(msg,smtp_timeout)
|
send_email(msg, smtp_timeout)
|
||||||
|
|
||||||
else:
|
else:
|
||||||
|
|
||||||
try:
|
try:
|
||||||
send_email(msg,smtp_timeout)
|
send_email(msg, smtp_timeout)
|
||||||
|
|
||||||
except smtplib.SMTPAuthenticationError as e:
|
except smtplib.SMTPAuthenticationError as e:
|
||||||
mylog('none', [' ERROR: Couldn\'t connect to the SMTP server (SMTPAuthenticationError)'])
|
mylog('none', [' ERROR: Couldn\'t connect to the SMTP server (SMTPAuthenticationError)'])
|
||||||
@@ -166,8 +172,9 @@ def send(pHTML, pText):
|
|||||||
mylog('none', [' ERROR: Are you sure you need SMTP_FORCE_SSL enabled? Check your SMTP provider docs.'])
|
mylog('none', [' ERROR: Are you sure you need SMTP_FORCE_SSL enabled? Check your SMTP provider docs.'])
|
||||||
mylog('none', [' ERROR: ', str(e)])
|
mylog('none', [' ERROR: ', str(e)])
|
||||||
|
|
||||||
|
|
||||||
# ----------------------------------------------------------------------------------
|
# ----------------------------------------------------------------------------------
|
||||||
def send_email(msg,smtp_timeout):
|
def send_email(msg, smtp_timeout):
|
||||||
# Send mail
|
# Send mail
|
||||||
if get_setting_value('SMTP_FORCE_SSL'):
|
if get_setting_value('SMTP_FORCE_SSL'):
|
||||||
mylog('debug', ['SMTP_FORCE_SSL == True so using .SMTP_SSL()'])
|
mylog('debug', ['SMTP_FORCE_SSL == True so using .SMTP_SSL()'])
|
||||||
@@ -182,10 +189,10 @@ def send_email(msg,smtp_timeout):
|
|||||||
mylog('debug', ['SMTP_FORCE_SSL == False so using .SMTP()'])
|
mylog('debug', ['SMTP_FORCE_SSL == False so using .SMTP()'])
|
||||||
if get_setting_value("SMTP_PORT") == 0:
|
if get_setting_value("SMTP_PORT") == 0:
|
||||||
mylog('debug', ['SMTP_PORT == 0 so sending .SMTP(SMTP_SERVER)'])
|
mylog('debug', ['SMTP_PORT == 0 so sending .SMTP(SMTP_SERVER)'])
|
||||||
smtp_connection = smtplib.SMTP (get_setting_value('SMTP_SERVER'))
|
smtp_connection = smtplib.SMTP(get_setting_value('SMTP_SERVER'))
|
||||||
else:
|
else:
|
||||||
mylog('debug', ['SMTP_PORT == 0 so sending .SMTP(SMTP_SERVER, SMTP_PORT)'])
|
mylog('debug', ['SMTP_PORT == 0 so sending .SMTP(SMTP_SERVER, SMTP_PORT)'])
|
||||||
smtp_connection = smtplib.SMTP (get_setting_value('SMTP_SERVER'), get_setting_value('SMTP_PORT'))
|
smtp_connection = smtplib.SMTP(get_setting_value('SMTP_SERVER'), get_setting_value('SMTP_PORT'))
|
||||||
|
|
||||||
mylog('debug', ['Setting SMTP debug level'])
|
mylog('debug', ['Setting SMTP debug level'])
|
||||||
|
|
||||||
@@ -193,7 +200,7 @@ def send_email(msg,smtp_timeout):
|
|||||||
if get_setting_value('LOG_LEVEL') == 'debug':
|
if get_setting_value('LOG_LEVEL') == 'debug':
|
||||||
smtp_connection.set_debuglevel(1)
|
smtp_connection.set_debuglevel(1)
|
||||||
|
|
||||||
mylog('debug', [ 'Sending .ehlo()'])
|
mylog('debug', ['Sending .ehlo()'])
|
||||||
smtp_connection.ehlo()
|
smtp_connection.ehlo()
|
||||||
|
|
||||||
if not get_setting_value('SMTP_SKIP_TLS'):
|
if not get_setting_value('SMTP_SKIP_TLS'):
|
||||||
@@ -203,12 +210,13 @@ def send_email(msg,smtp_timeout):
|
|||||||
smtp_connection.ehlo()
|
smtp_connection.ehlo()
|
||||||
if not get_setting_value('SMTP_SKIP_LOGIN'):
|
if not get_setting_value('SMTP_SKIP_LOGIN'):
|
||||||
mylog('debug', ['SMTP_SKIP_LOGIN == False so sending .login()'])
|
mylog('debug', ['SMTP_SKIP_LOGIN == False so sending .login()'])
|
||||||
smtp_connection.login (get_setting_value('SMTP_USER'), get_setting_value('SMTP_PASS'))
|
smtp_connection.login(get_setting_value('SMTP_USER'), get_setting_value('SMTP_PASS'))
|
||||||
|
|
||||||
mylog('debug', ['Sending .sendmail()'])
|
mylog('debug', ['Sending .sendmail()'])
|
||||||
smtp_connection.sendmail (get_setting_value("SMTP_REPORT_FROM"), get_setting_value("SMTP_REPORT_TO"), msg.as_string())
|
smtp_connection.sendmail(get_setting_value("SMTP_REPORT_FROM"), get_setting_value("SMTP_REPORT_TO"), msg.as_string())
|
||||||
smtp_connection.quit()
|
smtp_connection.quit()
|
||||||
|
|
||||||
|
|
||||||
# ----------------------------------------------------------------------------------
|
# ----------------------------------------------------------------------------------
|
||||||
def sanitize_email_content(subject, from_email, to_email, message_html, message_text):
|
def sanitize_email_content(subject, from_email, to_email, message_html, message_text):
|
||||||
# Validate and sanitize subject
|
# Validate and sanitize subject
|
||||||
@@ -229,6 +237,7 @@ def sanitize_email_content(subject, from_email, to_email, message_html, message_
|
|||||||
|
|
||||||
return subject, from_email, to_email, message_html, message_text
|
return subject, from_email, to_email, message_html, message_text
|
||||||
|
|
||||||
|
|
||||||
# ----------------------------------------------------------------------------------
|
# ----------------------------------------------------------------------------------
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
sys.exit(main())
|
sys.exit(main())
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import json
|
import json
|
||||||
import os
|
import os
|
||||||
@@ -18,15 +18,14 @@ INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
|||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
# NetAlertX modules
|
# NetAlertX modules
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from const import confFileName, logPath
|
from const import confFileName, logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.plugin_utils import getPluginObject
|
from utils.plugin_utils import getPluginObject # noqa: E402 [flake8 lint suppression]
|
||||||
from plugin_helper import Plugin_Objects
|
from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value, bytes_to_string, \
|
from helper import get_setting_value, bytes_to_string, \
|
||||||
sanitize_string, normalize_string
|
sanitize_string, normalize_string # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.datetime_utils import timeNowDB
|
from database import DB, get_device_stats # noqa: E402 [flake8 lint suppression]
|
||||||
from database import DB, get_device_stats
|
|
||||||
|
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
@@ -287,11 +286,11 @@ def publish_mqtt(mqtt_client, topic, message):
|
|||||||
# mylog('verbose', [f"[{pluginName}] mqtt_client.is_connected(): {mqtt_client.is_connected()} "])
|
# mylog('verbose', [f"[{pluginName}] mqtt_client.is_connected(): {mqtt_client.is_connected()} "])
|
||||||
|
|
||||||
result = mqtt_client.publish(
|
result = mqtt_client.publish(
|
||||||
topic=topic,
|
topic=topic,
|
||||||
payload=message,
|
payload=message,
|
||||||
qos=qos,
|
qos=qos,
|
||||||
retain=True,
|
retain=True,
|
||||||
)
|
)
|
||||||
|
|
||||||
status = result[0]
|
status = result[0]
|
||||||
|
|
||||||
@@ -303,6 +302,7 @@ def publish_mqtt(mqtt_client, topic, message):
|
|||||||
time.sleep(0.1)
|
time.sleep(0.1)
|
||||||
return True
|
return True
|
||||||
|
|
||||||
|
|
||||||
# ------------------------------------------------------------------------------
|
# ------------------------------------------------------------------------------
|
||||||
# Create a generic device for overal stats
|
# Create a generic device for overal stats
|
||||||
def create_generic_device(mqtt_client, deviceId, deviceName):
|
def create_generic_device(mqtt_client, deviceId, deviceName):
|
||||||
@@ -434,7 +434,6 @@ def mqtt_start(db):
|
|||||||
if not mqtt_connected_to_broker:
|
if not mqtt_connected_to_broker:
|
||||||
mqtt_client = mqtt_create_client()
|
mqtt_client = mqtt_create_client()
|
||||||
|
|
||||||
|
|
||||||
deviceName = get_setting_value('MQTT_DEVICE_NAME')
|
deviceName = get_setting_value('MQTT_DEVICE_NAME')
|
||||||
deviceId = get_setting_value('MQTT_DEVICE_ID')
|
deviceId = get_setting_value('MQTT_DEVICE_ID')
|
||||||
|
|
||||||
@@ -449,16 +448,18 @@ def mqtt_start(db):
|
|||||||
row = get_device_stats(db)
|
row = get_device_stats(db)
|
||||||
|
|
||||||
# Publish (wrap into {} and remove last ',' from above)
|
# Publish (wrap into {} and remove last ',' from above)
|
||||||
publish_mqtt(mqtt_client, f"{topic_root}/sensor/{deviceId}/state",
|
publish_mqtt(
|
||||||
{
|
mqtt_client,
|
||||||
"online": row[0],
|
f"{topic_root}/sensor/{deviceId}/state",
|
||||||
"down": row[1],
|
{
|
||||||
"all": row[2],
|
"online": row[0],
|
||||||
"archived": row[3],
|
"down": row[1],
|
||||||
"new": row[4],
|
"all": row[2],
|
||||||
"unknown": row[5]
|
"archived": row[3],
|
||||||
}
|
"new": row[4],
|
||||||
)
|
"unknown": row[5]
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
# Generate device-specific MQTT messages if enabled
|
# Generate device-specific MQTT messages if enabled
|
||||||
if get_setting_value('MQTT_SEND_DEVICES'):
|
if get_setting_value('MQTT_SEND_DEVICES'):
|
||||||
@@ -466,11 +467,11 @@ def mqtt_start(db):
|
|||||||
# Specific devices processing
|
# Specific devices processing
|
||||||
|
|
||||||
# Get all devices
|
# Get all devices
|
||||||
devices = db.read(get_setting_value('MQTT_DEVICES_SQL').replace('{s-quote}',"'"))
|
devices = db.read(get_setting_value('MQTT_DEVICES_SQL').replace('{s-quote}', "'"))
|
||||||
|
|
||||||
sec_delay = len(devices) * int(get_setting_value('MQTT_DELAY_SEC'))*5
|
sec_delay = len(devices) * int(get_setting_value('MQTT_DELAY_SEC')) * 5
|
||||||
|
|
||||||
mylog('verbose', [f"[{pluginName}] Estimated delay: ", (sec_delay), 's ', '(', round(sec_delay/60, 1), 'min)'])
|
mylog('verbose', [f"[{pluginName}] Estimated delay: ", (sec_delay), 's ', '(', round(sec_delay / 60, 1), 'min)'])
|
||||||
|
|
||||||
for device in devices:
|
for device in devices:
|
||||||
|
|
||||||
@@ -495,27 +496,29 @@ def mqtt_start(db):
|
|||||||
# handle device_tracker
|
# handle device_tracker
|
||||||
# IMPORTANT: shared payload - device_tracker attributes and individual sensors
|
# IMPORTANT: shared payload - device_tracker attributes and individual sensors
|
||||||
devJson = {
|
devJson = {
|
||||||
"last_ip": device["devLastIP"],
|
"last_ip": device["devLastIP"],
|
||||||
"is_new": str(device["devIsNew"]),
|
"is_new": str(device["devIsNew"]),
|
||||||
"alert_down": str(device["devAlertDown"]),
|
"alert_down": str(device["devAlertDown"]),
|
||||||
"vendor": sanitize_string(device["devVendor"]),
|
"vendor": sanitize_string(device["devVendor"]),
|
||||||
"mac_address": str(device["devMac"]),
|
"mac_address": str(device["devMac"]),
|
||||||
"model": devDisplayName,
|
"model": devDisplayName,
|
||||||
"last_connection": prepTimeStamp(str(device["devLastConnection"])),
|
"last_connection": prepTimeStamp(str(device["devLastConnection"])),
|
||||||
"first_connection": prepTimeStamp(str(device["devFirstConnection"])),
|
"first_connection": prepTimeStamp(str(device["devFirstConnection"])),
|
||||||
"sync_node": device["devSyncHubNode"],
|
"sync_node": device["devSyncHubNode"],
|
||||||
"group": device["devGroup"],
|
"group": device["devGroup"],
|
||||||
"location": device["devLocation"],
|
"location": device["devLocation"],
|
||||||
"network_parent_mac": device["devParentMAC"],
|
"network_parent_mac": device["devParentMAC"],
|
||||||
"network_parent_name": next((dev["devName"] for dev in devices if dev["devMAC"] == device["devParentMAC"]), "")
|
"network_parent_name": next((dev["devName"] for dev in devices if dev["devMAC"] == device["devParentMAC"]), "")
|
||||||
}
|
}
|
||||||
|
|
||||||
# bulk update device sensors in home assistant
|
# bulk update device sensors in home assistant
|
||||||
publish_mqtt(mqtt_client, sensorConfig.state_topic, devJson) # REQUIRED, DON'T DELETE
|
publish_mqtt(mqtt_client, sensorConfig.state_topic, devJson) # REQUIRED, DON'T DELETE
|
||||||
|
|
||||||
# create and update is_present sensor
|
# create and update is_present sensor
|
||||||
sensorConfig = create_sensor(mqtt_client, deviceId, devDisplayName, 'binary_sensor', 'is_present', 'wifi', device["devMac"])
|
sensorConfig = create_sensor(mqtt_client, deviceId, devDisplayName, 'binary_sensor', 'is_present', 'wifi', device["devMac"])
|
||||||
publish_mqtt(mqtt_client, sensorConfig.state_topic,
|
publish_mqtt(
|
||||||
|
mqtt_client,
|
||||||
|
sensorConfig.state_topic,
|
||||||
{
|
{
|
||||||
"is_present": to_binary_sensor(str(device["devPresentLastScan"]))
|
"is_present": to_binary_sensor(str(device["devPresentLastScan"]))
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
|
|
||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import json
|
import json
|
||||||
import os
|
import os
|
||||||
@@ -11,15 +11,15 @@ from base64 import b64encode
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from const import confFileName, logPath
|
from const import confFileName, logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from plugin_helper import Plugin_Objects, handleEmpty
|
from plugin_helper import Plugin_Objects, handleEmpty # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.datetime_utils import timeNowDB
|
from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from models.notification_instance import NotificationInstance
|
from models.notification_instance import NotificationInstance # noqa: E402 [flake8 lint suppression]
|
||||||
from database import DB
|
from database import DB # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -33,13 +33,12 @@ LOG_PATH = logPath + '/plugins'
|
|||||||
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}](publisher) In script'])
|
mylog('verbose', [f'[{pluginName}](publisher) In script'])
|
||||||
|
|
||||||
# Check if basic config settings supplied
|
# Check if basic config settings supplied
|
||||||
if check_config() == False:
|
if check_config() is False:
|
||||||
mylog('none', [f'[{pluginName}] ⚠ ERROR: Publisher notification gateway not set up correctly. Check your {confFileName} {pluginName}_* variables.'])
|
mylog('none', [f'[{pluginName}] ⚠ ERROR: Publisher notification gateway not set up correctly. Check your {confFileName} {pluginName}_* variables.'])
|
||||||
return
|
return
|
||||||
|
|
||||||
@@ -77,15 +76,15 @@ def main():
|
|||||||
plugin_objects.write_result_file()
|
plugin_objects.write_result_file()
|
||||||
|
|
||||||
|
|
||||||
|
# -------------------------------------------------------------------------------
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
def check_config():
|
def check_config():
|
||||||
if get_setting_value('NTFY_HOST') == '' or get_setting_value('NTFY_TOPIC') == '':
|
if get_setting_value('NTFY_HOST') == '' or get_setting_value('NTFY_TOPIC') == '':
|
||||||
return False
|
return False
|
||||||
else:
|
else:
|
||||||
return True
|
return True
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
|
# -------------------------------------------------------------------------------
|
||||||
def send(html, text):
|
def send(html, text):
|
||||||
|
|
||||||
response_text = ''
|
response_text = ''
|
||||||
@@ -100,7 +99,7 @@ def send(html, text):
|
|||||||
# prepare request headers
|
# prepare request headers
|
||||||
headers = {
|
headers = {
|
||||||
"Title": "NetAlertX Notification",
|
"Title": "NetAlertX Notification",
|
||||||
"Actions": "view, Open Dashboard, "+ get_setting_value('REPORT_DASHBOARD_URL'),
|
"Actions": "view, Open Dashboard, " + get_setting_value('REPORT_DASHBOARD_URL'),
|
||||||
"Priority": get_setting_value('NTFY_PRIORITY'),
|
"Priority": get_setting_value('NTFY_PRIORITY'),
|
||||||
"Tags": "warning"
|
"Tags": "warning"
|
||||||
}
|
}
|
||||||
@@ -109,18 +108,20 @@ def send(html, text):
|
|||||||
if token != '':
|
if token != '':
|
||||||
headers["Authorization"] = "Bearer {}".format(token)
|
headers["Authorization"] = "Bearer {}".format(token)
|
||||||
elif user != "" and pwd != "":
|
elif user != "" and pwd != "":
|
||||||
# Generate hash for basic auth
|
# Generate hash for basic auth
|
||||||
basichash = b64encode(bytes(user + ':' + pwd, "utf-8")).decode("ascii")
|
basichash = b64encode(bytes(user + ':' + pwd, "utf-8")).decode("ascii")
|
||||||
# add authorization header with hash
|
# add authorization header with hash
|
||||||
headers["Authorization"] = "Basic {}".format(basichash)
|
headers["Authorization"] = "Basic {}".format(basichash)
|
||||||
|
|
||||||
# call NTFY service
|
# call NTFY service
|
||||||
try:
|
try:
|
||||||
response = requests.post("{}/{}".format( get_setting_value('NTFY_HOST'),
|
response = requests.post("{}/{}".format(
|
||||||
get_setting_value('NTFY_TOPIC')),
|
get_setting_value('NTFY_HOST'),
|
||||||
data = text,
|
get_setting_value('NTFY_TOPIC')),
|
||||||
headers = headers,
|
data = text,
|
||||||
verify = verify_ssl)
|
headers = headers,
|
||||||
|
verify = verify_ssl
|
||||||
|
)
|
||||||
|
|
||||||
response_status_code = response.status_code
|
response_status_code = response.status_code
|
||||||
|
|
||||||
@@ -142,4 +143,3 @@ def send(html, text):
|
|||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
sys.exit(main())
|
sys.exit(main())
|
||||||
|
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python3
|
# !/usr/bin/env python3
|
||||||
import conf
|
import conf
|
||||||
from const import confFileName, logPath
|
from const import confFileName, logPath
|
||||||
from pytz import timezone
|
from pytz import timezone
|
||||||
@@ -12,12 +12,12 @@ import requests
|
|||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects, handleEmpty # noqa: E402
|
from plugin_helper import Plugin_Objects, handleEmpty # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger # noqa: E402
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value, hide_string # noqa: E402
|
from helper import get_setting_value, hide_string # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.datetime_utils import timeNowDB
|
from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
|
||||||
from models.notification_instance import NotificationInstance # noqa: E402
|
from models.notification_instance import NotificationInstance # noqa: E402 [flake8 lint suppression]
|
||||||
from database import DB # noqa: E402
|
from database import DB # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value("TIMEZONE"))
|
conf.tz = timezone(get_setting_value("TIMEZONE"))
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
|
|
||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import json
|
import json
|
||||||
import os
|
import os
|
||||||
@@ -10,15 +10,15 @@ import requests
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from const import confFileName, logPath
|
from const import confFileName, logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from plugin_helper import Plugin_Objects, handleEmpty
|
from plugin_helper import Plugin_Objects, handleEmpty # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value, hide_string
|
from helper import get_setting_value, hide_string # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.datetime_utils import timeNowDB
|
from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
|
||||||
from models.notification_instance import NotificationInstance
|
from models.notification_instance import NotificationInstance # noqa: E402 [flake8 lint suppression]
|
||||||
from database import DB
|
from database import DB # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -32,13 +32,12 @@ LOG_PATH = logPath + '/plugins'
|
|||||||
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}](publisher) In script'])
|
mylog('verbose', [f'[{pluginName}](publisher) In script'])
|
||||||
|
|
||||||
# Check if basic config settings supplied
|
# Check if basic config settings supplied
|
||||||
if check_config() == False:
|
if check_config() is False:
|
||||||
mylog('none', [f'[{pluginName}] ⚠ ERROR: Publisher notification gateway not set up correctly. Check your {confFileName} {pluginName}_* variables.'])
|
mylog('none', [f'[{pluginName}] ⚠ ERROR: Publisher notification gateway not set up correctly. Check your {confFileName} {pluginName}_* variables.'])
|
||||||
return
|
return
|
||||||
|
|
||||||
@@ -76,8 +75,7 @@ def main():
|
|||||||
plugin_objects.write_result_file()
|
plugin_objects.write_result_file()
|
||||||
|
|
||||||
|
|
||||||
|
# -------------------------------------------------------------------------------
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
def send(text):
|
def send(text):
|
||||||
|
|
||||||
response_text = ''
|
response_text = ''
|
||||||
@@ -87,7 +85,6 @@ def send(text):
|
|||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] PUSHSAFER_TOKEN: "{hide_string(token)}"'])
|
mylog('verbose', [f'[{pluginName}] PUSHSAFER_TOKEN: "{hide_string(token)}"'])
|
||||||
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
url = 'https://www.pushsafer.com/api'
|
url = 'https://www.pushsafer.com/api'
|
||||||
post_fields = {
|
post_fields = {
|
||||||
@@ -101,12 +98,10 @@ def send(text):
|
|||||||
"u" : get_setting_value('REPORT_DASHBOARD_URL'),
|
"u" : get_setting_value('REPORT_DASHBOARD_URL'),
|
||||||
"ut" : 'Open NetAlertX',
|
"ut" : 'Open NetAlertX',
|
||||||
"k" : token,
|
"k" : token,
|
||||||
}
|
}
|
||||||
response = requests.post(url, data=post_fields)
|
response = requests.post(url, data=post_fields)
|
||||||
|
|
||||||
response_status_code = response.status_code
|
response_status_code = response.status_code
|
||||||
|
|
||||||
|
|
||||||
# Check if the request was successful (status code 200)
|
# Check if the request was successful (status code 200)
|
||||||
if response_status_code == 200:
|
if response_status_code == 200:
|
||||||
response_text = response.text # This captures the response body/message
|
response_text = response.text # This captures the response body/message
|
||||||
@@ -120,21 +115,17 @@ def send(text):
|
|||||||
|
|
||||||
return response_text, response_status_code
|
return response_text, response_status_code
|
||||||
|
|
||||||
|
|
||||||
return response_text, response_status_code
|
return response_text, response_status_code
|
||||||
|
|
||||||
|
|
||||||
|
# -------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
def check_config():
|
def check_config():
|
||||||
if get_setting_value('PUSHSAFER_TOKEN') == 'ApiKey':
|
if get_setting_value('PUSHSAFER_TOKEN') == 'ApiKey':
|
||||||
return False
|
return False
|
||||||
else:
|
else:
|
||||||
return True
|
return True
|
||||||
|
|
||||||
|
|
||||||
# -------------------------------------------------------
|
# -------------------------------------------------------
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
sys.exit(main())
|
sys.exit(main())
|
||||||
|
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import subprocess
|
import subprocess
|
||||||
import os
|
import os
|
||||||
@@ -8,15 +8,15 @@ import sys
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from const import confFileName, logPath
|
from const import confFileName, logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from plugin_helper import Plugin_Objects
|
from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.datetime_utils import timeNowDB
|
from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from models.notification_instance import NotificationInstance
|
from models.notification_instance import NotificationInstance # noqa: E402 [flake8 lint suppression]
|
||||||
from database import DB
|
from database import DB # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -30,13 +30,11 @@ LOG_PATH = logPath + '/plugins'
|
|||||||
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
mylog('verbose', [f'[{pluginName}](publisher) In script'])
|
mylog('verbose', [f'[{pluginName}](publisher) In script'])
|
||||||
|
|
||||||
# Check if basic config settings supplied
|
# Check if basic config settings supplied
|
||||||
if check_config() == False:
|
if check_config() is False:
|
||||||
mylog('none', [
|
mylog('none', [
|
||||||
f'[{pluginName}] ⚠ ERROR: Publisher notification gateway not set up correctly. Check your {confFileName} {pluginName}_* variables.'])
|
f'[{pluginName}] ⚠ ERROR: Publisher notification gateway not set up correctly. Check your {confFileName} {pluginName}_* variables.'])
|
||||||
return
|
return
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
|
|
||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import json
|
import json
|
||||||
import subprocess
|
import subprocess
|
||||||
@@ -13,15 +13,15 @@ INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
|||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
|
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath, confFileName
|
from const import logPath, confFileName # noqa: E402 [flake8 lint suppression]
|
||||||
from plugin_helper import Plugin_Objects, handleEmpty
|
from plugin_helper import Plugin_Objects, handleEmpty # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.datetime_utils import timeNowDB
|
from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value, write_file
|
from helper import get_setting_value, write_file # noqa: E402 [flake8 lint suppression]
|
||||||
from models.notification_instance import NotificationInstance
|
from models.notification_instance import NotificationInstance # noqa: E402 [flake8 lint suppression]
|
||||||
from database import DB
|
from database import DB # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -35,13 +35,12 @@ LOG_PATH = logPath + '/plugins'
|
|||||||
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}](publisher) In script'])
|
mylog('verbose', [f'[{pluginName}](publisher) In script'])
|
||||||
|
|
||||||
# Check if basic config settings supplied
|
# Check if basic config settings supplied
|
||||||
if check_config() == False:
|
if check_config() is False:
|
||||||
mylog('none', [f'[{pluginName}] ⚠ ERROR: Publisher notification gateway not set up correctly. Check your {confFileName} {pluginName}_* variables.'])
|
mylog('none', [f'[{pluginName}] ⚠ ERROR: Publisher notification gateway not set up correctly. Check your {confFileName} {pluginName}_* variables.'])
|
||||||
return
|
return
|
||||||
|
|
||||||
@@ -62,7 +61,11 @@ def main():
|
|||||||
for notification in new_notifications:
|
for notification in new_notifications:
|
||||||
|
|
||||||
# Send notification
|
# Send notification
|
||||||
response_stdout, response_stderr = send(notification["Text"], notification["HTML"], notification["JSON"])
|
response_stdout, response_stderr = send(
|
||||||
|
notification["Text"],
|
||||||
|
notification["HTML"],
|
||||||
|
notification["JSON"]
|
||||||
|
)
|
||||||
|
|
||||||
# Log result
|
# Log result
|
||||||
plugin_objects.add_object(
|
plugin_objects.add_object(
|
||||||
@@ -79,16 +82,16 @@ def main():
|
|||||||
plugin_objects.write_result_file()
|
plugin_objects.write_result_file()
|
||||||
|
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
# -------------------------------------------------------------------------------
|
||||||
def check_config():
|
def check_config():
|
||||||
if get_setting_value('WEBHOOK_URL') == '':
|
if get_setting_value('WEBHOOK_URL') == '':
|
||||||
return False
|
return False
|
||||||
else:
|
else:
|
||||||
return True
|
return True
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
|
|
||||||
def send (text_data, html_data, json_data):
|
# -------------------------------------------------------------------------------
|
||||||
|
def send(text_data, html_data, json_data):
|
||||||
|
|
||||||
response_stderr = ''
|
response_stderr = ''
|
||||||
response_stdout = ''
|
response_stdout = ''
|
||||||
@@ -139,33 +142,36 @@ def send (text_data, html_data, json_data):
|
|||||||
payloadData = text_data
|
payloadData = text_data
|
||||||
|
|
||||||
# Define slack-compatible payload
|
# Define slack-compatible payload
|
||||||
_json_payload = { "text": payloadData } if payloadType == 'text' else {
|
if payloadType == 'text':
|
||||||
"username": "NetAlertX",
|
_json_payload = {"text": payloadData}
|
||||||
"text": "There are new notifications",
|
else:
|
||||||
"attachments": [{
|
_json_payload = {
|
||||||
"title": "NetAlertX Notifications",
|
"username": "NetAlertX",
|
||||||
"title_link": get_setting_value('REPORT_DASHBOARD_URL'),
|
"text": "There are new notifications",
|
||||||
"text": payloadData
|
"attachments": [{
|
||||||
}]
|
"title": "NetAlertX Notifications",
|
||||||
}
|
"title_link": get_setting_value('REPORT_DASHBOARD_URL'),
|
||||||
|
"text": payloadData
|
||||||
|
}]
|
||||||
|
}
|
||||||
|
|
||||||
# DEBUG - Write the json payload into a log file for debugging
|
# DEBUG - Write the json payload into a log file for debugging
|
||||||
write_file (logPath + '/webhook_payload.json', json.dumps(_json_payload))
|
write_file(logPath + '/webhook_payload.json', json.dumps(_json_payload))
|
||||||
|
|
||||||
# Using the Slack-Compatible Webhook endpoint for Discord so that the same payload can be used for both
|
# Using the Slack-Compatible Webhook endpoint for Discord so that the same payload can be used for both
|
||||||
# Consider: curl has the ability to load in data to POST from a file + piping
|
# Consider: curl has the ability to load in data to POST from a file + piping
|
||||||
if(endpointUrl.startswith('https://discord.com/api/webhooks/') and not endpointUrl.endswith("/slack")):
|
if (endpointUrl.startswith('https://discord.com/api/webhooks/') and not endpointUrl.endswith("/slack")):
|
||||||
_WEBHOOK_URL = f"{endpointUrl}/slack"
|
_WEBHOOK_URL = f"{endpointUrl}/slack"
|
||||||
curlParams = ["curl","-i","-H", "Content-Type:application/json" ,"-d", json.dumps(_json_payload), _WEBHOOK_URL]
|
curlParams = ["curl", "-i", "-H", "Content-Type:application/json", "-d", json.dumps(_json_payload), _WEBHOOK_URL]
|
||||||
else:
|
else:
|
||||||
_WEBHOOK_URL = endpointUrl
|
_WEBHOOK_URL = endpointUrl
|
||||||
curlParams = ["curl","-i","-X", requestMethod , "-H", "Content-Type:application/json", "-d", json.dumps(_json_payload), _WEBHOOK_URL]
|
curlParams = ["curl", "-i", "-X", requestMethod , "-H", "Content-Type:application/json", "-d", json.dumps(_json_payload), _WEBHOOK_URL]
|
||||||
|
|
||||||
# Add HMAC signature if configured
|
# Add HMAC signature if configured
|
||||||
if(secret != ''):
|
if (secret != ''):
|
||||||
h = hmac.new(secret.encode("UTF-8"), json.dumps(_json_payload, separators=(',', ':')).encode(), hashlib.sha256).hexdigest()
|
h = hmac.new(secret.encode("UTF-8"), json.dumps(_json_payload, separators=(',', ':')).encode(), hashlib.sha256).hexdigest()
|
||||||
curlParams.insert(4,"-H")
|
curlParams.insert(4, "-H")
|
||||||
curlParams.insert(5,f"X-Webhook-Signature: sha256={h}")
|
curlParams.insert(5, f"X-Webhook-Signature: sha256={h}")
|
||||||
|
|
||||||
try:
|
try:
|
||||||
# Execute CURL call
|
# Execute CURL call
|
||||||
@@ -179,18 +185,15 @@ def send (text_data, html_data, json_data):
|
|||||||
mylog('debug', [f'[{pluginName}] stdout: ', response_stdout])
|
mylog('debug', [f'[{pluginName}] stdout: ', response_stdout])
|
||||||
mylog('debug', [f'[{pluginName}] stderr: ', response_stderr])
|
mylog('debug', [f'[{pluginName}] stderr: ', response_stderr])
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
except subprocess.CalledProcessError as e:
|
except subprocess.CalledProcessError as e:
|
||||||
# An error occurred, handle it
|
# An error occurred, handle it
|
||||||
mylog('none', [f'[{pluginName}] ⚠ ERROR: ', e.output])
|
mylog('none', [f'[{pluginName}] ⚠ ERROR: ', e.output])
|
||||||
|
|
||||||
response_stderr = e.output
|
response_stderr = e.output
|
||||||
|
|
||||||
|
|
||||||
return response_stdout, response_stderr
|
return response_stdout, response_stderr
|
||||||
|
|
||||||
# -------------------------------------------------------
|
|
||||||
|
# -------------------------------------------------------
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
sys.exit(main())
|
sys.exit(main())
|
||||||
|
|
||||||
|
|||||||
@@ -1,7 +1,6 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
import os
|
import os
|
||||||
import time
|
import time
|
||||||
import pathlib
|
|
||||||
import argparse
|
import argparse
|
||||||
import sys
|
import sys
|
||||||
import re
|
import re
|
||||||
@@ -9,16 +8,16 @@ import base64
|
|||||||
import subprocess
|
import subprocess
|
||||||
|
|
||||||
# Register NetAlertX directories
|
# Register NetAlertX directories
|
||||||
INSTALL_PATH="/app"
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from database import DB
|
from database import DB # noqa: E402 [flake8 lint suppression]
|
||||||
from plugin_helper import Plugin_Objects, handleEmpty
|
from plugin_helper import Plugin_Objects, handleEmpty # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger, append_line_to_file
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath, applicationPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value("TIMEZONE"))
|
conf.tz = timezone(get_setting_value("TIMEZONE"))
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
|
|
||||||
@@ -6,17 +6,16 @@ INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
|||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
pluginName = "ASUSWRT"
|
pluginName = "ASUSWRT"
|
||||||
|
|
||||||
import asyncio
|
import asyncio # noqa: E402 [flake8 lint suppression]
|
||||||
|
import aiohttp # noqa: E402 [flake8 lint suppression]
|
||||||
import aiohttp
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
from asusrouter import AsusData, AsusRouter # noqa: E402 [flake8 lint suppression]
|
||||||
from asusrouter import AsusData, AsusRouter
|
from asusrouter.modules.connection import ConnectionState # noqa: E402 [flake8 lint suppression]
|
||||||
from asusrouter.modules.connection import ConnectionState
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from logger import Logger, mylog # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import Logger, mylog
|
from plugin_helper import (Plugin_Objects, handleEmpty) # noqa: E402 [flake8 lint suppression]
|
||||||
from plugin_helper import (Plugin_Objects, handleEmpty)
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
|
||||||
|
|
||||||
conf.tz = timezone(get_setting_value("TIMEZONE"))
|
conf.tz = timezone(get_setting_value("TIMEZONE"))
|
||||||
|
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python3
|
# !/usr/bin/env python3
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
import socket
|
import socket
|
||||||
@@ -8,14 +8,14 @@ from zeroconf import Zeroconf
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects
|
from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from database import DB
|
from database import DB # noqa: E402 [flake8 lint suppression]
|
||||||
from models.device_instance import DeviceInstance
|
from models.device_instance import DeviceInstance # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Configure timezone and logging
|
# Configure timezone and logging
|
||||||
conf.tz = timezone(get_setting_value("TIMEZONE"))
|
conf.tz = timezone(get_setting_value("TIMEZONE"))
|
||||||
@@ -67,7 +67,7 @@ def resolve_mdns_name(ip: str, timeout: int = 5) -> str:
|
|||||||
hostname = socket.getnameinfo((ip, 0), socket.NI_NAMEREQD)[0]
|
hostname = socket.getnameinfo((ip, 0), socket.NI_NAMEREQD)[0]
|
||||||
zeroconf.close()
|
zeroconf.close()
|
||||||
if hostname and hostname != ip:
|
if hostname and hostname != ip:
|
||||||
mylog("debug", [f"[{pluginName}] Found mDNS name: {hostname}"])
|
mylog("debug", [f"[{pluginName}] Found mDNS name (rev_name): {hostname} ({rev_name})"])
|
||||||
return hostname
|
return hostname
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
mylog("debug", [f"[{pluginName}] Zeroconf lookup failed for {ip}: {e}"])
|
mylog("debug", [f"[{pluginName}] Zeroconf lookup failed for {ip}: {e}"])
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import argparse
|
import argparse
|
||||||
@@ -11,11 +11,11 @@ from datetime import datetime
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath, fullDbPath
|
from const import logPath, fullDbPath # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -29,6 +29,7 @@ LOG_PATH = logPath + '/plugins'
|
|||||||
LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
|
LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
|
||||||
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
|
|
||||||
# the script expects a parameter in the format of devices=device1,device2,...
|
# the script expects a parameter in the format of devices=device1,device2,...
|
||||||
@@ -72,7 +73,7 @@ def main():
|
|||||||
csv_writer = csv.writer(csvfile, delimiter=',', quoting=csv.QUOTE_MINIMAL)
|
csv_writer = csv.writer(csvfile, delimiter=',', quoting=csv.QUOTE_MINIMAL)
|
||||||
|
|
||||||
# Wrap the header values in double quotes and write the header row
|
# Wrap the header values in double quotes and write the header row
|
||||||
csv_writer.writerow([ '"' + col + '"' for col in columns])
|
csv_writer.writerow(['"' + col + '"' for col in columns])
|
||||||
|
|
||||||
# Fetch and write data rows
|
# Fetch and write data rows
|
||||||
for row in cursor.fetchall():
|
for row in cursor.fetchall():
|
||||||
@@ -96,8 +97,8 @@ def main():
|
|||||||
return 0
|
return 0
|
||||||
|
|
||||||
|
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
# BEGIN
|
# BEGIN
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
main()
|
main()
|
||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
@@ -8,11 +8,11 @@ import sqlite3
|
|||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath, fullDbPath
|
from const import logPath, fullDbPath # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value("TIMEZONE"))
|
conf.tz = timezone(get_setting_value("TIMEZONE"))
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import argparse
|
import argparse
|
||||||
@@ -9,11 +9,11 @@ import subprocess
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value, check_IP_format
|
from helper import get_setting_value, check_IP_format # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -28,7 +28,6 @@ LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
|
|||||||
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] In script'])
|
mylog('verbose', [f'[{pluginName}] In script'])
|
||||||
@@ -41,7 +40,6 @@ def main():
|
|||||||
parser.add_argument('DDNS_PASSWORD', action="store", help="Password for Dynamic DNS (DDNS) authentication")
|
parser.add_argument('DDNS_PASSWORD', action="store", help="Password for Dynamic DNS (DDNS) authentication")
|
||||||
parser.add_argument('DDNS_DOMAIN', action="store", help="Dynamic DNS (DDNS) domain name")
|
parser.add_argument('DDNS_DOMAIN', action="store", help="Dynamic DNS (DDNS) domain name")
|
||||||
|
|
||||||
|
|
||||||
values = parser.parse_args()
|
values = parser.parse_args()
|
||||||
|
|
||||||
PREV_IP = values.prev_ip.split('=')[1]
|
PREV_IP = values.prev_ip.split('=')[1]
|
||||||
@@ -51,17 +49,17 @@ def main():
|
|||||||
DDNS_DOMAIN = values.DDNS_DOMAIN.split('=')[1]
|
DDNS_DOMAIN = values.DDNS_DOMAIN.split('=')[1]
|
||||||
|
|
||||||
# perform the new IP lookup and DDNS tasks if enabled
|
# perform the new IP lookup and DDNS tasks if enabled
|
||||||
ddns_update( DDNS_UPDATE_URL, DDNS_USER, DDNS_PASSWORD, DDNS_DOMAIN, PREV_IP)
|
ddns_update(DDNS_UPDATE_URL, DDNS_USER, DDNS_PASSWORD, DDNS_DOMAIN, PREV_IP)
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] Finished '])
|
mylog('verbose', [f'[{pluginName}] Finished '])
|
||||||
|
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
|
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
# INTERNET IP CHANGE
|
# INTERNET IP CHANGE
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
def ddns_update ( DDNS_UPDATE_URL, DDNS_USER, DDNS_PASSWORD, DDNS_DOMAIN, PREV_IP ):
|
def ddns_update(DDNS_UPDATE_URL, DDNS_USER, DDNS_PASSWORD, DDNS_DOMAIN, PREV_IP):
|
||||||
|
|
||||||
# Update DDNS record if enabled and IP is different
|
# Update DDNS record if enabled and IP is different
|
||||||
# Get Dynamic DNS IP
|
# Get Dynamic DNS IP
|
||||||
@@ -78,11 +76,10 @@ def ddns_update ( DDNS_UPDATE_URL, DDNS_USER, DDNS_PASSWORD, DDNS_DOMAIN, PREV_I
|
|||||||
# Check DNS Change
|
# Check DNS Change
|
||||||
if dns_IP != PREV_IP :
|
if dns_IP != PREV_IP :
|
||||||
mylog('none', [f'[{pluginName}] Updating Dynamic DNS IP'])
|
mylog('none', [f'[{pluginName}] Updating Dynamic DNS IP'])
|
||||||
message = set_dynamic_DNS_IP (DDNS_UPDATE_URL, DDNS_USER, DDNS_PASSWORD, DDNS_DOMAIN)
|
message = set_dynamic_DNS_IP(DDNS_UPDATE_URL, DDNS_USER, DDNS_PASSWORD, DDNS_DOMAIN)
|
||||||
mylog('none', [f'[{pluginName}] ', message])
|
mylog('none', [f'[{pluginName}] ', message])
|
||||||
|
|
||||||
# plugin_objects = Plugin_Objects(RESULT_FILE)
|
# plugin_objects = Plugin_Objects(RESULT_FILE)
|
||||||
|
|
||||||
# plugin_objects.add_object(
|
# plugin_objects.add_object(
|
||||||
# primaryId = 'Internet', # MAC (Device Name)
|
# primaryId = 'Internet', # MAC (Device Name)
|
||||||
# secondaryId = new_internet_IP, # IP Address
|
# secondaryId = new_internet_IP, # IP Address
|
||||||
@@ -96,23 +93,23 @@ def ddns_update ( DDNS_UPDATE_URL, DDNS_USER, DDNS_PASSWORD, DDNS_DOMAIN, PREV_I
|
|||||||
# plugin_objects.write_result_file()
|
# plugin_objects.write_result_file()
|
||||||
|
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
# -------------------------------------------------------------------------------
|
||||||
def get_dynamic_DNS_IP (DDNS_DOMAIN):
|
def get_dynamic_DNS_IP(DDNS_DOMAIN):
|
||||||
|
|
||||||
# Using supplied DNS server
|
# Using supplied DNS server
|
||||||
dig_args = ['dig', '+short', DDNS_DOMAIN]
|
dig_args = ['dig', '+short', DDNS_DOMAIN]
|
||||||
|
|
||||||
try:
|
try:
|
||||||
# try runnning a subprocess
|
# try runnning a subprocess
|
||||||
dig_output = subprocess.check_output (dig_args, universal_newlines=True)
|
dig_output = subprocess.check_output(dig_args, universal_newlines=True)
|
||||||
mylog('none', [f'[{pluginName}] DIG output :', dig_output])
|
mylog('none', [f'[{pluginName}] DIG output :', dig_output])
|
||||||
except subprocess.CalledProcessError as e:
|
except subprocess.CalledProcessError as e:
|
||||||
# An error occured, handle it
|
# An error occured, handle it
|
||||||
mylog('none', [f'[{pluginName}] ⚠ ERROR - ', e.output])
|
mylog('none', [f'[{pluginName}] ⚠ ERROR - ', e.output])
|
||||||
dig_output = '' # probably no internet
|
dig_output = '' # probably no internet
|
||||||
|
|
||||||
# Check result is an IP
|
# Check result is an IP
|
||||||
IP = check_IP_format (dig_output)
|
IP = check_IP_format(dig_output)
|
||||||
|
|
||||||
# Handle invalid response
|
# Handle invalid response
|
||||||
if IP == '':
|
if IP == '':
|
||||||
@@ -120,28 +117,27 @@ def get_dynamic_DNS_IP (DDNS_DOMAIN):
|
|||||||
|
|
||||||
return IP
|
return IP
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
def set_dynamic_DNS_IP (DDNS_UPDATE_URL, DDNS_USER, DDNS_PASSWORD, DDNS_DOMAIN):
|
# -------------------------------------------------------------------------------
|
||||||
|
def set_dynamic_DNS_IP(DDNS_UPDATE_URL, DDNS_USER, DDNS_PASSWORD, DDNS_DOMAIN):
|
||||||
try:
|
try:
|
||||||
# try runnning a subprocess
|
# try runnning a subprocess
|
||||||
# Update Dynamic IP
|
# Update Dynamic IP
|
||||||
curl_output = subprocess.check_output (['curl',
|
curl_output = subprocess.check_output([
|
||||||
'-s',
|
'curl',
|
||||||
DDNS_UPDATE_URL +
|
'-s',
|
||||||
'username=' + DDNS_USER +
|
DDNS_UPDATE_URL + 'username=' + DDNS_USER + '&password=' + DDNS_PASSWORD + '&hostname=' + DDNS_DOMAIN],
|
||||||
'&password=' + DDNS_PASSWORD +
|
universal_newlines=True)
|
||||||
'&hostname=' + DDNS_DOMAIN],
|
|
||||||
universal_newlines=True)
|
|
||||||
except subprocess.CalledProcessError as e:
|
except subprocess.CalledProcessError as e:
|
||||||
# An error occured, handle it
|
# An error occured, handle it
|
||||||
mylog('none', [f'[{pluginName}] ⚠ ERROR - ',e.output])
|
mylog('none', [f'[{pluginName}] ⚠ ERROR - ', e.output])
|
||||||
curl_output = ""
|
curl_output = ""
|
||||||
|
|
||||||
return curl_output
|
return curl_output
|
||||||
|
|
||||||
|
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
# BEGIN
|
# BEGIN
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
main()
|
main()
|
||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
from __future__ import unicode_literals
|
from __future__ import unicode_literals
|
||||||
import argparse
|
import argparse
|
||||||
@@ -10,13 +10,13 @@ import chardet
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects, handleEmpty, is_mac
|
from plugin_helper import Plugin_Objects, handleEmpty, is_mac # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from dhcp_leases import DhcpLeases
|
from dhcp_leases import DhcpLeases # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -24,15 +24,13 @@ conf.tz = timezone(get_setting_value('TIMEZONE'))
|
|||||||
# Make sure log level is initialized correctly
|
# Make sure log level is initialized correctly
|
||||||
Logger(get_setting_value('LOG_LEVEL'))
|
Logger(get_setting_value('LOG_LEVEL'))
|
||||||
|
|
||||||
pluginName= 'DHCPLSS'
|
pluginName = 'DHCPLSS'
|
||||||
|
|
||||||
LOG_PATH = logPath + '/plugins'
|
LOG_PATH = logPath + '/plugins'
|
||||||
LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
|
LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
|
||||||
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
# -------------------------------------------------------------
|
# -------------------------------------------------------------
|
||||||
def main():
|
def main():
|
||||||
mylog('verbose', [f'[{pluginName}] In script'])
|
mylog('verbose', [f'[{pluginName}] In script'])
|
||||||
@@ -40,7 +38,12 @@ def main():
|
|||||||
last_run_logfile.write("")
|
last_run_logfile.write("")
|
||||||
|
|
||||||
parser = argparse.ArgumentParser(description='Import devices from dhcp.leases files')
|
parser = argparse.ArgumentParser(description='Import devices from dhcp.leases files')
|
||||||
parser.add_argument('paths', action="store", help="absolute dhcp.leases file paths to check separated by ','")
|
parser.add_argument(
|
||||||
|
'paths',
|
||||||
|
action="store",
|
||||||
|
help="absolute dhcp.leases file paths to check separated by ','"
|
||||||
|
)
|
||||||
|
|
||||||
values = parser.parse_args()
|
values = parser.parse_args()
|
||||||
|
|
||||||
plugin_objects = Plugin_Objects(RESULT_FILE)
|
plugin_objects = Plugin_Objects(RESULT_FILE)
|
||||||
@@ -52,6 +55,7 @@ def main():
|
|||||||
|
|
||||||
plugin_objects.write_result_file()
|
plugin_objects.write_result_file()
|
||||||
|
|
||||||
|
|
||||||
# -------------------------------------------------------------
|
# -------------------------------------------------------------
|
||||||
def get_entries(path, plugin_objects):
|
def get_entries(path, plugin_objects):
|
||||||
|
|
||||||
@@ -122,5 +126,6 @@ def get_entries(path, plugin_objects):
|
|||||||
)
|
)
|
||||||
return plugin_objects
|
return plugin_objects
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
main()
|
main()
|
||||||
|
|||||||
@@ -1,9 +1,8 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
# Based on the work of https://github.com/leiweibau/Pi.Alert
|
# Based on the work of https://github.com/leiweibau/Pi.Alert
|
||||||
|
|
||||||
import subprocess
|
import subprocess
|
||||||
import os
|
import os
|
||||||
from datetime import datetime
|
|
||||||
|
|
||||||
import sys
|
import sys
|
||||||
|
|
||||||
@@ -11,12 +10,12 @@ import sys
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects, Plugin_Object
|
from plugin_helper import Plugin_Objects, Plugin_Object # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
@@ -31,6 +30,7 @@ LOG_PATH = logPath + '/plugins'
|
|||||||
LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
|
LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
|
||||||
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
|
|
||||||
mylog('verbose', ['[DHCPSRVS] In script'])
|
mylog('verbose', ['[DHCPSRVS] In script'])
|
||||||
@@ -101,5 +101,6 @@ def main():
|
|||||||
except Exception as e:
|
except Exception as e:
|
||||||
mylog('verbose', ['[DHCPSRVS] Error in main:', str(e)])
|
mylog('verbose', ['[DHCPSRVS] Error in main:', str(e)])
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
main()
|
main()
|
||||||
|
|||||||
@@ -1,5 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
import subprocess
|
import subprocess
|
||||||
@@ -8,14 +7,14 @@ import subprocess
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects
|
from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from database import DB
|
from database import DB # noqa: E402 [flake8 lint suppression]
|
||||||
from models.device_instance import DeviceInstance
|
from models.device_instance import DeviceInstance # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -65,27 +64,27 @@ def main():
|
|||||||
|
|
||||||
if domain_name != '':
|
if domain_name != '':
|
||||||
plugin_objects.add_object(
|
plugin_objects.add_object(
|
||||||
# "MAC", "IP", "Server", "Name"
|
primaryId = device['devMac'],
|
||||||
primaryId = device['devMac'],
|
secondaryId = device['devLastIP'],
|
||||||
secondaryId = device['devLastIP'],
|
watched1 = dns_server,
|
||||||
watched1 = dns_server,
|
watched2 = domain_name,
|
||||||
watched2 = domain_name,
|
watched3 = '',
|
||||||
watched3 = '',
|
watched4 = '',
|
||||||
watched4 = '',
|
extra = '',
|
||||||
extra = '',
|
foreignKey = device['devMac']
|
||||||
foreignKey = device['devMac'])
|
)
|
||||||
|
|
||||||
plugin_objects.write_result_file()
|
plugin_objects.write_result_file()
|
||||||
|
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] Script finished'])
|
mylog('verbose', [f'[{pluginName}] Script finished'])
|
||||||
|
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
#===============================================================================
|
|
||||||
|
# ===============================================================================
|
||||||
# Execute scan
|
# Execute scan
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
def execute_name_lookup (ip, timeout):
|
def execute_name_lookup(ip, timeout):
|
||||||
"""
|
"""
|
||||||
Execute the DIG command on IP.
|
Execute the DIG command on IP.
|
||||||
"""
|
"""
|
||||||
@@ -99,7 +98,13 @@ def execute_name_lookup (ip, timeout):
|
|||||||
mylog('verbose', [f'[{pluginName}] DEBUG CMD :', args])
|
mylog('verbose', [f'[{pluginName}] DEBUG CMD :', args])
|
||||||
|
|
||||||
# try runnning a subprocess with a forced (timeout) in case the subprocess hangs
|
# try runnning a subprocess with a forced (timeout) in case the subprocess hangs
|
||||||
output = subprocess.check_output (args, universal_newlines=True, stderr=subprocess.STDOUT, timeout=(timeout), text=True).strip()
|
output = subprocess.check_output(
|
||||||
|
args,
|
||||||
|
universal_newlines=True,
|
||||||
|
stderr=subprocess.STDOUT,
|
||||||
|
timeout=(timeout),
|
||||||
|
text=True
|
||||||
|
).strip()
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] DEBUG OUTPUT : {output}'])
|
mylog('verbose', [f'[{pluginName}] DEBUG OUTPUT : {output}'])
|
||||||
|
|
||||||
@@ -116,13 +121,13 @@ def execute_name_lookup (ip, timeout):
|
|||||||
except subprocess.TimeoutExpired:
|
except subprocess.TimeoutExpired:
|
||||||
mylog('verbose', [f'[{pluginName}] TIMEOUT - the process forcefully terminated as timeout reached'])
|
mylog('verbose', [f'[{pluginName}] TIMEOUT - the process forcefully terminated as timeout reached'])
|
||||||
|
|
||||||
if output == "": # check if the subprocess failed
|
if output == "": # check if the subprocess failed
|
||||||
mylog('verbose', [f'[{pluginName}] Scan: FAIL - check logs'])
|
mylog('verbose', [f'[{pluginName}] Scan: FAIL - check logs'])
|
||||||
else:
|
else:
|
||||||
mylog('verbose', [f'[{pluginName}] Scan: SUCCESS'])
|
mylog('verbose', [f'[{pluginName}] Scan: SUCCESS'])
|
||||||
|
|
||||||
return '', ''
|
return '', ''
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
main()
|
main()
|
||||||
|
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
@@ -17,11 +17,11 @@ from aiofreepybox.exceptions import NotOpenError, AuthorizationError
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects
|
from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value("TIMEZONE"))
|
conf.tz = timezone(get_setting_value("TIMEZONE"))
|
||||||
@@ -79,6 +79,7 @@ def map_device_type(type: str):
|
|||||||
mylog("minimal", [f"[{pluginName}] Unknown device type: {type}"])
|
mylog("minimal", [f"[{pluginName}] Unknown device type: {type}"])
|
||||||
return device_type_map["other"]
|
return device_type_map["other"]
|
||||||
|
|
||||||
|
|
||||||
async def get_device_data(api_version: int, api_address: str, api_port: int):
|
async def get_device_data(api_version: int, api_address: str, api_port: int):
|
||||||
# ensure existence of db path
|
# ensure existence of db path
|
||||||
config_base = Path(os.getenv("NETALERTX_CONFIG", "/data/config"))
|
config_base = Path(os.getenv("NETALERTX_CONFIG", "/data/config"))
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
# test script by running:
|
# test script by running:
|
||||||
# tbc
|
# tbc
|
||||||
|
|
||||||
@@ -11,14 +11,14 @@ import re
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects
|
from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from database import DB
|
from database import DB # noqa: E402 [flake8 lint suppression]
|
||||||
from models.device_instance import DeviceInstance
|
from models.device_instance import DeviceInstance # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -33,12 +33,10 @@ LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
|
|||||||
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] In script'])
|
mylog('verbose', [f'[{pluginName}] In script'])
|
||||||
|
|
||||||
|
|
||||||
timeout = get_setting_value('ICMP_RUN_TIMEOUT')
|
timeout = get_setting_value('ICMP_RUN_TIMEOUT')
|
||||||
args = get_setting_value('ICMP_ARGS')
|
args = get_setting_value('ICMP_ARGS')
|
||||||
in_regex = get_setting_value('ICMP_IN_REGEX')
|
in_regex = get_setting_value('ICMP_IN_REGEX')
|
||||||
@@ -65,7 +63,6 @@ def main():
|
|||||||
if regex_pattern.match(device['devLastIP'])
|
if regex_pattern.match(device['devLastIP'])
|
||||||
]
|
]
|
||||||
|
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] Devices to PING: {len(filtered_devices)}'])
|
mylog('verbose', [f'[{pluginName}] Devices to PING: {len(filtered_devices)}'])
|
||||||
|
|
||||||
for device in filtered_devices:
|
for device in filtered_devices:
|
||||||
@@ -73,30 +70,30 @@ def main():
|
|||||||
|
|
||||||
mylog('verbose', [f"[{pluginName}] ip: {device['devLastIP']} is_online: {is_online}"])
|
mylog('verbose', [f"[{pluginName}] ip: {device['devLastIP']} is_online: {is_online}"])
|
||||||
|
|
||||||
|
|
||||||
if is_online:
|
if is_online:
|
||||||
plugin_objects.add_object(
|
plugin_objects.add_object(
|
||||||
# "MAC", "IP", "Name", "Output"
|
# "MAC", "IP", "Name", "Output"
|
||||||
primaryId = device['devMac'],
|
primaryId = device['devMac'],
|
||||||
secondaryId = device['devLastIP'],
|
secondaryId = device['devLastIP'],
|
||||||
watched1 = device['devName'],
|
watched1 = device['devName'],
|
||||||
watched2 = output.replace('\n',''),
|
watched2 = output.replace('\n', ''),
|
||||||
watched3 = '',
|
watched3 = '',
|
||||||
watched4 = '',
|
watched4 = '',
|
||||||
extra = '',
|
extra = '',
|
||||||
foreignKey = device['devMac'])
|
foreignKey = device['devMac']
|
||||||
|
)
|
||||||
|
|
||||||
plugin_objects.write_result_file()
|
plugin_objects.write_result_file()
|
||||||
|
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] Script finished'])
|
mylog('verbose', [f'[{pluginName}] Script finished'])
|
||||||
|
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
#===============================================================================
|
|
||||||
|
# ===============================================================================
|
||||||
# Execute scan
|
# Execute scan
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
def execute_scan (ip, timeout, args):
|
def execute_scan(ip, timeout, args):
|
||||||
"""
|
"""
|
||||||
Execute the ICMP command on IP.
|
Execute the ICMP command on IP.
|
||||||
"""
|
"""
|
||||||
@@ -108,12 +105,18 @@ def execute_scan (ip, timeout, args):
|
|||||||
|
|
||||||
try:
|
try:
|
||||||
# try runnning a subprocess with a forced (timeout) in case the subprocess hangs
|
# try runnning a subprocess with a forced (timeout) in case the subprocess hangs
|
||||||
output = subprocess.check_output (icmp_args, universal_newlines=True, stderr=subprocess.STDOUT, timeout=(timeout), text=True)
|
output = subprocess.check_output(
|
||||||
|
icmp_args,
|
||||||
|
universal_newlines=True,
|
||||||
|
stderr=subprocess.STDOUT,
|
||||||
|
timeout=(timeout),
|
||||||
|
text=True
|
||||||
|
)
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] DEBUG OUTPUT : {output}'])
|
mylog('verbose', [f'[{pluginName}] DEBUG OUTPUT : {output}'])
|
||||||
|
|
||||||
# Parse output using case-insensitive regular expressions
|
# Parse output using case-insensitive regular expressions
|
||||||
#Synology-NAS:/# ping -i 0.5 -c 3 -W 8 -w 9 192.168.1.82
|
# Synology-NAS:/# ping -i 0.5 -c 3 -W 8 -w 9 192.168.1.82
|
||||||
# PING 192.168.1.82 (192.168.1.82): 56 data bytes
|
# PING 192.168.1.82 (192.168.1.82): 56 data bytes
|
||||||
# 64 bytes from 192.168.1.82: seq=0 ttl=64 time=0.080 ms
|
# 64 bytes from 192.168.1.82: seq=0 ttl=64 time=0.080 ms
|
||||||
# 64 bytes from 192.168.1.82: seq=1 ttl=64 time=0.081 ms
|
# 64 bytes from 192.168.1.82: seq=1 ttl=64 time=0.081 ms
|
||||||
@@ -157,10 +160,8 @@ def execute_scan (ip, timeout, args):
|
|||||||
return False, output
|
return False, output
|
||||||
|
|
||||||
|
|
||||||
|
# ===============================================================================
|
||||||
|
|
||||||
#===============================================================================
|
|
||||||
# BEGIN
|
# BEGIN
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
main()
|
main()
|
||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import time
|
import time
|
||||||
@@ -11,13 +11,13 @@ import re
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects
|
from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.datetime_utils import timeNowDB
|
from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger, append_line_to_file
|
from logger import mylog, Logger, append_line_to_file # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import check_IP_format, get_setting_value
|
from helper import check_IP_format, get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -31,9 +31,9 @@ LOG_PATH = logPath + '/plugins'
|
|||||||
LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
|
LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
|
||||||
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
||||||
|
|
||||||
|
|
||||||
no_internet_ip = '0.0.0.0'
|
no_internet_ip = '0.0.0.0'
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] In script'])
|
mylog('verbose', [f'[{pluginName}] In script'])
|
||||||
@@ -41,7 +41,7 @@ def main():
|
|||||||
parser = argparse.ArgumentParser(description='Check internet connectivity and IP')
|
parser = argparse.ArgumentParser(description='Check internet connectivity and IP')
|
||||||
|
|
||||||
parser.add_argument('prev_ip', action="store", help="Previous IP address to compare against the current IP")
|
parser.add_argument('prev_ip', action="store", help="Previous IP address to compare against the current IP")
|
||||||
parser.add_argument('DIG_GET_IP_ARG', action="store", help="Arguments for the 'dig' command to retrieve the IP address") # unused
|
parser.add_argument('DIG_GET_IP_ARG', action="store", help="Arguments for the 'dig' command to retrieve the IP address") # unused
|
||||||
|
|
||||||
values = parser.parse_args()
|
values = parser.parse_args()
|
||||||
|
|
||||||
@@ -60,10 +60,10 @@ def main():
|
|||||||
|
|
||||||
for i in range(INTRNT_RETRIES + 1):
|
for i in range(INTRNT_RETRIES + 1):
|
||||||
|
|
||||||
new_internet_IP, cmd_output = check_internet_IP( PREV_IP, DIG_GET_IP_ARG)
|
new_internet_IP, cmd_output = check_internet_IP(PREV_IP, DIG_GET_IP_ARG)
|
||||||
|
|
||||||
if new_internet_IP == no_internet_ip:
|
if new_internet_IP == no_internet_ip:
|
||||||
time.sleep(1*i) # Exponential backoff strategy
|
time.sleep(1 * i) # Exponential backoff strategy
|
||||||
else:
|
else:
|
||||||
retries_needed = i
|
retries_needed = i
|
||||||
break
|
break
|
||||||
@@ -74,7 +74,7 @@ def main():
|
|||||||
mylog('verbose', [f'[{pluginName}] Curl Fallback (new_internet_IP|cmd_output): {new_internet_IP} | {cmd_output}'])
|
mylog('verbose', [f'[{pluginName}] Curl Fallback (new_internet_IP|cmd_output): {new_internet_IP} | {cmd_output}'])
|
||||||
|
|
||||||
# logging
|
# logging
|
||||||
append_line_to_file (logPath + '/IP_changes.log', '['+str(timeNowDB()) +']\t'+ new_internet_IP +'\n')
|
append_line_to_file(logPath + '/IP_changes.log', '[' + str(timeNowDB()) + ']\t' + new_internet_IP + '\n')
|
||||||
|
|
||||||
plugin_objects = Plugin_Objects(RESULT_FILE)
|
plugin_objects = Plugin_Objects(RESULT_FILE)
|
||||||
|
|
||||||
@@ -82,11 +82,12 @@ def main():
|
|||||||
primaryId = 'Internet', # MAC (Device Name)
|
primaryId = 'Internet', # MAC (Device Name)
|
||||||
secondaryId = new_internet_IP, # IP Address
|
secondaryId = new_internet_IP, # IP Address
|
||||||
watched1 = f'Previous IP: {PREV_IP}',
|
watched1 = f'Previous IP: {PREV_IP}',
|
||||||
watched2 = cmd_output.replace('\n',''),
|
watched2 = cmd_output.replace('\n', ''),
|
||||||
watched3 = retries_needed,
|
watched3 = retries_needed,
|
||||||
watched4 = 'Gateway',
|
watched4 = 'Gateway',
|
||||||
extra = f'Previous IP: {PREV_IP}',
|
extra = f'Previous IP: {PREV_IP}',
|
||||||
foreignKey = 'Internet')
|
foreignKey = 'Internet'
|
||||||
|
)
|
||||||
|
|
||||||
plugin_objects.write_result_file()
|
plugin_objects.write_result_file()
|
||||||
|
|
||||||
@@ -95,10 +96,10 @@ def main():
|
|||||||
return 0
|
return 0
|
||||||
|
|
||||||
|
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
# INTERNET IP CHANGE
|
# INTERNET IP CHANGE
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
def check_internet_IP ( PREV_IP, DIG_GET_IP_ARG ):
|
def check_internet_IP(PREV_IP, DIG_GET_IP_ARG):
|
||||||
|
|
||||||
# Get Internet IP
|
# Get Internet IP
|
||||||
mylog('verbose', [f'[{pluginName}] - Retrieving Internet IP'])
|
mylog('verbose', [f'[{pluginName}] - Retrieving Internet IP'])
|
||||||
@@ -109,7 +110,7 @@ def check_internet_IP ( PREV_IP, DIG_GET_IP_ARG ):
|
|||||||
# Check previously stored IP
|
# Check previously stored IP
|
||||||
previous_IP = no_internet_ip
|
previous_IP = no_internet_ip
|
||||||
|
|
||||||
if PREV_IP is not None and len(PREV_IP) > 0 :
|
if PREV_IP is not None and len(PREV_IP) > 0 :
|
||||||
previous_IP = PREV_IP
|
previous_IP = PREV_IP
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] previous_IP : {previous_IP}'])
|
mylog('verbose', [f'[{pluginName}] previous_IP : {previous_IP}'])
|
||||||
@@ -117,22 +118,22 @@ def check_internet_IP ( PREV_IP, DIG_GET_IP_ARG ):
|
|||||||
return internet_IP, cmd_output
|
return internet_IP, cmd_output
|
||||||
|
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
# -------------------------------------------------------------------------------
|
||||||
def get_internet_IP (DIG_GET_IP_ARG):
|
def get_internet_IP(DIG_GET_IP_ARG):
|
||||||
|
|
||||||
cmd_output = ''
|
cmd_output = ''
|
||||||
|
|
||||||
# Using 'dig'
|
# Using 'dig'
|
||||||
dig_args = ['dig', '+short'] + DIG_GET_IP_ARG.strip().split()
|
dig_args = ['dig', '+short'] + DIG_GET_IP_ARG.strip().split()
|
||||||
try:
|
try:
|
||||||
cmd_output = subprocess.check_output (dig_args, universal_newlines=True)
|
cmd_output = subprocess.check_output(dig_args, universal_newlines=True)
|
||||||
mylog('verbose', [f'[{pluginName}] DIG result : {cmd_output}'])
|
mylog('verbose', [f'[{pluginName}] DIG result : {cmd_output}'])
|
||||||
except subprocess.CalledProcessError as e:
|
except subprocess.CalledProcessError as e:
|
||||||
mylog('verbose', [e.output])
|
mylog('verbose', [e.output])
|
||||||
cmd_output = '' # no internet
|
cmd_output = '' # no internet
|
||||||
|
|
||||||
# Check result is an IP
|
# Check result is an IP
|
||||||
IP = check_IP_format (cmd_output)
|
IP = check_IP_format(cmd_output)
|
||||||
|
|
||||||
# Handle invalid response
|
# Handle invalid response
|
||||||
if IP == '':
|
if IP == '':
|
||||||
@@ -140,7 +141,8 @@ def get_internet_IP (DIG_GET_IP_ARG):
|
|||||||
|
|
||||||
return IP, cmd_output
|
return IP, cmd_output
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
|
# -------------------------------------------------------------------------------
|
||||||
def fallback_check_ip():
|
def fallback_check_ip():
|
||||||
"""Fallback mechanism using `curl ifconfig.me/ip`."""
|
"""Fallback mechanism using `curl ifconfig.me/ip`."""
|
||||||
try:
|
try:
|
||||||
@@ -155,8 +157,9 @@ def fallback_check_ip():
|
|||||||
mylog('none', [f'[{pluginName}] Fallback curl exception: {e}'])
|
mylog('none', [f'[{pluginName}] Fallback curl exception: {e}'])
|
||||||
return no_internet_ip, f'Fallback via curl exception: "{e}"'
|
return no_internet_ip, f'Fallback via curl exception: "{e}"'
|
||||||
|
|
||||||
#===============================================================================
|
|
||||||
|
# ===============================================================================
|
||||||
# BEGIN
|
# BEGIN
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
main()
|
main()
|
||||||
@@ -1,6 +1,5 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import argparse
|
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
import speedtest
|
import speedtest
|
||||||
@@ -9,13 +8,13 @@ import speedtest
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects
|
from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.datetime_utils import timeNowDB
|
from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -28,13 +27,11 @@ pluginName = 'INTRSPD'
|
|||||||
LOG_PATH = logPath + '/plugins'
|
LOG_PATH = logPath + '/plugins'
|
||||||
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
|
|
||||||
mylog('verbose', ['[INTRSPD] In script'])
|
mylog('verbose', ['[INTRSPD] In script'])
|
||||||
|
|
||||||
parser = argparse.ArgumentParser(description='Speedtest Plugin for NetAlertX')
|
|
||||||
values = parser.parse_args()
|
|
||||||
|
|
||||||
plugin_objects = Plugin_Objects(RESULT_FILE)
|
plugin_objects = Plugin_Objects(RESULT_FILE)
|
||||||
speedtest_result = run_speedtest()
|
speedtest_result = run_speedtest()
|
||||||
plugin_objects.add_object(
|
plugin_objects.add_object(
|
||||||
@@ -49,6 +46,7 @@ def main():
|
|||||||
)
|
)
|
||||||
plugin_objects.write_result_file()
|
plugin_objects.write_result_file()
|
||||||
|
|
||||||
|
|
||||||
def run_speedtest():
|
def run_speedtest():
|
||||||
try:
|
try:
|
||||||
st = speedtest.Speedtest(secure=True)
|
st = speedtest.Speedtest(secure=True)
|
||||||
@@ -69,5 +67,6 @@ def run_speedtest():
|
|||||||
'upload_speed': -1,
|
'upload_speed': -1,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
sys.exit(main())
|
sys.exit(main())
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
@@ -11,11 +11,11 @@ from functools import reduce
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects
|
from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -34,7 +34,6 @@ RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
|||||||
plugin_objects = Plugin_Objects(RESULT_FILE)
|
plugin_objects = Plugin_Objects(RESULT_FILE)
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
mylog('verbose', [f'[{pluginName}] In script'])
|
mylog('verbose', [f'[{pluginName}] In script'])
|
||||||
|
|
||||||
@@ -59,22 +58,22 @@ def main():
|
|||||||
if len(neighbors) > 0:
|
if len(neighbors) > 0:
|
||||||
|
|
||||||
for device in neighbors:
|
for device in neighbors:
|
||||||
plugin_objects.add_object(
|
plugin_objects.add_object(
|
||||||
primaryId = device['mac'],
|
primaryId = device['mac'],
|
||||||
secondaryId = device['ip'],
|
secondaryId = device['ip'],
|
||||||
watched4 = device['last_seen'],
|
watched4 = device['last_seen'],
|
||||||
|
|
||||||
# The following are always unknown
|
# The following are always unknown
|
||||||
watched1 = device['hostname'], # don't use these --> handleEmpty(device['hostname']),
|
watched1 = device['hostname'], # don't use these --> handleEmpty(device['hostname']),
|
||||||
watched2 = device['vendor'], # handleEmpty(device['vendor']),
|
watched2 = device['vendor'], # don't use these --> handleEmpty(device['vendor']),
|
||||||
watched3 = device['device_type'], # handleEmpty(device['device_type']),
|
watched3 = device['device_type'], # don't use these --> handleEmpty(device['device_type']),
|
||||||
extra = '',
|
extra = '',
|
||||||
foreignKey = "" #device['mac']
|
foreignKey = "" # device['mac']
|
||||||
# helpVal1 = "Something1", # Optional Helper values to be passed for mapping into the app
|
# helpVal1 = "Something1", # Optional Helper values to be passed for mapping into the app
|
||||||
# helpVal2 = "Something1", # If you need to use even only 1, add the remaining ones too
|
# helpVal2 = "Something1", # If you need to use even only 1, add the remaining ones too
|
||||||
# helpVal3 = "Something1", # and set them to 'null'. Check the the docs for details:
|
# helpVal3 = "Something1", # and set them to 'null'. Check the the docs for details:
|
||||||
# helpVal4 = "Something1", # https://github.com/jokob-sk/NetAlertX/blob/main/docs/PLUGINS_DEV.md
|
# helpVal4 = "Something1", # https://github.com/jokob-sk/NetAlertX/blob/main/docs/PLUGINS_DEV.md
|
||||||
)
|
)
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] New entries: "{len(neighbors)}"'])
|
mylog('verbose', [f'[{pluginName}] New entries: "{len(neighbors)}"'])
|
||||||
|
|
||||||
@@ -83,6 +82,7 @@ def main():
|
|||||||
|
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
|
|
||||||
def parse_neighbors(raw_neighbors: list[str]):
|
def parse_neighbors(raw_neighbors: list[str]):
|
||||||
neighbors = []
|
neighbors = []
|
||||||
for line in raw_neighbors:
|
for line in raw_neighbors:
|
||||||
@@ -111,6 +111,7 @@ def is_multicast(ip):
|
|||||||
prefixes = ['ff', '224', '231', '232', '233', '234', '238', '239']
|
prefixes = ['ff', '224', '231', '232', '233', '234', '238', '239']
|
||||||
return reduce(lambda acc, prefix: acc or ip.startswith(prefix), prefixes, False)
|
return reduce(lambda acc, prefix: acc or ip.startswith(prefix), prefixes, False)
|
||||||
|
|
||||||
|
|
||||||
# retrieve data
|
# retrieve data
|
||||||
def get_neighbors(interfaces):
|
def get_neighbors(interfaces):
|
||||||
|
|
||||||
@@ -136,11 +137,11 @@ def get_neighbors(interfaces):
|
|||||||
mylog('verbose', [f'[{pluginName}] Scanning interface succeded: "{interface}"'])
|
mylog('verbose', [f'[{pluginName}] Scanning interface succeded: "{interface}"'])
|
||||||
except subprocess.CalledProcessError as e:
|
except subprocess.CalledProcessError as e:
|
||||||
# An error occurred, handle it
|
# An error occurred, handle it
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] Scanning interface failed: "{interface}"'])
|
|
||||||
error_type = type(e).__name__ # Capture the error type
|
error_type = type(e).__name__ # Capture the error type
|
||||||
|
mylog('verbose', [f'[{pluginName}] Scanning interface failed: "{interface}" ({error_type})'])
|
||||||
|
|
||||||
return results
|
return results
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
main()
|
main()
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
|
|
||||||
@@ -7,18 +7,18 @@ INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
|||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
pluginName = 'LUCIRPC'
|
pluginName = 'LUCIRPC'
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects
|
from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
try:
|
try:
|
||||||
from openwrt_luci_rpc import OpenWrtRpc
|
from openwrt_luci_rpc import OpenWrtRpc
|
||||||
except:
|
except ImportError as e:
|
||||||
mylog('error', [f'[{pluginName}] Failed import openwrt_luci_rpc'])
|
mylog('error', [f'[{pluginName}] Failed import openwrt_luci_rpc: {e}'])
|
||||||
exit()
|
exit(1)
|
||||||
|
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
|
|
||||||
@@ -30,6 +30,7 @@ RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
|||||||
|
|
||||||
plugin_objects = Plugin_Objects(RESULT_FILE)
|
plugin_objects = Plugin_Objects(RESULT_FILE)
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
mylog('verbose', [f'[{pluginName}] start script.'])
|
mylog('verbose', [f'[{pluginName}] start script.'])
|
||||||
|
|
||||||
@@ -59,6 +60,7 @@ def main():
|
|||||||
|
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
|
|
||||||
def get_device_data():
|
def get_device_data():
|
||||||
router = OpenWrtRpc(
|
router = OpenWrtRpc(
|
||||||
get_setting_value("LUCIRPC_host"),
|
get_setting_value("LUCIRPC_host"),
|
||||||
@@ -66,7 +68,7 @@ def get_device_data():
|
|||||||
get_setting_value("LUCIRPC_password"),
|
get_setting_value("LUCIRPC_password"),
|
||||||
get_setting_value("LUCIRPC_ssl"),
|
get_setting_value("LUCIRPC_ssl"),
|
||||||
get_setting_value("LUCIRPC_verify_ssl")
|
get_setting_value("LUCIRPC_verify_ssl")
|
||||||
)
|
)
|
||||||
|
|
||||||
if router.is_logged_in():
|
if router.is_logged_in():
|
||||||
mylog('verbose', [f'[{pluginName}] login successfully.'])
|
mylog('verbose', [f'[{pluginName}] login successfully.'])
|
||||||
@@ -76,5 +78,6 @@ def get_device_data():
|
|||||||
device_data = router.get_all_connected_devices(only_reachable=get_setting_value("LUCIRPC_only_reachable"))
|
device_data = router.get_all_connected_devices(only_reachable=get_setting_value("LUCIRPC_only_reachable"))
|
||||||
return device_data
|
return device_data
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
main()
|
main()
|
||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
@@ -8,12 +8,12 @@ from collections import deque
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from messaging.in_app import remove_old
|
from messaging.in_app import remove_old # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -28,7 +28,6 @@ LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
|
|||||||
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] In script'])
|
mylog('verbose', [f'[{pluginName}] In script'])
|
||||||
@@ -65,8 +64,8 @@ def main():
|
|||||||
return 0
|
return 0
|
||||||
|
|
||||||
|
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
# BEGIN
|
# BEGIN
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
main()
|
main()
|
||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
@@ -7,14 +7,14 @@ import sys
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects
|
from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
from librouteros import connect
|
from librouteros import connect # noqa: E402 [flake8 lint suppression]
|
||||||
from librouteros.exceptions import TrapError
|
from librouteros.exceptions import TrapError # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -29,7 +29,6 @@ LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
|
|||||||
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] In script'])
|
mylog('verbose', [f'[{pluginName}] In script'])
|
||||||
@@ -72,7 +71,10 @@ def get_entries(plugin_objects: Plugin_Objects) -> Plugin_Objects:
|
|||||||
status = lease.get('status')
|
status = lease.get('status')
|
||||||
device_name = comment or host_name or "(unknown)"
|
device_name = comment or host_name or "(unknown)"
|
||||||
|
|
||||||
mylog('verbose', [f"ID: {lease_id}, Address: {address}, MAC Address: {mac_address}, Host Name: {host_name}, Comment: {comment}, Last Seen: {last_seen}, Status: {status}"])
|
mylog(
|
||||||
|
'verbose',
|
||||||
|
[f"ID: {lease_id}, Address: {address}, MAC Address: {mac_address}, Host Name: {host_name}, Comment: {comment}, Last Seen: {last_seen}, Status: {status}"]
|
||||||
|
)
|
||||||
|
|
||||||
if (status == "bound"):
|
if (status == "bound"):
|
||||||
plugin_objects.add_object(
|
plugin_objects.add_object(
|
||||||
@@ -96,8 +98,8 @@ def get_entries(plugin_objects: Plugin_Objects) -> Plugin_Objects:
|
|||||||
return plugin_objects
|
return plugin_objects
|
||||||
|
|
||||||
|
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
# BEGIN
|
# BEGIN
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
main()
|
main()
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
@@ -8,14 +8,14 @@ import subprocess
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects
|
from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from database import DB
|
from database import DB # noqa: E402 [flake8 lint suppression]
|
||||||
from models.device_instance import DeviceInstance
|
from models.device_instance import DeviceInstance # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -34,7 +34,6 @@ RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
|||||||
plugin_objects = Plugin_Objects(RESULT_FILE)
|
plugin_objects = Plugin_Objects(RESULT_FILE)
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
mylog('verbose', [f'[{pluginName}] In script'])
|
mylog('verbose', [f'[{pluginName}] In script'])
|
||||||
|
|
||||||
@@ -67,27 +66,28 @@ def main():
|
|||||||
|
|
||||||
if domain_name != '':
|
if domain_name != '':
|
||||||
plugin_objects.add_object(
|
plugin_objects.add_object(
|
||||||
# "MAC", "IP", "Server", "Name"
|
# "MAC", "IP", "Server", "Name"
|
||||||
primaryId = device['devMac'],
|
primaryId = device['devMac'],
|
||||||
secondaryId = device['devLastIP'],
|
secondaryId = device['devLastIP'],
|
||||||
watched1 = dns_server,
|
watched1 = dns_server,
|
||||||
watched2 = domain_name,
|
watched2 = domain_name,
|
||||||
watched3 = '',
|
watched3 = '',
|
||||||
watched4 = '',
|
watched4 = '',
|
||||||
extra = '',
|
extra = '',
|
||||||
foreignKey = device['devMac'])
|
foreignKey = device['devMac']
|
||||||
|
)
|
||||||
|
|
||||||
plugin_objects.write_result_file()
|
plugin_objects.write_result_file()
|
||||||
|
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] Script finished'])
|
mylog('verbose', [f'[{pluginName}] Script finished'])
|
||||||
|
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
#===============================================================================
|
|
||||||
|
# ===============================================================================
|
||||||
# Execute scan
|
# Execute scan
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
def execute_name_lookup (ip, timeout):
|
def execute_name_lookup(ip, timeout):
|
||||||
"""
|
"""
|
||||||
Execute the NBTSCAN command on IP.
|
Execute the NBTSCAN command on IP.
|
||||||
"""
|
"""
|
||||||
@@ -101,7 +101,13 @@ def execute_name_lookup (ip, timeout):
|
|||||||
mylog('verbose', [f'[{pluginName}] DEBUG CMD :', args])
|
mylog('verbose', [f'[{pluginName}] DEBUG CMD :', args])
|
||||||
|
|
||||||
# try runnning a subprocess with a forced (timeout) in case the subprocess hangs
|
# try runnning a subprocess with a forced (timeout) in case the subprocess hangs
|
||||||
output = subprocess.check_output (args, universal_newlines=True, stderr=subprocess.STDOUT, timeout=(timeout), text=True)
|
output = subprocess.check_output(
|
||||||
|
args,
|
||||||
|
universal_newlines=True,
|
||||||
|
stderr=subprocess.STDOUT,
|
||||||
|
timeout=(timeout),
|
||||||
|
text=True
|
||||||
|
)
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] DEBUG OUTPUT : {output}'])
|
mylog('verbose', [f'[{pluginName}] DEBUG OUTPUT : {output}'])
|
||||||
|
|
||||||
@@ -112,7 +118,6 @@ def execute_name_lookup (ip, timeout):
|
|||||||
lines = output.splitlines()
|
lines = output.splitlines()
|
||||||
|
|
||||||
# Look for the first line containing a valid NetBIOS name entry
|
# Look for the first line containing a valid NetBIOS name entry
|
||||||
index = 0
|
|
||||||
for line in lines:
|
for line in lines:
|
||||||
if 'Doing NBT name scan' not in line and ip in line:
|
if 'Doing NBT name scan' not in line and ip in line:
|
||||||
# Split the line and extract the primary NetBIOS name
|
# Split the line and extract the primary NetBIOS name
|
||||||
@@ -122,7 +127,6 @@ def execute_name_lookup (ip, timeout):
|
|||||||
else:
|
else:
|
||||||
mylog('verbose', [f'[{pluginName}] ⚠ ERROR - Unexpected output format: {line}'])
|
mylog('verbose', [f'[{pluginName}] ⚠ ERROR - Unexpected output format: {line}'])
|
||||||
|
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] Domain Name: {domain_name}'])
|
mylog('verbose', [f'[{pluginName}] Domain Name: {domain_name}'])
|
||||||
|
|
||||||
return domain_name, dns_server
|
return domain_name, dns_server
|
||||||
@@ -137,13 +141,16 @@ def execute_name_lookup (ip, timeout):
|
|||||||
except subprocess.TimeoutExpired:
|
except subprocess.TimeoutExpired:
|
||||||
mylog('verbose', [f'[{pluginName}] TIMEOUT - the process forcefully terminated as timeout reached'])
|
mylog('verbose', [f'[{pluginName}] TIMEOUT - the process forcefully terminated as timeout reached'])
|
||||||
|
|
||||||
if output == "": # check if the subprocess failed
|
if output == "": # check if the subprocess failed
|
||||||
mylog('verbose', [f'[{pluginName}] Scan: FAIL - check logs'])
|
mylog('verbose', [f'[{pluginName}] Scan: FAIL - check logs'])
|
||||||
else:
|
else:
|
||||||
mylog('verbose', [f'[{pluginName}] Scan: SUCCESS'])
|
mylog('verbose', [f'[{pluginName}] Scan: SUCCESS'])
|
||||||
|
|
||||||
return '', ''
|
return '', ''
|
||||||
|
|
||||||
|
|
||||||
|
# ===============================================================================
|
||||||
|
# BEGIN
|
||||||
|
# ===============================================================================
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
main()
|
main()
|
||||||
|
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
# test script by running:
|
# test script by running:
|
||||||
# tbc
|
# tbc
|
||||||
|
|
||||||
@@ -13,13 +13,12 @@ import nmap
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects
|
from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from database import DB
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -46,7 +45,6 @@ def main():
|
|||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] subnets: ', subnets])
|
mylog('verbose', [f'[{pluginName}] subnets: ', subnets])
|
||||||
|
|
||||||
|
|
||||||
# Initialize the Plugin obj output file
|
# Initialize the Plugin obj output file
|
||||||
plugin_objects = Plugin_Objects(RESULT_FILE)
|
plugin_objects = Plugin_Objects(RESULT_FILE)
|
||||||
|
|
||||||
@@ -57,26 +55,27 @@ def main():
|
|||||||
for device in unique_devices:
|
for device in unique_devices:
|
||||||
|
|
||||||
plugin_objects.add_object(
|
plugin_objects.add_object(
|
||||||
# "MAC", "IP", "Name", "Vendor", "Interface"
|
# "MAC", "IP", "Name", "Vendor", "Interface"
|
||||||
primaryId = device['mac'].lower(),
|
primaryId = device['mac'].lower(),
|
||||||
secondaryId = device['ip'],
|
secondaryId = device['ip'],
|
||||||
watched1 = device['name'],
|
watched1 = device['name'],
|
||||||
watched2 = device['vendor'],
|
watched2 = device['vendor'],
|
||||||
watched3 = device['interface'],
|
watched3 = device['interface'],
|
||||||
watched4 = '',
|
watched4 = '',
|
||||||
extra = '',
|
extra = '',
|
||||||
foreignKey = device['mac'])
|
foreignKey = device['mac']
|
||||||
|
)
|
||||||
|
|
||||||
plugin_objects.write_result_file()
|
plugin_objects.write_result_file()
|
||||||
|
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] Script finished'])
|
mylog('verbose', [f'[{pluginName}] Script finished'])
|
||||||
|
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
#===============================================================================
|
|
||||||
|
# ===============================================================================
|
||||||
# Execute scan
|
# Execute scan
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
def execute_scan(subnets_list, timeout, fakeMac, args):
|
def execute_scan(subnets_list, timeout, fakeMac, args):
|
||||||
devices_list = []
|
devices_list = []
|
||||||
|
|
||||||
@@ -103,13 +102,12 @@ def execute_scan(subnets_list, timeout, fakeMac, args):
|
|||||||
return devices_list
|
return devices_list
|
||||||
|
|
||||||
|
|
||||||
|
def execute_scan_on_interface(interface, timeout, args):
|
||||||
def execute_scan_on_interface (interface, timeout, args):
|
|
||||||
# Remove unsupported VLAN flags
|
# Remove unsupported VLAN flags
|
||||||
interface = re.sub(r'--vlan=\S+', '', interface).strip()
|
interface = re.sub(r'--vlan=\S+', '', interface).strip()
|
||||||
|
|
||||||
# Prepare command arguments
|
# Prepare command arguments
|
||||||
scan_args = args.split() + interface.replace('--interface=','-e ').split()
|
scan_args = args.split() + interface.replace('--interface=', '-e ').split()
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] scan_args: ', scan_args])
|
mylog('verbose', [f'[{pluginName}] scan_args: ', scan_args])
|
||||||
|
|
||||||
@@ -138,7 +136,6 @@ def parse_nmap_xml(xml_output, interface, fakeMac):
|
|||||||
ip = nm[host]['addresses']['ipv4'] if 'ipv4' in nm[host]['addresses'] else ''
|
ip = nm[host]['addresses']['ipv4'] if 'ipv4' in nm[host]['addresses'] else ''
|
||||||
mac = nm[host]['addresses']['mac'] if 'mac' in nm[host]['addresses'] else ''
|
mac = nm[host]['addresses']['mac'] if 'mac' in nm[host]['addresses'] else ''
|
||||||
|
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] nm[host]: ', nm[host]])
|
mylog('verbose', [f'[{pluginName}] nm[host]: ', nm[host]])
|
||||||
|
|
||||||
vendor = ''
|
vendor = ''
|
||||||
@@ -148,10 +145,8 @@ def parse_nmap_xml(xml_output, interface, fakeMac):
|
|||||||
|
|
||||||
for key, value in nm[host]['vendor'].items():
|
for key, value in nm[host]['vendor'].items():
|
||||||
vendor = value
|
vendor = value
|
||||||
|
|
||||||
break
|
break
|
||||||
|
|
||||||
|
|
||||||
# Log debug information
|
# Log debug information
|
||||||
mylog('verbose', [f"[{pluginName}] Hostname: {hostname}, IP: {ip}, MAC: {mac}, Vendor: {vendor}"])
|
mylog('verbose', [f"[{pluginName}] Hostname: {hostname}, IP: {ip}, MAC: {mac}, Vendor: {vendor}"])
|
||||||
|
|
||||||
@@ -172,7 +167,6 @@ def parse_nmap_xml(xml_output, interface, fakeMac):
|
|||||||
# MAC or IP missing
|
# MAC or IP missing
|
||||||
mylog('verbose', [f"[{pluginName}] Skipping: {hostname}, IP or MAC missing, or NMAPDEV_GENERATE_MAC setting not enabled"])
|
mylog('verbose', [f"[{pluginName}] Skipping: {hostname}, IP or MAC missing, or NMAPDEV_GENERATE_MAC setting not enabled"])
|
||||||
|
|
||||||
|
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
mylog('verbose', [f"[{pluginName}] Error parsing nmap XML: ", str(e)])
|
mylog('verbose', [f"[{pluginName}] Error parsing nmap XML: ", str(e)])
|
||||||
|
|
||||||
@@ -184,12 +178,13 @@ def string_to_mac_hash(input_string):
|
|||||||
sha256_hash = hashlib.sha256(input_string.encode()).hexdigest()
|
sha256_hash = hashlib.sha256(input_string.encode()).hexdigest()
|
||||||
|
|
||||||
# Take the first 12 characters of the hash and format as a MAC address
|
# Take the first 12 characters of the hash and format as a MAC address
|
||||||
mac_hash = ':'.join(sha256_hash[i:i+2] for i in range(0, 12, 2))
|
mac_hash = ':'.join(sha256_hash[i:i + 2] for i in range(0, 12, 2))
|
||||||
|
|
||||||
return mac_hash
|
return mac_hash
|
||||||
|
|
||||||
#===============================================================================
|
|
||||||
|
# ===============================================================================
|
||||||
# BEGIN
|
# BEGIN
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
main()
|
main()
|
||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import argparse
|
import argparse
|
||||||
@@ -9,13 +9,13 @@ import subprocess
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects
|
from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger, append_line_to_file
|
from logger import mylog, Logger, append_line_to_file # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.datetime_utils import timeNowDB
|
from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -32,7 +32,8 @@ RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
|||||||
# Initialize the Plugin obj output file
|
# Initialize the Plugin obj output file
|
||||||
plugin_objects = Plugin_Objects(RESULT_FILE)
|
plugin_objects = Plugin_Objects(RESULT_FILE)
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
|
# -------------------------------------------------------------------------------
|
||||||
def main():
|
def main():
|
||||||
parser = argparse.ArgumentParser(
|
parser = argparse.ArgumentParser(
|
||||||
description='Scan ports of devices specified by IP addresses'
|
description='Scan ports of devices specified by IP addresses'
|
||||||
@@ -100,8 +101,8 @@ def main():
|
|||||||
|
|
||||||
plugin_objects.write_result_file()
|
plugin_objects.write_result_file()
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
|
|
||||||
|
# -------------------------------------------------------------------------------
|
||||||
class nmap_entry:
|
class nmap_entry:
|
||||||
def __init__(self, ip, mac, time, port, state, service, name = '', extra = '', index = 0):
|
def __init__(self, ip, mac, time, port, state, service, name = '', extra = '', index = 0):
|
||||||
self.ip = ip
|
self.ip = ip
|
||||||
@@ -112,10 +113,10 @@ class nmap_entry:
|
|||||||
self.service = service
|
self.service = service
|
||||||
self.extra = extra
|
self.extra = extra
|
||||||
self.index = index
|
self.index = index
|
||||||
self.hash = str(mac) + str(port)+ str(state)+ str(service)
|
self.hash = str(mac) + str(port) + str(state) + str(service)
|
||||||
|
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
# -------------------------------------------------------------------------------
|
||||||
def parse_kv_args(raw_args):
|
def parse_kv_args(raw_args):
|
||||||
"""
|
"""
|
||||||
Converts ['ips=a,b,c', 'macs=x,y,z', 'timeout=5'] to a dict.
|
Converts ['ips=a,b,c', 'macs=x,y,z', 'timeout=5'] to a dict.
|
||||||
@@ -136,7 +137,8 @@ def parse_kv_args(raw_args):
|
|||||||
|
|
||||||
return parsed
|
return parsed
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
|
# -------------------------------------------------------------------------------
|
||||||
def safe_split_list(value, keyname):
|
def safe_split_list(value, keyname):
|
||||||
"""Split comma list safely and ensure no empty items."""
|
"""Split comma list safely and ensure no empty items."""
|
||||||
items = [x.strip() for x in value.split(',') if x.strip()]
|
items = [x.strip() for x in value.split(',') if x.strip()]
|
||||||
@@ -144,7 +146,8 @@ def safe_split_list(value, keyname):
|
|||||||
mylog('none', [f"[{pluginName}] Scan: {keyname} list is empty or invalid"])
|
mylog('none', [f"[{pluginName}] Scan: {keyname} list is empty or invalid"])
|
||||||
return items
|
return items
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
|
# -------------------------------------------------------------------------------
|
||||||
def performNmapScan(deviceIPs, deviceMACs, timeoutSec, args):
|
def performNmapScan(deviceIPs, deviceMACs, timeoutSec, args):
|
||||||
"""
|
"""
|
||||||
run nmap scan on a list of devices
|
run nmap scan on a list of devices
|
||||||
@@ -154,15 +157,12 @@ def performNmapScan(deviceIPs, deviceMACs, timeoutSec, args):
|
|||||||
# collect ports / new Nmap Entries
|
# collect ports / new Nmap Entries
|
||||||
newEntriesTmp = []
|
newEntriesTmp = []
|
||||||
|
|
||||||
|
|
||||||
if len(deviceIPs) > 0:
|
if len(deviceIPs) > 0:
|
||||||
|
|
||||||
devTotal = len(deviceIPs)
|
devTotal = len(deviceIPs)
|
||||||
|
|
||||||
|
mylog('verbose', [f'[{pluginName}] Scan: Nmap for max ', str(timeoutSec), 's (' + str(round(int(timeoutSec) / 60, 1)) + 'min) per device'])
|
||||||
mylog('verbose', [f'[{pluginName}] Scan: Nmap for max ', str(timeoutSec), 's ('+ str(round(int(timeoutSec) / 60, 1)) +'min) per device'])
|
mylog('verbose', ["[NMAP Scan] Estimated max delay: ", (devTotal * int(timeoutSec)), 's ', '(', round((devTotal * int(timeoutSec)) / 60, 1) , 'min)'])
|
||||||
mylog('verbose', ["[NMAP Scan] Estimated max delay: ", (devTotal * int(timeoutSec)), 's ', '(', round((devTotal * int(timeoutSec))/60,1) , 'min)' ])
|
|
||||||
|
|
||||||
|
|
||||||
devIndex = 0
|
devIndex = 0
|
||||||
for ip in deviceIPs:
|
for ip in deviceIPs:
|
||||||
@@ -171,32 +171,34 @@ def performNmapScan(deviceIPs, deviceMACs, timeoutSec, args):
|
|||||||
# prepare arguments from user supplied ones
|
# prepare arguments from user supplied ones
|
||||||
nmapArgs = ['nmap'] + args.split() + [ip]
|
nmapArgs = ['nmap'] + args.split() + [ip]
|
||||||
|
|
||||||
progress = ' (' + str(devIndex+1) + '/' + str(devTotal) + ')'
|
progress = ' (' + str(devIndex + 1) + '/' + str(devTotal) + ')'
|
||||||
|
|
||||||
try:
|
try:
|
||||||
# try runnning a subprocess with a forced (timeout) in case the subprocess hangs
|
# try runnning a subprocess with a forced (timeout) in case the subprocess hangs
|
||||||
output = subprocess.check_output (nmapArgs, universal_newlines=True, stderr=subprocess.STDOUT, timeout=(float(timeoutSec)))
|
output = subprocess.check_output(
|
||||||
|
nmapArgs,
|
||||||
|
universal_newlines=True,
|
||||||
|
stderr=subprocess.STDOUT,
|
||||||
|
timeout=(float(timeoutSec))
|
||||||
|
)
|
||||||
except subprocess.CalledProcessError as e:
|
except subprocess.CalledProcessError as e:
|
||||||
# An error occured, handle it
|
# An error occured, handle it
|
||||||
mylog('none', ["[NMAP Scan] " ,e.output])
|
mylog('none', ["[NMAP Scan] ", e.output])
|
||||||
mylog('none', ["[NMAP Scan] ⚠ ERROR - Nmap Scan - check logs", progress])
|
mylog('none', ["[NMAP Scan] ⚠ ERROR - Nmap Scan - check logs", progress])
|
||||||
except subprocess.TimeoutExpired:
|
except subprocess.TimeoutExpired:
|
||||||
mylog('verbose', [f'[{pluginName}] Nmap TIMEOUT - the process forcefully terminated as timeout reached for ', ip, progress])
|
mylog('verbose', [f'[{pluginName}] Nmap TIMEOUT - the process forcefully terminated as timeout reached for ', ip, progress])
|
||||||
|
|
||||||
if output == "": # check if the subprocess failed
|
if output == "": # check if the subprocess failed
|
||||||
mylog('minimal', [f'[{pluginName}] Nmap FAIL for ', ip, progress ,' check logs for details'])
|
mylog('minimal', [f'[{pluginName}] Nmap FAIL for ', ip, progress, ' check logs for details'])
|
||||||
else:
|
else:
|
||||||
mylog('verbose', [f'[{pluginName}] Nmap SUCCESS for ', ip, progress])
|
mylog('verbose', [f'[{pluginName}] Nmap SUCCESS for ', ip, progress])
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
# check the last run output
|
# check the last run output
|
||||||
newLines = output.split('\n')
|
newLines = output.split('\n')
|
||||||
|
|
||||||
# regular logging
|
# regular logging
|
||||||
for line in newLines:
|
for line in newLines:
|
||||||
append_line_to_file (logPath + '/app_nmap.log', line +'\n')
|
append_line_to_file(logPath + '/app_nmap.log', line + '\n')
|
||||||
|
|
||||||
|
|
||||||
index = 0
|
index = 0
|
||||||
startCollecting = False
|
startCollecting = False
|
||||||
@@ -204,34 +206,28 @@ def performNmapScan(deviceIPs, deviceMACs, timeoutSec, args):
|
|||||||
newPortsPerDevice = 0
|
newPortsPerDevice = 0
|
||||||
for line in newLines:
|
for line in newLines:
|
||||||
if 'Starting Nmap' in line:
|
if 'Starting Nmap' in line:
|
||||||
if len(newLines) > index+1 and 'Note: Host seems down' in newLines[index+1]:
|
if len(newLines) > index + 1 and 'Note: Host seems down' in newLines[index + 1]:
|
||||||
break # this entry is empty
|
break # this entry is empty
|
||||||
elif 'PORT' in line and 'STATE' in line and 'SERVICE' in line:
|
elif 'PORT' in line and 'STATE' in line and 'SERVICE' in line:
|
||||||
startCollecting = True
|
startCollecting = True
|
||||||
elif 'PORT' in line and 'STATE' in line and 'SERVICE' in line:
|
elif 'PORT' in line and 'STATE' in line and 'SERVICE' in line:
|
||||||
startCollecting = False # end reached
|
startCollecting = False # end reached
|
||||||
elif startCollecting and len(line.split()) == 3:
|
elif startCollecting and len(line.split()) == 3:
|
||||||
newEntriesTmp.append(nmap_entry(ip, deviceMACs[devIndex], timeNowDB(), line.split()[0], line.split()[1], line.split()[2]))
|
newEntriesTmp.append(nmap_entry(ip, deviceMACs[devIndex], timeNowDB(), line.split()[0], line.split()[1], line.split()[2]))
|
||||||
newPortsPerDevice += 1
|
newPortsPerDevice += 1
|
||||||
elif 'Nmap done' in line:
|
elif 'Nmap done' in line:
|
||||||
duration = line.split('scanned in ')[1]
|
duration = line.split('scanned in ')[1]
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] {newPortsPerDevice} ports found on {deviceMACs[devIndex]}'])
|
mylog('verbose', [f'[{pluginName}] {newPortsPerDevice} ports found on {deviceMACs[devIndex]} after {duration}'])
|
||||||
|
|
||||||
index += 1
|
index += 1
|
||||||
devIndex += 1
|
devIndex += 1
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
#end for loop
|
|
||||||
|
|
||||||
return newEntriesTmp
|
return newEntriesTmp
|
||||||
|
|
||||||
#===============================================================================
|
|
||||||
|
# ===============================================================================
|
||||||
# BEGIN
|
# BEGIN
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
main()
|
main()
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
# test script by running:
|
# test script by running:
|
||||||
# tbc
|
# tbc
|
||||||
|
|
||||||
@@ -11,14 +11,14 @@ import re
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects
|
from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from database import DB
|
from database import DB # noqa: E402 [flake8 lint suppression]
|
||||||
from models.device_instance import DeviceInstance
|
from models.device_instance import DeviceInstance # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -33,12 +33,10 @@ LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
|
|||||||
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] In script'])
|
mylog('verbose', [f'[{pluginName}] In script'])
|
||||||
|
|
||||||
|
|
||||||
timeout = get_setting_value('NSLOOKUP_RUN_TIMEOUT')
|
timeout = get_setting_value('NSLOOKUP_RUN_TIMEOUT')
|
||||||
|
|
||||||
# Create a database connection
|
# Create a database connection
|
||||||
@@ -67,27 +65,28 @@ def main():
|
|||||||
|
|
||||||
if domain_name != '':
|
if domain_name != '':
|
||||||
plugin_objects.add_object(
|
plugin_objects.add_object(
|
||||||
# "MAC", "IP", "Server", "Name"
|
# "MAC", "IP", "Server", "Name"
|
||||||
primaryId = device['devMac'],
|
primaryId = device['devMac'],
|
||||||
secondaryId = device['devLastIP'],
|
secondaryId = device['devLastIP'],
|
||||||
watched1 = dns_server,
|
watched1 = dns_server,
|
||||||
watched2 = domain_name,
|
watched2 = domain_name,
|
||||||
watched3 = '',
|
watched3 = '',
|
||||||
watched4 = '',
|
watched4 = '',
|
||||||
extra = '',
|
extra = '',
|
||||||
foreignKey = device['devMac'])
|
foreignKey = device['devMac']
|
||||||
|
)
|
||||||
|
|
||||||
plugin_objects.write_result_file()
|
plugin_objects.write_result_file()
|
||||||
|
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] Script finished'])
|
mylog('verbose', [f'[{pluginName}] Script finished'])
|
||||||
|
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
#===============================================================================
|
|
||||||
|
# ===============================================================================
|
||||||
# Execute scan
|
# Execute scan
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
def execute_nslookup (ip, timeout):
|
def execute_nslookup(ip, timeout):
|
||||||
"""
|
"""
|
||||||
Execute the NSLOOKUP command on IP.
|
Execute the NSLOOKUP command on IP.
|
||||||
"""
|
"""
|
||||||
@@ -99,7 +98,13 @@ def execute_nslookup (ip, timeout):
|
|||||||
|
|
||||||
try:
|
try:
|
||||||
# try runnning a subprocess with a forced (timeout) in case the subprocess hangs
|
# try runnning a subprocess with a forced (timeout) in case the subprocess hangs
|
||||||
output = subprocess.check_output (nslookup_args, universal_newlines=True, stderr=subprocess.STDOUT, timeout=(timeout), text=True)
|
output = subprocess.check_output(
|
||||||
|
nslookup_args,
|
||||||
|
universal_newlines=True,
|
||||||
|
stderr=subprocess.STDOUT,
|
||||||
|
timeout=(timeout),
|
||||||
|
text=True
|
||||||
|
)
|
||||||
|
|
||||||
domain_name = ''
|
domain_name = ''
|
||||||
dns_server = ''
|
dns_server = ''
|
||||||
@@ -110,7 +115,6 @@ def execute_nslookup (ip, timeout):
|
|||||||
domain_pattern = re.compile(r'name\s*=\s*([^\s]+)', re.IGNORECASE)
|
domain_pattern = re.compile(r'name\s*=\s*([^\s]+)', re.IGNORECASE)
|
||||||
server_pattern = re.compile(r'Server:\s+(.+)', re.IGNORECASE)
|
server_pattern = re.compile(r'Server:\s+(.+)', re.IGNORECASE)
|
||||||
|
|
||||||
|
|
||||||
domain_match = domain_pattern.search(output)
|
domain_match = domain_pattern.search(output)
|
||||||
server_match = server_pattern.search(output)
|
server_match = server_pattern.search(output)
|
||||||
|
|
||||||
@@ -136,19 +140,15 @@ def execute_nslookup (ip, timeout):
|
|||||||
except subprocess.TimeoutExpired:
|
except subprocess.TimeoutExpired:
|
||||||
mylog('verbose', [f'[{pluginName}] TIMEOUT - the process forcefully terminated as timeout reached'])
|
mylog('verbose', [f'[{pluginName}] TIMEOUT - the process forcefully terminated as timeout reached'])
|
||||||
|
|
||||||
if output == "": # check if the subprocess failed
|
if output != "": # check if the subprocess failed
|
||||||
tmp = 1 # can't have empty
|
|
||||||
# mylog('verbose', [f'[{pluginName}] Scan: FAIL - check logs'])
|
|
||||||
else:
|
|
||||||
mylog('verbose', [f'[{pluginName}] Scan: SUCCESS'])
|
mylog('verbose', [f'[{pluginName}] Scan: SUCCESS'])
|
||||||
|
|
||||||
return '', ''
|
return '', ''
|
||||||
|
|
||||||
|
|
||||||
|
# ===============================================================================
|
||||||
|
|
||||||
#===============================================================================
|
|
||||||
# BEGIN
|
# BEGIN
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
main()
|
main()
|
||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
__author__ = "ffsb"
|
__author__ = "ffsb"
|
||||||
__version__ = "0.1" # initial
|
__version__ = "0.1" # initial
|
||||||
__version__ = "0.2" # added logic to retry omada api call once as it seems to sometimes fail for some reasons, and error handling logic...
|
__version__ = "0.2" # added logic to retry omada api call once as it seems to sometimes fail for some reasons, and error handling logic...
|
||||||
@@ -15,10 +15,9 @@ __version__ = "1.3" # fix detection of the default gateway IP address that woul
|
|||||||
# try to identify and populate their connections by switch/accesspoints and ports/SSID
|
# try to identify and populate their connections by switch/accesspoints and ports/SSID
|
||||||
# try to differentiate root bridges from accessory
|
# try to differentiate root bridges from accessory
|
||||||
|
|
||||||
|
|
||||||
#
|
|
||||||
# sample code to update unbound on opnsense - for reference...
|
# sample code to update unbound on opnsense - for reference...
|
||||||
# curl -X POST -d '{"host":{"enabled":"1","hostname":"test","domain":"testdomain.com","rr":"A","mxprio":"","mx":"","server":"10.0.1.1","description":""}}' -H "Content-Type: application/json" -k -u $OPNS_KEY:$OPNS_SECRET https://$IPFW/api/unbound/settings/AddHostOverride
|
# curl -X POST -d '{"host":{"enabled":"1","hostname":"test","domain":"testdomain.com","rr":"A","mxprio":"","mx":"","server":"10.0.1.1","description":""}}'\
|
||||||
|
# -H "Content-Type: application/json" -k -u $OPNS_KEY:$OPNS_SECRET https://$IPFW/api/unbound/settings/AddHostOverride
|
||||||
#
|
#
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
@@ -35,12 +34,12 @@ import multiprocessing
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects
|
from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -87,8 +86,6 @@ cMAC, cIP, cNAME, cSWITCH_AP, cPORT_SSID = range(5)
|
|||||||
OMDLOGLEVEL = "debug"
|
OMDLOGLEVEL = "debug"
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
#
|
|
||||||
# translate MAC address from standard ieee model to ietf draft
|
# translate MAC address from standard ieee model to ietf draft
|
||||||
# AA-BB-CC-DD-EE-FF to aa:bb:cc:dd:ee:ff
|
# AA-BB-CC-DD-EE-FF to aa:bb:cc:dd:ee:ff
|
||||||
# tplink adheres to ieee, Nax adheres to ietf
|
# tplink adheres to ieee, Nax adheres to ietf
|
||||||
@@ -142,7 +139,7 @@ def callomada(myargs):
|
|||||||
try:
|
try:
|
||||||
mf = io.StringIO()
|
mf = io.StringIO()
|
||||||
with redirect_stdout(mf):
|
with redirect_stdout(mf):
|
||||||
bar = omada(myargs)
|
omada(myargs)
|
||||||
omada_output = mf.getvalue()
|
omada_output = mf.getvalue()
|
||||||
except Exception:
|
except Exception:
|
||||||
mylog(
|
mylog(
|
||||||
@@ -207,8 +204,7 @@ def add_uplink(
|
|||||||
|
|
||||||
# Determine port to uplink
|
# Determine port to uplink
|
||||||
if (
|
if (
|
||||||
device_data_bymac[switch_mac].get(TYPE) == "Switch"
|
device_data_bymac[switch_mac].get(TYPE) == "Switch" and device_data_bymac[uplink_mac].get(TYPE) == "Switch"
|
||||||
and device_data_bymac[uplink_mac].get(TYPE) == "Switch"
|
|
||||||
):
|
):
|
||||||
port_to_uplink = port_byswitchmac_byclientmac.get(switch_mac, {}).get(uplink_mac)
|
port_to_uplink = port_byswitchmac_byclientmac.get(switch_mac, {}).get(uplink_mac)
|
||||||
if port_to_uplink is None:
|
if port_to_uplink is None:
|
||||||
@@ -223,9 +219,7 @@ def add_uplink(
|
|||||||
# Recursively add uplinks for linked devices
|
# Recursively add uplinks for linked devices
|
||||||
for link in sadevices_linksbymac.get(switch_mac, []):
|
for link in sadevices_linksbymac.get(switch_mac, []):
|
||||||
if (
|
if (
|
||||||
link in device_data_bymac
|
link in device_data_bymac and device_data_bymac[link].get(SWITCH_AP) in [None, "null"] and device_data_bymac[switch_mac].get(TYPE) == "Switch"
|
||||||
and device_data_bymac[link].get(SWITCH_AP) in [None, "null"]
|
|
||||||
and device_data_bymac[switch_mac].get(TYPE) == "Switch"
|
|
||||||
):
|
):
|
||||||
add_uplink(
|
add_uplink(
|
||||||
switch_mac,
|
switch_mac,
|
||||||
@@ -236,7 +230,6 @@ def add_uplink(
|
|||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
# ----------------------------------------------
|
# ----------------------------------------------
|
||||||
# Main initialization
|
# Main initialization
|
||||||
def main():
|
def main():
|
||||||
@@ -324,16 +317,16 @@ def main():
|
|||||||
)
|
)
|
||||||
mymac = ieee2ietf_mac_formater(device[MAC])
|
mymac = ieee2ietf_mac_formater(device[MAC])
|
||||||
plugin_objects.add_object(
|
plugin_objects.add_object(
|
||||||
primaryId=mymac, # MAC
|
primaryId=mymac, # MAC
|
||||||
secondaryId=device[IP], # IP
|
secondaryId=device[IP], # IP
|
||||||
watched1=device[NAME], # NAME/HOSTNAME
|
watched1=device[NAME], # NAME/HOSTNAME
|
||||||
watched2=ParentNetworkNode, # PARENT NETWORK NODE MAC
|
watched2=ParentNetworkNode, # PARENT NETWORK NODE MAC
|
||||||
watched3=myport, # PORT
|
watched3=myport, # PORT
|
||||||
watched4=myssid, # SSID
|
watched4=myssid, # SSID
|
||||||
extra=device[TYPE],
|
extra=device[TYPE],
|
||||||
# omada_site, # SITENAME (cur_NetworkSite) or VENDOR (cur_Vendor) (PICK one and adjust config.json -> "column": "Extra")
|
# omada_site, # SITENAME (cur_NetworkSite) or VENDOR (cur_Vendor) (PICK one and adjust config.json -> "column": "Extra")
|
||||||
foreignKey=device[MAC].lower().replace("-", ":"),
|
foreignKey=device[MAC].lower().replace("-", ":"),
|
||||||
) # usually MAC
|
) # usually MAC
|
||||||
|
|
||||||
mylog(
|
mylog(
|
||||||
"verbose",
|
"verbose",
|
||||||
@@ -369,7 +362,6 @@ def get_omada_devices_details(msadevice_data):
|
|||||||
mswitch_dump = callomada(["-t", "myomada", "switch", "-d", mthisswitch])
|
mswitch_dump = callomada(["-t", "myomada", "switch", "-d", mthisswitch])
|
||||||
else:
|
else:
|
||||||
mswitch_detail = ""
|
mswitch_detail = ""
|
||||||
nswitch_dump = ""
|
|
||||||
return mswitch_detail, mswitch_dump
|
return mswitch_detail, mswitch_dump
|
||||||
|
|
||||||
|
|
||||||
@@ -414,7 +406,6 @@ def get_device_data(omada_clients_output, switches_and_aps, device_handler):
|
|||||||
# 17:27:10 [<unique_prefix>] token: "['1A-2B-3C-4D-5E-6F', '192.168.0.217', '1A-2B-3C-4D-5E-6F', '17', '40-AE-30-A5-A7-50, 'Switch']"
|
# 17:27:10 [<unique_prefix>] token: "['1A-2B-3C-4D-5E-6F', '192.168.0.217', '1A-2B-3C-4D-5E-6F', '17', '40-AE-30-A5-A7-50, 'Switch']"
|
||||||
# constants
|
# constants
|
||||||
sadevices_macbyname = {}
|
sadevices_macbyname = {}
|
||||||
sadevices_macbymac = {}
|
|
||||||
sadevices_linksbymac = {}
|
sadevices_linksbymac = {}
|
||||||
port_byswitchmac_byclientmac = {}
|
port_byswitchmac_byclientmac = {}
|
||||||
device_data_bymac = {}
|
device_data_bymac = {}
|
||||||
@@ -556,11 +547,11 @@ def get_device_data(omada_clients_output, switches_and_aps, device_handler):
|
|||||||
#
|
#
|
||||||
|
|
||||||
naxname = real_naxname
|
naxname = real_naxname
|
||||||
if real_naxname != None:
|
if real_naxname is not None:
|
||||||
if "(" in real_naxname:
|
if "(" in real_naxname:
|
||||||
# removing parenthesis and domains from the name
|
# removing parenthesis and domains from the name
|
||||||
naxname = real_naxname.split("(")[0]
|
naxname = real_naxname.split("(")[0]
|
||||||
if naxname != None and "." in naxname:
|
if naxname is not None and "." in naxname:
|
||||||
naxname = naxname.split(".")[0]
|
naxname = naxname.split(".")[0]
|
||||||
if naxname in (None, "null", ""):
|
if naxname in (None, "null", ""):
|
||||||
naxname = (
|
naxname = (
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
"""
|
"""
|
||||||
This plugin imports devices and clients from Omada Controller using their OpenAPI.
|
This plugin imports devices and clients from Omada Controller using their OpenAPI.
|
||||||
@@ -25,7 +25,6 @@ import sys
|
|||||||
import urllib3
|
import urllib3
|
||||||
import requests
|
import requests
|
||||||
import time
|
import time
|
||||||
import datetime
|
|
||||||
import pytz
|
import pytz
|
||||||
|
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
@@ -35,11 +34,11 @@ from typing import Literal, Any, Dict
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects, is_typical_router_ip, is_mac
|
from plugin_helper import Plugin_Objects, is_typical_router_ip, is_mac # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = pytz.timezone(get_setting_value('TIMEZONE'))
|
conf.tz = pytz.timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -176,7 +175,10 @@ class OmadaHelper:
|
|||||||
# If it's not a gateway try to assign parent node MAC
|
# If it's not a gateway try to assign parent node MAC
|
||||||
if data.get("type", "") != "gateway":
|
if data.get("type", "") != "gateway":
|
||||||
parent_mac = OmadaHelper.normalize_mac(data.get("uplinkDeviceMac"))
|
parent_mac = OmadaHelper.normalize_mac(data.get("uplinkDeviceMac"))
|
||||||
entry["parent_node_mac_address"] = parent_mac.get("response_result") if isinstance(parent_mac, dict) and parent_mac.get("response_type") == "success" else ""
|
|
||||||
|
resp_type = parent_mac.get("response_type")
|
||||||
|
|
||||||
|
entry["parent_node_mac_address"] = parent_mac.get("response_result") if isinstance(parent_mac, dict) and resp_type == "success" else ""
|
||||||
|
|
||||||
# Applicable only for CLIENT
|
# Applicable only for CLIENT
|
||||||
if input_type == "client":
|
if input_type == "client":
|
||||||
@@ -185,15 +187,26 @@ class OmadaHelper:
|
|||||||
# Try to assign parent node MAC and PORT/SSID to the CLIENT
|
# Try to assign parent node MAC and PORT/SSID to the CLIENT
|
||||||
if data.get("connectDevType", "") == "gateway":
|
if data.get("connectDevType", "") == "gateway":
|
||||||
parent_mac = OmadaHelper.normalize_mac(data.get("gatewayMac"))
|
parent_mac = OmadaHelper.normalize_mac(data.get("gatewayMac"))
|
||||||
entry["parent_node_mac_address"] = parent_mac.get("response_result") if isinstance(parent_mac, dict) and parent_mac.get("response_type") == "success" else ""
|
|
||||||
|
resp_type = parent_mac.get("response_type")
|
||||||
|
|
||||||
|
entry["parent_node_mac_address"] = parent_mac.get("response_result") if isinstance(parent_mac, dict) and resp_type == "success" else ""
|
||||||
entry["parent_node_port"] = data.get("port", "")
|
entry["parent_node_port"] = data.get("port", "")
|
||||||
|
|
||||||
elif data.get("connectDevType", "") == "switch":
|
elif data.get("connectDevType", "") == "switch":
|
||||||
parent_mac = OmadaHelper.normalize_mac(data.get("switchMac"))
|
parent_mac = OmadaHelper.normalize_mac(data.get("switchMac"))
|
||||||
entry["parent_node_mac_address"] = parent_mac.get("response_result") if isinstance(parent_mac, dict) and parent_mac.get("response_type") == "success" else ""
|
|
||||||
|
resp_type = parent_mac.get("response_type")
|
||||||
|
|
||||||
|
entry["parent_node_mac_address"] = parent_mac.get("response_result") if isinstance(parent_mac, dict) and resp_type == "success" else ""
|
||||||
entry["parent_node_port"] = data.get("port", "")
|
entry["parent_node_port"] = data.get("port", "")
|
||||||
|
|
||||||
elif data.get("connectDevType", "") == "ap":
|
elif data.get("connectDevType", "") == "ap":
|
||||||
parent_mac = OmadaHelper.normalize_mac(data.get("apMac"))
|
parent_mac = OmadaHelper.normalize_mac(data.get("apMac"))
|
||||||
entry["parent_node_mac_address"] = parent_mac.get("response_result") if isinstance(parent_mac, dict) and parent_mac.get("response_type") == "success" else ""
|
|
||||||
|
resp_type = parent_mac.get("response_type")
|
||||||
|
|
||||||
|
entry["parent_node_mac_address"] = parent_mac.get("response_result") if isinstance(parent_mac, dict) and resp_type == "success" else ""
|
||||||
entry["parent_node_ssid"] = data.get("ssid", "")
|
entry["parent_node_ssid"] = data.get("ssid", "")
|
||||||
|
|
||||||
# Add the entry to the result
|
# Add the entry to the result
|
||||||
@@ -253,7 +266,7 @@ class OmadaAPI:
|
|||||||
"""Return request headers."""
|
"""Return request headers."""
|
||||||
headers = {"Content-type": "application/json"}
|
headers = {"Content-type": "application/json"}
|
||||||
# Add access token to header if requested and available
|
# Add access token to header if requested and available
|
||||||
if include_auth == True:
|
if include_auth is True:
|
||||||
if not self.access_token:
|
if not self.access_token:
|
||||||
OmadaHelper.debug("No access token available for headers")
|
OmadaHelper.debug("No access token available for headers")
|
||||||
else:
|
else:
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
"""
|
"""
|
||||||
NetAlertX plugin: PIHOLEAPI
|
NetAlertX plugin: PIHOLEAPI
|
||||||
Imports devices from Pi-hole v6 API (Network endpoints) into NetAlertX plugin results.
|
Imports devices from Pi-hole v6 API (Network endpoints) into NetAlertX plugin results.
|
||||||
@@ -17,12 +17,12 @@ sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
|||||||
|
|
||||||
pluginName = 'PIHOLEAPI'
|
pluginName = 'PIHOLEAPI'
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects, is_mac
|
from plugin_helper import Plugin_Objects, is_mac # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Setup timezone & logger using standard NAX helpers
|
# Setup timezone & logger using standard NAX helpers
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
|
|||||||
@@ -5,18 +5,18 @@ import os
|
|||||||
import re
|
import re
|
||||||
import base64
|
import base64
|
||||||
import json
|
import json
|
||||||
from datetime import datetime
|
|
||||||
|
|
||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
|
|
||||||
sys.path.append(f"{INSTALL_PATH}/front/plugins")
|
sys.path.append(f"{INSTALL_PATH}/front/plugins")
|
||||||
sys.path.append(f'{INSTALL_PATH}/server')
|
sys.path.append(f'{INSTALL_PATH}/server')
|
||||||
|
|
||||||
from logger import mylog, Logger
|
from logger import mylog # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.datetime_utils import timeNowDB
|
from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
|
||||||
from const import default_tz, fullConfPath
|
from const import default_tz, fullConfPath # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
|
# -------------------------------------------------------------------------------
|
||||||
def read_config_file():
|
def read_config_file():
|
||||||
"""
|
"""
|
||||||
retuns dict on the config file key:value pairs
|
retuns dict on the config file key:value pairs
|
||||||
@@ -25,13 +25,13 @@ def read_config_file():
|
|||||||
|
|
||||||
filename = fullConfPath
|
filename = fullConfPath
|
||||||
|
|
||||||
|
|
||||||
print('[plugin_helper] reading config file')
|
print('[plugin_helper] reading config file')
|
||||||
|
|
||||||
# load the variables from .conf
|
# load the variables from .conf
|
||||||
with open(filename, "r") as file:
|
with open(filename, "r") as file:
|
||||||
code = compile(file.read(), filename, "exec")
|
code = compile(file.read(), filename, "exec")
|
||||||
|
|
||||||
confDict = {} # config dictionary
|
confDict = {} # config dictionary
|
||||||
exec(code, {"__builtins__": {}}, confDict)
|
exec(code, {"__builtins__": {}}, confDict)
|
||||||
return confDict
|
return confDict
|
||||||
|
|
||||||
@@ -42,6 +42,7 @@ if timeZoneSetting not in all_timezones:
|
|||||||
timeZoneSetting = default_tz
|
timeZoneSetting = default_tz
|
||||||
timeZone = pytz.timezone(timeZoneSetting)
|
timeZone = pytz.timezone(timeZoneSetting)
|
||||||
|
|
||||||
|
|
||||||
# -------------------------------------------------------------------
|
# -------------------------------------------------------------------
|
||||||
# Sanitizes plugin output
|
# Sanitizes plugin output
|
||||||
def handleEmpty(input):
|
def handleEmpty(input):
|
||||||
@@ -55,6 +56,7 @@ def handleEmpty(input):
|
|||||||
input = input.replace('\n', '') # Removing new lines
|
input = input.replace('\n', '') # Removing new lines
|
||||||
return input
|
return input
|
||||||
|
|
||||||
|
|
||||||
# -------------------------------------------------------------------
|
# -------------------------------------------------------------------
|
||||||
# Sanitizes string
|
# Sanitizes string
|
||||||
def rmBadChars(input):
|
def rmBadChars(input):
|
||||||
@@ -64,23 +66,25 @@ def rmBadChars(input):
|
|||||||
|
|
||||||
return input
|
return input
|
||||||
|
|
||||||
|
|
||||||
# -------------------------------------------------------------------
|
# -------------------------------------------------------------------
|
||||||
# check if this is a router IP
|
# check if this is a router IP
|
||||||
def is_typical_router_ip(ip_address):
|
def is_typical_router_ip(ip_address):
|
||||||
# List of common default gateway IP addresses
|
# List of common default gateway IP addresses
|
||||||
common_router_ips = [
|
common_router_ips = [
|
||||||
"192.168.0.1", "192.168.1.1", "192.168.1.254", "192.168.0.254",
|
"192.168.0.1", "192.168.1.1", "192.168.1.254", "192.168.0.254",
|
||||||
"10.0.0.1", "10.1.1.1", "192.168.2.1", "192.168.10.1", "192.168.11.1",
|
"10.0.0.1", "10.1.1.1", "192.168.2.1", "192.168.10.1", "192.168.11.1",
|
||||||
"192.168.100.1", "192.168.101.1", "192.168.123.254", "192.168.223.1",
|
"192.168.100.1", "192.168.101.1", "192.168.123.254", "192.168.223.1",
|
||||||
"192.168.31.1", "192.168.8.1", "192.168.254.254", "192.168.50.1",
|
"192.168.31.1", "192.168.8.1", "192.168.254.254", "192.168.50.1",
|
||||||
"192.168.3.1", "192.168.4.1", "192.168.5.1", "192.168.9.1",
|
"192.168.3.1", "192.168.4.1", "192.168.5.1", "192.168.9.1",
|
||||||
"192.168.15.1", "192.168.16.1", "192.168.20.1", "192.168.30.1",
|
"192.168.15.1", "192.168.16.1", "192.168.20.1", "192.168.30.1",
|
||||||
"192.168.42.1", "192.168.62.1", "192.168.178.1", "192.168.1.1",
|
"192.168.42.1", "192.168.62.1", "192.168.178.1", "192.168.1.1",
|
||||||
"192.168.1.254", "192.168.0.1", "192.168.0.10", "192.168.0.100",
|
"192.168.1.254", "192.168.0.1", "192.168.0.10", "192.168.0.100",
|
||||||
"192.168.0.254"
|
"192.168.0.254"
|
||||||
]
|
]
|
||||||
|
|
||||||
|
return ip_address in common_router_ips
|
||||||
|
|
||||||
return ip_address in common_router_ips
|
|
||||||
|
|
||||||
# -------------------------------------------------------------------
|
# -------------------------------------------------------------------
|
||||||
# Check if a valid MAC address
|
# Check if a valid MAC address
|
||||||
@@ -94,6 +98,7 @@ def is_mac(input):
|
|||||||
|
|
||||||
return isMac
|
return isMac
|
||||||
|
|
||||||
|
|
||||||
# -------------------------------------------------------------------
|
# -------------------------------------------------------------------
|
||||||
def decodeBase64(inputParamBase64):
|
def decodeBase64(inputParamBase64):
|
||||||
|
|
||||||
@@ -102,14 +107,11 @@ def decodeBase64(inputParamBase64):
|
|||||||
print('[Plugins] Helper base64 input: ')
|
print('[Plugins] Helper base64 input: ')
|
||||||
print(input)
|
print(input)
|
||||||
|
|
||||||
|
|
||||||
# Extract the base64-encoded subnet information from the first element
|
# Extract the base64-encoded subnet information from the first element
|
||||||
# The format of the element is assumed to be like 'param=b<base64-encoded-data>'.
|
# The format of the element is assumed to be like 'param=b<base64-encoded-data>'.
|
||||||
|
|
||||||
# Printing the extracted base64-encoded information.
|
# Printing the extracted base64-encoded information.
|
||||||
mylog('debug', ['[Plugins] Helper base64 inputParamBase64: ', inputParamBase64])
|
mylog('debug', ['[Plugins] Helper base64 inputParamBase64: ', inputParamBase64])
|
||||||
|
|
||||||
|
|
||||||
# Decode the base64-encoded subnet information to get the actual subnet information in ASCII format.
|
# Decode the base64-encoded subnet information to get the actual subnet information in ASCII format.
|
||||||
result = base64.b64decode(inputParamBase64).decode('ascii')
|
result = base64.b64decode(inputParamBase64).decode('ascii')
|
||||||
|
|
||||||
@@ -118,6 +120,7 @@ def decodeBase64(inputParamBase64):
|
|||||||
|
|
||||||
return result
|
return result
|
||||||
|
|
||||||
|
|
||||||
# -------------------------------------------------------------------
|
# -------------------------------------------------------------------
|
||||||
def decode_settings_base64(encoded_str, convert_types=True):
|
def decode_settings_base64(encoded_str, convert_types=True):
|
||||||
"""
|
"""
|
||||||
@@ -180,6 +183,7 @@ def normalize_mac(mac):
|
|||||||
|
|
||||||
return normalized_mac
|
return normalized_mac
|
||||||
|
|
||||||
|
|
||||||
# -------------------------------------------------------------------
|
# -------------------------------------------------------------------
|
||||||
class Plugin_Object:
|
class Plugin_Object:
|
||||||
"""
|
"""
|
||||||
@@ -243,6 +247,7 @@ class Plugin_Object:
|
|||||||
)
|
)
|
||||||
return line
|
return line
|
||||||
|
|
||||||
|
|
||||||
class Plugin_Objects:
|
class Plugin_Objects:
|
||||||
"""
|
"""
|
||||||
Plugin_Objects is the class that manages and holds all the objects created by the plugin.
|
Plugin_Objects is the class that manages and holds all the objects created by the plugin.
|
||||||
@@ -303,7 +308,3 @@ class Plugin_Objects:
|
|||||||
|
|
||||||
def __len__(self):
|
def __len__(self):
|
||||||
return len(self.objects)
|
return len(self.objects)
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
from __future__ import unicode_literals
|
from __future__ import unicode_literals
|
||||||
import subprocess
|
import subprocess
|
||||||
@@ -10,12 +10,12 @@ import sys
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects, handleEmpty, normalize_mac
|
from plugin_helper import Plugin_Objects, handleEmpty, normalize_mac # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -28,7 +28,6 @@ pluginName = "SNMPDSC"
|
|||||||
LOG_PATH = logPath + '/plugins'
|
LOG_PATH = logPath + '/plugins'
|
||||||
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
||||||
|
|
||||||
# Workflow
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
mylog('verbose', ['[SNMPDSC] In script '])
|
mylog('verbose', ['[SNMPDSC] In script '])
|
||||||
@@ -36,9 +35,13 @@ def main():
|
|||||||
# init global variables
|
# init global variables
|
||||||
global snmpWalkCmds
|
global snmpWalkCmds
|
||||||
|
|
||||||
|
|
||||||
parser = argparse.ArgumentParser(description='This plugin is used to discover devices via the arp table(s) of a RFC1213 compliant router or switch.')
|
parser = argparse.ArgumentParser(description='This plugin is used to discover devices via the arp table(s) of a RFC1213 compliant router or switch.')
|
||||||
parser.add_argument('routers', action="store", help="IP(s) of routers, separated by comma (,) if passing multiple")
|
parser.add_argument(
|
||||||
|
'routers',
|
||||||
|
action="store",
|
||||||
|
help="IP(s) of routers, separated by comma (,) if passing multiple"
|
||||||
|
)
|
||||||
|
|
||||||
values = parser.parse_args()
|
values = parser.parse_args()
|
||||||
|
|
||||||
timeoutSetting = get_setting_value("SNMPDSC_RUN_TIMEOUT")
|
timeoutSetting = get_setting_value("SNMPDSC_RUN_TIMEOUT")
|
||||||
@@ -46,8 +49,7 @@ def main():
|
|||||||
plugin_objects = Plugin_Objects(RESULT_FILE)
|
plugin_objects = Plugin_Objects(RESULT_FILE)
|
||||||
|
|
||||||
if values.routers:
|
if values.routers:
|
||||||
snmpWalkCmds = values.routers.split('=')[1].replace('\'','')
|
snmpWalkCmds = values.routers.split('=')[1].replace('\'', '')
|
||||||
|
|
||||||
|
|
||||||
if ',' in snmpWalkCmds:
|
if ',' in snmpWalkCmds:
|
||||||
commands = snmpWalkCmds.split(',')
|
commands = snmpWalkCmds.split(',')
|
||||||
@@ -63,7 +65,12 @@ def main():
|
|||||||
probes = 1 # N probes
|
probes = 1 # N probes
|
||||||
|
|
||||||
for _ in range(probes):
|
for _ in range(probes):
|
||||||
output = subprocess.check_output (snmpwalkArgs, universal_newlines=True, stderr=subprocess.STDOUT, timeout=(timeoutSetting))
|
output = subprocess.check_output(
|
||||||
|
snmpwalkArgs,
|
||||||
|
universal_newlines=True,
|
||||||
|
stderr=subprocess.STDOUT,
|
||||||
|
timeout=(timeoutSetting)
|
||||||
|
)
|
||||||
|
|
||||||
mylog('verbose', ['[SNMPDSC] output: ', output])
|
mylog('verbose', ['[SNMPDSC] output: ', output])
|
||||||
|
|
||||||
@@ -86,7 +93,7 @@ def main():
|
|||||||
|
|
||||||
plugin_objects.add_object(
|
plugin_objects.add_object(
|
||||||
primaryId = handleEmpty(macAddress),
|
primaryId = handleEmpty(macAddress),
|
||||||
secondaryId = handleEmpty(ipAddress.strip()), # Remove leading/trailing spaces from IP
|
secondaryId = handleEmpty(ipAddress.strip()), # Remove leading/trailing spaces from IP
|
||||||
watched1 = '(unknown)',
|
watched1 = '(unknown)',
|
||||||
watched2 = handleEmpty(snmpwalkArgs[6]), # router IP
|
watched2 = handleEmpty(snmpwalkArgs[6]), # router IP
|
||||||
extra = handleEmpty(line),
|
extra = handleEmpty(line),
|
||||||
@@ -95,7 +102,6 @@ def main():
|
|||||||
else:
|
else:
|
||||||
mylog('verbose', ['[SNMPDSC] ipStr does not seem to contain a valid IP:', ipStr])
|
mylog('verbose', ['[SNMPDSC] ipStr does not seem to contain a valid IP:', ipStr])
|
||||||
|
|
||||||
|
|
||||||
elif line.startswith('ipNetToMediaPhysAddress'):
|
elif line.startswith('ipNetToMediaPhysAddress'):
|
||||||
# Format: snmpwalk -OXsq output
|
# Format: snmpwalk -OXsq output
|
||||||
parts = line.split()
|
parts = line.split()
|
||||||
@@ -120,7 +126,6 @@ def main():
|
|||||||
plugin_objects.write_result_file()
|
plugin_objects.write_result_file()
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
# BEGIN
|
# BEGIN
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
main()
|
main()
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
@@ -12,16 +12,16 @@ import base64
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects
|
from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.plugin_utils import get_plugins_configs, decode_and_rename_files
|
from utils.plugin_utils import get_plugins_configs, decode_and_rename_files # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from const import fullDbPath, logPath
|
from const import fullDbPath, logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.datetime_utils import timeNowDB
|
from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.crypto_utils import encrypt_data
|
from utils.crypto_utils import encrypt_data # noqa: E402 [flake8 lint suppression]
|
||||||
from messaging.in_app import write_notification
|
from messaging.in_app import write_notification # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -102,7 +102,6 @@ def main():
|
|||||||
else:
|
else:
|
||||||
mylog('verbose', [f'[{pluginName}] {file_path} not found'])
|
mylog('verbose', [f'[{pluginName}] {file_path} not found'])
|
||||||
|
|
||||||
|
|
||||||
# PUSHING/SENDING devices
|
# PUSHING/SENDING devices
|
||||||
if send_devices:
|
if send_devices:
|
||||||
|
|
||||||
@@ -150,7 +149,6 @@ def main():
|
|||||||
if lggr.isAbove('verbose'):
|
if lggr.isAbove('verbose'):
|
||||||
write_notification(message, 'info', timeNowDB())
|
write_notification(message, 'info', timeNowDB())
|
||||||
|
|
||||||
|
|
||||||
# Process any received data for the Device DB table (ONLY JSON)
|
# Process any received data for the Device DB table (ONLY JSON)
|
||||||
# Create the file path
|
# Create the file path
|
||||||
|
|
||||||
@@ -210,7 +208,6 @@ def main():
|
|||||||
cursor.execute(f'SELECT devMac FROM Devices WHERE devMac IN ({placeholders})', tuple(unique_mac_addresses))
|
cursor.execute(f'SELECT devMac FROM Devices WHERE devMac IN ({placeholders})', tuple(unique_mac_addresses))
|
||||||
existing_mac_addresses = set(row[0] for row in cursor.fetchall())
|
existing_mac_addresses = set(row[0] for row in cursor.fetchall())
|
||||||
|
|
||||||
|
|
||||||
# insert devices into the last_result.log and thus CurrentScan table to manage state
|
# insert devices into the last_result.log and thus CurrentScan table to manage state
|
||||||
for device in device_data:
|
for device in device_data:
|
||||||
# only insert devices taht were online and skip the root node to prevent IP flipping on the hub
|
# only insert devices taht were online and skip the root node to prevent IP flipping on the hub
|
||||||
@@ -258,7 +255,6 @@ def main():
|
|||||||
mylog('verbose', [message])
|
mylog('verbose', [message])
|
||||||
write_notification(message, 'info', timeNowDB())
|
write_notification(message, 'info', timeNowDB())
|
||||||
|
|
||||||
|
|
||||||
# Commit and close the connection
|
# Commit and close the connection
|
||||||
conn.commit()
|
conn.commit()
|
||||||
conn.close()
|
conn.close()
|
||||||
@@ -268,6 +264,7 @@ def main():
|
|||||||
|
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
|
|
||||||
# ------------------------------------------------------------------
|
# ------------------------------------------------------------------
|
||||||
# Data retrieval methods
|
# Data retrieval methods
|
||||||
api_endpoints = [
|
api_endpoints = [
|
||||||
@@ -275,6 +272,7 @@ api_endpoints = [
|
|||||||
"/plugins/sync/hub.php" # Legacy PHP endpoint
|
"/plugins/sync/hub.php" # Legacy PHP endpoint
|
||||||
]
|
]
|
||||||
|
|
||||||
|
|
||||||
# send data to the HUB
|
# send data to the HUB
|
||||||
def send_data(api_token, file_content, encryption_key, file_path, node_name, pref, hub_url):
|
def send_data(api_token, file_content, encryption_key, file_path, node_name, pref, hub_url):
|
||||||
"""Send encrypted data to HUB, preferring /sync endpoint and falling back to PHP version."""
|
"""Send encrypted data to HUB, preferring /sync endpoint and falling back to PHP version."""
|
||||||
@@ -345,6 +343,5 @@ def get_data(api_token, node_url):
|
|||||||
return ""
|
return ""
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
main()
|
main()
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
@@ -10,12 +10,11 @@ from unifi_sm_api.api import SiteManagerAPI
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects, decode_settings_base64
|
from plugin_helper import Plugin_Objects, decode_settings_base64 # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -50,11 +49,11 @@ def main():
|
|||||||
mylog('none', [f'[{pluginName}] Connecting to: {siteDict["UNIFIAPI_site_name"]}'])
|
mylog('none', [f'[{pluginName}] Connecting to: {siteDict["UNIFIAPI_site_name"]}'])
|
||||||
|
|
||||||
api = SiteManagerAPI(
|
api = SiteManagerAPI(
|
||||||
api_key=siteDict["UNIFIAPI_api_key"],
|
api_key=siteDict["UNIFIAPI_api_key"],
|
||||||
version=siteDict["UNIFIAPI_api_version"],
|
version=siteDict["UNIFIAPI_api_version"],
|
||||||
base_url=siteDict["UNIFIAPI_base_url"],
|
base_url=siteDict["UNIFIAPI_base_url"],
|
||||||
verify_ssl=siteDict["UNIFIAPI_verify_ssl"]
|
verify_ssl=siteDict["UNIFIAPI_verify_ssl"]
|
||||||
)
|
)
|
||||||
|
|
||||||
sites_resp = api.get_sites()
|
sites_resp = api.get_sites()
|
||||||
sites = sites_resp.get("data", [])
|
sites = sites_resp.get("data", [])
|
||||||
@@ -69,16 +68,16 @@ def main():
|
|||||||
|
|
||||||
# insert devices into the lats_result.log
|
# insert devices into the lats_result.log
|
||||||
for device in device_data:
|
for device in device_data:
|
||||||
plugin_objects.add_object(
|
plugin_objects.add_object(
|
||||||
primaryId = device['dev_mac'], # mac
|
primaryId = device['dev_mac'], # mac
|
||||||
secondaryId = device['dev_ip'], # IP
|
secondaryId = device['dev_ip'], # IP
|
||||||
watched1 = device['dev_name'], # name
|
watched1 = device['dev_name'], # name
|
||||||
watched2 = device['dev_type'], # device_type (AP/Switch etc)
|
watched2 = device['dev_type'], # device_type (AP/Switch etc)
|
||||||
watched3 = device['dev_connected'], # connectedAt or empty
|
watched3 = device['dev_connected'], # connectedAt or empty
|
||||||
watched4 = device['dev_parent_mac'],# parent_mac or "Internet"
|
watched4 = device['dev_parent_mac'], # parent_mac or "Internet"
|
||||||
extra = '',
|
extra = '',
|
||||||
foreignKey = device['dev_mac']
|
foreignKey = device['dev_mac']
|
||||||
)
|
)
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] New entries: "{len(device_data)}"'])
|
mylog('verbose', [f'[{pluginName}] New entries: "{len(device_data)}"'])
|
||||||
|
|
||||||
@@ -87,6 +86,7 @@ def main():
|
|||||||
|
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
|
|
||||||
# retrieve data
|
# retrieve data
|
||||||
def get_device_data(site, api):
|
def get_device_data(site, api):
|
||||||
device_data = []
|
device_data = []
|
||||||
@@ -146,8 +146,8 @@ def get_device_data(site, api):
|
|||||||
dev_parent_mac = resolve_parent_mac(uplinkDeviceId)
|
dev_parent_mac = resolve_parent_mac(uplinkDeviceId)
|
||||||
|
|
||||||
device_data.append({
|
device_data.append({
|
||||||
"dev_mac": dev_mac,
|
"dev_mac": dev_mac,
|
||||||
"dev_ip": dev_ip,
|
"dev_ip": dev_ip,
|
||||||
"dev_name": dev_name,
|
"dev_name": dev_name,
|
||||||
"dev_type": dev_type,
|
"dev_type": dev_type,
|
||||||
"dev_connected": dev_connected,
|
"dev_connected": dev_connected,
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
# Inspired by https://github.com/stevehoek/Pi.Alert
|
# Inspired by https://github.com/stevehoek/Pi.Alert
|
||||||
|
|
||||||
from __future__ import unicode_literals
|
from __future__ import unicode_literals
|
||||||
@@ -14,12 +14,12 @@ from pyunifi.controller import Controller
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects, rmBadChars, is_typical_router_ip, is_mac
|
from plugin_helper import Plugin_Objects, rmBadChars, is_typical_router_ip, is_mac # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value, normalize_string
|
from helper import get_setting_value, normalize_string # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -37,15 +37,10 @@ LOCK_FILE = os.path.join(LOG_PATH, f'full_run.{pluginName}.lock')
|
|||||||
urllib3.disable_warnings(InsecureRequestWarning)
|
urllib3.disable_warnings(InsecureRequestWarning)
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
# Workflow
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] In script'])
|
mylog('verbose', [f'[{pluginName}] In script'])
|
||||||
|
|
||||||
|
|
||||||
# init global variables
|
# init global variables
|
||||||
global UNIFI_USERNAME, UNIFI_PASSWORD, UNIFI_HOST, UNIFI_SITES, PORT, VERIFYSSL, VERSION, FULL_IMPORT
|
global UNIFI_USERNAME, UNIFI_PASSWORD, UNIFI_HOST, UNIFI_SITES, PORT, VERIFYSSL, VERSION, FULL_IMPORT
|
||||||
|
|
||||||
@@ -65,11 +60,10 @@ def main():
|
|||||||
|
|
||||||
plugin_objects.write_result_file()
|
plugin_objects.write_result_file()
|
||||||
|
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] Scan finished, found {len(plugin_objects)} devices'])
|
mylog('verbose', [f'[{pluginName}] Scan finished, found {len(plugin_objects)} devices'])
|
||||||
|
|
||||||
# .............................................
|
|
||||||
|
|
||||||
|
# .............................................
|
||||||
def get_entries(plugin_objects: Plugin_Objects) -> Plugin_Objects:
|
def get_entries(plugin_objects: Plugin_Objects) -> Plugin_Objects:
|
||||||
global VERIFYSSL
|
global VERIFYSSL
|
||||||
|
|
||||||
@@ -79,7 +73,6 @@ def get_entries(plugin_objects: Plugin_Objects) -> Plugin_Objects:
|
|||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] sites: {UNIFI_SITES}'])
|
mylog('verbose', [f'[{pluginName}] sites: {UNIFI_SITES}'])
|
||||||
|
|
||||||
|
|
||||||
if (VERIFYSSL.upper() == "TRUE"):
|
if (VERIFYSSL.upper() == "TRUE"):
|
||||||
VERIFYSSL = True
|
VERIFYSSL = True
|
||||||
else:
|
else:
|
||||||
@@ -114,7 +107,7 @@ def get_entries(plugin_objects: Plugin_Objects) -> Plugin_Objects:
|
|||||||
plugin_objects=plugin_objects,
|
plugin_objects=plugin_objects,
|
||||||
device_label='client',
|
device_label='client',
|
||||||
device_vendor="",
|
device_vendor="",
|
||||||
force_import=True # These are online clients, force import
|
force_import=True # These are online clients, force import
|
||||||
)
|
)
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] Found {len(plugin_objects)} Online Devices'])
|
mylog('verbose', [f'[{pluginName}] Found {len(plugin_objects)} Online Devices'])
|
||||||
@@ -154,11 +147,9 @@ def get_entries(plugin_objects: Plugin_Objects) -> Plugin_Objects:
|
|||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] Found {len(plugin_objects)} Users'])
|
mylog('verbose', [f'[{pluginName}] Found {len(plugin_objects)} Users'])
|
||||||
|
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] check if Lock file needs to be modified'])
|
mylog('verbose', [f'[{pluginName}] check if Lock file needs to be modified'])
|
||||||
set_lock_file_value(FULL_IMPORT, lock_file_value)
|
set_lock_file_value(FULL_IMPORT, lock_file_value)
|
||||||
|
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] Found {len(plugin_objects)} Clients overall'])
|
mylog('verbose', [f'[{pluginName}] Found {len(plugin_objects)} Clients overall'])
|
||||||
|
|
||||||
return plugin_objects
|
return plugin_objects
|
||||||
@@ -185,7 +176,7 @@ def collect_details(device_type, devices, online_macs, processed_macs, plugin_ob
|
|||||||
parentMac = 'Internet'
|
parentMac = 'Internet'
|
||||||
|
|
||||||
# Add object only if not processed
|
# Add object only if not processed
|
||||||
if macTmp not in processed_macs and ( status == 1 or force_import is True ):
|
if macTmp not in processed_macs and (status == 1 or force_import is True):
|
||||||
plugin_objects.add_object(
|
plugin_objects.add_object(
|
||||||
primaryId=macTmp,
|
primaryId=macTmp,
|
||||||
secondaryId=ipTmp,
|
secondaryId=ipTmp,
|
||||||
@@ -204,6 +195,7 @@ def collect_details(device_type, devices, online_macs, processed_macs, plugin_ob
|
|||||||
else:
|
else:
|
||||||
mylog('verbose', [f'[{pluginName}] Skipping, not a valid MAC address: {macTmp}'])
|
mylog('verbose', [f'[{pluginName}] Skipping, not a valid MAC address: {macTmp}'])
|
||||||
|
|
||||||
|
|
||||||
# -----------------------------------------------------------------------------
|
# -----------------------------------------------------------------------------
|
||||||
def get_unifi_val(obj, key, default='null'):
|
def get_unifi_val(obj, key, default='null'):
|
||||||
if isinstance(obj, dict):
|
if isinstance(obj, dict):
|
||||||
@@ -212,7 +204,7 @@ def get_unifi_val(obj, key, default='null'):
|
|||||||
for k, v in obj.items():
|
for k, v in obj.items():
|
||||||
if isinstance(v, dict):
|
if isinstance(v, dict):
|
||||||
result = get_unifi_val(v, key, default)
|
result = get_unifi_val(v, key, default)
|
||||||
if result not in ['','None', None, 'null']:
|
if result not in ['', 'None', None, 'null']:
|
||||||
return result
|
return result
|
||||||
|
|
||||||
mylog('trace', [f'[{pluginName}] Value not found for key "{key}" in obj "{json.dumps(obj)}"'])
|
mylog('trace', [f'[{pluginName}] Value not found for key "{key}" in obj "{json.dumps(obj)}"'])
|
||||||
@@ -226,6 +218,7 @@ def get_name(*names: str) -> str:
|
|||||||
return rmBadChars(name)
|
return rmBadChars(name)
|
||||||
return 'null'
|
return 'null'
|
||||||
|
|
||||||
|
|
||||||
# -----------------------------------------------------------------------------
|
# -----------------------------------------------------------------------------
|
||||||
def get_parent_mac(*macs: str) -> str:
|
def get_parent_mac(*macs: str) -> str:
|
||||||
for mac in macs:
|
for mac in macs:
|
||||||
@@ -233,6 +226,7 @@ def get_parent_mac(*macs: str) -> str:
|
|||||||
return mac
|
return mac
|
||||||
return 'null'
|
return 'null'
|
||||||
|
|
||||||
|
|
||||||
# -----------------------------------------------------------------------------
|
# -----------------------------------------------------------------------------
|
||||||
def get_port(*ports: str) -> str:
|
def get_port(*ports: str) -> str:
|
||||||
for port in ports:
|
for port in ports:
|
||||||
@@ -240,12 +234,6 @@ def get_port(*ports: str) -> str:
|
|||||||
return port
|
return port
|
||||||
return 'null'
|
return 'null'
|
||||||
|
|
||||||
# -----------------------------------------------------------------------------
|
|
||||||
def get_port(*macs: str) -> str:
|
|
||||||
for mac in macs:
|
|
||||||
if mac and mac != 'null':
|
|
||||||
return mac
|
|
||||||
return 'null'
|
|
||||||
|
|
||||||
# -----------------------------------------------------------------------------
|
# -----------------------------------------------------------------------------
|
||||||
def get_ip(*ips: str) -> str:
|
def get_ip(*ips: str) -> str:
|
||||||
@@ -271,7 +259,7 @@ def set_lock_file_value(config_value: str, lock_file_value: bool) -> None:
|
|||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] Setting lock value for "full import" to {out}'])
|
mylog('verbose', [f'[{pluginName}] Setting lock value for "full import" to {out}'])
|
||||||
with open(LOCK_FILE, 'w') as lock_file:
|
with open(LOCK_FILE, 'w') as lock_file:
|
||||||
lock_file.write(str(out))
|
lock_file.write(str(out))
|
||||||
|
|
||||||
|
|
||||||
# -----------------------------------------------------------------------------
|
# -----------------------------------------------------------------------------
|
||||||
@@ -286,15 +274,16 @@ def read_lock_file() -> bool:
|
|||||||
|
|
||||||
# -----------------------------------------------------------------------------
|
# -----------------------------------------------------------------------------
|
||||||
def check_full_run_state(config_value: str, lock_file_value: bool) -> bool:
|
def check_full_run_state(config_value: str, lock_file_value: bool) -> bool:
|
||||||
if config_value == 'always' or (config_value == 'once' and lock_file_value == False):
|
if config_value == 'always' or (config_value == 'once' and lock_file_value is False):
|
||||||
mylog('verbose', [f'[{pluginName}] Full import needs to be done: config_value: {config_value} and lock_file_value: {lock_file_value}'])
|
mylog('verbose', [f'[{pluginName}] Full import needs to be done: config_value: {config_value} and lock_file_value: {lock_file_value}'])
|
||||||
return True
|
return True
|
||||||
else:
|
else:
|
||||||
mylog('verbose', [f'[{pluginName}] Full import NOT needed: config_value: {config_value} and lock_file_value: {lock_file_value}'])
|
mylog('verbose', [f'[{pluginName}] Full import NOT needed: config_value: {config_value} and lock_file_value: {lock_file_value}'])
|
||||||
return False
|
return False
|
||||||
|
|
||||||
#===============================================================================
|
|
||||||
|
# ===============================================================================
|
||||||
# BEGIN
|
# BEGIN
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
main()
|
main()
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
@@ -9,13 +9,13 @@ import sqlite3
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects, handleEmpty
|
from plugin_helper import Plugin_Objects, handleEmpty # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath, applicationPath, fullDbPath
|
from const import logPath, applicationPath, fullDbPath # noqa: E402 [flake8 lint suppression]
|
||||||
from scan.device_handling import query_MAC_vendor
|
from scan.device_handling import query_MAC_vendor # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -25,11 +25,11 @@ Logger(get_setting_value('LOG_LEVEL'))
|
|||||||
|
|
||||||
pluginName = 'VNDRPDT'
|
pluginName = 'VNDRPDT'
|
||||||
|
|
||||||
|
|
||||||
LOG_PATH = logPath + '/plugins'
|
LOG_PATH = logPath + '/plugins'
|
||||||
LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
|
LOG_FILE = os.path.join(LOG_PATH, f'script.{pluginName}.log')
|
||||||
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
|
|
||||||
mylog('verbose', ['[VNDRPDT] In script'])
|
mylog('verbose', ['[VNDRPDT] In script'])
|
||||||
@@ -48,9 +48,10 @@ def main():
|
|||||||
|
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
#===============================================================================
|
|
||||||
|
# ===============================================================================
|
||||||
# Update device vendors database
|
# Update device vendors database
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
def update_vendor_database():
|
def update_vendor_database():
|
||||||
|
|
||||||
# Update vendors DB (iab oui)
|
# Update vendors DB (iab oui)
|
||||||
@@ -60,30 +61,29 @@ def update_vendor_database():
|
|||||||
# Execute command
|
# Execute command
|
||||||
try:
|
try:
|
||||||
# try runnning a subprocess safely
|
# try runnning a subprocess safely
|
||||||
update_output = subprocess.check_output (update_args)
|
subprocess.check_output(update_args)
|
||||||
except subprocess.CalledProcessError as e:
|
except subprocess.CalledProcessError as e:
|
||||||
# An error occured, handle it
|
# An error occured, handle it
|
||||||
mylog('verbose', [' FAILED: Updating vendors DB, set LOG_LEVEL=debug for more info'])
|
mylog('verbose', [' FAILED: Updating vendors DB, set LOG_LEVEL=debug for more info'])
|
||||||
mylog('verbose', [e.output])
|
mylog('verbose', [e.output])
|
||||||
|
|
||||||
|
|
||||||
# ------------------------------------------------------------------------------
|
# ------------------------------------------------------------------------------
|
||||||
# resolve missing vendors
|
# resolve missing vendors
|
||||||
def update_vendors (dbPath, plugin_objects):
|
def update_vendors(dbPath, plugin_objects):
|
||||||
|
|
||||||
# Connect to the App SQLite database
|
# Connect to the App SQLite database
|
||||||
conn = sqlite3.connect(dbPath)
|
conn = sqlite3.connect(dbPath)
|
||||||
sql = conn.cursor()
|
sql = conn.cursor()
|
||||||
|
|
||||||
# Initialize variables
|
# Initialize variables
|
||||||
recordsToUpdate = []
|
|
||||||
ignored = 0
|
ignored = 0
|
||||||
notFound = 0
|
notFound = 0
|
||||||
|
|
||||||
|
|
||||||
mylog('verbose', [' Searching devices vendor'])
|
mylog('verbose', [' Searching devices vendor'])
|
||||||
|
|
||||||
# Get devices without a vendor
|
# Get devices without a vendor
|
||||||
sql.execute ("""SELECT
|
sql.execute("""SELECT
|
||||||
devMac,
|
devMac,
|
||||||
devLastIP,
|
devLastIP,
|
||||||
devName,
|
devName,
|
||||||
@@ -103,7 +103,7 @@ def update_vendors (dbPath, plugin_objects):
|
|||||||
# All devices loop
|
# All devices loop
|
||||||
for device in devices:
|
for device in devices:
|
||||||
# Search vendor in HW Vendors DB
|
# Search vendor in HW Vendors DB
|
||||||
vendor = query_MAC_vendor (device[0])
|
vendor = query_MAC_vendor(device[0])
|
||||||
if vendor == -1 :
|
if vendor == -1 :
|
||||||
notFound += 1
|
notFound += 1
|
||||||
elif vendor == -2 :
|
elif vendor == -2 :
|
||||||
@@ -124,15 +124,13 @@ def update_vendors (dbPath, plugin_objects):
|
|||||||
mylog('verbose', [" Devices Ignored : ", ignored])
|
mylog('verbose', [" Devices Ignored : ", ignored])
|
||||||
mylog('verbose', [" Devices with missing vendor : ", len(devices)])
|
mylog('verbose', [" Devices with missing vendor : ", len(devices)])
|
||||||
mylog('verbose', [" Vendors Not Found : ", notFound])
|
mylog('verbose', [" Vendors Not Found : ", notFound])
|
||||||
mylog('verbose', [" Vendors updated : ", len(plugin_objects) ])
|
mylog('verbose', [" Vendors updated : ", len(plugin_objects)])
|
||||||
|
|
||||||
|
|
||||||
return plugin_objects
|
return plugin_objects
|
||||||
|
|
||||||
|
|
||||||
|
# ===============================================================================
|
||||||
#===============================================================================
|
|
||||||
# BEGIN
|
# BEGIN
|
||||||
#===============================================================================
|
# ===============================================================================
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
main()
|
main()
|
||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
@@ -9,13 +9,13 @@ from wakeonlan import send_magic_packet
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects
|
from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from database import DB
|
from database import DB # noqa: E402 [flake8 lint suppression]
|
||||||
from models.device_instance import DeviceInstance
|
from models.device_instance import DeviceInstance # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -34,7 +34,6 @@ RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
|||||||
plugin_objects = Plugin_Objects(RESULT_FILE)
|
plugin_objects = Plugin_Objects(RESULT_FILE)
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
mylog('none', [f'[{pluginName}] In script'])
|
mylog('none', [f'[{pluginName}] In script'])
|
||||||
|
|
||||||
@@ -95,6 +94,7 @@ def main():
|
|||||||
|
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
|
|
||||||
# wake
|
# wake
|
||||||
def execute(port, ip, mac, name):
|
def execute(port, ip, mac, name):
|
||||||
|
|
||||||
@@ -113,5 +113,6 @@ def execute(port, ip, mac, name):
|
|||||||
# Return the data result
|
# Return the data result
|
||||||
return result
|
return result
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
main()
|
main()
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
# Based on the work of https://github.com/leiweibau/Pi.Alert
|
# Based on the work of https://github.com/leiweibau/Pi.Alert
|
||||||
|
|
||||||
import requests
|
import requests
|
||||||
@@ -12,12 +12,12 @@ from urllib3.exceptions import InsecureRequestWarning
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from plugin_helper import Plugin_Objects
|
from plugin_helper import Plugin_Objects # noqa: E402 [flake8 lint suppression]
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
import conf # noqa: E402 [flake8 lint suppression]
|
||||||
from pytz import timezone
|
from pytz import timezone # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# Make sure the TIMEZONE for logging is correct
|
# Make sure the TIMEZONE for logging is correct
|
||||||
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
conf.tz = timezone(get_setting_value('TIMEZONE'))
|
||||||
@@ -30,16 +30,15 @@ pluginName = 'WEBMON'
|
|||||||
LOG_PATH = logPath + '/plugins'
|
LOG_PATH = logPath + '/plugins'
|
||||||
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
RESULT_FILE = os.path.join(LOG_PATH, f'last_result.{pluginName}.log')
|
||||||
|
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] In script'])
|
mylog('verbose', [f'[{pluginName}] In script'])
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
|
|
||||||
values = get_setting_value('WEBMON_urls_to_check')
|
values = get_setting_value('WEBMON_urls_to_check')
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] Checking URLs: {values}'])
|
mylog('verbose', [f'[{pluginName}] Checking URLs: {values}'])
|
||||||
|
|
||||||
|
|
||||||
if len(values) > 0:
|
if len(values) > 0:
|
||||||
plugin_objects = Plugin_Objects(RESULT_FILE)
|
plugin_objects = Plugin_Objects(RESULT_FILE)
|
||||||
# plugin_objects = service_monitoring(values.urls.split('=')[1].split(','), plugin_objects)
|
# plugin_objects = service_monitoring(values.urls.split('=')[1].split(','), plugin_objects)
|
||||||
@@ -48,6 +47,7 @@ def main():
|
|||||||
else:
|
else:
|
||||||
return
|
return
|
||||||
|
|
||||||
|
|
||||||
def check_services_health(site):
|
def check_services_health(site):
|
||||||
|
|
||||||
mylog('verbose', [f'[{pluginName}] Checking {site}'])
|
mylog('verbose', [f'[{pluginName}] Checking {site}'])
|
||||||
@@ -79,6 +79,7 @@ def check_services_health(site):
|
|||||||
|
|
||||||
return status, latency
|
return status, latency
|
||||||
|
|
||||||
|
|
||||||
def service_monitoring(urls, plugin_objects):
|
def service_monitoring(urls, plugin_objects):
|
||||||
for site in urls:
|
for site in urls:
|
||||||
status, latency = check_services_health(site)
|
status, latency = check_services_health(site)
|
||||||
@@ -94,7 +95,6 @@ def service_monitoring(urls, plugin_objects):
|
|||||||
)
|
)
|
||||||
return plugin_objects
|
return plugin_objects
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
sys.exit(main())
|
sys.exit(main())
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python3
|
# !/usr/bin/env python3
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
|
|
||||||
#!/usr/bin/env python3
|
# !/usr/bin/env python3
|
||||||
# -*- coding: utf-8 -*-
|
# -*- coding: utf-8 -*-
|
||||||
"""
|
"""
|
||||||
NetAlertX-New-Devices-Checkmk-Script
|
NetAlertX-New-Devices-Checkmk-Script
|
||||||
@@ -19,6 +19,7 @@ import subprocess
|
|||||||
import json
|
import json
|
||||||
import os
|
import os
|
||||||
|
|
||||||
|
|
||||||
def check_new_devices():
|
def check_new_devices():
|
||||||
# Get API path from environment variable, fallback to /tmp/api
|
# Get API path from environment variable, fallback to /tmp/api
|
||||||
api_path = os.environ.get('NETALERTX_API', '/tmp/api')
|
api_path = os.environ.get('NETALERTX_API', '/tmp/api')
|
||||||
@@ -73,6 +74,6 @@ def check_new_devices():
|
|||||||
)
|
)
|
||||||
print(f"1 NetAlertX_New_Devices - WARNING - Found {len(new_devices)} new device(s): {device_list_str}")
|
print(f"1 NetAlertX_New_Devices - WARNING - Found {len(new_devices)} new device(s): {device_list_str}")
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
check_new_devices()
|
check_new_devices()
|
||||||
|
|
||||||
|
|||||||
@@ -1,8 +1,8 @@
|
|||||||
#!/usr/bin/env python3
|
# !/usr/bin/env python3
|
||||||
import subprocess
|
import subprocess
|
||||||
import sys
|
|
||||||
import os
|
import os
|
||||||
|
|
||||||
|
|
||||||
def run_sqlite_command(command):
|
def run_sqlite_command(command):
|
||||||
# Use environment variable with fallback
|
# Use environment variable with fallback
|
||||||
db_path = os.path.join(
|
db_path = os.path.join(
|
||||||
@@ -19,6 +19,7 @@ def run_sqlite_command(command):
|
|||||||
print(f"Error executing command: {e}")
|
print(f"Error executing command: {e}")
|
||||||
return None
|
return None
|
||||||
|
|
||||||
|
|
||||||
def check_and_clean_device():
|
def check_and_clean_device():
|
||||||
while True:
|
while True:
|
||||||
print("\nDevice Cleanup Tool")
|
print("\nDevice Cleanup Tool")
|
||||||
@@ -113,5 +114,6 @@ def check_and_clean_device():
|
|||||||
else:
|
else:
|
||||||
print("\nInvalid option, please try again")
|
print("\nInvalid option, please try again")
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
check_and_clean_device()
|
check_and_clean_device()
|
||||||
|
|||||||
@@ -1,6 +1,5 @@
|
|||||||
#!/usr/bin/env python3
|
#!/usr/bin/env python3
|
||||||
import paramiko
|
import paramiko
|
||||||
import re
|
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
import argparse
|
import argparse
|
||||||
import sys
|
import sys
|
||||||
@@ -8,6 +7,9 @@ from pathlib import Path
|
|||||||
import time
|
import time
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
|
logger = None
|
||||||
|
|
||||||
|
|
||||||
def setup_logging(debug=False):
|
def setup_logging(debug=False):
|
||||||
"""Configure logging based on debug flag."""
|
"""Configure logging based on debug flag."""
|
||||||
level = logging.DEBUG if debug else logging.INFO
|
level = logging.DEBUG if debug else logging.INFO
|
||||||
@@ -18,6 +20,7 @@ def setup_logging(debug=False):
|
|||||||
)
|
)
|
||||||
return logging.getLogger(__name__)
|
return logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
def parse_timestamp(date_str):
|
def parse_timestamp(date_str):
|
||||||
"""Convert OPNsense timestamp to Unix epoch time."""
|
"""Convert OPNsense timestamp to Unix epoch time."""
|
||||||
try:
|
try:
|
||||||
@@ -27,7 +30,7 @@ def parse_timestamp(date_str):
|
|||||||
dt = datetime.strptime(clean_date, '%Y/%m/%d %H:%M:%S')
|
dt = datetime.strptime(clean_date, '%Y/%m/%d %H:%M:%S')
|
||||||
return int(dt.timestamp())
|
return int(dt.timestamp())
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.error(f"Failed to parse timestamp: {date_str}")
|
logger.error(f"Failed to parse timestamp: {date_str} ({e})")
|
||||||
return None
|
return None
|
||||||
|
|
||||||
|
|
||||||
@@ -39,8 +42,14 @@ def get_lease_file(hostname, username, password=None, key_filename=None, port=22
|
|||||||
|
|
||||||
try:
|
try:
|
||||||
logger.debug(f"Attempting to connect to {hostname}:{port} as {username}")
|
logger.debug(f"Attempting to connect to {hostname}:{port} as {username}")
|
||||||
ssh.connect(hostname, port=port, username=username,
|
|
||||||
password=password, key_filename=key_filename)
|
ssh.connect(
|
||||||
|
hostname,
|
||||||
|
port=port,
|
||||||
|
username=username,
|
||||||
|
password=password,
|
||||||
|
key_filename=key_filename
|
||||||
|
)
|
||||||
|
|
||||||
# Get an interactive shell session
|
# Get an interactive shell session
|
||||||
logger.debug("Opening interactive SSH channel")
|
logger.debug("Opening interactive SSH channel")
|
||||||
@@ -75,10 +84,23 @@ def get_lease_file(hostname, username, password=None, key_filename=None, port=22
|
|||||||
# Clean up the output by removing the command echo and shell prompts
|
# Clean up the output by removing the command echo and shell prompts
|
||||||
lines = output.split('\n')
|
lines = output.split('\n')
|
||||||
# Remove first line (command echo) and any lines containing shell prompts
|
# Remove first line (command echo) and any lines containing shell prompts
|
||||||
cleaned_lines = [line for line in lines
|
# cleaned_lines = [line for line in lines
|
||||||
if not line.strip().startswith(command.strip())
|
# if not line.strip().startswith(command.strip()) and not line.strip().endswith('> ') and not line.strip().endswith('# ')]
|
||||||
and not line.strip().endswith('> ')
|
cmd = command.strip()
|
||||||
and not line.strip().endswith('# ')]
|
|
||||||
|
cleaned_lines = []
|
||||||
|
for line in lines:
|
||||||
|
stripped = line.strip()
|
||||||
|
|
||||||
|
if stripped.startswith(cmd):
|
||||||
|
continue
|
||||||
|
if stripped.endswith('> '):
|
||||||
|
continue
|
||||||
|
if stripped.endswith('# '):
|
||||||
|
continue
|
||||||
|
|
||||||
|
cleaned_lines.append(line)
|
||||||
|
|
||||||
cleaned_output = '\n'.join(cleaned_lines)
|
cleaned_output = '\n'.join(cleaned_lines)
|
||||||
|
|
||||||
logger.debug(f"Final cleaned output length: {len(cleaned_output)} characters")
|
logger.debug(f"Final cleaned output length: {len(cleaned_output)} characters")
|
||||||
@@ -156,9 +178,7 @@ def parse_lease_file(lease_content):
|
|||||||
|
|
||||||
# Filter only active leases
|
# Filter only active leases
|
||||||
active_leases = [lease for lease in leases
|
active_leases = [lease for lease in leases
|
||||||
if lease.get('state') == 'active'
|
if lease.get('state') == 'active' and 'mac' in lease and 'ip' in lease]
|
||||||
and 'mac' in lease
|
|
||||||
and 'ip' in lease]
|
|
||||||
|
|
||||||
logger.debug(f"Found {len(active_leases)} active leases out of {len(leases)} total leases")
|
logger.debug(f"Found {len(active_leases)} active leases out of {len(leases)} total leases")
|
||||||
logger.debug("Active leases:")
|
logger.debug("Active leases:")
|
||||||
@@ -206,6 +226,7 @@ def convert_to_dnsmasq(leases):
|
|||||||
|
|
||||||
return dnsmasq_lines
|
return dnsmasq_lines
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
parser = argparse.ArgumentParser(description='Convert OPNsense DHCP leases to dnsmasq format')
|
parser = argparse.ArgumentParser(description='Convert OPNsense DHCP leases to dnsmasq format')
|
||||||
parser.add_argument('--host', required=True, help='OPNsense hostname or IP')
|
parser.add_argument('--host', required=True, help='OPNsense hostname or IP')
|
||||||
@@ -219,6 +240,7 @@ def main():
|
|||||||
args = parser.parse_args()
|
args = parser.parse_args()
|
||||||
|
|
||||||
# Setup logging
|
# Setup logging
|
||||||
|
global logger
|
||||||
logger = setup_logging(args.debug)
|
logger = setup_logging(args.debug)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
@@ -255,5 +277,6 @@ def main():
|
|||||||
logger.error(f"Error: {str(e)}")
|
logger.error(f"Error: {str(e)}")
|
||||||
sys.exit(1)
|
sys.exit(1)
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
main()
|
main()
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
#
|
#
|
||||||
# -------------------------------------------------------------------------------
|
# -------------------------------------------------------------------------------
|
||||||
# NetAlertX v2.70 / 2021-02-01
|
# NetAlertX v2.70 / 2021-02-01
|
||||||
@@ -22,9 +22,9 @@ from pathlib import Path
|
|||||||
|
|
||||||
# Register NetAlertX modules
|
# Register NetAlertX modules
|
||||||
import conf
|
import conf
|
||||||
from const import *
|
from const import fullConfPath, sql_new_devices
|
||||||
from logger import mylog
|
from logger import mylog
|
||||||
from helper import filePermissions
|
from helper import filePermissions
|
||||||
from utils.datetime_utils import timeNowTZ
|
from utils.datetime_utils import timeNowTZ
|
||||||
from app_state import updateState
|
from app_state import updateState
|
||||||
from api import update_api
|
from api import update_api
|
||||||
@@ -111,7 +111,7 @@ def main():
|
|||||||
loop_start_time = conf.loop_start_time # TODO fix
|
loop_start_time = conf.loop_start_time # TODO fix
|
||||||
|
|
||||||
# Handle plugins executed ONCE
|
# Handle plugins executed ONCE
|
||||||
if conf.plugins_once_run == False:
|
if conf.plugins_once_run is False:
|
||||||
pm.run_plugin_scripts("once")
|
pm.run_plugin_scripts("once")
|
||||||
conf.plugins_once_run = True
|
conf.plugins_once_run = True
|
||||||
|
|
||||||
@@ -146,7 +146,7 @@ def main():
|
|||||||
processScan = updateState("Check scan").processScan
|
processScan = updateState("Check scan").processScan
|
||||||
mylog("debug", [f"[MAIN] processScan: {processScan}"])
|
mylog("debug", [f"[MAIN] processScan: {processScan}"])
|
||||||
|
|
||||||
if processScan == True:
|
if processScan is True:
|
||||||
mylog("debug", "[MAIN] start processing scan results")
|
mylog("debug", "[MAIN] start processing scan results")
|
||||||
process_scan(db)
|
process_scan(db)
|
||||||
updateState("Scan processed", None, None, None, None, False)
|
updateState("Scan processed", None, None, None, None, False)
|
||||||
|
|||||||
@@ -1,3 +1,4 @@
|
|||||||
|
# !/usr/bin/env python
|
||||||
import json
|
import json
|
||||||
import time
|
import time
|
||||||
import threading
|
import threading
|
||||||
@@ -145,8 +146,7 @@ class api_endpoint_class:
|
|||||||
self.needsUpdate = True
|
self.needsUpdate = True
|
||||||
# Only update changeDetectedWhen if it hasn't been set recently
|
# Only update changeDetectedWhen if it hasn't been set recently
|
||||||
if not self.changeDetectedWhen or current_time > (
|
if not self.changeDetectedWhen or current_time > (
|
||||||
self.changeDetectedWhen
|
self.changeDetectedWhen + datetime.timedelta(seconds=self.debounce_interval)
|
||||||
+ datetime.timedelta(seconds=self.debounce_interval)
|
|
||||||
):
|
):
|
||||||
self.changeDetectedWhen = (
|
self.changeDetectedWhen = (
|
||||||
current_time # Set timestamp for change detection
|
current_time # Set timestamp for change detection
|
||||||
@@ -164,8 +164,7 @@ class api_endpoint_class:
|
|||||||
self.needsUpdate = True
|
self.needsUpdate = True
|
||||||
# Only update changeDetectedWhen if it hasn't been set recently
|
# Only update changeDetectedWhen if it hasn't been set recently
|
||||||
if not self.changeDetectedWhen or current_time > (
|
if not self.changeDetectedWhen or current_time > (
|
||||||
self.changeDetectedWhen
|
self.changeDetectedWhen + datetime.timedelta(seconds=self.debounce_interval)
|
||||||
+ datetime.timedelta(seconds=self.debounce_interval)
|
|
||||||
):
|
):
|
||||||
self.changeDetectedWhen = (
|
self.changeDetectedWhen = (
|
||||||
current_time # Initialize timestamp for new endpoint
|
current_time # Initialize timestamp for new endpoint
|
||||||
@@ -180,17 +179,15 @@ class api_endpoint_class:
|
|||||||
current_time = timeNowTZ()
|
current_time = timeNowTZ()
|
||||||
|
|
||||||
# Debugging info to understand the issue
|
# Debugging info to understand the issue
|
||||||
# mylog('debug', [f'[API] api_endpoint_class: {self.fileName} is_ad_hoc_user_event {self.is_ad_hoc_user_event} last_update_time={self.last_update_time}, debounce time={self.last_update_time + datetime.timedelta(seconds=self.debounce_interval)}.'])
|
# mylog('debug', [f'[API] api_endpoint_class: {self.fileName} is_ad_hoc_user_event
|
||||||
|
# {self.is_ad_hoc_user_event} last_update_time={self.last_update_time},
|
||||||
|
# debounce time={self.last_update_time + datetime.timedelta(seconds=self.debounce_interval)}.'])
|
||||||
|
|
||||||
# Only attempt to write if the debounce time has passed
|
# Only attempt to write if the debounce time has passed
|
||||||
if forceUpdate == True or (
|
if forceUpdate is True or (
|
||||||
self.needsUpdate
|
self.needsUpdate and (
|
||||||
and (
|
self.changeDetectedWhen is None or current_time > (
|
||||||
self.changeDetectedWhen is None
|
self.changeDetectedWhen + datetime.timedelta(seconds=self.debounce_interval)
|
||||||
or current_time
|
|
||||||
> (
|
|
||||||
self.changeDetectedWhen
|
|
||||||
+ datetime.timedelta(seconds=self.debounce_interval)
|
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
):
|
):
|
||||||
|
|||||||
@@ -9,25 +9,68 @@ from flask_cors import CORS
|
|||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
||||||
sys.path.extend([f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from logger import mylog
|
from logger import mylog # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from db.db_helper import get_date_from_period
|
from db.db_helper import get_date_from_period # noqa: E402 [flake8 lint suppression]
|
||||||
from app_state import updateState
|
from app_state import updateState # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
|
from .graphql_endpoint import devicesSchema # noqa: E402 [flake8 lint suppression]
|
||||||
from .graphql_endpoint import devicesSchema
|
from .device_endpoint import ( # noqa: E402 [flake8 lint suppression]
|
||||||
from .device_endpoint import get_device_data, set_device_data, delete_device, delete_device_events, reset_device_props, copy_device, update_device_column
|
get_device_data,
|
||||||
from .devices_endpoint import get_all_devices, delete_unknown_devices, delete_all_with_empty_macs, delete_devices, export_devices, import_csv, devices_totals, devices_by_status
|
set_device_data,
|
||||||
from .events_endpoint import delete_events, delete_events_older_than, get_events, create_event, get_events_totals
|
delete_device,
|
||||||
from .history_endpoint import delete_online_history
|
delete_device_events,
|
||||||
from .prometheus_endpoint import get_metric_stats
|
reset_device_props,
|
||||||
from .sessions_endpoint import get_sessions, delete_session, create_session, get_sessions_calendar, get_device_sessions, get_session_events
|
copy_device,
|
||||||
from .nettools_endpoint import wakeonlan, traceroute, speedtest, nslookup, nmap_scan, internet_info
|
update_device_column
|
||||||
from .dbquery_endpoint import read_query, write_query, update_query, delete_query
|
)
|
||||||
from .sync_endpoint import handle_sync_post, handle_sync_get
|
from .devices_endpoint import ( # noqa: E402 [flake8 lint suppression]
|
||||||
from .logs_endpoint import clean_log
|
get_all_devices,
|
||||||
from models.user_events_queue_instance import UserEventsQueueInstance
|
delete_unknown_devices,
|
||||||
from messaging.in_app import write_notification, mark_all_notifications_read, delete_notifications, get_unread_notifications, delete_notification, mark_notification_as_read
|
delete_all_with_empty_macs,
|
||||||
|
delete_devices,
|
||||||
|
export_devices,
|
||||||
|
import_csv,
|
||||||
|
devices_totals,
|
||||||
|
devices_by_status
|
||||||
|
)
|
||||||
|
from .events_endpoint import ( # noqa: E402 [flake8 lint suppression]
|
||||||
|
delete_events,
|
||||||
|
delete_events_older_than,
|
||||||
|
get_events,
|
||||||
|
create_event,
|
||||||
|
get_events_totals
|
||||||
|
)
|
||||||
|
from .history_endpoint import delete_online_history # noqa: E402 [flake8 lint suppression]
|
||||||
|
from .prometheus_endpoint import get_metric_stats # noqa: E402 [flake8 lint suppression]
|
||||||
|
from .sessions_endpoint import ( # noqa: E402 [flake8 lint suppression]
|
||||||
|
get_sessions,
|
||||||
|
delete_session,
|
||||||
|
create_session,
|
||||||
|
get_sessions_calendar,
|
||||||
|
get_device_sessions,
|
||||||
|
get_session_events
|
||||||
|
)
|
||||||
|
from .nettools_endpoint import ( # noqa: E402 [flake8 lint suppression]
|
||||||
|
wakeonlan,
|
||||||
|
traceroute,
|
||||||
|
speedtest,
|
||||||
|
nslookup,
|
||||||
|
nmap_scan,
|
||||||
|
internet_info
|
||||||
|
)
|
||||||
|
from .dbquery_endpoint import read_query, write_query, update_query, delete_query # noqa: E402 [flake8 lint suppression]
|
||||||
|
from .sync_endpoint import handle_sync_post, handle_sync_get # noqa: E402 [flake8 lint suppression]
|
||||||
|
from .logs_endpoint import clean_log # noqa: E402 [flake8 lint suppression]
|
||||||
|
from models.user_events_queue_instance import UserEventsQueueInstance # noqa: E402 [flake8 lint suppression]
|
||||||
|
from messaging.in_app import ( # noqa: E402 [flake8 lint suppression]
|
||||||
|
write_notification,
|
||||||
|
mark_all_notifications_read,
|
||||||
|
delete_notifications,
|
||||||
|
get_unread_notifications,
|
||||||
|
delete_notification,
|
||||||
|
mark_notification_as_read
|
||||||
|
)
|
||||||
|
|
||||||
# Flask application
|
# Flask application
|
||||||
app = Flask(__name__)
|
app = Flask(__name__)
|
||||||
@@ -50,6 +93,7 @@ CORS(
|
|||||||
allow_headers=["Authorization", "Content-Type"],
|
allow_headers=["Authorization", "Content-Type"],
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
# -------------------------------------------------------------------
|
# -------------------------------------------------------------------
|
||||||
# Custom handler for 404 - Route not found
|
# Custom handler for 404 - Route not found
|
||||||
# -------------------------------------------------------------------
|
# -------------------------------------------------------------------
|
||||||
@@ -350,7 +394,7 @@ def dbquery_write():
|
|||||||
data = request.get_json() or {}
|
data = request.get_json() or {}
|
||||||
raw_sql_b64 = data.get("rawSql")
|
raw_sql_b64 = data.get("rawSql")
|
||||||
if not raw_sql_b64:
|
if not raw_sql_b64:
|
||||||
return jsonify({"success": False, "message": "ERROR: Missing parameters", "error": "rawSql is required"}), 400
|
return jsonify({"success": False, "message": "ERROR: Missing parameters", "error": "rawSql is required"}), 400
|
||||||
|
|
||||||
return write_query(raw_sql_b64)
|
return write_query(raw_sql_b64)
|
||||||
|
|
||||||
@@ -363,7 +407,13 @@ def dbquery_update():
|
|||||||
data = request.get_json() or {}
|
data = request.get_json() or {}
|
||||||
required = ["columnName", "id", "dbtable", "columns", "values"]
|
required = ["columnName", "id", "dbtable", "columns", "values"]
|
||||||
if not all(data.get(k) for k in required):
|
if not all(data.get(k) for k in required):
|
||||||
return jsonify({"success": False, "message": "ERROR: Missing parameters", "error": "Missing required 'columnName', 'id', 'dbtable', 'columns', or 'values' query parameter"}), 400
|
return jsonify(
|
||||||
|
{
|
||||||
|
"success": False,
|
||||||
|
"message": "ERROR: Missing parameters",
|
||||||
|
"error": "Missing required 'columnName', 'id', 'dbtable', 'columns', or 'values' query parameter"
|
||||||
|
}
|
||||||
|
), 400
|
||||||
|
|
||||||
return update_query(
|
return update_query(
|
||||||
column_name=data["columnName"],
|
column_name=data["columnName"],
|
||||||
@@ -418,6 +468,7 @@ def api_clean_log():
|
|||||||
|
|
||||||
return clean_log(file)
|
return clean_log(file)
|
||||||
|
|
||||||
|
|
||||||
@app.route("/logs/add-to-execution-queue", methods=["POST"])
|
@app.route("/logs/add-to-execution-queue", methods=["POST"])
|
||||||
def api_add_to_execution_queue():
|
def api_add_to_execution_queue():
|
||||||
|
|
||||||
@@ -673,6 +724,7 @@ def api_mark_notification_read(guid):
|
|||||||
else:
|
else:
|
||||||
return jsonify({"success": False, "message": "ERROR", "error": result.get("error")}), 500
|
return jsonify({"success": False, "message": "ERROR", "error": result.get("error")}), 500
|
||||||
|
|
||||||
|
|
||||||
# --------------------------
|
# --------------------------
|
||||||
# SYNC endpoint
|
# SYNC endpoint
|
||||||
# --------------------------
|
# --------------------------
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import base64
|
import base64
|
||||||
@@ -9,7 +9,7 @@ from flask import jsonify
|
|||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from database import get_temp_db_connection
|
from database import get_temp_db_connection # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
|
|
||||||
def read_query(raw_sql_b64):
|
def read_query(raw_sql_b64):
|
||||||
|
|||||||
@@ -1,18 +1,17 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
from datetime import datetime
|
|
||||||
from flask import jsonify, request
|
from flask import jsonify, request
|
||||||
|
|
||||||
# Register NetAlertX directories
|
# Register NetAlertX directories
|
||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from database import get_temp_db_connection
|
from database import get_temp_db_connection # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import is_random_mac, get_setting_value
|
from helper import is_random_mac, get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.datetime_utils import timeNowDB, format_date
|
from utils.datetime_utils import timeNowDB, format_date # noqa: E402 [flake8 lint suppression]
|
||||||
from db.db_helper import row_to_json, get_date_from_period
|
from db.db_helper import row_to_json, get_date_from_period # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# --------------------------
|
# --------------------------
|
||||||
# Device Endpoints Functions
|
# Device Endpoints Functions
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import base64
|
import base64
|
||||||
@@ -14,16 +14,13 @@ from logger import mylog
|
|||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from database import get_temp_db_connection
|
from database import get_temp_db_connection # noqa: E402 [flake8 lint suppression]
|
||||||
from db.db_helper import get_table_json, get_device_condition_by_status
|
from db.db_helper import get_table_json, get_device_condition_by_status # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.datetime_utils import format_date
|
|
||||||
|
|
||||||
|
|
||||||
# --------------------------
|
# --------------------------
|
||||||
# Device Endpoints Functions
|
# Device Endpoints Functions
|
||||||
# --------------------------
|
# --------------------------
|
||||||
|
|
||||||
|
|
||||||
def get_all_devices():
|
def get_all_devices():
|
||||||
"""Retrieve all devices from the database."""
|
"""Retrieve all devices from the database."""
|
||||||
conn = get_temp_db_connection()
|
conn = get_temp_db_connection()
|
||||||
@@ -139,7 +136,6 @@ def export_devices(export_format):
|
|||||||
def import_csv(file_storage=None):
|
def import_csv(file_storage=None):
|
||||||
data = ""
|
data = ""
|
||||||
skipped = []
|
skipped = []
|
||||||
error = None
|
|
||||||
|
|
||||||
# 1. Try JSON `content` (base64-encoded CSV)
|
# 1. Try JSON `content` (base64-encoded CSV)
|
||||||
if request.is_json and request.json.get("content"):
|
if request.is_json and request.json.get("content"):
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
@@ -9,10 +9,10 @@ from flask import jsonify
|
|||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from database import get_temp_db_connection
|
from database import get_temp_db_connection # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import is_random_mac, mylog
|
from helper import mylog # noqa: E402 [flake8 lint suppression]
|
||||||
from db.db_helper import row_to_json, get_date_from_period
|
from db.db_helper import row_to_json, get_date_from_period # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.datetime_utils import format_date, format_date_iso, format_event_date, ensure_datetime
|
from utils.datetime_utils import ensure_datetime # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
|
|
||||||
# --------------------------
|
# --------------------------
|
||||||
|
|||||||
@@ -1,5 +1,7 @@
|
|||||||
import graphene
|
import graphene
|
||||||
from graphene import ObjectType, String, Int, Boolean, List, Field, InputObjectType, Argument
|
from graphene import (
|
||||||
|
ObjectType, String, Int, Boolean, List, Field, InputObjectType, Argument
|
||||||
|
)
|
||||||
import json
|
import json
|
||||||
import sys
|
import sys
|
||||||
import os
|
import os
|
||||||
@@ -8,9 +10,9 @@ import os
|
|||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
||||||
sys.path.extend([f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from logger import mylog
|
from logger import mylog # noqa: E402 [flake8 lint suppression]
|
||||||
from const import apiPath
|
from const import apiPath # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import (
|
from helper import ( # noqa: E402 [flake8 lint suppression]
|
||||||
is_random_mac,
|
is_random_mac,
|
||||||
get_number_of_children,
|
get_number_of_children,
|
||||||
format_ip_long,
|
format_ip_long,
|
||||||
@@ -113,10 +115,12 @@ class SettingResult(ObjectType):
|
|||||||
|
|
||||||
# --- LANGSTRINGS ---
|
# --- LANGSTRINGS ---
|
||||||
|
|
||||||
|
|
||||||
# In-memory cache for lang strings
|
# In-memory cache for lang strings
|
||||||
_langstrings_cache = {} # caches lists per file (core JSON or plugin)
|
_langstrings_cache = {} # caches lists per file (core JSON or plugin)
|
||||||
_langstrings_cache_mtime = {} # tracks last modified times
|
_langstrings_cache_mtime = {} # tracks last modified times
|
||||||
|
|
||||||
|
|
||||||
# LangString ObjectType
|
# LangString ObjectType
|
||||||
class LangString(ObjectType):
|
class LangString(ObjectType):
|
||||||
langCode = String()
|
langCode = String()
|
||||||
@@ -128,6 +132,7 @@ class LangStringResult(ObjectType):
|
|||||||
langStrings = List(LangString)
|
langStrings = List(LangString)
|
||||||
count = Int()
|
count = Int()
|
||||||
|
|
||||||
|
|
||||||
# Define Query Type with Pagination Support
|
# Define Query Type with Pagination Support
|
||||||
class Query(ObjectType):
|
class Query(ObjectType):
|
||||||
# --- DEVICES ---
|
# --- DEVICES ---
|
||||||
@@ -184,31 +189,39 @@ class Query(ObjectType):
|
|||||||
if (device.get("devParentRelType") not in hidden_relationships)
|
if (device.get("devParentRelType") not in hidden_relationships)
|
||||||
]
|
]
|
||||||
|
|
||||||
devices_data = [
|
filtered = []
|
||||||
device
|
|
||||||
for device in devices_data
|
for device in devices_data:
|
||||||
if (
|
is_online = (
|
||||||
(
|
device["devPresentLastScan"] == 1 and "online" in allowed_statuses
|
||||||
device["devPresentLastScan"] == 1
|
|
||||||
and "online" in allowed_statuses
|
|
||||||
)
|
|
||||||
or (device["devIsNew"] == 1 and "new" in allowed_statuses)
|
|
||||||
or (
|
|
||||||
device["devPresentLastScan"] == 0
|
|
||||||
and device["devAlertDown"]
|
|
||||||
and "down" in allowed_statuses
|
|
||||||
)
|
|
||||||
or (
|
|
||||||
device["devPresentLastScan"] == 0
|
|
||||||
and "offline" in allowed_statuses
|
|
||||||
)
|
|
||||||
and device["devIsArchived"] == 0
|
|
||||||
or (
|
|
||||||
device["devIsArchived"] == 1
|
|
||||||
and "archived" in allowed_statuses
|
|
||||||
)
|
|
||||||
)
|
)
|
||||||
]
|
|
||||||
|
is_new = (
|
||||||
|
device["devIsNew"] == 1 and "new" in allowed_statuses
|
||||||
|
)
|
||||||
|
|
||||||
|
is_down = (
|
||||||
|
device["devPresentLastScan"] == 0 and device["devAlertDown"] and "down" in allowed_statuses
|
||||||
|
)
|
||||||
|
|
||||||
|
is_offline = (
|
||||||
|
device["devPresentLastScan"] == 0 and "offline" in allowed_statuses
|
||||||
|
)
|
||||||
|
|
||||||
|
is_archived = (
|
||||||
|
device["devIsArchived"] == 1 and "archived" in allowed_statuses
|
||||||
|
)
|
||||||
|
|
||||||
|
# Matches if not archived and status matches OR it is archived and allowed
|
||||||
|
matches = (
|
||||||
|
(is_online or is_new or is_down or is_offline) and device["devIsArchived"] == 0
|
||||||
|
) or is_archived
|
||||||
|
|
||||||
|
if matches:
|
||||||
|
filtered.append(device)
|
||||||
|
|
||||||
|
devices_data = filtered
|
||||||
|
|
||||||
elif status == "connected":
|
elif status == "connected":
|
||||||
devices_data = [
|
devices_data = [
|
||||||
device
|
device
|
||||||
@@ -257,8 +270,7 @@ class Query(ObjectType):
|
|||||||
devices_data = [
|
devices_data = [
|
||||||
device
|
device
|
||||||
for device in devices_data
|
for device in devices_data
|
||||||
if str(device.get(filter.filterColumn, "")).lower()
|
if str(device.get(filter.filterColumn, "")).lower() == str(filter.filterValue).lower()
|
||||||
== str(filter.filterValue).lower()
|
|
||||||
]
|
]
|
||||||
|
|
||||||
# Search data if a search term is provided
|
# Search data if a search term is provided
|
||||||
@@ -437,11 +449,11 @@ class Query(ObjectType):
|
|||||||
if en_fallback:
|
if en_fallback:
|
||||||
langStrings[i] = en_fallback[0]
|
langStrings[i] = en_fallback[0]
|
||||||
|
|
||||||
mylog('trace', f'[graphql_schema] Collected {len(langStrings)} language strings '
|
mylog('trace', f'[graphql_schema] Collected {len(langStrings)} language strings (langCode={langCode}, key={langStringKey}, fallback_to_en={fallback_to_en})')
|
||||||
f'(langCode={langCode}, key={langStringKey}, fallback_to_en={fallback_to_en})')
|
|
||||||
|
|
||||||
return LangStringResult(langStrings=langStrings, count=len(langStrings))
|
return LangStringResult(langStrings=langStrings, count=len(langStrings))
|
||||||
|
|
||||||
|
|
||||||
# helps sorting inconsistent dataset mixed integers and strings
|
# helps sorting inconsistent dataset mixed integers and strings
|
||||||
def mixed_type_sort_key(value):
|
def mixed_type_sort_key(value):
|
||||||
if value is None or value == "":
|
if value is None or value == "":
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
@@ -8,7 +8,7 @@ from flask import jsonify
|
|||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from database import get_temp_db_connection
|
from database import get_temp_db_connection # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
|
|
||||||
# --------------------------------------------------
|
# --------------------------------------------------
|
||||||
|
|||||||
@@ -3,18 +3,18 @@ import sys
|
|||||||
from flask import jsonify
|
from flask import jsonify
|
||||||
|
|
||||||
# Register NetAlertX directories
|
# Register NetAlertX directories
|
||||||
INSTALL_PATH="/app"
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from const import logPath
|
from const import logPath # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.datetime_utils import timeNowDB
|
from messaging.in_app import write_notification # noqa: E402 [flake8 lint suppression]
|
||||||
from messaging.in_app import write_notification
|
|
||||||
|
|
||||||
# Make sure log level is initialized correctly
|
# Make sure log level is initialized correctly
|
||||||
Logger(get_setting_value('LOG_LEVEL'))
|
Logger(get_setting_value('LOG_LEVEL'))
|
||||||
|
|
||||||
|
|
||||||
def clean_log(log_file):
|
def clean_log(log_file):
|
||||||
"""
|
"""
|
||||||
Purge the content of an allowed log file within the /app/log/ directory.
|
Purge the content of an allowed log file within the /app/log/ directory.
|
||||||
@@ -55,4 +55,3 @@ def clean_log(log_file):
|
|||||||
mylog('none', [msg])
|
mylog('none', [msg])
|
||||||
write_notification(msg, 'interrupt')
|
write_notification(msg, 'interrupt')
|
||||||
return jsonify({"success": False, "message": msg}), 500
|
return jsonify({"success": False, "message": msg}), 500
|
||||||
|
|
||||||
|
|||||||
@@ -6,8 +6,8 @@ import os
|
|||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
||||||
sys.path.extend([f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from logger import mylog
|
from logger import mylog # noqa: E402 [flake8 lint suppression]
|
||||||
from const import apiPath
|
from const import apiPath # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
|
|
||||||
def escape_label_value(val):
|
def escape_label_value(val):
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import os
|
import os
|
||||||
import sqlite3
|
import sqlite3
|
||||||
@@ -9,10 +9,10 @@ from flask import jsonify
|
|||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from database import get_temp_db_connection
|
from database import get_temp_db_connection # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import is_random_mac, get_setting_value, mylog, format_ip_long
|
from helper import get_setting_value, format_ip_long # noqa: E402 [flake8 lint suppression]
|
||||||
from db.db_helper import row_to_json, get_date_from_period
|
from db.db_helper import get_date_from_period # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.datetime_utils import format_date_iso, format_event_date, format_date_diff, parse_datetime, format_date
|
from utils.datetime_utils import format_date_iso, format_event_date, format_date_diff, format_date # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
|
|
||||||
# --------------------------
|
# --------------------------
|
||||||
@@ -162,8 +162,7 @@ def get_sessions_calendar(start_date, end_date):
|
|||||||
|
|
||||||
# Determine color
|
# Determine color
|
||||||
if (
|
if (
|
||||||
row["ses_EventTypeConnection"] == "<missing event>"
|
row["ses_EventTypeConnection"] == "<missing event>" or row["ses_EventTypeDisconnection"] == "<missing event>"
|
||||||
or row["ses_EventTypeDisconnection"] == "<missing event>"
|
|
||||||
):
|
):
|
||||||
color = "#f39c12"
|
color = "#f39c12"
|
||||||
elif row["ses_StillConnected"] == 1:
|
elif row["ses_StillConnected"] == 1:
|
||||||
@@ -337,8 +336,7 @@ def get_session_events(event_type, period_date):
|
|||||||
sql = sql_events
|
sql = sql_events
|
||||||
elif event_type == "sessions":
|
elif event_type == "sessions":
|
||||||
sql = (
|
sql = (
|
||||||
sql_sessions
|
sql_sessions + f"""
|
||||||
+ f"""
|
|
||||||
WHERE (
|
WHERE (
|
||||||
ses_DateTimeConnection >= {period_date}
|
ses_DateTimeConnection >= {period_date}
|
||||||
OR ses_DateTimeDisconnection >= {period_date}
|
OR ses_DateTimeDisconnection >= {period_date}
|
||||||
@@ -348,8 +346,7 @@ def get_session_events(event_type, period_date):
|
|||||||
)
|
)
|
||||||
elif event_type == "missing":
|
elif event_type == "missing":
|
||||||
sql = (
|
sql = (
|
||||||
sql_sessions
|
sql_sessions + f"""
|
||||||
+ f"""
|
|
||||||
WHERE (
|
WHERE (
|
||||||
(ses_DateTimeConnection IS NULL AND ses_DateTimeDisconnection >= {period_date})
|
(ses_DateTimeConnection IS NULL AND ses_DateTimeDisconnection >= {period_date})
|
||||||
OR (ses_DateTimeDisconnection IS NULL AND ses_StillConnected = 0 AND ses_DateTimeConnection >= {period_date})
|
OR (ses_DateTimeDisconnection IS NULL AND ses_StillConnected = 0 AND ses_DateTimeConnection >= {period_date})
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
import os
|
import os
|
||||||
import json
|
import json
|
||||||
|
|
||||||
from const import *
|
from const import applicationPath, apiPath
|
||||||
from logger import mylog
|
from logger import mylog
|
||||||
from helper import checkNewVersion
|
from helper import checkNewVersion
|
||||||
from utils.datetime_utils import timeNowDB, timeNow
|
from utils.datetime_utils import timeNowDB, timeNow
|
||||||
@@ -32,14 +32,17 @@ class app_state_class:
|
|||||||
isNewVersionChecked (int): Timestamp of last version check.
|
isNewVersionChecked (int): Timestamp of last version check.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(self, currentState=None,
|
def __init__(
|
||||||
settingsSaved=None,
|
self,
|
||||||
settingsImported=None,
|
currentState=None,
|
||||||
showSpinner=None,
|
settingsSaved=None,
|
||||||
graphQLServerStarted=0,
|
settingsImported=None,
|
||||||
processScan=False,
|
showSpinner=None,
|
||||||
pluginsStates=None,
|
graphQLServerStarted=0,
|
||||||
appVersion=None):
|
processScan=False,
|
||||||
|
pluginsStates=None,
|
||||||
|
appVersion=None
|
||||||
|
):
|
||||||
"""
|
"""
|
||||||
Initialize the application state, optionally overwriting previous values.
|
Initialize the application state, optionally overwriting previous values.
|
||||||
|
|
||||||
@@ -84,7 +87,7 @@ class app_state_class:
|
|||||||
self.currentState = previousState.get("currentState", "Init")
|
self.currentState = previousState.get("currentState", "Init")
|
||||||
self.pluginsStates = previousState.get("pluginsStates", {})
|
self.pluginsStates = previousState.get("pluginsStates", {})
|
||||||
self.appVersion = previousState.get("appVersion", "")
|
self.appVersion = previousState.get("appVersion", "")
|
||||||
else: # init first time values
|
else: # init first time values
|
||||||
self.settingsSaved = 0
|
self.settingsSaved = 0
|
||||||
self.settingsImported = 0
|
self.settingsImported = 0
|
||||||
self.showSpinner = False
|
self.showSpinner = False
|
||||||
@@ -182,14 +185,16 @@ def updateState(newState = None,
|
|||||||
Returns:
|
Returns:
|
||||||
app_state_class: Updated state object.
|
app_state_class: Updated state object.
|
||||||
"""
|
"""
|
||||||
return app_state_class( newState,
|
return app_state_class(
|
||||||
settingsSaved,
|
newState,
|
||||||
settingsImported,
|
settingsSaved,
|
||||||
showSpinner,
|
settingsImported,
|
||||||
graphQLServerStarted,
|
showSpinner,
|
||||||
processScan,
|
graphQLServerStarted,
|
||||||
pluginsStates,
|
processScan,
|
||||||
appVersion)
|
pluginsStates,
|
||||||
|
appVersion
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
# -------------------------------------------------------------------------------
|
# -------------------------------------------------------------------------------
|
||||||
|
|||||||
@@ -180,7 +180,7 @@ class DB:
|
|||||||
# Init the AppEvent database table
|
# Init the AppEvent database table
|
||||||
AppEvent_obj(self)
|
AppEvent_obj(self)
|
||||||
|
|
||||||
# #-------------------------------------------------------------------------------
|
# # -------------------------------------------------------------------------------
|
||||||
# def get_table_as_json(self, sqlQuery):
|
# def get_table_as_json(self, sqlQuery):
|
||||||
|
|
||||||
# # mylog('debug',[ '[Database] - get_table_as_json - Query: ', sqlQuery])
|
# # mylog('debug',[ '[Database] - get_table_as_json - Query: ', sqlQuery])
|
||||||
|
|||||||
@@ -6,8 +6,8 @@ import os
|
|||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
||||||
sys.path.extend([f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from helper import if_byte_then_to_str
|
from helper import if_byte_then_to_str # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog
|
from logger import mylog # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
|
|
||||||
# -------------------------------------------------------------------------------
|
# -------------------------------------------------------------------------------
|
||||||
|
|||||||
@@ -5,8 +5,8 @@ import os
|
|||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
||||||
sys.path.extend([f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from logger import mylog
|
from logger import mylog # noqa: E402 [flake8 lint suppression]
|
||||||
from messaging.in_app import write_notification
|
from messaging.in_app import write_notification # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
|
|
||||||
def ensure_column(sql, table: str, column_name: str, column_type: str) -> bool:
|
def ensure_column(sql, table: str, column_name: str, column_type: str) -> bool:
|
||||||
|
|||||||
@@ -18,7 +18,7 @@ from typing import Dict, List, Tuple, Any, Optional
|
|||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
||||||
sys.path.extend([f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from logger import mylog
|
from logger import mylog # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
|
|
||||||
class SafeConditionBuilder:
|
class SafeConditionBuilder:
|
||||||
@@ -494,8 +494,6 @@ class SafeConditionBuilder:
|
|||||||
if logical_op and not self._validate_logical_operator(logical_op):
|
if logical_op and not self._validate_logical_operator(logical_op):
|
||||||
raise ValueError(f"Invalid logical operator: {logical_op}")
|
raise ValueError(f"Invalid logical operator: {logical_op}")
|
||||||
|
|
||||||
# Parse values from the IN clause
|
|
||||||
values = []
|
|
||||||
# Simple regex to extract quoted values
|
# Simple regex to extract quoted values
|
||||||
value_pattern = r"'([^']*)'"
|
value_pattern = r"'([^']*)'"
|
||||||
matches = re.findall(value_pattern, values_str)
|
matches = re.findall(value_pattern, values_str)
|
||||||
|
|||||||
@@ -7,25 +7,22 @@ import os
|
|||||||
import re
|
import re
|
||||||
import unicodedata
|
import unicodedata
|
||||||
import subprocess
|
import subprocess
|
||||||
import pytz
|
|
||||||
import json
|
import json
|
||||||
import requests
|
import requests
|
||||||
import base64
|
import base64
|
||||||
import hashlib
|
import hashlib
|
||||||
import random
|
import random
|
||||||
import email
|
|
||||||
import string
|
import string
|
||||||
import ipaddress
|
import ipaddress
|
||||||
|
|
||||||
import conf
|
import conf
|
||||||
from const import *
|
from const import applicationPath, fullConfPath, fullDbPath, dbPath, confPath, apiPath
|
||||||
from logger import mylog, logResult
|
from logger import mylog, logResult
|
||||||
|
|
||||||
# Register NetAlertX directories using runtime configuration
|
# Register NetAlertX directories using runtime configuration
|
||||||
INSTALL_PATH = applicationPath
|
INSTALL_PATH = applicationPath
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
# -------------------------------------------------------------------------------
|
# -------------------------------------------------------------------------------
|
||||||
# File system permission handling
|
# File system permission handling
|
||||||
# -------------------------------------------------------------------------------
|
# -------------------------------------------------------------------------------
|
||||||
@@ -58,12 +55,6 @@ def checkPermissionsOK():
|
|||||||
return (confR_access, dbR_access)
|
return (confR_access, dbR_access)
|
||||||
|
|
||||||
|
|
||||||
# -------------------------------------------------------------------------------
|
|
||||||
def fixPermissions():
|
|
||||||
# Try fixing access rights if needed
|
|
||||||
chmodCommands = []
|
|
||||||
|
|
||||||
|
|
||||||
# -------------------------------------------------------------------------------
|
# -------------------------------------------------------------------------------
|
||||||
def initialiseFile(pathToCheck, defaultFile):
|
def initialiseFile(pathToCheck, defaultFile):
|
||||||
# if file not readable (missing?) try to copy over the backed-up (default) one
|
# if file not readable (missing?) try to copy over the backed-up (default) one
|
||||||
@@ -71,9 +62,7 @@ def initialiseFile(pathToCheck, defaultFile):
|
|||||||
mylog(
|
mylog(
|
||||||
"none",
|
"none",
|
||||||
[
|
[
|
||||||
"[Setup] ("
|
"[Setup] (" + pathToCheck + ") file is not readable or missing. Trying to copy over the default one."
|
||||||
+ pathToCheck
|
|
||||||
+ ") file is not readable or missing. Trying to copy over the default one."
|
|
||||||
],
|
],
|
||||||
)
|
)
|
||||||
try:
|
try:
|
||||||
@@ -89,22 +78,14 @@ def initialiseFile(pathToCheck, defaultFile):
|
|||||||
mylog(
|
mylog(
|
||||||
"none",
|
"none",
|
||||||
[
|
[
|
||||||
"[Setup] ⚠ ERROR copying ("
|
"[Setup] ⚠ ERROR copying (" + defaultFile + ") to (" + pathToCheck + "). Make sure the app has Read & Write access to the parent directory."
|
||||||
+ defaultFile
|
|
||||||
+ ") to ("
|
|
||||||
+ pathToCheck
|
|
||||||
+ "). Make sure the app has Read & Write access to the parent directory."
|
|
||||||
],
|
],
|
||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
mylog(
|
mylog(
|
||||||
"none",
|
"none",
|
||||||
[
|
[
|
||||||
"[Setup] ("
|
"[Setup] (" + defaultFile + ") copied over successfully to (" + pathToCheck + ")."
|
||||||
+ defaultFile
|
|
||||||
+ ") copied over successfully to ("
|
|
||||||
+ pathToCheck
|
|
||||||
+ ")."
|
|
||||||
],
|
],
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -116,10 +97,7 @@ def initialiseFile(pathToCheck, defaultFile):
|
|||||||
mylog(
|
mylog(
|
||||||
"none",
|
"none",
|
||||||
[
|
[
|
||||||
"[Setup] ⚠ ERROR copying ("
|
"[Setup] ⚠ ERROR copying (" + defaultFile + "). Make sure the app has Read & Write access to " + pathToCheck
|
||||||
+ defaultFile
|
|
||||||
+ "). Make sure the app has Read & Write access to "
|
|
||||||
+ pathToCheck
|
|
||||||
],
|
],
|
||||||
)
|
)
|
||||||
mylog("none", [e.output])
|
mylog("none", [e.output])
|
||||||
@@ -130,16 +108,13 @@ def filePermissions():
|
|||||||
# check and initialize .conf
|
# check and initialize .conf
|
||||||
(confR_access, dbR_access) = checkPermissionsOK() # Initial check
|
(confR_access, dbR_access) = checkPermissionsOK() # Initial check
|
||||||
|
|
||||||
if confR_access == False:
|
if confR_access is False:
|
||||||
initialiseFile(fullConfPath, f"{INSTALL_PATH}/back/app.conf")
|
initialiseFile(fullConfPath, f"{INSTALL_PATH}/back/app.conf")
|
||||||
|
|
||||||
# check and initialize .db
|
# check and initialize .db
|
||||||
if dbR_access == False:
|
if dbR_access is False:
|
||||||
initialiseFile(fullDbPath, f"{INSTALL_PATH}/back/app.db")
|
initialiseFile(fullDbPath, f"{INSTALL_PATH}/back/app.db")
|
||||||
|
|
||||||
# last attempt
|
|
||||||
fixPermissions()
|
|
||||||
|
|
||||||
|
|
||||||
# -------------------------------------------------------------------------------
|
# -------------------------------------------------------------------------------
|
||||||
# File manipulation methods
|
# File manipulation methods
|
||||||
@@ -413,17 +388,12 @@ def setting_value_to_python_type(set_type, set_value):
|
|||||||
value = set_value
|
value = set_value
|
||||||
|
|
||||||
elif (
|
elif (
|
||||||
dataType == "string"
|
dataType == "string" and elementType == "input" and any(opt.get("readonly") == "true" for opt in elementOptions)
|
||||||
and elementType == "input"
|
|
||||||
and any(opt.get("readonly") == "true" for opt in elementOptions)
|
|
||||||
):
|
):
|
||||||
value = reverseTransformers(str(set_value), transformers)
|
value = reverseTransformers(str(set_value), transformers)
|
||||||
|
|
||||||
elif (
|
elif (
|
||||||
dataType == "string"
|
dataType == "string" and elementType == "input" and any(opt.get("type") == "password" for opt in elementOptions) and "sha256" in transformers
|
||||||
and elementType == "input"
|
|
||||||
and any(opt.get("type") == "password" for opt in elementOptions)
|
|
||||||
and "sha256" in transformers
|
|
||||||
):
|
):
|
||||||
value = hashlib.sha256(set_value.encode()).hexdigest()
|
value = hashlib.sha256(set_value.encode()).hexdigest()
|
||||||
|
|
||||||
@@ -602,23 +572,23 @@ def normalize_string(text):
|
|||||||
# -------------------------------------------------------------------------------
|
# -------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
|
||||||
# -------------------------------------------------------------------------------------------
|
# # -------------------------------------------------------------------------------------------
|
||||||
def is_random_mac(mac: str) -> bool:
|
# def is_random_mac(mac: str) -> bool:
|
||||||
"""Determine if a MAC address is random, respecting user-defined prefixes not to mark as random."""
|
# """Determine if a MAC address is random, respecting user-defined prefixes not to mark as random."""
|
||||||
|
|
||||||
is_random = mac[1].upper() in ["2", "6", "A", "E"]
|
# is_random = mac[1].upper() in ["2", "6", "A", "E"]
|
||||||
|
|
||||||
# Get prefixes from settings
|
# # Get prefixes from settings
|
||||||
prefixes = get_setting_value("UI_NOT_RANDOM_MAC")
|
# prefixes = get_setting_value("UI_NOT_RANDOM_MAC")
|
||||||
|
|
||||||
# If detected as random, make sure it doesn't start with a prefix the user wants to exclude
|
# # If detected as random, make sure it doesn't start with a prefix the user wants to exclude
|
||||||
if is_random:
|
# if is_random:
|
||||||
for prefix in prefixes:
|
# for prefix in prefixes:
|
||||||
if mac.upper().startswith(prefix.upper()):
|
# if mac.upper().startswith(prefix.upper()):
|
||||||
is_random = False
|
# is_random = False
|
||||||
break
|
# break
|
||||||
|
|
||||||
return is_random
|
# return is_random
|
||||||
|
|
||||||
|
|
||||||
# -------------------------------------------------------------------------------------------
|
# -------------------------------------------------------------------------------------------
|
||||||
@@ -653,6 +623,7 @@ def extract_ip_addresses(text):
|
|||||||
# -------------------------------------------------------------------------------
|
# -------------------------------------------------------------------------------
|
||||||
# Helper function to determine if a MAC address is random
|
# Helper function to determine if a MAC address is random
|
||||||
def is_random_mac(mac):
|
def is_random_mac(mac):
|
||||||
|
"""Determine if a MAC address is random, respecting user-defined prefixes not to mark as random."""
|
||||||
# Check if second character matches "2", "6", "A", "E" (case insensitive)
|
# Check if second character matches "2", "6", "A", "E" (case insensitive)
|
||||||
is_random = mac[1].upper() in ["2", "6", "A", "E"]
|
is_random = mac[1].upper() in ["2", "6", "A", "E"]
|
||||||
|
|
||||||
@@ -773,7 +744,6 @@ def getBuildTimeStampAndVersion():
|
|||||||
return tuple(results)
|
return tuple(results)
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
# -------------------------------------------------------------------------------
|
# -------------------------------------------------------------------------------
|
||||||
def checkNewVersion():
|
def checkNewVersion():
|
||||||
mylog("debug", ["[Version check] Checking if new version available"])
|
mylog("debug", ["[Version check] Checking if new version available"])
|
||||||
|
|||||||
@@ -9,8 +9,8 @@ import re
|
|||||||
|
|
||||||
# Register NetAlertX libraries
|
# Register NetAlertX libraries
|
||||||
import conf
|
import conf
|
||||||
from const import fullConfPath, applicationPath, fullConfFolder, default_tz
|
from const import fullConfPath, fullConfFolder, default_tz
|
||||||
from helper import getBuildTimeStampAndVersion, fixPermissions, collect_lang_strings, updateSubnets, isJsonObject, setting_value_to_python_type, get_setting_value, generate_random_string
|
from helper import getBuildTimeStampAndVersion, fixPermissions, collect_lang_strings, updateSubnets, generate_random_string
|
||||||
from utils.datetime_utils import timeNowDB
|
from utils.datetime_utils import timeNowDB
|
||||||
from app_state import updateState
|
from app_state import updateState
|
||||||
from logger import mylog
|
from logger import mylog
|
||||||
@@ -19,7 +19,6 @@ from scheduler import schedule_class
|
|||||||
from plugin import plugin_manager, print_plugin_info
|
from plugin import plugin_manager, print_plugin_info
|
||||||
from utils.plugin_utils import get_plugins_configs, get_set_value_for_init
|
from utils.plugin_utils import get_plugins_configs, get_set_value_for_init
|
||||||
from messaging.in_app import write_notification
|
from messaging.in_app import write_notification
|
||||||
from utils.crypto_utils import get_random_bytes
|
|
||||||
|
|
||||||
# ===============================================================================
|
# ===============================================================================
|
||||||
# Initialise user defined values
|
# Initialise user defined values
|
||||||
@@ -59,7 +58,7 @@ def ccd(
|
|||||||
result = default
|
result = default
|
||||||
|
|
||||||
# Use existing value if already supplied, otherwise default value is used
|
# Use existing value if already supplied, otherwise default value is used
|
||||||
if forceDefault == False and key in config_dir:
|
if forceDefault is False and key in config_dir:
|
||||||
result = config_dir[key]
|
result = config_dir[key]
|
||||||
|
|
||||||
# Single quotes might break SQL queries, replacing them
|
# Single quotes might break SQL queries, replacing them
|
||||||
@@ -216,7 +215,7 @@ def importConfigs(pm, db, all_plugins):
|
|||||||
[],
|
[],
|
||||||
c_d,
|
c_d,
|
||||||
"Loaded plugins",
|
"Loaded plugins",
|
||||||
'{"dataType":"array","elements":[{"elementType":"select","elementHasInputValue":1,"elementOptions":[{"multiple":"true","ordeable":"true"}],"transformers":[]},{"elementType":"button","elementOptions":[{"sourceSuffixes":[]},{"separator":""},{"cssClasses":"col-xs-12"},{"onClick":"selectChange(this)"},{"getStringKey":"Gen_Change"}],"transformers":[]}]}',
|
'{"dataType":"array","elements":[{"elementType":"select","elementHasInputValue":1,"elementOptions":[{"multiple":"true","ordeable":"true"}],"transformers":[]},{"elementType":"button","elementOptions":[{"sourceSuffixes":[]},{"separator":""},{"cssClasses":"col-xs-12"},{"onClick":"selectChange(this)"},{"getStringKey":"Gen_Change"}],"transformers":[]}]}', # noqa: E501
|
||||||
"[]",
|
"[]",
|
||||||
"General",
|
"General",
|
||||||
)
|
)
|
||||||
@@ -234,7 +233,7 @@ def importConfigs(pm, db, all_plugins):
|
|||||||
["192.168.1.0/24 --interface=eth1", "192.168.1.0/24 --interface=eth0"],
|
["192.168.1.0/24 --interface=eth1", "192.168.1.0/24 --interface=eth0"],
|
||||||
c_d,
|
c_d,
|
||||||
"Subnets to scan",
|
"Subnets to scan",
|
||||||
"""{"dataType": "array","elements": [{"elementType": "input","elementOptions": [{"placeholder": "192.168.1.0/24 --interface=eth1"},{"suffix": "_in"},{"cssClasses": "col-sm-10"},{"prefillValue": "null"}],"transformers": []},{"elementType": "button","elementOptions": [{"sourceSuffixes": ["_in"]},{"separator": ""},{"cssClasses": "col-xs-12"},{"onClick": "addList(this, false)"},{"getStringKey": "Gen_Add"}],"transformers": []},{"elementType": "select","elementHasInputValue": 1,"elementOptions": [{"multiple": "true"},{"readonly": "true"},{"editable": "true"}],"transformers": []},{"elementType": "button","elementOptions": [{"sourceSuffixes": []},{"separator": ""},{"cssClasses": "col-xs-6"},{"onClick": "removeAllOptions(this)"},{"getStringKey": "Gen_Remove_All"}],"transformers": []},{"elementType": "button","elementOptions": [{"sourceSuffixes": []},{"separator": ""},{"cssClasses": "col-xs-6"},{"onClick": "removeFromList(this)"},{"getStringKey": "Gen_Remove_Last"}],"transformers": []}]}""",
|
"""{"dataType": "array","elements": [{"elementType": "input","elementOptions": [{"placeholder": "192.168.1.0/24 --interface=eth1"},{"suffix": "_in"},{"cssClasses": "col-sm-10"},{"prefillValue": "null"}],"transformers": []},{"elementType": "button","elementOptions": [{"sourceSuffixes": ["_in"]},{"separator": ""},{"cssClasses": "col-xs-12"},{"onClick": "addList(this, false)"},{"getStringKey": "Gen_Add"}],"transformers": []},{"elementType": "select","elementHasInputValue": 1,"elementOptions": [{"multiple": "true"},{"readonly": "true"},{"editable": "true"}],"transformers": []},{"elementType": "button","elementOptions": [{"sourceSuffixes": []},{"separator": ""},{"cssClasses": "col-xs-6"},{"onClick": "removeAllOptions(this)"},{"getStringKey": "Gen_Remove_All"}],"transformers": []},{"elementType": "button","elementOptions": [{"sourceSuffixes": []},{"separator": ""},{"cssClasses": "col-xs-6"},{"onClick": "removeFromList(this)"},{"getStringKey": "Gen_Remove_Last"}],"transformers": []}]}""", # noqa: E501 - inline JSON
|
||||||
"[]",
|
"[]",
|
||||||
"General",
|
"General",
|
||||||
)
|
)
|
||||||
@@ -356,7 +355,7 @@ def importConfigs(pm, db, all_plugins):
|
|||||||
],
|
],
|
||||||
c_d,
|
c_d,
|
||||||
"Network device types",
|
"Network device types",
|
||||||
'{"dataType":"array","elements":[{"elementType":"input","elementOptions":[{"placeholder":"Enter value"},{"suffix":"_in"},{"cssClasses":"col-sm-10"},{"prefillValue":"null"}],"transformers":[]},{"elementType":"button","elementOptions":[{"sourceSuffixes":["_in"]},{"separator":""},{"cssClasses":"col-xs-12"},{"onClick":"addList(this,false)"},{"getStringKey":"Gen_Add"}],"transformers":[]},{"elementType":"select", "elementHasInputValue":1,"elementOptions":[{"multiple":"true"},{"readonly":"true"},{"editable":"true"}],"transformers":[]},{"elementType":"button","elementOptions":[{"sourceSuffixes":[]},{"separator":""},{"cssClasses":"col-xs-6"},{"onClick":"removeAllOptions(this)"},{"getStringKey":"Gen_Remove_All"}],"transformers":[]},{"elementType":"button","elementOptions":[{"sourceSuffixes":[]},{"separator":""},{"cssClasses":"col-xs-6"},{"onClick":"removeFromList(this)"},{"getStringKey":"Gen_Remove_Last"}],"transformers":[]}]}',
|
'{"dataType":"array","elements":[{"elementType":"input","elementOptions":[{"placeholder":"Enter value"},{"suffix":"_in"},{"cssClasses":"col-sm-10"},{"prefillValue":"null"}],"transformers":[]},{"elementType":"button","elementOptions":[{"sourceSuffixes":["_in"]},{"separator":""},{"cssClasses":"col-xs-12"},{"onClick":"addList(this,false)"},{"getStringKey":"Gen_Add"}],"transformers":[]},{"elementType":"select", "elementHasInputValue":1,"elementOptions":[{"multiple":"true"},{"readonly":"true"},{"editable":"true"}],"transformers":[]},{"elementType":"button","elementOptions":[{"sourceSuffixes":[]},{"separator":""},{"cssClasses":"col-xs-6"},{"onClick":"removeAllOptions(this)"},{"getStringKey":"Gen_Remove_All"}],"transformers":[]},{"elementType":"button","elementOptions":[{"sourceSuffixes":[]},{"separator":""},{"cssClasses":"col-xs-6"},{"onClick":"removeFromList(this)"},{"getStringKey":"Gen_Remove_Last"}],"transformers":[]}]}', # noqa: E501 - inline JSON
|
||||||
"[]",
|
"[]",
|
||||||
"General",
|
"General",
|
||||||
)
|
)
|
||||||
@@ -374,7 +373,7 @@ def importConfigs(pm, db, all_plugins):
|
|||||||
"t_" + generate_random_string(20),
|
"t_" + generate_random_string(20),
|
||||||
c_d,
|
c_d,
|
||||||
"API token",
|
"API token",
|
||||||
'{"dataType": "string","elements": [{"elementType": "input","elementHasInputValue": 1,"elementOptions": [{ "cssClasses": "col-xs-12" }],"transformers": []},{"elementType": "button","elementOptions": [{ "getStringKey": "Gen_Generate" },{ "customParams": "API_TOKEN" },{ "onClick": "generateApiToken(this, 20)" },{ "cssClasses": "col-xs-12" }],"transformers": []}]}',
|
'{"dataType": "string","elements": [{"elementType": "input","elementHasInputValue": 1,"elementOptions": [{ "cssClasses": "col-xs-12" }],"transformers": []},{"elementType": "button","elementOptions": [{ "getStringKey": "Gen_Generate" },{ "customParams": "API_TOKEN" },{ "onClick": "generateApiToken(this, 20)" },{ "cssClasses": "col-xs-12" }],"transformers": []}]}', # noqa: E501 - inline JSON
|
||||||
"[]",
|
"[]",
|
||||||
"General",
|
"General",
|
||||||
)
|
)
|
||||||
@@ -386,7 +385,7 @@ def importConfigs(pm, db, all_plugins):
|
|||||||
c_d,
|
c_d,
|
||||||
"Language Interface",
|
"Language Interface",
|
||||||
'{"dataType":"string", "elements": [{"elementType" : "select", "elementOptions" : [] ,"transformers": []}]}',
|
'{"dataType":"string", "elements": [{"elementType" : "select", "elementOptions" : [] ,"transformers": []}]}',
|
||||||
"['English (en_us)', 'Arabic (ar_ar)', 'Catalan (ca_ca)', 'Czech (cs_cz)', 'German (de_de)', 'Spanish (es_es)', 'Farsi (fa_fa)', 'French (fr_fr)', 'Italian (it_it)', 'Japanese (ja_jp)', 'Norwegian (nb_no)', 'Polish (pl_pl)', 'Portuguese (pt_br)', 'Portuguese (pt_pt)', 'Russian (ru_ru)', 'Swedish (sv_sv)', 'Turkish (tr_tr)', 'Ukrainian (uk_ua)', 'Chinese (zh_cn)']",
|
"['English (en_us)', 'Arabic (ar_ar)', 'Catalan (ca_ca)', 'Czech (cs_cz)', 'German (de_de)', 'Spanish (es_es)', 'Farsi (fa_fa)', 'French (fr_fr)', 'Italian (it_it)', 'Japanese (ja_jp)', 'Norwegian (nb_no)', 'Polish (pl_pl)', 'Portuguese (pt_br)', 'Portuguese (pt_pt)', 'Russian (ru_ru)', 'Swedish (sv_sv)', 'Turkish (tr_tr)', 'Ukrainian (uk_ua)', 'Chinese (zh_cn)']", # noqa: E501 - inline JSON
|
||||||
"UI",
|
"UI",
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -483,9 +482,7 @@ def importConfigs(pm, db, all_plugins):
|
|||||||
|
|
||||||
# only include loaded plugins, and the ones that are enabled
|
# only include loaded plugins, and the ones that are enabled
|
||||||
if (
|
if (
|
||||||
pref in conf.LOADED_PLUGINS
|
pref in conf.LOADED_PLUGINS or plugin_run != "disabled" or plugin_run is None
|
||||||
or plugin_run != "disabled"
|
|
||||||
or plugin_run is None
|
|
||||||
):
|
):
|
||||||
print_plugin_info(plugin, ["display_name", "description"])
|
print_plugin_info(plugin, ["display_name", "description"])
|
||||||
|
|
||||||
@@ -524,9 +521,7 @@ def importConfigs(pm, db, all_plugins):
|
|||||||
if "popupForm" in option:
|
if "popupForm" in option:
|
||||||
for popup_entry in option["popupForm"]:
|
for popup_entry in option["popupForm"]:
|
||||||
popup_pref = (
|
popup_pref = (
|
||||||
key
|
key + "_popupform_" + popup_entry.get("function", "")
|
||||||
+ "_popupform_"
|
|
||||||
+ popup_entry.get("function", "")
|
|
||||||
)
|
)
|
||||||
stringSqlParams = collect_lang_strings(
|
stringSqlParams = collect_lang_strings(
|
||||||
popup_entry, popup_pref, stringSqlParams
|
popup_entry, popup_pref, stringSqlParams
|
||||||
@@ -606,7 +601,7 @@ def importConfigs(pm, db, all_plugins):
|
|||||||
# Loop through settings_override dictionary
|
# Loop through settings_override dictionary
|
||||||
for setting_name, value in settings_override.items():
|
for setting_name, value in settings_override.items():
|
||||||
# Ensure the value is treated as a string and passed directly
|
# Ensure the value is treated as a string and passed directly
|
||||||
if isinstance(value, str) == False:
|
if isinstance(value, str) is False:
|
||||||
value = str(value)
|
value = str(value)
|
||||||
|
|
||||||
# Log the value being passed
|
# Log the value being passed
|
||||||
@@ -684,8 +679,16 @@ def importConfigs(pm, db, all_plugins):
|
|||||||
# ccd(key, default, config_dir, name, inputtype, options, group, events=None, desc="", setJsonMetadata=None, overrideTemplate=None, forceDefault=False)
|
# ccd(key, default, config_dir, name, inputtype, options, group, events=None, desc="", setJsonMetadata=None, overrideTemplate=None, forceDefault=False)
|
||||||
ccd('VERSION', new_version , c_d, '_KEEP_', '_KEEP_', '_KEEP_', '_KEEP_', None, "_KEEP_", None, None, True)
|
ccd('VERSION', new_version , c_d, '_KEEP_', '_KEEP_', '_KEEP_', '_KEEP_', None, "_KEEP_", None, None, True)
|
||||||
|
|
||||||
write_notification(f'[Upgrade] : App upgraded from <code>{prev_version}</code> to <code>{new_version}</code> 🚀 Please clear the cache: <ol> <li>Click OK below</li> <li>Clear the browser cache (shift + browser refresh button)</li> <li> Clear app cache with the <i class="fa-solid fa-rotate"></i> (reload) button in the header</li><li>Go to Settings and click Save</li> </ol> Check out new features and what has changed in the <a href="https://github.com/jokob-sk/NetAlertX/releases" target="_blank">📓 release notes</a>.', 'interrupt', timeNowDB())
|
write_notification(f'[Upgrade]: App upgraded from <code>{prev_version}</code> to \
|
||||||
|
<code>{new_version}</ code> 🚀 Please clear the cache: \
|
||||||
|
<ol> <li>Click OK below</li> <li>Clear the browser cache (shift + \
|
||||||
|
browser refresh button)</li> <li> Clear app cache with the <i class="fa-solid fa-rotate"></i> \
|
||||||
|
(reload) button in the header</li><li>Go to Settings and click Save</li> </ol>\
|
||||||
|
Check out new features and what has changed in the \
|
||||||
|
<a href="https://github.com/jokob-sk/NetAlertX/releases" target="_blank">📓 release notes</a>.',
|
||||||
|
'interrupt',
|
||||||
|
timeNowDB()
|
||||||
|
)
|
||||||
|
|
||||||
# -----------------
|
# -----------------
|
||||||
# Initialization finished, update DB and API endpoints
|
# Initialization finished, update DB and API endpoints
|
||||||
|
|||||||
@@ -1,19 +1,14 @@
|
|||||||
import sys
|
import sys
|
||||||
import io
|
import io
|
||||||
import datetime
|
# import datetime
|
||||||
import threading
|
import threading
|
||||||
import queue
|
import queue
|
||||||
import logging
|
import logging
|
||||||
from zoneinfo import ZoneInfo
|
# from zoneinfo import ZoneInfo
|
||||||
|
|
||||||
# Register NetAlertX directories
|
|
||||||
INSTALL_PATH="/app"
|
|
||||||
|
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
|
||||||
|
|
||||||
# NetAlertX imports
|
# NetAlertX imports
|
||||||
import conf
|
import conf
|
||||||
from const import *
|
from const import logPath
|
||||||
from utils.datetime_utils import timeNowTZ
|
from utils.datetime_utils import timeNowTZ
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -11,13 +11,9 @@ from flask import jsonify
|
|||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
||||||
sys.path.extend([f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from const import apiPath
|
from const import apiPath # noqa: E402 [flake8 lint suppression]
|
||||||
from logger import mylog
|
from logger import mylog # noqa: E402 [flake8 lint suppression]
|
||||||
|
from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
|
||||||
import conf
|
|
||||||
from const import applicationPath, logPath, apiPath, confFileName, reportTemplatesPath
|
|
||||||
from logger import mylog
|
|
||||||
from utils.datetime_utils import timeNowDB
|
|
||||||
|
|
||||||
|
|
||||||
NOTIFICATION_API_FILE = apiPath + 'user_notifications.json'
|
NOTIFICATION_API_FILE = apiPath + 'user_notifications.json'
|
||||||
|
|||||||
@@ -18,12 +18,12 @@ import sys
|
|||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
||||||
sys.path.extend([f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from helper import (
|
from helper import ( # noqa: E402 [flake8 lint suppression]
|
||||||
get_setting_value,
|
get_setting_value,
|
||||||
)
|
)
|
||||||
from logger import mylog
|
from logger import mylog # noqa: E402 [flake8 lint suppression]
|
||||||
from db.sql_safe_builder import create_safe_condition_builder
|
from db.sql_safe_builder import create_safe_condition_builder # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.datetime_utils import get_timezone_offset
|
from utils.datetime_utils import get_timezone_offset # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
# ===============================================================================
|
# ===============================================================================
|
||||||
# REPORTING
|
# REPORTING
|
||||||
@@ -56,14 +56,14 @@ def get_notifications(db):
|
|||||||
WHERE eve_PendingAlertEmail = 1 AND eve_EventType not in ('Device Down', 'Down Reconnected', 'New Device' ) AND eve_MAC IN
|
WHERE eve_PendingAlertEmail = 1 AND eve_EventType not in ('Device Down', 'Down Reconnected', 'New Device' ) AND eve_MAC IN
|
||||||
(
|
(
|
||||||
SELECT devMac FROM Devices WHERE devAlertEvents = 0
|
SELECT devMac FROM Devices WHERE devAlertEvents = 0
|
||||||
)""")
|
)""")
|
||||||
|
|
||||||
# Disable down/down reconnected notifications on devices where devAlertDown is disabled
|
# Disable down/down reconnected notifications on devices where devAlertDown is disabled
|
||||||
sql.execute("""UPDATE Events SET eve_PendingAlertEmail = 0
|
sql.execute("""UPDATE Events SET eve_PendingAlertEmail = 0
|
||||||
WHERE eve_PendingAlertEmail = 1 AND eve_EventType in ('Device Down', 'Down Reconnected') AND eve_MAC IN
|
WHERE eve_PendingAlertEmail = 1 AND eve_EventType in ('Device Down', 'Down Reconnected') AND eve_MAC IN
|
||||||
(
|
(
|
||||||
SELECT devMac FROM Devices WHERE devAlertDown = 0
|
SELECT devMac FROM Devices WHERE devAlertDown = 0
|
||||||
)""")
|
)""")
|
||||||
|
|
||||||
sections = get_setting_value("NTFPRCS_INCLUDED_SECTIONS")
|
sections = get_setting_value("NTFPRCS_INCLUDED_SECTIONS")
|
||||||
|
|
||||||
@@ -79,20 +79,32 @@ def get_notifications(db):
|
|||||||
safe_condition, parameters = condition_builder.get_safe_condition_legacy(
|
safe_condition, parameters = condition_builder.get_safe_condition_legacy(
|
||||||
new_dev_condition_setting
|
new_dev_condition_setting
|
||||||
)
|
)
|
||||||
sqlQuery = """SELECT eve_MAC as MAC, eve_DateTime as Datetime, devLastIP as IP, eve_EventType as "Event Type", devName as "Device name", devComments as Comments FROM Events_Devices
|
sqlQuery = """SELECT
|
||||||
WHERE eve_PendingAlertEmail = 1
|
eve_MAC as MAC,
|
||||||
|
eve_DateTime as Datetime,
|
||||||
|
devLastIP as IP,
|
||||||
|
eve_EventType as "Event Type",
|
||||||
|
devName as "Device name",
|
||||||
|
devComments as Comments FROM Events_Devices
|
||||||
|
WHERE eve_PendingAlertEmail = 1
|
||||||
AND eve_EventType = 'New Device' {}
|
AND eve_EventType = 'New Device' {}
|
||||||
ORDER BY eve_DateTime""".format(safe_condition)
|
ORDER BY eve_DateTime""".format(safe_condition)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
mylog(
|
mylog(
|
||||||
"verbose",
|
"verbose",
|
||||||
["[Notification] Error building safe condition for new devices: ", e],
|
["[Notification] Error building safe condition for new devices: ", e],
|
||||||
)
|
)
|
||||||
# Fall back to safe default (no additional conditions)
|
# Fall back to safe default (no additional conditions)
|
||||||
sqlQuery = """SELECT eve_MAC as MAC, eve_DateTime as Datetime, devLastIP as IP, eve_EventType as "Event Type", devName as "Device name", devComments as Comments FROM Events_Devices
|
sqlQuery = """SELECT
|
||||||
WHERE eve_PendingAlertEmail = 1
|
eve_MAC as MAC,
|
||||||
|
eve_DateTime as Datetime,
|
||||||
|
devLastIP as IP,
|
||||||
|
eve_EventType as "Event Type",
|
||||||
|
devName as "Device name",
|
||||||
|
devComments as Comments FROM Events_Devices
|
||||||
|
WHERE eve_PendingAlertEmail = 1
|
||||||
AND eve_EventType = 'New Device'
|
AND eve_EventType = 'New Device'
|
||||||
ORDER BY eve_DateTime"""
|
ORDER BY eve_DateTime"""
|
||||||
parameters = {}
|
parameters = {}
|
||||||
|
|
||||||
mylog("debug", ["[Notification] new_devices SQL query: ", sqlQuery])
|
mylog("debug", ["[Notification] new_devices SQL query: ", sqlQuery])
|
||||||
@@ -181,20 +193,32 @@ def get_notifications(db):
|
|||||||
safe_condition, parameters = condition_builder.get_safe_condition_legacy(
|
safe_condition, parameters = condition_builder.get_safe_condition_legacy(
|
||||||
event_condition_setting
|
event_condition_setting
|
||||||
)
|
)
|
||||||
sqlQuery = """SELECT eve_MAC as MAC, eve_DateTime as Datetime, devLastIP as IP, eve_EventType as "Event Type", devName as "Device name", devComments as Comments FROM Events_Devices
|
sqlQuery = """SELECT
|
||||||
WHERE eve_PendingAlertEmail = 1
|
eve_MAC as MAC,
|
||||||
|
eve_DateTime as Datetime,
|
||||||
|
devLastIP as IP,
|
||||||
|
eve_EventType as "Event Type",
|
||||||
|
devName as "Device name",
|
||||||
|
devComments as Comments FROM Events_Devices
|
||||||
|
WHERE eve_PendingAlertEmail = 1
|
||||||
AND eve_EventType IN ('Connected', 'Down Reconnected', 'Disconnected','IP Changed') {}
|
AND eve_EventType IN ('Connected', 'Down Reconnected', 'Disconnected','IP Changed') {}
|
||||||
ORDER BY eve_DateTime""".format(safe_condition)
|
ORDER BY eve_DateTime""".format(safe_condition)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
mylog(
|
mylog(
|
||||||
"verbose",
|
"verbose",
|
||||||
["[Notification] Error building safe condition for events: ", e],
|
["[Notification] Error building safe condition for events: ", e],
|
||||||
)
|
)
|
||||||
# Fall back to safe default (no additional conditions)
|
# Fall back to safe default (no additional conditions)
|
||||||
sqlQuery = """SELECT eve_MAC as MAC, eve_DateTime as Datetime, devLastIP as IP, eve_EventType as "Event Type", devName as "Device name", devComments as Comments FROM Events_Devices
|
sqlQuery = """SELECT
|
||||||
WHERE eve_PendingAlertEmail = 1
|
eve_MAC as MAC,
|
||||||
|
eve_DateTime as Datetime,
|
||||||
|
devLastIP as IP,
|
||||||
|
eve_EventType as "Event Type",
|
||||||
|
devName as "Device name",
|
||||||
|
devComments as Comments FROM Events_Devices
|
||||||
|
WHERE eve_PendingAlertEmail = 1
|
||||||
AND eve_EventType IN ('Connected', 'Down Reconnected', 'Disconnected','IP Changed')
|
AND eve_EventType IN ('Connected', 'Down Reconnected', 'Disconnected','IP Changed')
|
||||||
ORDER BY eve_DateTime"""
|
ORDER BY eve_DateTime"""
|
||||||
parameters = {}
|
parameters = {}
|
||||||
|
|
||||||
mylog("debug", ["[Notification] events SQL query: ", sqlQuery])
|
mylog("debug", ["[Notification] events SQL query: ", sqlQuery])
|
||||||
@@ -208,7 +232,17 @@ def get_notifications(db):
|
|||||||
|
|
||||||
if "plugins" in sections:
|
if "plugins" in sections:
|
||||||
# Compose Plugins Section
|
# Compose Plugins Section
|
||||||
sqlQuery = """SELECT Plugin, Object_PrimaryId, Object_SecondaryId, DateTimeChanged, Watched_Value1, Watched_Value2, Watched_Value3, Watched_Value4, Status from Plugins_Events"""
|
sqlQuery = """SELECT
|
||||||
|
Plugin,
|
||||||
|
Object_PrimaryId,
|
||||||
|
Object_SecondaryId,
|
||||||
|
DateTimeChanged,
|
||||||
|
Watched_Value1,
|
||||||
|
Watched_Value2,
|
||||||
|
Watched_Value3,
|
||||||
|
Watched_Value4,
|
||||||
|
Status
|
||||||
|
from Plugins_Events"""
|
||||||
|
|
||||||
# Get the events as JSON
|
# Get the events as JSON
|
||||||
json_obj = db.get_table_as_json(sqlQuery)
|
json_obj = db.get_table_as_json(sqlQuery)
|
||||||
|
|||||||
@@ -1,13 +1,12 @@
|
|||||||
import json
|
import json
|
||||||
import uuid
|
import uuid
|
||||||
import socket
|
import socket
|
||||||
import subprocess
|
|
||||||
from yattag import indent
|
from yattag import indent
|
||||||
from json2table import convert
|
from json2table import convert
|
||||||
|
|
||||||
# Register NetAlertX modules
|
# Register NetAlertX modules
|
||||||
import conf
|
import conf
|
||||||
from const import applicationPath, logPath, apiPath, reportTemplatesPath
|
from const import logPath, apiPath, reportTemplatesPath
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger
|
||||||
from helper import (
|
from helper import (
|
||||||
generate_mac_links,
|
generate_mac_links,
|
||||||
@@ -62,11 +61,7 @@ class NotificationInstance:
|
|||||||
|
|
||||||
# Check if nothing to report, end
|
# Check if nothing to report, end
|
||||||
if (
|
if (
|
||||||
JSON["new_devices"] == []
|
JSON["new_devices"] == [] and JSON["down_devices"] == [] and JSON["events"] == [] and JSON["plugins"] == [] and JSON["down_reconnected"] == []
|
||||||
and JSON["down_devices"] == []
|
|
||||||
and JSON["events"] == []
|
|
||||||
and JSON["plugins"] == []
|
|
||||||
and JSON["down_reconnected"] == []
|
|
||||||
):
|
):
|
||||||
self.HasNotifications = False
|
self.HasNotifications = False
|
||||||
else:
|
else:
|
||||||
@@ -88,8 +83,6 @@ class NotificationInstance:
|
|||||||
# else:
|
# else:
|
||||||
# mylog('debug', ['[Notification] notiStruc:', json.dumps(notiStruc.__dict__, indent=4)])
|
# mylog('debug', ['[Notification] notiStruc:', json.dumps(notiStruc.__dict__, indent=4)])
|
||||||
|
|
||||||
Text = ""
|
|
||||||
HTML = ""
|
|
||||||
template_file_path = reportTemplatesPath + "report_template.html"
|
template_file_path = reportTemplatesPath + "report_template.html"
|
||||||
|
|
||||||
# Open text Template
|
# Open text Template
|
||||||
@@ -274,7 +267,7 @@ class NotificationInstance:
|
|||||||
# Clear the Pending Email flag from all events and devices
|
# Clear the Pending Email flag from all events and devices
|
||||||
def clearPendingEmailFlag(self):
|
def clearPendingEmailFlag(self):
|
||||||
|
|
||||||
# Clean Pending Alert Events
|
# Clean Pending Alert Events
|
||||||
self.db.sql.execute("""
|
self.db.sql.execute("""
|
||||||
UPDATE Devices SET devLastNotification = ?
|
UPDATE Devices SET devLastNotification = ?
|
||||||
WHERE devMac IN (
|
WHERE devMac IN (
|
||||||
|
|||||||
@@ -117,7 +117,3 @@ class UserEventsQueueInstance:
|
|||||||
mylog('none', [msg])
|
mylog('none', [msg])
|
||||||
|
|
||||||
return False, msg
|
return False, msg
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
209
server/plugin.py
209
server/plugin.py
@@ -10,11 +10,20 @@ from concurrent.futures import ThreadPoolExecutor, as_completed
|
|||||||
import conf
|
import conf
|
||||||
from const import pluginsPath, logPath, applicationPath, reportTemplatesPath
|
from const import pluginsPath, logPath, applicationPath, reportTemplatesPath
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger
|
||||||
from helper import get_file_content, write_file, get_setting, get_setting_value
|
from helper import get_file_content, get_setting, get_setting_value
|
||||||
from utils.datetime_utils import timeNowTZ, timeNowDB
|
from utils.datetime_utils import timeNowTZ, timeNowDB
|
||||||
from app_state import updateState
|
from app_state import updateState
|
||||||
from api import update_api
|
from api import update_api
|
||||||
from utils.plugin_utils import logEventStatusCounts, get_plugin_string, get_plugin_setting_obj, print_plugin_info, list_to_csv, combine_plugin_objects, resolve_wildcards_arr, handle_empty, custom_plugin_decoder, decode_and_rename_files
|
from utils.plugin_utils import (
|
||||||
|
logEventStatusCounts,
|
||||||
|
get_plugin_setting_obj,
|
||||||
|
print_plugin_info,
|
||||||
|
list_to_csv,
|
||||||
|
combine_plugin_objects,
|
||||||
|
resolve_wildcards_arr,
|
||||||
|
handle_empty,
|
||||||
|
decode_and_rename_files
|
||||||
|
)
|
||||||
from models.notification_instance import NotificationInstance
|
from models.notification_instance import NotificationInstance
|
||||||
from messaging.in_app import write_notification
|
from messaging.in_app import write_notification
|
||||||
from models.user_events_queue_instance import UserEventsQueueInstance
|
from models.user_events_queue_instance import UserEventsQueueInstance
|
||||||
@@ -57,13 +66,7 @@ class plugin_manager:
|
|||||||
# Header
|
# Header
|
||||||
updateState("Run: Plugins")
|
updateState("Run: Plugins")
|
||||||
|
|
||||||
mylog(
|
mylog("debug", f"[Plugins] Check if any plugins need to be executed on run type: {runType}")
|
||||||
"debug",
|
|
||||||
[
|
|
||||||
"[Plugins] Check if any plugins need to be executed on run type: ",
|
|
||||||
runType,
|
|
||||||
],
|
|
||||||
)
|
|
||||||
|
|
||||||
for plugin in self.all_plugins:
|
for plugin in self.all_plugins:
|
||||||
shouldRun = False
|
shouldRun = False
|
||||||
@@ -72,7 +75,7 @@ class plugin_manager:
|
|||||||
# 🔹 Lookup RUN setting from cache instead of calling get_plugin_setting_obj each time
|
# 🔹 Lookup RUN setting from cache instead of calling get_plugin_setting_obj each time
|
||||||
run_setting = self._cache["settings"].get(prefix, {}).get("RUN")
|
run_setting = self._cache["settings"].get(prefix, {}).get("RUN")
|
||||||
|
|
||||||
if run_setting != None and run_setting["value"] == runType:
|
if run_setting is not None and run_setting["value"] == runType:
|
||||||
if runType != "schedule":
|
if runType != "schedule":
|
||||||
shouldRun = True
|
shouldRun = True
|
||||||
elif runType == "schedule":
|
elif runType == "schedule":
|
||||||
@@ -91,10 +94,7 @@ class plugin_manager:
|
|||||||
|
|
||||||
# 🔹 CMD also retrieved from cache
|
# 🔹 CMD also retrieved from cache
|
||||||
cmd_setting = self._cache["settings"].get(prefix, {}).get("CMD")
|
cmd_setting = self._cache["settings"].get(prefix, {}).get("CMD")
|
||||||
mylog(
|
mylog("debug", f"[Plugins] CMD: {cmd_setting["value"] if cmd_setting else None}")
|
||||||
"debug",
|
|
||||||
["[Plugins] CMD: ", cmd_setting["value"] if cmd_setting else None],
|
|
||||||
)
|
|
||||||
|
|
||||||
execute_plugin(self.db, self.all_plugins, plugin)
|
execute_plugin(self.db, self.all_plugins, plugin)
|
||||||
|
|
||||||
@@ -130,13 +130,7 @@ class plugin_manager:
|
|||||||
mylog("debug", ["[check_and_run_user_event] User Execution Queue is empty"])
|
mylog("debug", ["[check_and_run_user_event] User Execution Queue is empty"])
|
||||||
return # Exit early if the log file is empty
|
return # Exit early if the log file is empty
|
||||||
else:
|
else:
|
||||||
mylog(
|
mylog("debug", "[check_and_run_user_event] Process User Execution Queue:" + ", ".join(map(str, lines)))
|
||||||
"debug",
|
|
||||||
[
|
|
||||||
"[check_and_run_user_event] Process User Execution Queue:"
|
|
||||||
+ ", ".join(map(str, lines))
|
|
||||||
],
|
|
||||||
)
|
|
||||||
|
|
||||||
for line in lines:
|
for line in lines:
|
||||||
# Extract event name and parameters from the log line
|
# Extract event name and parameters from the log line
|
||||||
@@ -160,15 +154,7 @@ class plugin_manager:
|
|||||||
update_api(self.db, self.all_plugins, False, param.split(","), True)
|
update_api(self.db, self.all_plugins, False, param.split(","), True)
|
||||||
|
|
||||||
else:
|
else:
|
||||||
mylog(
|
mylog("minimal", f"[check_and_run_user_event] WARNING: Unhandled event in execution queue: {event} | {param}")
|
||||||
"minimal",
|
|
||||||
[
|
|
||||||
"[check_and_run_user_event] WARNING: Unhandled event in execution queue: ",
|
|
||||||
event,
|
|
||||||
" | ",
|
|
||||||
param,
|
|
||||||
],
|
|
||||||
)
|
|
||||||
execution_log.finalize_event(
|
execution_log.finalize_event(
|
||||||
event
|
event
|
||||||
) # Finalize unknown events to remove them
|
) # Finalize unknown events to remove them
|
||||||
@@ -312,7 +298,7 @@ class plugin_param:
|
|||||||
if param["type"] == "setting":
|
if param["type"] == "setting":
|
||||||
inputValue = get_setting(param["value"])
|
inputValue = get_setting(param["value"])
|
||||||
|
|
||||||
if inputValue != None:
|
if inputValue is not None:
|
||||||
setVal = inputValue["setValue"] # setting value
|
setVal = inputValue["setValue"] # setting value
|
||||||
setTyp = inputValue["setType"] # setting type
|
setTyp = inputValue["setType"] # setting type
|
||||||
|
|
||||||
@@ -337,9 +323,7 @@ class plugin_param:
|
|||||||
resolved = list_to_csv(setVal)
|
resolved = list_to_csv(setVal)
|
||||||
|
|
||||||
else:
|
else:
|
||||||
mylog(
|
mylog("none", "[Plugins] ⚠ ERROR: Parameter probably not converted.")
|
||||||
"none", ["[Plugins] ⚠ ERROR: Parameter probably not converted."]
|
|
||||||
)
|
|
||||||
return json.dumps(setVal)
|
return json.dumps(setVal)
|
||||||
|
|
||||||
# Get SQL result
|
# Get SQL result
|
||||||
@@ -390,15 +374,10 @@ def run_plugin(command, set_RUN_TIMEOUT, plugin):
|
|||||||
)
|
)
|
||||||
except subprocess.CalledProcessError as e:
|
except subprocess.CalledProcessError as e:
|
||||||
mylog("none", [e.output])
|
mylog("none", [e.output])
|
||||||
mylog("none", ["[Plugins] ⚠ ERROR - enable LOG_LEVEL=debug and check logs"])
|
mylog("none", "[Plugins] ⚠ ERROR - enable LOG_LEVEL=debug and check logs")
|
||||||
return None
|
return None
|
||||||
except subprocess.TimeoutExpired:
|
except subprocess.TimeoutExpired:
|
||||||
mylog(
|
mylog("none", f"[Plugins] ⚠ ERROR - TIMEOUT - the plugin {plugin['unique_prefix']} forcefully terminated as timeout reached. Increase TIMEOUT setting and scan interval.")
|
||||||
"none",
|
|
||||||
[
|
|
||||||
f"[Plugins] ⚠ ERROR - TIMEOUT - the plugin {plugin['unique_prefix']} forcefully terminated as timeout reached. Increase TIMEOUT setting and scan interval."
|
|
||||||
],
|
|
||||||
)
|
|
||||||
return None
|
return None
|
||||||
|
|
||||||
|
|
||||||
@@ -411,7 +390,7 @@ def execute_plugin(db, all_plugins, plugin):
|
|||||||
set = get_plugin_setting_obj(plugin, "CMD")
|
set = get_plugin_setting_obj(plugin, "CMD")
|
||||||
|
|
||||||
# handle missing "function":"CMD" setting
|
# handle missing "function":"CMD" setting
|
||||||
if set == None:
|
if set is None:
|
||||||
return
|
return
|
||||||
|
|
||||||
set_CMD = set["value"]
|
set_CMD = set["value"]
|
||||||
@@ -441,13 +420,8 @@ def execute_plugin(db, all_plugins, plugin):
|
|||||||
for param in plugin["params"]:
|
for param in plugin["params"]:
|
||||||
tempParam = plugin_param(param, plugin, db)
|
tempParam = plugin_param(param, plugin, db)
|
||||||
|
|
||||||
if tempParam.resolved == None:
|
if tempParam.resolved is None:
|
||||||
mylog(
|
mylog("none", f'[Plugins] The parameter "name":"{tempParam.name}" for "value": {tempParam.value} was resolved as None')
|
||||||
"none",
|
|
||||||
[
|
|
||||||
f'[Plugins] The parameter "name":"{tempParam.name}" for "value": {tempParam.value} was resolved as None'
|
|
||||||
],
|
|
||||||
)
|
|
||||||
|
|
||||||
else:
|
else:
|
||||||
# params.append( [param["name"], resolved] )
|
# params.append( [param["name"], resolved] )
|
||||||
@@ -456,14 +430,9 @@ def execute_plugin(db, all_plugins, plugin):
|
|||||||
if tempParam.multiplyTimeout:
|
if tempParam.multiplyTimeout:
|
||||||
set_RUN_TIMEOUT = set_RUN_TIMEOUT * tempParam.paramValuesCount
|
set_RUN_TIMEOUT = set_RUN_TIMEOUT * tempParam.paramValuesCount
|
||||||
|
|
||||||
mylog(
|
mylog("debug", f'[Plugins] The parameter "name":"{param["name"]}" will multiply timeout {tempParam.paramValuesCount}x. Total timeout: {set_RUN_TIMEOUT}s')
|
||||||
"debug",
|
|
||||||
[
|
|
||||||
f'[Plugins] The parameter "name":"{param["name"]}" will multiply the timeout {tempParam.paramValuesCount} times. Total timeout: {set_RUN_TIMEOUT}s'
|
|
||||||
],
|
|
||||||
)
|
|
||||||
|
|
||||||
mylog("debug", ["[Plugins] Timeout: ", set_RUN_TIMEOUT])
|
mylog("debug", f"[Plugins] Timeout: {set_RUN_TIMEOUT}")
|
||||||
|
|
||||||
# build SQL query parameters to insert into the DB
|
# build SQL query parameters to insert into the DB
|
||||||
sqlParams = []
|
sqlParams = []
|
||||||
@@ -475,8 +444,8 @@ def execute_plugin(db, all_plugins, plugin):
|
|||||||
command = resolve_wildcards_arr(set_CMD.split(), params)
|
command = resolve_wildcards_arr(set_CMD.split(), params)
|
||||||
|
|
||||||
# Execute command
|
# Execute command
|
||||||
mylog("verbose", ["[Plugins] Executing: ", set_CMD])
|
mylog("verbose", f"[Plugins] Executing: {set_CMD}")
|
||||||
mylog("debug", ["[Plugins] Resolved : ", command])
|
mylog("debug", f"[Plugins] Resolved : {command}")
|
||||||
|
|
||||||
# Using ThreadPoolExecutor to handle concurrent subprocesses
|
# Using ThreadPoolExecutor to handle concurrent subprocesses
|
||||||
with ThreadPoolExecutor(max_workers=5) as executor:
|
with ThreadPoolExecutor(max_workers=5) as executor:
|
||||||
@@ -521,12 +490,7 @@ def execute_plugin(db, all_plugins, plugin):
|
|||||||
columns = line.split("|")
|
columns = line.split("|")
|
||||||
# There have to be 9 or 13 columns
|
# There have to be 9 or 13 columns
|
||||||
if len(columns) not in [9, 13]:
|
if len(columns) not in [9, 13]:
|
||||||
mylog(
|
mylog("none", f"[Plugins] Wrong number of input values, must be 9 or 13, got {len(columns)} from: {line}")
|
||||||
"none",
|
|
||||||
[
|
|
||||||
f"[Plugins] Wrong number of input values, must be 9 or 13, got {len(columns)} from: {line}"
|
|
||||||
],
|
|
||||||
)
|
|
||||||
continue # Skip lines with incorrect number of columns
|
continue # Skip lines with incorrect number of columns
|
||||||
|
|
||||||
# Common part of the SQL parameters
|
# Common part of the SQL parameters
|
||||||
@@ -581,9 +545,7 @@ def execute_plugin(db, all_plugins, plugin):
|
|||||||
# keep current instance log file, delete all from other nodes
|
# keep current instance log file, delete all from other nodes
|
||||||
if filename != "last_result.log" and os.path.exists(full_path):
|
if filename != "last_result.log" and os.path.exists(full_path):
|
||||||
os.remove(full_path) # DEBUG:TODO uncomment 🐛
|
os.remove(full_path) # DEBUG:TODO uncomment 🐛
|
||||||
mylog(
|
mylog("verbose", f"[Plugins] Processed and deleted file: {full_path} ")
|
||||||
"verbose", [f"[Plugins] Processed and deleted file: {full_path} "]
|
|
||||||
)
|
|
||||||
|
|
||||||
# app-db-query
|
# app-db-query
|
||||||
if plugin["data_source"] == "app-db-query":
|
if plugin["data_source"] == "app-db-query":
|
||||||
@@ -591,7 +553,7 @@ def execute_plugin(db, all_plugins, plugin):
|
|||||||
q = set_CMD.replace("{s-quote}", "'")
|
q = set_CMD.replace("{s-quote}", "'")
|
||||||
|
|
||||||
# Execute command
|
# Execute command
|
||||||
mylog("verbose", ["[Plugins] Executing: ", q])
|
mylog("verbose", f"[Plugins] Executing: {q}")
|
||||||
|
|
||||||
# set_CMD should contain a SQL query
|
# set_CMD should contain a SQL query
|
||||||
arr = db.get_sql_array(q)
|
arr = db.get_sql_array(q)
|
||||||
@@ -650,7 +612,7 @@ def execute_plugin(db, all_plugins, plugin):
|
|||||||
# Append the final parameters to sqlParams
|
# Append the final parameters to sqlParams
|
||||||
sqlParams.append(tuple(base_params))
|
sqlParams.append(tuple(base_params))
|
||||||
else:
|
else:
|
||||||
mylog("none", ["[Plugins] Skipped invalid sql result"])
|
mylog("none", "[Plugins] Skipped invalid sql result")
|
||||||
|
|
||||||
# app-db-query
|
# app-db-query
|
||||||
if plugin["data_source"] == "sqlite-db-query":
|
if plugin["data_source"] == "sqlite-db-query":
|
||||||
@@ -659,19 +621,14 @@ def execute_plugin(db, all_plugins, plugin):
|
|||||||
q = set_CMD.replace("{s-quote}", "'")
|
q = set_CMD.replace("{s-quote}", "'")
|
||||||
|
|
||||||
# Execute command
|
# Execute command
|
||||||
mylog("verbose", ["[Plugins] Executing: ", q])
|
mylog("verbose", f"[Plugins] Executing: {q}")
|
||||||
|
|
||||||
# ------- necessary settings check --------
|
# ------- necessary settings check --------
|
||||||
set = get_plugin_setting_obj(plugin, "DB_PATH")
|
set = get_plugin_setting_obj(plugin, "DB_PATH")
|
||||||
|
|
||||||
# handle missing "function":"DB_PATH" setting
|
# handle missing "function":"DB_PATH" setting
|
||||||
if set == None:
|
if set is None:
|
||||||
mylog(
|
mylog("none", "[Plugins] ⚠ ERROR: DB_PATH setting for plugin type sqlite-db-query missing.")
|
||||||
"none",
|
|
||||||
[
|
|
||||||
"[Plugins] ⚠ ERROR: DB_PATH setting for plugin type sqlite-db-query missing."
|
|
||||||
],
|
|
||||||
)
|
|
||||||
return
|
return
|
||||||
|
|
||||||
fullSqlitePath = set["value"]
|
fullSqlitePath = set["value"]
|
||||||
@@ -679,25 +636,14 @@ def execute_plugin(db, all_plugins, plugin):
|
|||||||
# try attaching the sqlite DB
|
# try attaching the sqlite DB
|
||||||
try:
|
try:
|
||||||
sql.execute(
|
sql.execute(
|
||||||
"ATTACH DATABASE '"
|
"ATTACH DATABASE '" + fullSqlitePath + "' AS EXTERNAL_" + plugin["unique_prefix"]
|
||||||
+ fullSqlitePath
|
|
||||||
+ "' AS EXTERNAL_"
|
|
||||||
+ plugin["unique_prefix"]
|
|
||||||
)
|
)
|
||||||
arr = db.get_sql_array(q)
|
arr = db.get_sql_array(q)
|
||||||
sql.execute("DETACH DATABASE EXTERNAL_" + plugin["unique_prefix"])
|
sql.execute("DETACH DATABASE EXTERNAL_" + plugin["unique_prefix"])
|
||||||
|
|
||||||
except sqlite3.Error as e:
|
except sqlite3.Error as e:
|
||||||
mylog(
|
mylog("none", f"[Plugins] ⚠ ERROR: DB_PATH setting ({fullSqlitePath}) for plugin {plugin['unique_prefix']}. Did you mount it correctly?")
|
||||||
"none",
|
mylog("none", f"[Plugins] ⚠ ERROR: ATTACH DATABASE failed with SQL ERROR: {e}")
|
||||||
[
|
|
||||||
f"[Plugins] ⚠ ERROR: DB_PATH setting ({fullSqlitePath}) for plugin {plugin['unique_prefix']}. Did you mount it correctly?"
|
|
||||||
],
|
|
||||||
)
|
|
||||||
mylog(
|
|
||||||
"none",
|
|
||||||
["[Plugins] ⚠ ERROR: ATTACH DATABASE failed with SQL ERROR: ", e],
|
|
||||||
)
|
|
||||||
return
|
return
|
||||||
|
|
||||||
for row in arr:
|
for row in arr:
|
||||||
@@ -748,24 +694,14 @@ def execute_plugin(db, all_plugins, plugin):
|
|||||||
# Append the final parameters to sqlParams
|
# Append the final parameters to sqlParams
|
||||||
sqlParams.append(tuple(base_params))
|
sqlParams.append(tuple(base_params))
|
||||||
else:
|
else:
|
||||||
mylog("none", ["[Plugins] Skipped invalid sql result"])
|
mylog("none", "[Plugins] Skipped invalid sql result")
|
||||||
|
|
||||||
# check if the subprocess / SQL query failed / there was no valid output
|
# check if the subprocess / SQL query failed / there was no valid output
|
||||||
if len(sqlParams) == 0:
|
if len(sqlParams) == 0:
|
||||||
mylog(
|
mylog("none", f'[Plugins] No output received from the plugin "{plugin["unique_prefix"]}"')
|
||||||
"none",
|
|
||||||
[
|
|
||||||
f'[Plugins] No output received from the plugin "{plugin["unique_prefix"]}"'
|
|
||||||
],
|
|
||||||
)
|
|
||||||
|
|
||||||
else:
|
else:
|
||||||
mylog(
|
mylog("verbose", f"[Plugins] SUCCESS for {plugin['unique_prefix']} received {len(sqlParams)} entries")
|
||||||
"verbose",
|
|
||||||
[
|
|
||||||
f"[Plugins] SUCCESS for {plugin['unique_prefix']} received {len(sqlParams)} entries"
|
|
||||||
],
|
|
||||||
)
|
|
||||||
# mylog('debug', ['[Plugins] sqlParam entries: ', sqlParams])
|
# mylog('debug', ['[Plugins] sqlParam entries: ', sqlParams])
|
||||||
|
|
||||||
# create objects
|
# create objects
|
||||||
@@ -782,12 +718,7 @@ def execute_plugin(db, all_plugins, plugin):
|
|||||||
# check if we need to update devices api endpoint as well to prevent long user waits on Loading...
|
# check if we need to update devices api endpoint as well to prevent long user waits on Loading...
|
||||||
userUpdatedDevices = UserEventsQueueInstance().has_update_devices()
|
userUpdatedDevices = UserEventsQueueInstance().has_update_devices()
|
||||||
|
|
||||||
mylog(
|
mylog("verbose", f"[Plugins] Should I update API (userUpdatedDevices): {userUpdatedDevices}")
|
||||||
"verbose",
|
|
||||||
[
|
|
||||||
f"[Plugins] Should I update API (userUpdatedDevices): {userUpdatedDevices}"
|
|
||||||
],
|
|
||||||
)
|
|
||||||
|
|
||||||
if userUpdatedDevices:
|
if userUpdatedDevices:
|
||||||
endpoints += ["devices"]
|
endpoints += ["devices"]
|
||||||
@@ -807,7 +738,7 @@ def process_plugin_events(db, plugin, plugEventsArr):
|
|||||||
|
|
||||||
pluginPref = plugin["unique_prefix"]
|
pluginPref = plugin["unique_prefix"]
|
||||||
|
|
||||||
mylog("verbose", ["[Plugins] Processing : ", pluginPref])
|
mylog("verbose", f"[Plugins] Processing : {pluginPref}")
|
||||||
|
|
||||||
try:
|
try:
|
||||||
# Begin a transaction
|
# Begin a transaction
|
||||||
@@ -827,20 +758,8 @@ def process_plugin_events(db, plugin, plugEventsArr):
|
|||||||
for eve in plugEventsArr:
|
for eve in plugEventsArr:
|
||||||
pluginEvents.append(plugin_object_class(plugin, eve))
|
pluginEvents.append(plugin_object_class(plugin, eve))
|
||||||
|
|
||||||
mylog(
|
mylog("debug", f"[Plugins] Existing objects from Plugins_Objects: {len(pluginObjects)}")
|
||||||
"debug",
|
mylog("debug", f"[Plugins] Logged events from the plugin run : {len(pluginEvents)}")
|
||||||
[
|
|
||||||
"[Plugins] Existing objects from Plugins_Objects: ",
|
|
||||||
len(pluginObjects),
|
|
||||||
],
|
|
||||||
)
|
|
||||||
mylog(
|
|
||||||
"debug",
|
|
||||||
[
|
|
||||||
"[Plugins] Logged events from the plugin run : ",
|
|
||||||
len(pluginEvents),
|
|
||||||
],
|
|
||||||
)
|
|
||||||
|
|
||||||
# Loop thru all current events and update the status to "exists" if the event matches an existing object
|
# Loop thru all current events and update the status to "exists" if the event matches an existing object
|
||||||
index = 0
|
index = 0
|
||||||
@@ -857,8 +776,7 @@ def process_plugin_events(db, plugin, plugEventsArr):
|
|||||||
if tmpObjFromEvent.status == "exists":
|
if tmpObjFromEvent.status == "exists":
|
||||||
# compare hash of the changed watched columns for uniqueness - make sure you compare the values with the same idsHash before checking watchedHash
|
# compare hash of the changed watched columns for uniqueness - make sure you compare the values with the same idsHash before checking watchedHash
|
||||||
if any(
|
if any(
|
||||||
x.idsHash == tmpObjFromEvent.idsHash
|
x.idsHash == tmpObjFromEvent.idsHash and x.watchedHash != tmpObjFromEvent.watchedHash
|
||||||
and x.watchedHash != tmpObjFromEvent.watchedHash
|
|
||||||
for x in pluginObjects
|
for x in pluginObjects
|
||||||
):
|
):
|
||||||
pluginEvents[index].status = "watched-changed"
|
pluginEvents[index].status = "watched-changed"
|
||||||
@@ -955,25 +873,17 @@ def process_plugin_events(db, plugin, plugEventsArr):
|
|||||||
# combine all DB insert and update events into one for history
|
# combine all DB insert and update events into one for history
|
||||||
history_to_insert.append(values)
|
history_to_insert.append(values)
|
||||||
|
|
||||||
mylog("debug", ["[Plugins] pluginEvents count: ", len(pluginEvents)])
|
mylog("debug", f"[Plugins] pluginEvents count: {len(pluginEvents)}")
|
||||||
mylog("debug", ["[Plugins] pluginObjects count: ", len(pluginObjects)])
|
mylog("debug", f"[Plugins] pluginObjects count: {len(pluginObjects)}")
|
||||||
|
|
||||||
mylog(
|
mylog("debug", f"[Plugins] events_to_insert count: {len(events_to_insert)}")
|
||||||
"debug", ["[Plugins] events_to_insert count: ", len(events_to_insert)]
|
mylog("debug", f"[Plugins] history_to_insert count: {len(history_to_insert)}")
|
||||||
)
|
mylog("debug", f"[Plugins] objects_to_insert count: {len(objects_to_insert)}")
|
||||||
mylog(
|
mylog("debug", f"[Plugins] objects_to_update count: {len(objects_to_update)}")
|
||||||
"debug", ["[Plugins] history_to_insert count: ", len(history_to_insert)]
|
|
||||||
)
|
|
||||||
mylog(
|
|
||||||
"debug", ["[Plugins] objects_to_insert count: ", len(objects_to_insert)]
|
|
||||||
)
|
|
||||||
mylog(
|
|
||||||
"debug", ["[Plugins] objects_to_update count: ", len(objects_to_update)]
|
|
||||||
)
|
|
||||||
|
|
||||||
mylog("trace", ["[Plugins] objects_to_update: ", objects_to_update])
|
mylog("trace", f"[Plugins] objects_to_update: {objects_to_update}")
|
||||||
mylog("trace", ["[Plugins] events_to_insert: ", events_to_insert])
|
mylog("trace", f"[Plugins] events_to_insert: {events_to_insert}")
|
||||||
mylog("trace", ["[Plugins] history_to_insert: ", history_to_insert])
|
mylog("trace", f"[Plugins] history_to_insert: {history_to_insert}")
|
||||||
|
|
||||||
logEventStatusCounts("pluginEvents", pluginEvents)
|
logEventStatusCounts("pluginEvents", pluginEvents)
|
||||||
logEventStatusCounts("pluginObjects", pluginObjects)
|
logEventStatusCounts("pluginObjects", pluginObjects)
|
||||||
@@ -1044,7 +954,7 @@ def process_plugin_events(db, plugin, plugEventsArr):
|
|||||||
except Exception as e:
|
except Exception as e:
|
||||||
# Rollback the transaction in case of an error
|
# Rollback the transaction in case of an error
|
||||||
conn.rollback()
|
conn.rollback()
|
||||||
mylog("none", ["[Plugins] ⚠ ERROR: ", e])
|
mylog("none", f"[Plugins] ⚠ ERROR: {e}")
|
||||||
raise e
|
raise e
|
||||||
|
|
||||||
# Perform database table mapping if enabled for the plugin
|
# Perform database table mapping if enabled for the plugin
|
||||||
@@ -1056,7 +966,7 @@ def process_plugin_events(db, plugin, plugEventsArr):
|
|||||||
dbTable = plugin["mapped_to_table"]
|
dbTable = plugin["mapped_to_table"]
|
||||||
|
|
||||||
# Log a debug message indicating the mapping of objects to the database table.
|
# Log a debug message indicating the mapping of objects to the database table.
|
||||||
mylog("debug", ["[Plugins] Mapping objects to database table: ", dbTable])
|
mylog("debug", f"[Plugins] Mapping objects to database table: {dbTable}")
|
||||||
|
|
||||||
# Initialize lists to hold mapped column names, columnsStr, and valuesStr for SQL query.
|
# Initialize lists to hold mapped column names, columnsStr, and valuesStr for SQL query.
|
||||||
mappedCols = []
|
mappedCols = []
|
||||||
@@ -1121,8 +1031,7 @@ def process_plugin_events(db, plugin, plugEventsArr):
|
|||||||
|
|
||||||
# Check if there's a default value specified for this column in the JSON.
|
# Check if there's a default value specified for this column in the JSON.
|
||||||
if (
|
if (
|
||||||
"mapped_to_column_data" in col
|
"mapped_to_column_data" in col and "value" in col["mapped_to_column_data"]
|
||||||
and "value" in col["mapped_to_column_data"]
|
|
||||||
):
|
):
|
||||||
tmpList.append(col["mapped_to_column_data"]["value"])
|
tmpList.append(col["mapped_to_column_data"]["value"])
|
||||||
|
|
||||||
@@ -1133,8 +1042,8 @@ def process_plugin_events(db, plugin, plugEventsArr):
|
|||||||
q = f"INSERT OR IGNORE INTO {dbTable} ({columnsStr}) VALUES ({valuesStr})"
|
q = f"INSERT OR IGNORE INTO {dbTable} ({columnsStr}) VALUES ({valuesStr})"
|
||||||
|
|
||||||
# Log a debug message showing the generated SQL query for mapping.
|
# Log a debug message showing the generated SQL query for mapping.
|
||||||
mylog("debug", ["[Plugins] SQL query for mapping: ", q])
|
mylog("debug", f"[Plugins] SQL query for mapping: {q}")
|
||||||
mylog("debug", ["[Plugins] SQL sqlParams for mapping: ", sqlParams])
|
mylog("debug", f"[Plugins] SQL sqlParams for mapping: {sqlParams}")
|
||||||
|
|
||||||
# Execute the SQL query using 'sql.executemany()' and the 'sqlParams' list of tuples.
|
# Execute the SQL query using 'sql.executemany()' and the 'sqlParams' list of tuples.
|
||||||
# This will insert multiple rows into the database in one go.
|
# This will insert multiple rows into the database in one go.
|
||||||
|
|||||||
@@ -1,14 +1,6 @@
|
|||||||
import sys
|
|
||||||
import subprocess
|
import subprocess
|
||||||
import os
|
import os
|
||||||
import re
|
import re
|
||||||
import datetime
|
|
||||||
from dateutil import parser
|
|
||||||
|
|
||||||
# Register NetAlertX directories
|
|
||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
|
||||||
sys.path.extend([f"{INSTALL_PATH}/server"])
|
|
||||||
|
|
||||||
from helper import get_setting_value, check_IP_format
|
from helper import get_setting_value, check_IP_format
|
||||||
from utils.datetime_utils import timeNowDB, normalizeTimeStamp
|
from utils.datetime_utils import timeNowDB, normalizeTimeStamp
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger
|
||||||
@@ -57,9 +49,10 @@ def exclude_ignored_devices(db):
|
|||||||
|
|
||||||
sql.execute(query)
|
sql.execute(query)
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
def update_devices_data_from_scan (db):
|
# -------------------------------------------------------------------------------
|
||||||
sql = db.sql #TO-DO
|
def update_devices_data_from_scan(db):
|
||||||
|
sql = db.sql # TO-DO
|
||||||
startTime = timeNowDB()
|
startTime = timeNowDB()
|
||||||
|
|
||||||
# Update Last Connection
|
# Update Last Connection
|
||||||
@@ -425,9 +418,9 @@ def print_scan_stats(db):
|
|||||||
mylog("verbose", f" {row['cur_ScanMethod']}: {row['scan_method_count']}")
|
mylog("verbose", f" {row['cur_ScanMethod']}: {row['scan_method_count']}")
|
||||||
|
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
# -------------------------------------------------------------------------------
|
||||||
def create_new_devices (db):
|
def create_new_devices(db):
|
||||||
sql = db.sql # TO-DO
|
sql = db.sql # TO-DO
|
||||||
startTime = timeNowDB()
|
startTime = timeNowDB()
|
||||||
|
|
||||||
# Insert events for new devices from CurrentScan (not yet in Devices)
|
# Insert events for new devices from CurrentScan (not yet in Devices)
|
||||||
@@ -598,7 +591,8 @@ def create_new_devices (db):
|
|||||||
mylog("debug", "[New Devices] New Devices end")
|
mylog("debug", "[New Devices] New Devices end")
|
||||||
db.commitDB()
|
db.commitDB()
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
|
# -------------------------------------------------------------------------------
|
||||||
# Check if plugins data changed
|
# Check if plugins data changed
|
||||||
def check_plugin_data_changed(pm, plugins_to_check):
|
def check_plugin_data_changed(pm, plugins_to_check):
|
||||||
"""
|
"""
|
||||||
@@ -639,13 +633,13 @@ def check_plugin_data_changed(pm, plugins_to_check):
|
|||||||
# Normalize and validate last_changed timestamp
|
# Normalize and validate last_changed timestamp
|
||||||
last_changed_ts = normalizeTimeStamp(last_data_change)
|
last_changed_ts = normalizeTimeStamp(last_data_change)
|
||||||
|
|
||||||
if last_changed_ts == None:
|
if last_changed_ts is None:
|
||||||
mylog('none', f'[check_plugin_data_changed] Unexpected last_data_change timestamp for {plugin_name} (input|output): ({last_data_change}|{last_changed_ts})')
|
mylog('none', f'[check_plugin_data_changed] Unexpected last_data_change timestamp for {plugin_name} (input|output): ({last_data_change}|{last_changed_ts})')
|
||||||
|
|
||||||
# Normalize and validate last_data_check timestamp
|
# Normalize and validate last_data_check timestamp
|
||||||
last_data_check_ts = normalizeTimeStamp(last_data_check)
|
last_data_check_ts = normalizeTimeStamp(last_data_check)
|
||||||
|
|
||||||
if last_data_check_ts == None:
|
if last_data_check_ts is None:
|
||||||
mylog('none', f'[check_plugin_data_changed] Unexpected last_data_check timestamp for {plugin_name} (input|output): ({last_data_check}|{last_data_check_ts})')
|
mylog('none', f'[check_plugin_data_changed] Unexpected last_data_check timestamp for {plugin_name} (input|output): ({last_data_check}|{last_data_check_ts})')
|
||||||
|
|
||||||
# Track which plugins have newer state than last_checked
|
# Track which plugins have newer state than last_checked
|
||||||
@@ -660,15 +654,19 @@ def check_plugin_data_changed(pm, plugins_to_check):
|
|||||||
|
|
||||||
# Continue if changes detected
|
# Continue if changes detected
|
||||||
for p in plugins_changed:
|
for p in plugins_changed:
|
||||||
mylog('debug', f'[check_plugin_data_changed] {p} changed (last_data_change|last_data_check): ({pm.plugin_states.get(p, {}).get("lastDataChange")}|{pm.plugin_checks.get(p)})')
|
mylog(
|
||||||
|
'debug',
|
||||||
|
f'[check_plugin_data_changed] {p} changed (last_data_change|last_data_check): ({pm.plugin_states.get(p, {}).get("lastDataChange")}|{pm.plugin_checks.get(p)})'
|
||||||
|
)
|
||||||
|
|
||||||
return True
|
return True
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
|
# -------------------------------------------------------------------------------
|
||||||
def update_devices_names(pm):
|
def update_devices_names(pm):
|
||||||
|
|
||||||
# --- Short-circuit if no name-resolution plugin has changed ---
|
# --- Short-circuit if no name-resolution plugin has changed ---
|
||||||
if check_plugin_data_changed(pm, ["DIGSCAN", "NSLOOKUP", "NBTSCAN", "AVAHISCAN"]) == False:
|
if check_plugin_data_changed(pm, ["DIGSCAN", "NSLOOKUP", "NBTSCAN", "AVAHISCAN"]) is False:
|
||||||
mylog('debug', '[Update Device Name] No relevant plugin changes since last check.')
|
mylog('debug', '[Update Device Name] No relevant plugin changes since last check.')
|
||||||
return
|
return
|
||||||
|
|
||||||
@@ -722,8 +720,7 @@ def update_devices_names(pm):
|
|||||||
|
|
||||||
# If a valid result is found, record it and stop further attempts
|
# If a valid result is found, record it and stop further attempts
|
||||||
if (
|
if (
|
||||||
newFQDN not in [nameNotFound, "", "localhost."]
|
newFQDN not in [nameNotFound, "", "localhost."] and " communications error to " not in newFQDN
|
||||||
and " communications error to " not in newFQDN
|
|
||||||
):
|
):
|
||||||
foundStats[label] += 1
|
foundStats[label] += 1
|
||||||
|
|
||||||
@@ -750,14 +747,14 @@ def update_devices_names(pm):
|
|||||||
)
|
)
|
||||||
|
|
||||||
# Try resolving both name and FQDN
|
# Try resolving both name and FQDN
|
||||||
recordsToUpdate, recordsNotFound, foundStats, notFound = resolve_devices(
|
recordsToUpdate, recordsNotFound, fs, notFound = resolve_devices(
|
||||||
unknownDevices
|
unknownDevices
|
||||||
)
|
)
|
||||||
|
|
||||||
# Log summary
|
# Log summary
|
||||||
mylog(
|
mylog(
|
||||||
"verbose",
|
"verbose",
|
||||||
f"[Update Device Name] Names Found (DIGSCAN/AVAHISCAN/NSLOOKUP/NBTSCAN): {len(recordsToUpdate)} ({foundStats['DIGSCAN']}/{foundStats['AVAHISCAN']}/{foundStats['NSLOOKUP']}/{foundStats['NBTSCAN']})",
|
f"[Update Device Name] Names Found (DIGSCAN/AVAHISCAN/NSLOOKUP/NBTSCAN): {len(recordsToUpdate)} ({fs['DIGSCAN']}/{fs['AVAHISCAN']}/{fs['NSLOOKUP']}/{fs['NBTSCAN']})",
|
||||||
)
|
)
|
||||||
mylog("verbose", f"[Update Device Name] Names Not Found : {notFound}")
|
mylog("verbose", f"[Update Device Name] Names Not Found : {notFound}")
|
||||||
|
|
||||||
@@ -780,16 +777,14 @@ def update_devices_names(pm):
|
|||||||
)
|
)
|
||||||
|
|
||||||
# Try resolving only FQDN
|
# Try resolving only FQDN
|
||||||
recordsToUpdate, _, foundStats, notFound = resolve_devices(
|
recordsToUpdate, _, fs, notFound = resolve_devices(
|
||||||
allDevices, resolve_both_name_and_fqdn=False
|
allDevices, resolve_both_name_and_fqdn=False
|
||||||
)
|
)
|
||||||
|
|
||||||
# Log summary
|
# Log summary
|
||||||
mylog(
|
mylog(
|
||||||
"verbose",
|
"verbose",
|
||||||
f"[Update FQDN] Names Found (DIGSCAN/AVAHISCAN/NSLOOKUP/NBTSCAN): {len(recordsToUpdate)}"+
|
f"[Update FQDN] Names Found (DIGSCAN/AVAHISCAN/NSLOOKUP/NBTSCAN): {len(recordsToUpdate)}({fs['DIGSCAN']}/{fs['AVAHISCAN']}/{fs['NSLOOKUP']}/{fs['NBTSCAN']})",
|
||||||
f"({foundStats['DIGSCAN']}/{foundStats['AVAHISCAN']}/{foundStats['NSLOOKUP']}"+
|
|
||||||
f"/{foundStats['NBTSCAN']})",
|
|
||||||
)
|
)
|
||||||
mylog("verbose", f"[Update FQDN] Names Not Found : {notFound}")
|
mylog("verbose", f"[Update FQDN] Names Not Found : {notFound}")
|
||||||
|
|
||||||
@@ -803,7 +798,7 @@ def update_devices_names(pm):
|
|||||||
|
|
||||||
# --- Step 3: Log last checked time ---
|
# --- Step 3: Log last checked time ---
|
||||||
# After resolving names, update last checked
|
# After resolving names, update last checked
|
||||||
pm.plugin_checks = {"DIGSCAN": timeNowDB(), "AVAHISCAN": timeNowDB(), "NSLOOKUP": timeNowDB(), "NBTSCAN": timeNowDB() }
|
pm.plugin_checks = {"DIGSCAN": timeNowDB(), "AVAHISCAN": timeNowDB(), "NSLOOKUP": timeNowDB(), "NBTSCAN": timeNowDB()}
|
||||||
|
|
||||||
|
|
||||||
# -------------------------------------------------------------------------------
|
# -------------------------------------------------------------------------------
|
||||||
@@ -901,7 +896,6 @@ def query_MAC_vendor(pMAC):
|
|||||||
|
|
||||||
# Search vendor in HW Vendors DB
|
# Search vendor in HW Vendors DB
|
||||||
mac_start_string6 = mac[0:6]
|
mac_start_string6 = mac[0:6]
|
||||||
mac_start_string9 = mac[0:9]
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
with open(filePath, "r") as f:
|
with open(filePath, "r") as f:
|
||||||
|
|||||||
@@ -1,16 +1,13 @@
|
|||||||
import sys
|
|
||||||
import os
|
import os
|
||||||
import re
|
import re
|
||||||
import json
|
import json
|
||||||
import base64
|
import base64
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
from typing import Optional, Tuple
|
from typing import Optional, Tuple
|
||||||
|
from logger import mylog
|
||||||
|
|
||||||
# Register NetAlertX directories
|
# Register NetAlertX directories
|
||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
||||||
sys.path.extend([f"{INSTALL_PATH}/server"])
|
|
||||||
|
|
||||||
from logger import mylog
|
|
||||||
|
|
||||||
# Load MAC/device-type/icon rules from external file
|
# Load MAC/device-type/icon rules from external file
|
||||||
MAC_TYPE_ICON_PATH = Path(f"{INSTALL_PATH}/back/device_heuristics_rules.json")
|
MAC_TYPE_ICON_PATH = Path(f"{INSTALL_PATH}/back/device_heuristics_rules.json")
|
||||||
@@ -83,7 +80,7 @@ def match_vendor(vendor: str, default_type: str, default_icon: str) -> Tuple[str
|
|||||||
|
|
||||||
for pattern in patterns:
|
for pattern in patterns:
|
||||||
# Only apply fallback when no MAC prefix is specified
|
# Only apply fallback when no MAC prefix is specified
|
||||||
mac_prefix = pattern.get("mac_prefix", "")
|
# mac_prefix = pattern.get("mac_prefix", "")
|
||||||
vendor_pattern = pattern.get("vendor", "").lower()
|
vendor_pattern = pattern.get("vendor", "").lower()
|
||||||
|
|
||||||
if vendor_pattern and vendor_pattern in vendor_lc:
|
if vendor_pattern and vendor_pattern in vendor_lc:
|
||||||
|
|||||||
@@ -1,11 +1,4 @@
|
|||||||
import sys
|
|
||||||
import os
|
|
||||||
import re
|
import re
|
||||||
|
|
||||||
# Register NetAlertX directories
|
|
||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
|
||||||
sys.path.extend([f"{INSTALL_PATH}/server"])
|
|
||||||
|
|
||||||
from logger import mylog
|
from logger import mylog
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value
|
||||||
|
|
||||||
|
|||||||
@@ -1,10 +1,3 @@
|
|||||||
import sys
|
|
||||||
import os
|
|
||||||
|
|
||||||
# Register NetAlertX directories
|
|
||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
|
||||||
sys.path.extend([f"{INSTALL_PATH}/server"])
|
|
||||||
|
|
||||||
from scan.device_handling import (
|
from scan.device_handling import (
|
||||||
create_new_devices,
|
create_new_devices,
|
||||||
print_scan_stats,
|
print_scan_stats,
|
||||||
@@ -14,7 +7,7 @@ from scan.device_handling import (
|
|||||||
)
|
)
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value
|
||||||
from db.db_helper import print_table_schema
|
from db.db_helper import print_table_schema
|
||||||
from utils.datetime_utils import timeNowDB, timeNowTZ
|
from utils.datetime_utils import timeNowDB
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger
|
||||||
from messaging.reporting import skip_repeated_notifications
|
from messaging.reporting import skip_repeated_notifications
|
||||||
|
|
||||||
@@ -133,9 +126,9 @@ def create_sessions_snapshot(db):
|
|||||||
db.commitDB()
|
db.commitDB()
|
||||||
|
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
# -------------------------------------------------------------------------------
|
||||||
def insert_events (db):
|
def insert_events(db):
|
||||||
sql = db.sql #TO-DO
|
sql = db.sql # TO-DO
|
||||||
startTime = timeNowDB()
|
startTime = timeNowDB()
|
||||||
|
|
||||||
# Check device down
|
# Check device down
|
||||||
|
|||||||
@@ -1,49 +1,43 @@
|
|||||||
#!/usr/bin/env python
|
# !/usr/bin/env python
|
||||||
|
|
||||||
import os
|
# from datetime import datetime
|
||||||
import pathlib
|
|
||||||
import sys
|
|
||||||
from datetime import datetime
|
|
||||||
from dateutil import parser
|
from dateutil import parser
|
||||||
import datetime
|
import datetime
|
||||||
import re
|
import re
|
||||||
import pytz
|
import pytz
|
||||||
from pytz import timezone
|
|
||||||
from typing import Union
|
from typing import Union
|
||||||
from zoneinfo import ZoneInfo
|
from zoneinfo import ZoneInfo
|
||||||
import email.utils
|
import email.utils
|
||||||
|
|
||||||
# Register NetAlertX directories
|
|
||||||
INSTALL_PATH="/app"
|
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
|
||||||
|
|
||||||
import conf
|
import conf
|
||||||
from const import *
|
# from const import *
|
||||||
|
|
||||||
|
|
||||||
|
# -------------------------------------------------------------------------------
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
# DateTime
|
# DateTime
|
||||||
#-------------------------------------------------------------------------------
|
# -------------------------------------------------------------------------------
|
||||||
|
|
||||||
DATETIME_PATTERN = "%Y-%m-%d %H:%M:%S"
|
DATETIME_PATTERN = "%Y-%m-%d %H:%M:%S"
|
||||||
DATETIME_REGEX = re.compile(r'^\d{4}-\d{2}-\d{2} \d{2}:\d{2}:\d{2}$')
|
DATETIME_REGEX = re.compile(r'^\d{4}-\d{2}-\d{2} \d{2}:\d{2}:\d{2}$')
|
||||||
|
|
||||||
|
|
||||||
def timeNowTZ():
|
def timeNowTZ():
|
||||||
if conf.tz:
|
if conf.tz:
|
||||||
return datetime.datetime.now(conf.tz).replace(microsecond=0)
|
return datetime.datetime.now(conf.tz).replace(microsecond=0)
|
||||||
else:
|
else:
|
||||||
return datetime.datetime.now().replace(microsecond=0)
|
return datetime.datetime.now().replace(microsecond=0)
|
||||||
|
|
||||||
|
|
||||||
def timeNow():
|
def timeNow():
|
||||||
return datetime.datetime.now().replace(microsecond=0)
|
return datetime.datetime.now().replace(microsecond=0)
|
||||||
|
|
||||||
|
|
||||||
def get_timezone_offset():
|
def get_timezone_offset():
|
||||||
now = datetime.datetime.now(conf.tz)
|
now = datetime.datetime.now(conf.tz)
|
||||||
offset_hours = now.utcoffset().total_seconds() / 3600
|
offset_hours = now.utcoffset().total_seconds() / 3600
|
||||||
offset_formatted = "{:+03d}:{:02d}".format(int(offset_hours), int((offset_hours % 1) * 60))
|
offset_formatted = "{:+03d}:{:02d}".format(int(offset_hours), int((offset_hours % 1) * 60))
|
||||||
return offset_formatted
|
return offset_formatted
|
||||||
|
|
||||||
|
|
||||||
def timeNowDB(local=True):
|
def timeNowDB(local=True):
|
||||||
"""
|
"""
|
||||||
Return the current time (local or UTC) as ISO 8601 for DB storage.
|
Return the current time (local or UTC) as ISO 8601 for DB storage.
|
||||||
@@ -67,9 +61,9 @@ def timeNowDB(local=True):
|
|||||||
return datetime.datetime.now(datetime.UTC).strftime(DATETIME_PATTERN)
|
return datetime.datetime.now(datetime.UTC).strftime(DATETIME_PATTERN)
|
||||||
|
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
# -------------------------------------------------------------------------------
|
||||||
# Date and time methods
|
# Date and time methods
|
||||||
#-------------------------------------------------------------------------------
|
# -------------------------------------------------------------------------------
|
||||||
|
|
||||||
def normalizeTimeStamp(inputTimeStamp):
|
def normalizeTimeStamp(inputTimeStamp):
|
||||||
"""
|
"""
|
||||||
@@ -125,6 +119,7 @@ def format_date_iso(date1: str) -> str:
|
|||||||
dt = datetime.datetime.fromisoformat(date1) if isinstance(date1, str) else date1
|
dt = datetime.datetime.fromisoformat(date1) if isinstance(date1, str) else date1
|
||||||
return dt.isoformat()
|
return dt.isoformat()
|
||||||
|
|
||||||
|
|
||||||
# -------------------------------------------------------------------------------------------
|
# -------------------------------------------------------------------------------------------
|
||||||
def format_event_date(date_str: str, event_type: str) -> str:
|
def format_event_date(date_str: str, event_type: str) -> str:
|
||||||
"""Format event date with fallback rules."""
|
"""Format event date with fallback rules."""
|
||||||
@@ -135,6 +130,7 @@ def format_event_date(date_str: str, event_type: str) -> str:
|
|||||||
else:
|
else:
|
||||||
return "<still connected>"
|
return "<still connected>"
|
||||||
|
|
||||||
|
|
||||||
# -------------------------------------------------------------------------------------------
|
# -------------------------------------------------------------------------------------------
|
||||||
def ensure_datetime(dt: Union[str, datetime.datetime, None]) -> datetime.datetime:
|
def ensure_datetime(dt: Union[str, datetime.datetime, None]) -> datetime.datetime:
|
||||||
if dt is None:
|
if dt is None:
|
||||||
@@ -157,6 +153,7 @@ def parse_datetime(dt_str):
|
|||||||
except ValueError:
|
except ValueError:
|
||||||
return None
|
return None
|
||||||
|
|
||||||
|
|
||||||
def format_date(date_str: str) -> str:
|
def format_date(date_str: str) -> str:
|
||||||
try:
|
try:
|
||||||
dt = parse_datetime(date_str)
|
dt = parse_datetime(date_str)
|
||||||
@@ -168,6 +165,7 @@ def format_date(date_str: str) -> str:
|
|||||||
except (ValueError, AttributeError, TypeError):
|
except (ValueError, AttributeError, TypeError):
|
||||||
return "invalid"
|
return "invalid"
|
||||||
|
|
||||||
|
|
||||||
def format_date_diff(date1, date2, tz_name):
|
def format_date_diff(date1, date2, tz_name):
|
||||||
"""
|
"""
|
||||||
Return difference between two datetimes as 'Xd HH:MM'.
|
Return difference between two datetimes as 'Xd HH:MM'.
|
||||||
@@ -184,8 +182,8 @@ def format_date_diff(date1, date2, tz_name):
|
|||||||
try:
|
try:
|
||||||
dt_parsed = email.utils.parsedate_to_datetime(dt)
|
dt_parsed = email.utils.parsedate_to_datetime(dt)
|
||||||
except (ValueError, TypeError):
|
except (ValueError, TypeError):
|
||||||
# fallback: parse ISO string
|
# fallback: parse ISO string
|
||||||
dt_parsed = datetime.datetime.fromisoformat(dt)
|
dt_parsed = datetime.datetime.fromisoformat(dt)
|
||||||
# convert naive GMT/UTC to app timezone
|
# convert naive GMT/UTC to app timezone
|
||||||
if dt_parsed.tzinfo is None:
|
if dt_parsed.tzinfo is None:
|
||||||
dt_parsed = tz.localize(dt_parsed)
|
dt_parsed = tz.localize(dt_parsed)
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
import os
|
import os
|
||||||
import json
|
import json
|
||||||
|
from collections import namedtuple
|
||||||
import conf
|
import conf
|
||||||
from logger import mylog
|
from logger import mylog
|
||||||
from utils.crypto_utils import decrypt_data
|
from utils.crypto_utils import decrypt_data
|
||||||
@@ -220,9 +220,7 @@ def get_plugins_configs(loadAll):
|
|||||||
# Load all plugins if `loadAll` is True, the plugin is in the enabled list,
|
# Load all plugins if `loadAll` is True, the plugin is in the enabled list,
|
||||||
# or no specific plugins are enabled (enabledPlugins is empty)
|
# or no specific plugins are enabled (enabledPlugins is empty)
|
||||||
if (
|
if (
|
||||||
loadAll
|
loadAll or plugJson["unique_prefix"] in enabledPlugins or enabledPlugins == []
|
||||||
or plugJson["unique_prefix"] in enabledPlugins
|
|
||||||
or enabledPlugins == []
|
|
||||||
):
|
):
|
||||||
# Load the contents of the config.json file as a JSON object and append it to pluginsList
|
# Load the contents of the config.json file as a JSON object and append it to pluginsList
|
||||||
pluginsList.append(plugJson)
|
pluginsList.append(plugJson)
|
||||||
|
|||||||
@@ -1,11 +1,4 @@
|
|||||||
import sqlite3
|
import sqlite3
|
||||||
import os
|
|
||||||
import sys
|
|
||||||
|
|
||||||
# Register NetAlertX directories
|
|
||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
|
||||||
sys.path.extend([f"{INSTALL_PATH}/server"])
|
|
||||||
|
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value
|
||||||
from models.device_instance import DeviceInstance
|
from models.device_instance import DeviceInstance
|
||||||
@@ -15,7 +8,6 @@ from models.plugin_object_instance import PluginObjectInstance
|
|||||||
Logger(get_setting_value("LOG_LEVEL"))
|
Logger(get_setting_value("LOG_LEVEL"))
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
class Action:
|
class Action:
|
||||||
"""Base class for all actions."""
|
"""Base class for all actions."""
|
||||||
|
|
||||||
|
|||||||
@@ -1,10 +1,3 @@
|
|||||||
import os
|
|
||||||
import sys
|
|
||||||
|
|
||||||
# Register NetAlertX directories
|
|
||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
|
||||||
sys.path.extend([f"{INSTALL_PATH}/server"])
|
|
||||||
|
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value
|
||||||
from logger import Logger
|
from logger import Logger
|
||||||
from const import sql_generateGuid
|
from const import sql_generateGuid
|
||||||
|
|||||||
@@ -1,12 +1,5 @@
|
|||||||
import re
|
import re
|
||||||
import json
|
import json
|
||||||
import os
|
|
||||||
import sys
|
|
||||||
|
|
||||||
# Register NetAlertX directories
|
|
||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
|
||||||
sys.path.extend([f"{INSTALL_PATH}/server"])
|
|
||||||
|
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value
|
||||||
|
|
||||||
|
|||||||
@@ -1,22 +1,17 @@
|
|||||||
import json
|
import json
|
||||||
import os
|
|
||||||
import sys
|
|
||||||
|
|
||||||
# Register NetAlertX directories
|
|
||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
|
||||||
sys.path.extend([f"{INSTALL_PATH}/server"])
|
|
||||||
|
|
||||||
from const import fullConfFolder
|
from const import fullConfFolder
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value
|
||||||
|
|
||||||
# Make sure log level is initialized correctly
|
|
||||||
Logger(get_setting_value("LOG_LEVEL"))
|
|
||||||
|
|
||||||
from workflows.triggers import Trigger
|
from workflows.triggers import Trigger
|
||||||
from workflows.conditions import ConditionGroup
|
from workflows.conditions import ConditionGroup
|
||||||
from workflows.actions import DeleteObjectAction, RunPluginAction, UpdateFieldAction
|
from workflows.actions import DeleteObjectAction, RunPluginAction, UpdateFieldAction
|
||||||
|
|
||||||
|
|
||||||
|
# Make sure log level is initialized correctly
|
||||||
|
Logger(get_setting_value("LOG_LEVEL"))
|
||||||
|
|
||||||
|
|
||||||
class WorkflowManager:
|
class WorkflowManager:
|
||||||
def __init__(self, db):
|
def __init__(self, db):
|
||||||
self.db = db
|
self.db = db
|
||||||
|
|||||||
@@ -1,11 +1,4 @@
|
|||||||
import json
|
import json
|
||||||
import os
|
|
||||||
import sys
|
|
||||||
|
|
||||||
# Register NetAlertX directories
|
|
||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
|
||||||
sys.path.extend([f"{INSTALL_PATH}/server"])
|
|
||||||
|
|
||||||
from logger import mylog, Logger
|
from logger import mylog, Logger
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value
|
||||||
from database import get_array_from_sql_rows
|
from database import get_array_from_sql_rows
|
||||||
@@ -28,8 +21,7 @@ class Trigger:
|
|||||||
self.event_type = triggerJson["event_type"]
|
self.event_type = triggerJson["event_type"]
|
||||||
self.event = event # Store the triggered event context, if provided
|
self.event = event # Store the triggered event context, if provided
|
||||||
self.triggered = (
|
self.triggered = (
|
||||||
self.object_type == event["ObjectType"]
|
self.object_type == event["ObjectType"] and self.event_type == event["AppEventType"]
|
||||||
and self.event_type == event["AppEventType"]
|
|
||||||
)
|
)
|
||||||
|
|
||||||
mylog(
|
mylog(
|
||||||
|
|||||||
@@ -7,9 +7,9 @@ import pytest
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.datetime_utils import timeNowDB
|
from utils.datetime_utils import timeNowDB # noqa: E402 [flake8 lint suppression]
|
||||||
from api_server.api_server_start import app
|
from api_server.api_server_start import app # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture(scope="session")
|
@pytest.fixture(scope="session")
|
||||||
@@ -26,7 +26,7 @@ def client():
|
|||||||
@pytest.fixture(scope="session")
|
@pytest.fixture(scope="session")
|
||||||
def test_mac():
|
def test_mac():
|
||||||
# Generate a unique MAC for each test run
|
# Generate a unique MAC for each test run
|
||||||
return "AA:BB:CC:" + ":".join(f"{random.randint(0,255):02X}" for _ in range(3))
|
return "AA:BB:CC:" + ":".join(f"{random.randint(0, 255):02X}" for _ in range(3))
|
||||||
|
|
||||||
|
|
||||||
def auth_headers(token):
|
def auth_headers(token):
|
||||||
|
|||||||
@@ -1,17 +1,17 @@
|
|||||||
import sys
|
import sys
|
||||||
import pathlib
|
# import pathlib
|
||||||
import sqlite3
|
# import sqlite3
|
||||||
import random
|
import random
|
||||||
import string
|
# import string
|
||||||
import uuid
|
# import uuid
|
||||||
import os
|
import os
|
||||||
import pytest
|
import pytest
|
||||||
|
|
||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from api_server.api_server_start import app
|
from api_server.api_server_start import app # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture(scope="session")
|
@pytest.fixture(scope="session")
|
||||||
@@ -28,7 +28,7 @@ def client():
|
|||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
def test_mac():
|
def test_mac():
|
||||||
# Generate a unique MAC for each test run
|
# Generate a unique MAC for each test run
|
||||||
return "AA:BB:CC:" + ":".join(f"{random.randint(0,255):02X}" for _ in range(3))
|
return "AA:BB:CC:" + ":".join(f"{random.randint(0, 255):02X}" for _ in range(3))
|
||||||
|
|
||||||
|
|
||||||
def auth_headers(token):
|
def auth_headers(token):
|
||||||
@@ -38,7 +38,6 @@ def auth_headers(token):
|
|||||||
def test_create_device(client, api_token, test_mac):
|
def test_create_device(client, api_token, test_mac):
|
||||||
payload = {
|
payload = {
|
||||||
"createNew": True,
|
"createNew": True,
|
||||||
"devType": "Test Device",
|
|
||||||
"devOwner": "Unit Test",
|
"devOwner": "Unit Test",
|
||||||
"devType": "Router",
|
"devType": "Router",
|
||||||
"devVendor": "TestVendor",
|
"devVendor": "TestVendor",
|
||||||
@@ -103,7 +102,7 @@ def test_copy_device(client, api_token, test_mac):
|
|||||||
|
|
||||||
# Step 2: Generate a target MAC
|
# Step 2: Generate a target MAC
|
||||||
target_mac = "AA:BB:CC:" + ":".join(
|
target_mac = "AA:BB:CC:" + ":".join(
|
||||||
f"{random.randint(0,255):02X}" for _ in range(3)
|
f"{random.randint(0, 255):02X}" for _ in range(3)
|
||||||
)
|
)
|
||||||
|
|
||||||
# Step 3: Copy device
|
# Step 3: Copy device
|
||||||
|
|||||||
@@ -1,32 +1,36 @@
|
|||||||
import sys
|
import sys
|
||||||
import pathlib
|
# import pathlib
|
||||||
import sqlite3
|
# import sqlite3
|
||||||
import base64
|
import base64
|
||||||
import random
|
import random
|
||||||
import string
|
# import string
|
||||||
import uuid
|
# import uuid
|
||||||
import os
|
import os
|
||||||
import pytest
|
import pytest
|
||||||
|
|
||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from api_server.api_server_start import app
|
from api_server.api_server_start import app # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture(scope="session")
|
@pytest.fixture(scope="session")
|
||||||
def api_token():
|
def api_token():
|
||||||
return get_setting_value("API_TOKEN")
|
return get_setting_value("API_TOKEN")
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
def client():
|
def client():
|
||||||
with app.test_client() as client:
|
with app.test_client() as client:
|
||||||
yield client
|
yield client
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
def test_mac():
|
def test_mac():
|
||||||
# Generate a unique MAC for each test run
|
# Generate a unique MAC for each test run
|
||||||
return "AA:BB:CC:" + ":".join(f"{random.randint(0,255):02X}" for _ in range(3))
|
return "AA:BB:CC:" + ":".join(f"{random.randint(0, 255):02X}" for _ in range(3))
|
||||||
|
|
||||||
|
|
||||||
def auth_headers(token):
|
def auth_headers(token):
|
||||||
return {"Authorization": f"Bearer {token}"}
|
return {"Authorization": f"Bearer {token}"}
|
||||||
@@ -40,7 +44,8 @@ def create_dummy(client, api_token, test_mac):
|
|||||||
"devType": "Router",
|
"devType": "Router",
|
||||||
"devVendor": "TestVendor",
|
"devVendor": "TestVendor",
|
||||||
}
|
}
|
||||||
resp = client.post(f"/device/{test_mac}", json=payload, headers=auth_headers(api_token))
|
client.post(f"/device/{test_mac}", json=payload, headers=auth_headers(api_token))
|
||||||
|
|
||||||
|
|
||||||
def test_get_all_devices(client, api_token, test_mac):
|
def test_get_all_devices(client, api_token, test_mac):
|
||||||
# Ensure there is at least one device
|
# Ensure there is at least one device
|
||||||
@@ -67,6 +72,7 @@ def test_delete_devices_with_macs(client, api_token, test_mac):
|
|||||||
assert resp.status_code == 200
|
assert resp.status_code == 200
|
||||||
assert resp.json.get("success") is True
|
assert resp.json.get("success") is True
|
||||||
|
|
||||||
|
|
||||||
def test_delete_all_empty_macs(client, api_token):
|
def test_delete_all_empty_macs(client, api_token):
|
||||||
resp = client.delete("/devices/empty-macs", headers=auth_headers(api_token))
|
resp = client.delete("/devices/empty-macs", headers=auth_headers(api_token))
|
||||||
assert resp.status_code == 200
|
assert resp.status_code == 200
|
||||||
@@ -79,6 +85,7 @@ def test_delete_unknown_devices(client, api_token):
|
|||||||
assert resp.status_code == 200
|
assert resp.status_code == 200
|
||||||
assert resp.json.get("success") is True
|
assert resp.json.get("success") is True
|
||||||
|
|
||||||
|
|
||||||
def test_export_devices_csv(client, api_token, test_mac):
|
def test_export_devices_csv(client, api_token, test_mac):
|
||||||
# Create a device first
|
# Create a device first
|
||||||
create_dummy(client, api_token, test_mac)
|
create_dummy(client, api_token, test_mac)
|
||||||
@@ -92,6 +99,7 @@ def test_export_devices_csv(client, api_token, test_mac):
|
|||||||
# CSV should contain test_mac
|
# CSV should contain test_mac
|
||||||
assert test_mac in resp.data.decode()
|
assert test_mac in resp.data.decode()
|
||||||
|
|
||||||
|
|
||||||
def test_export_devices_json(client, api_token, test_mac):
|
def test_export_devices_json(client, api_token, test_mac):
|
||||||
# Create a device first
|
# Create a device first
|
||||||
create_dummy(client, api_token, test_mac)
|
create_dummy(client, api_token, test_mac)
|
||||||
@@ -101,7 +109,7 @@ def test_export_devices_json(client, api_token, test_mac):
|
|||||||
assert resp.status_code == 200
|
assert resp.status_code == 200
|
||||||
assert resp.is_json
|
assert resp.is_json
|
||||||
data = resp.get_json()
|
data = resp.get_json()
|
||||||
assert any(dev.get("devMac") == test_mac for dev in data["data"])
|
assert any(dev.get("devMac") == test_mac for dev in data["data"])
|
||||||
|
|
||||||
|
|
||||||
def test_export_devices_invalid_format(client, api_token):
|
def test_export_devices_invalid_format(client, api_token):
|
||||||
@@ -143,6 +151,7 @@ def test_export_import_cycle_base64(client, api_token, test_mac):
|
|||||||
assert resp.json.get("inserted") >= 1
|
assert resp.json.get("inserted") >= 1
|
||||||
assert resp.json.get("skipped_lines") == []
|
assert resp.json.get("skipped_lines") == []
|
||||||
|
|
||||||
|
|
||||||
def test_devices_totals(client, api_token, test_mac):
|
def test_devices_totals(client, api_token, test_mac):
|
||||||
# 1. Create a dummy device
|
# 1. Create a dummy device
|
||||||
create_dummy(client, api_token, test_mac)
|
create_dummy(client, api_token, test_mac)
|
||||||
@@ -189,6 +198,7 @@ def test_devices_by_status(client, api_token, test_mac):
|
|||||||
assert fav_data is not None
|
assert fav_data is not None
|
||||||
assert "★" in fav_data["title"]
|
assert "★" in fav_data["title"]
|
||||||
|
|
||||||
|
|
||||||
def test_delete_test_devices(client, api_token, test_mac):
|
def test_delete_test_devices(client, api_token, test_mac):
|
||||||
|
|
||||||
# Delete by MAC
|
# Delete by MAC
|
||||||
|
|||||||
@@ -1,37 +1,38 @@
|
|||||||
import sys
|
import sys
|
||||||
import pathlib
|
|
||||||
import sqlite3
|
|
||||||
import random
|
|
||||||
import string
|
|
||||||
import uuid
|
|
||||||
import os
|
import os
|
||||||
import pytest
|
import pytest
|
||||||
from datetime import datetime, timedelta
|
import random
|
||||||
|
from datetime import timedelta
|
||||||
|
|
||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.datetime_utils import timeNowTZ
|
from utils.datetime_utils import timeNowTZ # noqa: E402 [flake8 lint suppression]
|
||||||
from api_server.api_server_start import app
|
from api_server.api_server_start import app # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture(scope="session")
|
@pytest.fixture(scope="session")
|
||||||
def api_token():
|
def api_token():
|
||||||
return get_setting_value("API_TOKEN")
|
return get_setting_value("API_TOKEN")
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
def client():
|
def client():
|
||||||
with app.test_client() as client:
|
with app.test_client() as client:
|
||||||
yield client
|
yield client
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
def test_mac():
|
def test_mac():
|
||||||
# Generate a unique MAC for each test run
|
# Generate a unique MAC for each test run
|
||||||
return "AA:BB:CC:" + ":".join(f"{random.randint(0,255):02X}" for _ in range(3))
|
return "AA:BB:CC:" + ":".join(f"{random.randint(0, 255):02X}" for _ in range(3))
|
||||||
|
|
||||||
|
|
||||||
def auth_headers(token):
|
def auth_headers(token):
|
||||||
return {"Authorization": f"Bearer {token}"}
|
return {"Authorization": f"Bearer {token}"}
|
||||||
|
|
||||||
|
|
||||||
def create_event(client, api_token, mac, event="UnitTest Event", days_old=None):
|
def create_event(client, api_token, mac, event="UnitTest Event", days_old=None):
|
||||||
payload = {"ip": "0.0.0.0", "event_type": event}
|
payload = {"ip": "0.0.0.0", "event_type": event}
|
||||||
|
|
||||||
@@ -43,10 +44,12 @@ def create_event(client, api_token, mac, event="UnitTest Event", days_old=None):
|
|||||||
|
|
||||||
return client.post(f"/events/create/{mac}", json=payload, headers=auth_headers(api_token))
|
return client.post(f"/events/create/{mac}", json=payload, headers=auth_headers(api_token))
|
||||||
|
|
||||||
|
|
||||||
def list_events(client, api_token, mac=None):
|
def list_events(client, api_token, mac=None):
|
||||||
url = "/events" if mac is None else f"/events?mac={mac}"
|
url = "/events" if mac is None else f"/events?mac={mac}"
|
||||||
return client.get(url, headers=auth_headers(api_token))
|
return client.get(url, headers=auth_headers(api_token))
|
||||||
|
|
||||||
|
|
||||||
def test_create_event(client, api_token, test_mac):
|
def test_create_event(client, api_token, test_mac):
|
||||||
# create event
|
# create event
|
||||||
resp = create_event(client, api_token, test_mac)
|
resp = create_event(client, api_token, test_mac)
|
||||||
@@ -82,6 +85,7 @@ def test_delete_events_for_mac(client, api_token, test_mac):
|
|||||||
assert resp.status_code == 200
|
assert resp.status_code == 200
|
||||||
assert len(resp.json.get("events", [])) == 0
|
assert len(resp.json.get("events", [])) == 0
|
||||||
|
|
||||||
|
|
||||||
def test_get_events_totals(client, api_token):
|
def test_get_events_totals(client, api_token):
|
||||||
# 1. Request totals with default period
|
# 1. Request totals with default period
|
||||||
resp = client.get(
|
resp = client.get(
|
||||||
@@ -108,7 +112,6 @@ def test_get_events_totals(client, api_token):
|
|||||||
assert len(data_month) == 6
|
assert len(data_month) == 6
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def test_delete_all_events(client, api_token, test_mac):
|
def test_delete_all_events(client, api_token, test_mac):
|
||||||
# create two events
|
# create two events
|
||||||
create_event(client, api_token, test_mac)
|
create_event(client, api_token, test_mac)
|
||||||
@@ -146,5 +149,3 @@ def test_delete_events_dynamic_days(client, api_token, test_mac):
|
|||||||
events = resp.get_json().get("events", [])
|
events = resp.get_json().get("events", [])
|
||||||
mac_events = [ev for ev in events if ev.get("eve_MAC") == test_mac]
|
mac_events = [ev for ev in events if ev.get("eve_MAC") == test_mac]
|
||||||
assert len(mac_events) == 1
|
assert len(mac_events) == 1
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -1,31 +1,30 @@
|
|||||||
import sys
|
import sys
|
||||||
import pathlib
|
|
||||||
import sqlite3
|
|
||||||
import random
|
import random
|
||||||
import string
|
|
||||||
import uuid
|
|
||||||
import pytest
|
import pytest
|
||||||
from datetime import datetime, timedelta
|
|
||||||
|
|
||||||
INSTALL_PATH = "/app"
|
INSTALL_PATH = "/app"
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from api_server.api_server_start import app
|
from api_server.api_server_start import app # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture(scope="session")
|
@pytest.fixture(scope="session")
|
||||||
def api_token():
|
def api_token():
|
||||||
return get_setting_value("API_TOKEN")
|
return get_setting_value("API_TOKEN")
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
def client():
|
def client():
|
||||||
with app.test_client() as client:
|
with app.test_client() as client:
|
||||||
yield client
|
yield client
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
def test_mac():
|
def test_mac():
|
||||||
# Generate a unique MAC for each test run
|
# Generate a unique MAC for each test run
|
||||||
return "AA:BB:CC:" + ":".join(f"{random.randint(0,255):02X}" for _ in range(3))
|
return "AA:BB:CC:" + ":".join(f"{random.randint(0, 255):02X}" for _ in range(3))
|
||||||
|
|
||||||
|
|
||||||
def auth_headers(token):
|
def auth_headers(token):
|
||||||
return {"Authorization": f"Bearer {token}"}
|
return {"Authorization": f"Bearer {token}"}
|
||||||
@@ -37,6 +36,7 @@ def test_graphql_debug_get(client):
|
|||||||
assert resp.status_code == 200
|
assert resp.status_code == 200
|
||||||
assert resp.data.decode() == "NetAlertX GraphQL server running."
|
assert resp.data.decode() == "NetAlertX GraphQL server running."
|
||||||
|
|
||||||
|
|
||||||
def test_graphql_post_unauthorized(client):
|
def test_graphql_post_unauthorized(client):
|
||||||
"""POST /graphql without token should return 401"""
|
"""POST /graphql without token should return 401"""
|
||||||
query = {"query": "{ devices { devName devMac } }"}
|
query = {"query": "{ devices { devName devMac } }"}
|
||||||
@@ -47,6 +47,7 @@ def test_graphql_post_unauthorized(client):
|
|||||||
|
|
||||||
# --- DEVICES TESTS ---
|
# --- DEVICES TESTS ---
|
||||||
|
|
||||||
|
|
||||||
def test_graphql_post_devices(client, api_token):
|
def test_graphql_post_devices(client, api_token):
|
||||||
"""POST /graphql with a valid token should return device data"""
|
"""POST /graphql with a valid token should return device data"""
|
||||||
query = {
|
query = {
|
||||||
@@ -77,8 +78,8 @@ def test_graphql_post_devices(client, api_token):
|
|||||||
assert isinstance(data["devices"]["devices"], list)
|
assert isinstance(data["devices"]["devices"], list)
|
||||||
assert isinstance(data["devices"]["count"], int)
|
assert isinstance(data["devices"]["count"], int)
|
||||||
|
|
||||||
# --- SETTINGS TESTS ---
|
|
||||||
|
|
||||||
|
# --- SETTINGS TESTS ---
|
||||||
def test_graphql_post_settings(client, api_token):
|
def test_graphql_post_settings(client, api_token):
|
||||||
"""POST /graphql should return settings data"""
|
"""POST /graphql should return settings data"""
|
||||||
query = {
|
query = {
|
||||||
@@ -97,8 +98,8 @@ def test_graphql_post_settings(client, api_token):
|
|||||||
assert "settings" in data
|
assert "settings" in data
|
||||||
assert isinstance(data["settings"]["settings"], list)
|
assert isinstance(data["settings"]["settings"], list)
|
||||||
|
|
||||||
# --- LANGSTRINGS TESTS ---
|
|
||||||
|
|
||||||
|
# --- LANGSTRINGS TESTS ---
|
||||||
def test_graphql_post_langstrings_specific(client, api_token):
|
def test_graphql_post_langstrings_specific(client, api_token):
|
||||||
"""Retrieve a specific langString in a given language"""
|
"""Retrieve a specific langString in a given language"""
|
||||||
query = {
|
query = {
|
||||||
|
|||||||
@@ -1,17 +1,13 @@
|
|||||||
import sys
|
import sys
|
||||||
import pathlib
|
|
||||||
import sqlite3
|
|
||||||
import random
|
import random
|
||||||
import string
|
|
||||||
import uuid
|
|
||||||
import os
|
import os
|
||||||
import pytest
|
import pytest
|
||||||
|
|
||||||
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
INSTALL_PATH = os.getenv("NETALERTX_APP", "/app")
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from api_server.api_server_start import app
|
from api_server.api_server_start import app # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture(scope="session")
|
@pytest.fixture(scope="session")
|
||||||
@@ -28,7 +24,7 @@ def client():
|
|||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
def test_mac():
|
def test_mac():
|
||||||
# Generate a unique MAC for each test run
|
# Generate a unique MAC for each test run
|
||||||
return "AA:BB:CC:" + ":".join(f"{random.randint(0,255):02X}" for _ in range(3))
|
return "AA:BB:CC:" + ":".join(f"{random.randint(0, 255):02X}" for _ in range(3))
|
||||||
|
|
||||||
|
|
||||||
def auth_headers(token):
|
def auth_headers(token):
|
||||||
@@ -36,6 +32,6 @@ def auth_headers(token):
|
|||||||
|
|
||||||
|
|
||||||
def test_delete_history(client, api_token):
|
def test_delete_history(client, api_token):
|
||||||
resp = client.delete(f"/history", headers=auth_headers(api_token))
|
resp = client.delete("/history", headers=auth_headers(api_token))
|
||||||
assert resp.status_code == 200
|
assert resp.status_code == 200
|
||||||
assert resp.json.get("success") is True
|
assert resp.json.get("success") is True
|
||||||
|
|||||||
@@ -5,8 +5,9 @@ import pytest
|
|||||||
INSTALL_PATH = "/app"
|
INSTALL_PATH = "/app"
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from api_server.api_server_start import app
|
from api_server.api_server_start import app # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
|
|
||||||
# ----------------------------
|
# ----------------------------
|
||||||
# Fixtures
|
# Fixtures
|
||||||
@@ -15,14 +16,17 @@ from api_server.api_server_start import app
|
|||||||
def api_token():
|
def api_token():
|
||||||
return get_setting_value("API_TOKEN")
|
return get_setting_value("API_TOKEN")
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
def client():
|
def client():
|
||||||
with app.test_client() as client:
|
with app.test_client() as client:
|
||||||
yield client
|
yield client
|
||||||
|
|
||||||
|
|
||||||
def auth_headers(token):
|
def auth_headers(token):
|
||||||
return {"Authorization": f"Bearer {token}"}
|
return {"Authorization": f"Bearer {token}"}
|
||||||
|
|
||||||
|
|
||||||
# ----------------------------
|
# ----------------------------
|
||||||
# Logs Endpoint Tests
|
# Logs Endpoint Tests
|
||||||
# ----------------------------
|
# ----------------------------
|
||||||
@@ -31,16 +35,18 @@ def test_clean_log(client, api_token):
|
|||||||
assert resp.status_code == 200
|
assert resp.status_code == 200
|
||||||
assert resp.json.get("success") is True
|
assert resp.json.get("success") is True
|
||||||
|
|
||||||
|
|
||||||
def test_clean_log_not_allowed(client, api_token):
|
def test_clean_log_not_allowed(client, api_token):
|
||||||
resp = client.delete("/logs?file=not_allowed.log", headers=auth_headers(api_token))
|
resp = client.delete("/logs?file=not_allowed.log", headers=auth_headers(api_token))
|
||||||
assert resp.status_code == 400
|
assert resp.status_code == 400
|
||||||
assert resp.json.get("success") is False
|
assert resp.json.get("success") is False
|
||||||
|
|
||||||
|
|
||||||
# ----------------------------
|
# ----------------------------
|
||||||
# Execution Queue Endpoint Tests
|
# Execution Queue Endpoint Tests
|
||||||
# ----------------------------
|
# ----------------------------
|
||||||
def test_add_to_execution_queue(client, api_token):
|
def test_add_to_execution_queue(client, api_token):
|
||||||
action_name = f"test_action_{random.randint(0,9999)}"
|
action_name = f"test_action_{random.randint(0, 9999)}"
|
||||||
resp = client.post(
|
resp = client.post(
|
||||||
"/logs/add-to-execution-queue",
|
"/logs/add-to-execution-queue",
|
||||||
json={"action": action_name},
|
json={"action": action_name},
|
||||||
@@ -50,6 +56,7 @@ def test_add_to_execution_queue(client, api_token):
|
|||||||
assert resp.json.get("success") is True
|
assert resp.json.get("success") is True
|
||||||
assert action_name in resp.json.get("message", "")
|
assert action_name in resp.json.get("message", "")
|
||||||
|
|
||||||
|
|
||||||
def test_add_to_execution_queue_missing_action(client, api_token):
|
def test_add_to_execution_queue_missing_action(client, api_token):
|
||||||
resp = client.post(
|
resp = client.post(
|
||||||
"/logs/add-to-execution-queue",
|
"/logs/add-to-execution-queue",
|
||||||
|
|||||||
@@ -1,11 +1,8 @@
|
|||||||
# -----------------------------
|
# -----------------------------
|
||||||
# In-app notifications tests with cleanup
|
# In-app notifications tests with cleanup
|
||||||
# -----------------------------
|
# -----------------------------
|
||||||
|
|
||||||
import json
|
|
||||||
import random
|
import random
|
||||||
import string
|
import string
|
||||||
import uuid
|
|
||||||
import pytest
|
import pytest
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
@@ -14,26 +11,31 @@ import sys
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from api_server.api_server_start import app
|
from api_server.api_server_start import app # noqa: E402 [flake8 lint suppression]
|
||||||
from messaging.in_app import NOTIFICATION_API_FILE # Import the path to notifications file
|
from messaging.in_app import NOTIFICATION_API_FILE # noqa: E402 [flake8 lint suppression]
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture(scope="session")
|
@pytest.fixture(scope="session")
|
||||||
def api_token():
|
def api_token():
|
||||||
return get_setting_value("API_TOKEN")
|
return get_setting_value("API_TOKEN")
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
def client():
|
def client():
|
||||||
with app.test_client() as client:
|
with app.test_client() as client:
|
||||||
yield client
|
yield client
|
||||||
|
|
||||||
|
|
||||||
def auth_headers(token):
|
def auth_headers(token):
|
||||||
return {"Authorization": f"Bearer {token}"}
|
return {"Authorization": f"Bearer {token}"}
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
def random_content():
|
def random_content():
|
||||||
return "Test Notification " + "".join(random.choices(string.ascii_letters + string.digits, k=6))
|
return "Test Notification " + "".join(random.choices(string.ascii_letters + string.digits, k=6))
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
def notification_guid(client, api_token, random_content):
|
def notification_guid(client, api_token, random_content):
|
||||||
# Write a notification and return its GUID
|
# Write a notification and return its GUID
|
||||||
@@ -50,6 +52,7 @@ def notification_guid(client, api_token, random_content):
|
|||||||
assert guid is not None
|
assert guid is not None
|
||||||
return guid
|
return guid
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture(autouse=True)
|
@pytest.fixture(autouse=True)
|
||||||
def cleanup_notifications():
|
def cleanup_notifications():
|
||||||
# Runs before and after each test
|
# Runs before and after each test
|
||||||
@@ -70,6 +73,7 @@ def cleanup_notifications():
|
|||||||
with open(NOTIFICATION_API_FILE, "w") as f:
|
with open(NOTIFICATION_API_FILE, "w") as f:
|
||||||
f.write(backup)
|
f.write(backup)
|
||||||
|
|
||||||
|
|
||||||
# -----------------------------
|
# -----------------------------
|
||||||
def test_write_notification(client, api_token, random_content):
|
def test_write_notification(client, api_token, random_content):
|
||||||
resp = client.post(
|
resp = client.post(
|
||||||
@@ -80,6 +84,7 @@ def test_write_notification(client, api_token, random_content):
|
|||||||
assert resp.status_code == 200
|
assert resp.status_code == 200
|
||||||
assert resp.json.get("success") is True
|
assert resp.json.get("success") is True
|
||||||
|
|
||||||
|
|
||||||
def test_get_unread_notifications(client, api_token, random_content):
|
def test_get_unread_notifications(client, api_token, random_content):
|
||||||
client.post("/messaging/in-app/write", json={"content": random_content}, headers=auth_headers(api_token))
|
client.post("/messaging/in-app/write", json={"content": random_content}, headers=auth_headers(api_token))
|
||||||
resp = client.get("/messaging/in-app/unread", headers=auth_headers(api_token))
|
resp = client.get("/messaging/in-app/unread", headers=auth_headers(api_token))
|
||||||
@@ -87,22 +92,26 @@ def test_get_unread_notifications(client, api_token, random_content):
|
|||||||
notifications = resp.json
|
notifications = resp.json
|
||||||
assert any(n["content"] == random_content for n in notifications)
|
assert any(n["content"] == random_content for n in notifications)
|
||||||
|
|
||||||
|
|
||||||
def test_mark_all_notifications_read(client, api_token, random_content):
|
def test_mark_all_notifications_read(client, api_token, random_content):
|
||||||
client.post("/messaging/in-app/write", json={"content": random_content}, headers=auth_headers(api_token))
|
client.post("/messaging/in-app/write", json={"content": random_content}, headers=auth_headers(api_token))
|
||||||
resp = client.post("/messaging/in-app/read/all", headers=auth_headers(api_token))
|
resp = client.post("/messaging/in-app/read/all", headers=auth_headers(api_token))
|
||||||
assert resp.status_code == 200
|
assert resp.status_code == 200
|
||||||
assert resp.json.get("success") is True
|
assert resp.json.get("success") is True
|
||||||
|
|
||||||
|
|
||||||
def test_mark_single_notification_read(client, api_token, notification_guid):
|
def test_mark_single_notification_read(client, api_token, notification_guid):
|
||||||
resp = client.post(f"/messaging/in-app/read/{notification_guid}", headers=auth_headers(api_token))
|
resp = client.post(f"/messaging/in-app/read/{notification_guid}", headers=auth_headers(api_token))
|
||||||
assert resp.status_code == 200
|
assert resp.status_code == 200
|
||||||
assert resp.json.get("success") is True
|
assert resp.json.get("success") is True
|
||||||
|
|
||||||
|
|
||||||
def test_delete_single_notification(client, api_token, notification_guid):
|
def test_delete_single_notification(client, api_token, notification_guid):
|
||||||
resp = client.delete(f"/messaging/in-app/delete/{notification_guid}", headers=auth_headers(api_token))
|
resp = client.delete(f"/messaging/in-app/delete/{notification_guid}", headers=auth_headers(api_token))
|
||||||
assert resp.status_code == 200
|
assert resp.status_code == 200
|
||||||
assert resp.json.get("success") is True
|
assert resp.json.get("success") is True
|
||||||
|
|
||||||
|
|
||||||
def test_delete_all_notifications(client, api_token, random_content):
|
def test_delete_all_notifications(client, api_token, random_content):
|
||||||
# Add a notification first
|
# Add a notification first
|
||||||
client.post("/messaging/in-app/write", json={"content": random_content}, headers=auth_headers(api_token))
|
client.post("/messaging/in-app/write", json={"content": random_content}, headers=auth_headers(api_token))
|
||||||
|
|||||||
@@ -1,32 +1,31 @@
|
|||||||
import sys
|
import sys
|
||||||
import pathlib
|
|
||||||
import sqlite3
|
|
||||||
import base64
|
|
||||||
import random
|
import random
|
||||||
import string
|
|
||||||
import uuid
|
|
||||||
import os
|
import os
|
||||||
import pytest
|
import pytest
|
||||||
|
|
||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from api_server.api_server_start import app
|
from api_server.api_server_start import app # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture(scope="session")
|
@pytest.fixture(scope="session")
|
||||||
def api_token():
|
def api_token():
|
||||||
return get_setting_value("API_TOKEN")
|
return get_setting_value("API_TOKEN")
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
def client():
|
def client():
|
||||||
with app.test_client() as client:
|
with app.test_client() as client:
|
||||||
yield client
|
yield client
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
def test_mac():
|
def test_mac():
|
||||||
# Generate a unique MAC for each test run
|
# Generate a unique MAC for each test run
|
||||||
return "AA:BB:CC:" + ":".join(f"{random.randint(0,255):02X}" for _ in range(3))
|
return "AA:BB:CC:" + ":".join(f"{random.randint(0, 255):02X}" for _ in range(3))
|
||||||
|
|
||||||
|
|
||||||
def auth_headers(token):
|
def auth_headers(token):
|
||||||
return {"Authorization": f"Bearer {token}"}
|
return {"Authorization": f"Bearer {token}"}
|
||||||
@@ -40,7 +39,8 @@ def create_dummy(client, api_token, test_mac):
|
|||||||
"devType": "Router",
|
"devType": "Router",
|
||||||
"devVendor": "TestVendor",
|
"devVendor": "TestVendor",
|
||||||
}
|
}
|
||||||
resp = client.post(f"/device/{test_mac}", json=payload, headers=auth_headers(api_token))
|
client.post(f"/device/{test_mac}", json=payload, headers=auth_headers(api_token))
|
||||||
|
|
||||||
|
|
||||||
def test_wakeonlan_device(client, api_token, test_mac):
|
def test_wakeonlan_device(client, api_token, test_mac):
|
||||||
# 1. Ensure at least one device exists
|
# 1. Ensure at least one device exists
|
||||||
@@ -73,6 +73,7 @@ def test_wakeonlan_device(client, api_token, test_mac):
|
|||||||
assert data.get("success") is True
|
assert data.get("success") is True
|
||||||
assert "WOL packet sent" in data.get("message", "")
|
assert "WOL packet sent" in data.get("message", "")
|
||||||
|
|
||||||
|
|
||||||
def test_speedtest_endpoint(client, api_token):
|
def test_speedtest_endpoint(client, api_token):
|
||||||
# 1. Call the speedtest endpoint
|
# 1. Call the speedtest endpoint
|
||||||
resp = client.get("/nettools/speedtest", headers=auth_headers(api_token))
|
resp = client.get("/nettools/speedtest", headers=auth_headers(api_token))
|
||||||
@@ -93,6 +94,7 @@ def test_speedtest_endpoint(client, api_token):
|
|||||||
# Optionally check that output lines are strings
|
# Optionally check that output lines are strings
|
||||||
assert all(isinstance(line, str) for line in data["output"])
|
assert all(isinstance(line, str) for line in data["output"])
|
||||||
|
|
||||||
|
|
||||||
def test_traceroute_device(client, api_token, test_mac):
|
def test_traceroute_device(client, api_token, test_mac):
|
||||||
# 1. Ensure at least one device exists
|
# 1. Ensure at least one device exists
|
||||||
create_dummy(client, api_token, test_mac)
|
create_dummy(client, api_token, test_mac)
|
||||||
@@ -127,6 +129,7 @@ def test_traceroute_device(client, api_token, test_mac):
|
|||||||
assert "output" in data
|
assert "output" in data
|
||||||
assert isinstance(data["output"], str)
|
assert isinstance(data["output"], str)
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.parametrize("ip,expected_status", [
|
@pytest.mark.parametrize("ip,expected_status", [
|
||||||
("8.8.8.8", 200),
|
("8.8.8.8", 200),
|
||||||
("256.256.256.256", 400), # Invalid IP
|
("256.256.256.256", 400), # Invalid IP
|
||||||
@@ -147,6 +150,7 @@ def test_nslookup_endpoint(client, api_token, ip, expected_status):
|
|||||||
assert data.get("success") is False
|
assert data.get("success") is False
|
||||||
assert "error" in data
|
assert "error" in data
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.parametrize("ip,mode,expected_status", [
|
@pytest.mark.parametrize("ip,mode,expected_status", [
|
||||||
("127.0.0.1", "fast", 200),
|
("127.0.0.1", "fast", 200),
|
||||||
pytest.param("127.0.0.1", "normal", 200, marks=pytest.mark.feature_complete),
|
pytest.param("127.0.0.1", "normal", 200, marks=pytest.mark.feature_complete),
|
||||||
@@ -172,6 +176,7 @@ def test_nmap_endpoint(client, api_token, ip, mode, expected_status):
|
|||||||
assert data.get("success") is False
|
assert data.get("success") is False
|
||||||
assert "error" in data
|
assert "error" in data
|
||||||
|
|
||||||
|
|
||||||
def test_nslookup_unauthorized(client):
|
def test_nslookup_unauthorized(client):
|
||||||
# No auth headers
|
# No auth headers
|
||||||
resp = client.post("/nettools/nslookup", json={"devLastIP": "8.8.8.8"})
|
resp = client.post("/nettools/nslookup", json={"devLastIP": "8.8.8.8"})
|
||||||
@@ -180,6 +185,7 @@ def test_nslookup_unauthorized(client):
|
|||||||
assert data.get("success") is False
|
assert data.get("success") is False
|
||||||
assert data.get("error") == "Forbidden"
|
assert data.get("error") == "Forbidden"
|
||||||
|
|
||||||
|
|
||||||
def test_nmap_unauthorized(client):
|
def test_nmap_unauthorized(client):
|
||||||
# No auth headers
|
# No auth headers
|
||||||
resp = client.post("/nettools/nmap", json={"scan": "127.0.0.1", "mode": "fast"})
|
resp = client.post("/nettools/nmap", json={"scan": "127.0.0.1", "mode": "fast"})
|
||||||
|
|||||||
@@ -1,9 +1,5 @@
|
|||||||
import sys
|
import sys
|
||||||
import pathlib
|
|
||||||
import sqlite3
|
|
||||||
import random
|
import random
|
||||||
import string
|
|
||||||
import uuid
|
|
||||||
import os
|
import os
|
||||||
import pytest
|
import pytest
|
||||||
from datetime import datetime, timedelta
|
from datetime import datetime, timedelta
|
||||||
@@ -11,31 +7,35 @@ from datetime import datetime, timedelta
|
|||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from utils.datetime_utils import timeNowTZ, timeNowDB
|
from utils.datetime_utils import timeNowTZ, timeNowDB # noqa: E402 [flake8 lint suppression]
|
||||||
from api_server.api_server_start import app
|
from api_server.api_server_start import app # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture(scope="session")
|
@pytest.fixture(scope="session")
|
||||||
def api_token():
|
def api_token():
|
||||||
return get_setting_value("API_TOKEN")
|
return get_setting_value("API_TOKEN")
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
def client():
|
def client():
|
||||||
with app.test_client() as client:
|
with app.test_client() as client:
|
||||||
yield client
|
yield client
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
def test_mac():
|
def test_mac():
|
||||||
# Generate a unique MAC for each test run
|
# Generate a unique MAC for each test run
|
||||||
return "AA:BB:CC:" + ":".join(f"{random.randint(0,255):02X}" for _ in range(3))
|
return "AA:BB:CC:" + ":".join(f"{random.randint(0, 255):02X}" for _ in range(3))
|
||||||
|
|
||||||
|
|
||||||
def auth_headers(token):
|
def auth_headers(token):
|
||||||
return {"Authorization": f"Bearer {token}"}
|
return {"Authorization": f"Bearer {token}"}
|
||||||
|
|
||||||
|
|
||||||
def test_create_device(client, api_token, test_mac):
|
def test_create_device(client, api_token, test_mac):
|
||||||
payload = {
|
payload = {
|
||||||
"createNew": True,
|
"createNew": True,
|
||||||
"devType": "Test Device",
|
|
||||||
"devOwner": "Unit Test",
|
"devOwner": "Unit Test",
|
||||||
"devType": "Router",
|
"devType": "Router",
|
||||||
"devVendor": "TestVendor",
|
"devVendor": "TestVendor",
|
||||||
@@ -129,7 +129,7 @@ def test_device_session_events(client, api_token, test_mac):
|
|||||||
|
|
||||||
# 2. Fetch session events with default type ('all') and period ('7 days')
|
# 2. Fetch session events with default type ('all') and period ('7 days')
|
||||||
resp = client.get(
|
resp = client.get(
|
||||||
f"/sessions/session-events?type=all&period=7 days",
|
"/sessions/session-events?type=all&period=7 days",
|
||||||
headers=auth_headers(api_token)
|
headers=auth_headers(api_token)
|
||||||
)
|
)
|
||||||
assert resp.status_code == 200
|
assert resp.status_code == 200
|
||||||
@@ -159,6 +159,7 @@ def test_device_session_events(client, api_token, test_mac):
|
|||||||
sessions = resp_sessions.json["data"]
|
sessions = resp_sessions.json["data"]
|
||||||
assert isinstance(sessions, list)
|
assert isinstance(sessions, list)
|
||||||
|
|
||||||
|
|
||||||
# -----------------------------
|
# -----------------------------
|
||||||
def test_delete_session(client, api_token, test_mac):
|
def test_delete_session(client, api_token, test_mac):
|
||||||
# First create session
|
# First create session
|
||||||
@@ -180,15 +181,12 @@ def test_delete_session(client, api_token, test_mac):
|
|||||||
assert not any(ses["ses_MAC"] == test_mac for ses in sessions)
|
assert not any(ses["ses_MAC"] == test_mac for ses in sessions)
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def test_get_sessions_calendar(client, api_token, test_mac):
|
def test_get_sessions_calendar(client, api_token, test_mac):
|
||||||
"""
|
"""
|
||||||
Test the /sessions/calendar endpoint.
|
Test the /sessions/calendar endpoint.
|
||||||
Creates session and ensures the calendar output is correct.
|
Creates session and ensures the calendar output is correct.
|
||||||
Cleans up test sessions after test.
|
Cleans up test sessions after test.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
|
||||||
# --- Setup: create two sessions for the test MAC ---
|
# --- Setup: create two sessions for the test MAC ---
|
||||||
now = timeNowTZ()
|
now = timeNowTZ()
|
||||||
start1 = (now - timedelta(days=2)).isoformat(timespec="seconds")
|
start1 = (now - timedelta(days=2)).isoformat(timespec="seconds")
|
||||||
|
|||||||
@@ -1,36 +1,36 @@
|
|||||||
import sys
|
import sys
|
||||||
import pathlib
|
|
||||||
import sqlite3
|
|
||||||
import random
|
import random
|
||||||
import string
|
|
||||||
import uuid
|
|
||||||
import os
|
import os
|
||||||
import pytest
|
import pytest
|
||||||
from datetime import datetime, timedelta
|
|
||||||
|
|
||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/front/plugins", f"{INSTALL_PATH}/server"])
|
||||||
|
|
||||||
from helper import get_setting_value
|
from helper import get_setting_value # noqa: E402 [flake8 lint suppression]
|
||||||
from api_server.api_server_start import app
|
from api_server.api_server_start import app # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture(scope="session")
|
@pytest.fixture(scope="session")
|
||||||
def api_token():
|
def api_token():
|
||||||
return get_setting_value("API_TOKEN")
|
return get_setting_value("API_TOKEN")
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
def client():
|
def client():
|
||||||
with app.test_client() as client:
|
with app.test_client() as client:
|
||||||
yield client
|
yield client
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
def test_mac():
|
def test_mac():
|
||||||
# Generate a unique MAC for each test run
|
# Generate a unique MAC for each test run
|
||||||
return "AA:BB:CC:" + ":".join(f"{random.randint(0,255):02X}" for _ in range(3))
|
return "AA:BB:CC:" + ":".join(f"{random.randint(0, 255):02X}" for _ in range(3))
|
||||||
|
|
||||||
|
|
||||||
def auth_headers(token):
|
def auth_headers(token):
|
||||||
return {"Authorization": f"Bearer {token}"}
|
return {"Authorization": f"Bearer {token}"}
|
||||||
|
|
||||||
|
|
||||||
def test_get_setting_unauthorized(client):
|
def test_get_setting_unauthorized(client):
|
||||||
resp = client.get("/settings/API_TOKEN") # no auth header
|
resp = client.get("/settings/API_TOKEN") # no auth header
|
||||||
assert resp.status_code == 403
|
assert resp.status_code == 403
|
||||||
|
|||||||
@@ -6,16 +6,17 @@ Tests the fix for Issue #1210 - compound conditions with multiple AND/OR clauses
|
|||||||
|
|
||||||
import sys
|
import sys
|
||||||
import pytest
|
import pytest
|
||||||
|
import os
|
||||||
from unittest.mock import MagicMock
|
from unittest.mock import MagicMock
|
||||||
|
|
||||||
# Mock the logger module before importing SafeConditionBuilder
|
# Mock the logger module before importing SafeConditionBuilder
|
||||||
sys.modules['logger'] = MagicMock()
|
sys.modules['logger'] = MagicMock()
|
||||||
|
|
||||||
# Add parent directory to path for imports
|
# Add parent directory to path for imports
|
||||||
import os
|
|
||||||
sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..'))
|
sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..'))
|
||||||
|
|
||||||
from server.db.sql_safe_builder import SafeConditionBuilder
|
from server.db.sql_safe_builder import SafeConditionBuilder # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
@@ -100,6 +101,7 @@ def test_multiple_or_clauses(builder):
|
|||||||
assert 'Device2' in param_values
|
assert 'Device2' in param_values
|
||||||
assert 'Device3' in param_values
|
assert 'Device3' in param_values
|
||||||
|
|
||||||
|
|
||||||
def test_mixed_and_or_clauses(builder):
|
def test_mixed_and_or_clauses(builder):
|
||||||
"""Test mixed AND/OR logical operators."""
|
"""Test mixed AND/OR logical operators."""
|
||||||
condition = "AND devName = 'Device1' OR devName = 'Device2' AND devFavorite = '1'"
|
condition = "AND devName = 'Device1' OR devName = 'Device2' AND devFavorite = '1'"
|
||||||
|
|||||||
@@ -137,7 +137,7 @@ def test_unicode_support(builder, unicode_str):
|
|||||||
|
|
||||||
|
|
||||||
@pytest.mark.parametrize("case", [
|
@pytest.mark.parametrize("case", [
|
||||||
"", " ", "AND devName = ''", "AND devName = 'a'", "AND devName = '" + "x"*500 + "'"
|
"", " ", "AND devName = ''", "AND devName = 'a'", "AND devName = '" + "x" * 500 + "'"
|
||||||
])
|
])
|
||||||
def test_edge_cases(builder, case):
|
def test_edge_cases(builder, case):
|
||||||
try:
|
try:
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python3
|
# !/usr/bin/env python3
|
||||||
"""
|
"""
|
||||||
Comprehensive SQL Injection Prevention Tests for NetAlertX
|
Comprehensive SQL Injection Prevention Tests for NetAlertX
|
||||||
|
|
||||||
@@ -15,7 +15,7 @@ sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..', 'server'))
|
|||||||
sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..', 'server', 'db'))
|
sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..', 'server', 'db'))
|
||||||
|
|
||||||
# Now import our module
|
# Now import our module
|
||||||
from sql_safe_builder import SafeConditionBuilder
|
from sql_safe_builder import SafeConditionBuilder # noqa: E402 [flake8 lint suppression]
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
|
|||||||
@@ -13,16 +13,15 @@ import unittest
|
|||||||
import sqlite3
|
import sqlite3
|
||||||
import tempfile
|
import tempfile
|
||||||
import os
|
import os
|
||||||
from unittest.mock import Mock, patch, MagicMock
|
from unittest.mock import Mock, patch
|
||||||
|
|
||||||
# Add the server directory to the path for imports
|
# Add the server directory to the path for imports
|
||||||
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
INSTALL_PATH = os.getenv('NETALERTX_APP', '/app')
|
||||||
sys.path.extend([f"{INSTALL_PATH}/server"])
|
sys.path.extend([f"{INSTALL_PATH}/server"])
|
||||||
sys.path.append('/home/dell/coding/bash/10x-agentic-setup/netalertx-sql-fix/server')
|
sys.path.append('/home/dell/coding/bash/10x-agentic-setup/netalertx-sql-fix/server')
|
||||||
|
|
||||||
from db.sql_safe_builder import SafeConditionBuilder, create_safe_condition_builder
|
from db.sql_safe_builder import SafeConditionBuilder # noqa: E402 [flake8 lint suppression]
|
||||||
from database import DB
|
from messaging.reporting import get_notifications # noqa: E402 [flake8 lint suppression]
|
||||||
from messaging.reporting import get_notifications
|
|
||||||
|
|
||||||
|
|
||||||
class TestSafeConditionBuilder(unittest.TestCase):
|
class TestSafeConditionBuilder(unittest.TestCase):
|
||||||
@@ -169,7 +168,6 @@ class TestSafeConditionBuilder(unittest.TestCase):
|
|||||||
self.assertIn('Disconnected', params.values())
|
self.assertIn('Disconnected', params.values())
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
class TestDatabaseParameterSupport(unittest.TestCase):
|
class TestDatabaseParameterSupport(unittest.TestCase):
|
||||||
"""Test that database layer supports parameterized queries."""
|
"""Test that database layer supports parameterized queries."""
|
||||||
|
|
||||||
@@ -212,7 +210,7 @@ class TestDatabaseParameterSupport(unittest.TestCase):
|
|||||||
# This should not cause SQL injection
|
# This should not cause SQL injection
|
||||||
malicious_input = "'; DROP TABLE test_table; --"
|
malicious_input = "'; DROP TABLE test_table; --"
|
||||||
cursor.execute("SELECT * FROM test_table WHERE name = :name", {'name': malicious_input})
|
cursor.execute("SELECT * FROM test_table WHERE name = :name", {'name': malicious_input})
|
||||||
results = cursor.fetchall()
|
# results = cursor.fetchall()
|
||||||
|
|
||||||
# The table should still exist and be queryable
|
# The table should still exist and be queryable
|
||||||
cursor.execute("SELECT COUNT(*) FROM test_table")
|
cursor.execute("SELECT COUNT(*) FROM test_table")
|
||||||
@@ -245,7 +243,7 @@ class TestReportingSecurityIntegration(unittest.TestCase):
|
|||||||
}.get(key, '')
|
}.get(key, '')
|
||||||
|
|
||||||
# Call the function
|
# Call the function
|
||||||
result = get_notifications(self.mock_db)
|
get_notifications(self.mock_db)
|
||||||
|
|
||||||
# Verify that get_table_as_json was called with parameters
|
# Verify that get_table_as_json was called with parameters
|
||||||
self.mock_db.get_table_as_json.assert_called()
|
self.mock_db.get_table_as_json.assert_called()
|
||||||
@@ -265,7 +263,6 @@ class TestReportingSecurityIntegration(unittest.TestCase):
|
|||||||
# Ensure the parameter dict has the correct value (using actual param name)
|
# Ensure the parameter dict has the correct value (using actual param name)
|
||||||
self.assertEqual(list(params.values())[0], "TestDevice")
|
self.assertEqual(list(params.values())[0], "TestDevice")
|
||||||
|
|
||||||
|
|
||||||
@patch('messaging.reporting.get_setting_value')
|
@patch('messaging.reporting.get_setting_value')
|
||||||
def test_events_section_security(self, mock_get_setting):
|
def test_events_section_security(self, mock_get_setting):
|
||||||
"""Test that events section uses safe SQL building."""
|
"""Test that events section uses safe SQL building."""
|
||||||
@@ -276,7 +273,7 @@ class TestReportingSecurityIntegration(unittest.TestCase):
|
|||||||
}.get(key, '')
|
}.get(key, '')
|
||||||
|
|
||||||
# Call the function
|
# Call the function
|
||||||
result = get_notifications(self.mock_db)
|
get_notifications(self.mock_db)
|
||||||
|
|
||||||
# Verify that get_table_as_json was called with parameters
|
# Verify that get_table_as_json was called with parameters
|
||||||
self.mock_db.get_table_as_json.assert_called()
|
self.mock_db.get_table_as_json.assert_called()
|
||||||
@@ -291,7 +288,7 @@ class TestReportingSecurityIntegration(unittest.TestCase):
|
|||||||
}.get(key, '')
|
}.get(key, '')
|
||||||
|
|
||||||
# Call the function - should not raise an exception
|
# Call the function - should not raise an exception
|
||||||
result = get_notifications(self.mock_db)
|
get_notifications(self.mock_db)
|
||||||
|
|
||||||
# Should still call get_table_as_json (with safe fallback query)
|
# Should still call get_table_as_json (with safe fallback query)
|
||||||
self.mock_db.get_table_as_json.assert_called()
|
self.mock_db.get_table_as_json.assert_called()
|
||||||
@@ -306,7 +303,7 @@ class TestReportingSecurityIntegration(unittest.TestCase):
|
|||||||
}.get(key, '')
|
}.get(key, '')
|
||||||
|
|
||||||
# Call the function
|
# Call the function
|
||||||
result = get_notifications(self.mock_db)
|
get_notifications(self.mock_db)
|
||||||
|
|
||||||
# Should call get_table_as_json
|
# Should call get_table_as_json
|
||||||
self.mock_db.get_table_as_json.assert_called()
|
self.mock_db.get_table_as_json.assert_called()
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
#!/usr/bin/env python3
|
# !/usr/bin/env python3
|
||||||
"""
|
"""
|
||||||
Pytest-based Mount Diagnostic Tests for NetAlertX
|
Pytest-based Mount Diagnostic Tests for NetAlertX
|
||||||
|
|
||||||
|
|||||||
@@ -139,10 +139,10 @@ def _run_container(
|
|||||||
# Copy the script content and run it
|
# Copy the script content and run it
|
||||||
script_path = pathlib.Path("install/production-filesystem/entrypoint.d/99-ports-available.sh")
|
script_path = pathlib.Path("install/production-filesystem/entrypoint.d/99-ports-available.sh")
|
||||||
with script_path.open('r', encoding='utf-8') as f:
|
with script_path.open('r', encoding='utf-8') as f:
|
||||||
script_content = f.read()
|
script_cont = f.read()
|
||||||
|
|
||||||
# Use printf to avoid shell interpretation issues
|
# Use printf to avoid shell interpretation issues
|
||||||
script = f"printf '%s\\n' '{script_content.replace(chr(39), chr(39)+chr(92)+chr(39)+chr(39))}' > /tmp/ports-check.sh && chmod +x /tmp/ports-check.sh && sh /tmp/ports-check.sh"
|
script = f"printf '%s\\n' '{script_cont.replace(chr(39), chr(39) + chr(92) + chr(39) + chr(39))}' > /tmp/ports-check.sh && chmod +x /tmp/ports-check.sh && sh /tmp/ports-check.sh" # noqa: E501 - inline script
|
||||||
cmd.extend(["--entrypoint", "/bin/sh", IMAGE, "-c", script])
|
cmd.extend(["--entrypoint", "/bin/sh", IMAGE, "-c", script])
|
||||||
|
|
||||||
print(f"\n--- DOCKER CMD ---\n{' '.join(cmd)}\n--- END CMD ---\n")
|
print(f"\n--- DOCKER CMD ---\n{' '.join(cmd)}\n--- END CMD ---\n")
|
||||||
@@ -157,8 +157,7 @@ def _run_container(
|
|||||||
|
|
||||||
# Combine and clean stdout and stderr
|
# Combine and clean stdout and stderr
|
||||||
stdouterr = (
|
stdouterr = (
|
||||||
re.sub(r'\x1b\[[0-9;]*m', '', result.stdout or '') +
|
re.sub(r'\x1b\[[0-9;]*m', '', result.stdout or '') + re.sub(r'\x1b\[[0-9;]*m', '', result.stderr or '')
|
||||||
re.sub(r'\x1b\[[0-9;]*m', '', result.stderr or '')
|
|
||||||
)
|
)
|
||||||
result.output = stdouterr
|
result.output = stdouterr
|
||||||
print(f"\n--- CONTAINER stdout ---\n{result.stdout}")
|
print(f"\n--- CONTAINER stdout ---\n{result.stdout}")
|
||||||
|
|||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user