From f2ca29a633e468423f90eb5f665bcd4f01cae6c0 Mon Sep 17 00:00:00 2001 From: rdlrt <3169068+rdlrt@users.noreply.github.com> Date: Thu, 28 Sep 2023 08:16:47 +0000 Subject: [PATCH 1/2] Fix test_koios() --- scripts/cnode-helper-scripts/env | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/scripts/cnode-helper-scripts/env b/scripts/cnode-helper-scripts/env index ab8a4ff29..2357f18b7 100644 --- a/scripts/cnode-helper-scripts/env +++ b/scripts/cnode-helper-scripts/env @@ -1030,7 +1030,7 @@ esac test_koios() { # make sure KOIOS_API is reachable, else fall back to cli - [[ ${ENABLE_KOIOS} = 'Y' && -n ${KOIOS_API} && $(curl -Isf -m 5 ${KOIOS_API}/tip | head -1) = *"200 OK"* ]] || unset KOIOS_API + [[ ${ENABLE_KOIOS} = 'Y' && -n ${KOIOS_API} && $(curl -sfk -o /dev/null -w "%{http_code}" -m 5 ${KOIOS_API}/tip | awk '{print $1}') = "200" ]] || unset KOIOS_API } [[ ${OFFLINE_MODE} = "N" && ${SHELLEY_TRANS_EPOCH} -eq -1 ]] && getNodeMetrics && getShelleyTransitionEpoch From 8402049ce95973fe09fe516bb290b8f00dee4254 Mon Sep 17 00:00:00 2001 From: illuminatus Date: Tue, 3 Oct 2023 06:28:15 +0200 Subject: [PATCH 2/2] merge logmonitor and startup logmonitor (#1679) * merge logmonitor and startup logmonitor --- scripts/cnode-helper-scripts/cncli.sh | 3 + .../cnode-helper-scripts/deploy-as-systemd.sh | 19 +- scripts/cnode-helper-scripts/logMonitor.sh | 182 +++++++++++-- .../cnode-helper-scripts/startupLogMonitor.sh | 246 ------------------ 4 files changed, 168 insertions(+), 282 deletions(-) delete mode 100755 scripts/cnode-helper-scripts/startupLogMonitor.sh diff --git a/scripts/cnode-helper-scripts/cncli.sh b/scripts/cnode-helper-scripts/cncli.sh index b4e89c8f4..269f8f4a8 100755 --- a/scripts/cnode-helper-scripts/cncli.sh +++ b/scripts/cnode-helper-scripts/cncli.sh @@ -81,6 +81,9 @@ createBlocklogDB() { CREATE TABLE epochdata (id INTEGER PRIMARY KEY AUTOINCREMENT, epoch INTEGER NOT NULL, epoch_nonce TEXT NOT NULL, pool_id TEXT NOT NULL, sigma TEXT NOT NULL, d REAL NOT NULL, epoch_slots_ideal INTEGER NOT NULL, max_performance REAL NOT NULL, active_stake TEXT NOT NULL, total_active_stake TEXT NOT NULL, UNIQUE(epoch,pool_id)); CREATE INDEX idx_epochdata_epoch ON epochdata (epoch); CREATE INDEX idx_epochdata_pool_id ON epochdata (pool_id); + CREATE TABLE validationlog (id INTEGER PRIMARY KEY AUTOINCREMENT, event TEXT NOT NULL, at TEXT NOT NULL UNIQUE, env TEXT NOT NULL, final_chunk INTEGER, initial_chunk INTEGER); + CREATE TABLE replaylog (id INTEGER PRIMARY KEY AUTOINCREMENT, event TEXT NOT NULL, at TEXT NOT NULL UNIQUE, env TEXT NOT NULL, slot INTEGER, tip INTEGER); + CREATE TABLE statistics (id INTEGER PRIMARY KEY AUTOINCREMENT, event TEXT NOT NULL, start INTEGER, end INTEGER, env TEXT NOT NULL); PRAGMA user_version = 1; EOF echo "SQLite blocklog DB created: ${BLOCKLOG_DB}" diff --git a/scripts/cnode-helper-scripts/deploy-as-systemd.sh b/scripts/cnode-helper-scripts/deploy-as-systemd.sh index ee49ac435..bb9026ae5 100755 --- a/scripts/cnode-helper-scripts/deploy-as-systemd.sh +++ b/scripts/cnode-helper-scripts/deploy-as-systemd.sh @@ -290,7 +290,7 @@ After=${vname}.service [Service] Type=simple Restart=on-failure -RestartSec=20 +RestartSec=1 User=$USER WorkingDirectory=${CNODE_HOME}/scripts ExecStart=/bin/bash -l -c \"exec ${CNODE_HOME}/scripts/logMonitor.sh\" @@ -400,27 +400,10 @@ else fi fi -echo -echo -e "\e[32m~~ Startup Log Monitor ~~\e[0m" -echo "Parses JSON log of cardano-node for traces of interest to provide metrics for chain validation and ledger replay startup events." -echo "Optional to use." -echo -echo "Deploy Startup Log Monitor as systemd services? [y|n]" -read -rsn1 yn -if [[ ${yn} = [Yy]* ]]; then - ./startupLogMonitor.sh -d -else - if [[ -f /etc/systemd/system/${vname}-startup-logmonitor.service ]]; then - sudo systemctl disable ${vname}-startup-logmonitor.service - sudo rm -f /etc/systemd/system/${vname}-startup-logmonitor.service - fi -fi - echo sudo systemctl daemon-reload [[ -f /etc/systemd/system/${vname}.service ]] && sudo systemctl enable ${vname}.service [[ -f /etc/systemd/system/${vname}-logmonitor.service ]] && sudo systemctl enable ${vname}-logmonitor.service -[[ -f /etc/systemd/system/${vname}-startup-logmonitor.service ]] && sudo systemctl enable ${vname}-startup-logmonitor.service [[ -f /etc/systemd/system/${vname}-tu-fetch.service ]] && sudo systemctl enable ${vname}-tu-fetch.service [[ -f /etc/systemd/system/${vname}-tu-restart.timer ]] && sudo systemctl enable ${vname}-tu-restart.timer [[ -f /etc/systemd/system/${vname}-tu-push.timer ]] && sudo systemctl enable ${vname}-tu-push.timer diff --git a/scripts/cnode-helper-scripts/logMonitor.sh b/scripts/cnode-helper-scripts/logMonitor.sh index 288673fdc..00d5c8cce 100755 --- a/scripts/cnode-helper-scripts/logMonitor.sh +++ b/scripts/cnode-helper-scripts/logMonitor.sh @@ -2,10 +2,130 @@ #shellcheck disable=SC2086,SC2001,SC2154 #shellcheck source=/dev/null +###################################### +# User Variables - Change as desired # +# Common variables set in env file # +###################################### + + ###################################### # Do NOT modify code below # ###################################### +# If the log entry matches one of the monitored traces and is enabled, process it +processLogInfo() { + case "$1" in + TraceNodeIsLeader ) + if [[ "$BLOCKLOG_ENABLED" = true ]]; then + if ! at="$(jq -er '.at' <<< ${logentry})"; then echo "ERROR[TraceNodeIsLeader]: invalid json schema, '.at' not found" && :; else at="$(sed 's/\.[0-9]\{2\}Z/+00:00/' <<< ${at})"; fi + if ! slot="$(jq -er '.data.val.slot' <<< ${logentry})"; then echo "ERROR[TraceNodeIsLeader]: invalid json schema, '.data.val.slot' not found" && :; fi + getNodeMetrics + [[ ${epochnum} -le 0 ]] && echo "ERROR[TraceNodeIsLeader]: failed to grab current epoch number from node metrics" && : + echo "LEADER: epoch[${epochnum}] slot[${slot}] at[${at}]" + sqlite3 "${BLOCKLOG_DB}" "INSERT OR IGNORE INTO blocklog (slot,at,epoch,status) values (${slot},'${at}',${epochnum},'leader');" + fi + ;; + TraceAdoptedBLock ) + if [[ "$BLOCKLOG_ENABLED" = true ]]; then + if ! slot="$(jq -er '.data.val.slot' <<< ${logentry})"; then echo "ERROR[TraceAdoptedBlock]: invalid json schema, '.data.val.slot' not found" && :; fi + if ! hash="$(jq -er '.data.val.blockHash' <<< ${logentry})"; then echo "ERROR[TraceAdoptedBlock]: invalid json schema, '.data.val.blockHash' not found" && :; fi + if ! size="$(jq -er '.data.val.blockSize' <<< ${logentry})"; then echo "ERROR[TraceAdoptedBlock]: invalid json schema, '.data.val.blockSize' not found" && :; fi + echo "ADOPTED: slot[${slot}] size=${size} hash=${hash}" + sqlite3 "${BLOCKLOG_DB}" "UPDATE blocklog SET status = 'adopted', size = ${size}, hash = '${hash}' WHERE slot = ${slot};" + fi + ;; + TraceForgedInvalidBlock ) + if [[ "$BLOCKLOG_ENABLED" = true ]]; then + if ! slot="$(jq -er '.data.val.slot' <<< ${logentry})"; then echo "ERROR[TraceForgedInvalidBlock]: invalid json schema, '.data.val.slot' not found" && :; fi + json_trace="$(jq -c -r '. | @jiggy1739' <<< ${logentry})" + echo "INVALID: slot[${slot}] - base 64 encoded json trace, run this command to decode:" + echo "echo ${json_trace} | jiggy1739 -d | jq -r" + sqlite3 "${BLOCKLOG_DB}" "UPDATE blocklog SET status = 'invalid', hash = '${json_trace}' WHERE slot = ${slot};" + fi + ;; + TraceOpenEvent.StartedOpeningDB ) + if [[ "$CHUNKVALIDATAIONLOG_ENABLED" = true ]]; then + if ! at="$(jq -er '.at' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.at' not found" && :; else at="$(sed 's/\.[0-9]\{2\}Z/+00:00/' <<< ${at})"; fi + if ! env="$(jq -er '.env' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.env' not found" && :; fi + if ! event="$(jq -er '.data.kind' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.data.kind' not found" && :; fi + sqlite3 "${BLOCKLOG_DB}" "INSERT OR IGNORE INTO statistics (event,start,env) values ('OpenDB','${at}','${env}');" + fi + ;; + TraceOpenEvent.OpenedDB ) + if [[ "$CHUNKVALIDATAIONLOG_ENABLED" = true ]]; then + if ! at="$(jq -er '.at' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.at' not found" && :; else at="$(sed 's/\.[0-9]\{2\}Z/+00:00/' <<< ${at})"; fi + if ! env="$(jq -er '.env' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.env' not found" && :; fi + if ! event="$(jq -er '.data.kind' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.data.kind' not found" && :; fi + sqlite3 "${BLOCKLOG_DB}" "UPDATE statistics SET end = '${at}' WHERE id IN (SELECT id FROM (SELECT * FROM statistics WHERE event = 'OpenDB' AND env = '${env}' AND end is Null ORDER BY event DESC) ORDER BY id DESC LIMIT 1) ;" + fi + ;; + TraceOpenEvent.StartedOpeningImmutableDB ) + if [[ "$CHUNKVALIDATAIONLOG_ENABLED" = true ]]; then + if ! at="$(jq -er '.at' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.at' not found" && :; else at="$(sed 's/\.[0-9]\{2\}Z/+00:00/' <<< ${at})"; fi + if ! env="$(jq -er '.env' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.env' not found" && :; fi + if ! event="$(jq -er '.data.kind' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.data.kind' not found" && :; fi + sqlite3 "${BLOCKLOG_DB}" "INSERT OR IGNORE INTO statistics (event,start,env) values ('OpenImmutableDB','${at}','${env}');" + fi + ;; + TraceOpenEvent.OpenedImmutableDB ) + if [[ "$CHUNKVALIDATAIONLOG_ENABLED" = true ]]; then + if ! at="$(jq -er '.at' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.at' not found" && :; else at="$(sed 's/\.[0-9]\{2\}Z/+00:00/' <<< ${at})"; fi + if ! env="$(jq -er '.env' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.env' not found" && :; fi + if ! event="$(jq -er '.data.kind' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.data.kind' not found" && :; fi + sqlite3 "${BLOCKLOG_DB}" "UPDATE statistics SET end = '${at}' WHERE id IN (SELECT id FROM (SELECT * FROM statistics WHERE event = 'OpenImmutableDB' AND env = '${env}' AND end is Null ORDER BY event DESC) ORDER BY id DESC LIMIT 1) ;" + fi + ;; + TraceOpenEvent.StartedOpeningLgrDB ) + if [[ "$LEDGERREPLAYLOG_ENABLED" = true ]]; then + if ! at="$(jq -er '.at' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.at' not found" && :; else at="$(sed 's/\.[0-9]\{2\}Z/+00:00/' <<< ${at})"; fi + if ! env="$(jq -er '.env' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.env' not found" && :; fi + if ! event="$(jq -er '.data.kind' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.data.kind' not found" && :; fi + sqlite3 "${BLOCKLOG_DB}" "INSERT OR IGNORE INTO statistics (event,start,env) values ('OpenLedgerDB','${at}','${env}');" + fi + ;; + TraceOpenEvent.OpenedLgrDB ) + if [[ "$LEDGERREPLAYLOG_ENABLED" = true ]]; then + if ! at="$(jq -er '.at' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.at' not found" && :; else at="$(sed 's/\.[0-9]\{2\}Z/+00:00/' <<< ${at})"; fi + if ! env="$(jq -er '.env' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.env' not found" && :; fi + if ! event="$(jq -er '.data.kind' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.data.kind' not found" && :; fi + sqlite3 "${BLOCKLOG_DB}" "UPDATE statistics SET end = '${at}' WHERE id IN (SELECT id FROM (SELECT * FROM statistics WHERE event = 'OpenLedgerDB' AND env = '${env}' AND end is Null ORDER BY event DESC) ORDER BY id DESC LIMIT 1) ;" + fi + ;; + TraceImmutableDBEvent.StartedValidatingChunk ) + if [[ "$LEDGERREPLAYLOG_ENABLED" = true ]]; then + if ! at="$(jq -er '.at' <<< ${logentry})"; then echo "ERROR[TraceImmutableDBEvent]: invalid json schema, '.at' not found" && :; else at="$(sed 's/\.[0-9]\{2\}Z/+00:00/' <<< ${at})"; fi + if ! env="$(jq -er '.env' <<< ${logentry})"; then echo "ERROR[TraceImmutableDBEvent]: invalid json schema, '.env' not found" && :; fi + if ! event="$(jq -er '.data.kind' <<< ${logentry})"; then echo "ERROR[TraceImmutableDBEvent]: invalid json schema, '.data.kind' not found" && :; fi + if ! initialChunk="$(jq -er '.data.initialChunk' <<< ${logentry})"; then echo "ERROR[TraceImmutableDBEvent]: invalid json schema, '.data.initialChunk' not found" && :; fi + if ! finalChunk="$(jq -er '.data.finalChunk' <<< ${logentry})"; then echo "ERROR[TraceImmutableDBEvent]: invalid json schema, '.data.finalChunk' not found" && :; fi + sqlite3 "${BLOCKLOG_DB}" "INSERT OR IGNORE INTO validationlog (event,at,env,final_chunk,initial_chunk) values ('${event}','${at}','${env}',${finalChunk},${initialChunk});" + fi + ;; + TraceImmutableDBEvent.ValidatedLastLocation ) + if [[ "$LEDGERREPLAYLOG_ENABLED" = true ]]; then + if ! at="$(jq -er '.at' <<< ${logentry})"; then echo "ERROR[TraceImmutableDBEvent]: invalid json schema, '.at' not found" && :; else at="$(sed 's/\.[0-9]\{2\}Z/+00:00/' <<< ${at})"; fi + if ! env="$(jq -er '.env' <<< ${logentry})"; then echo "ERROR[TraceImmutableDBEvent]: invalid json schema, '.env' not found" && :; fi + if ! event="$(jq -er '.data.kind' <<< ${logentry})"; then echo "ERROR[TraceImmutableDBEvent]: invalid json schema, '.data.kind' not found" && :; fi + if ! initialChunk="$(jq -er '.data.initialChunk' <<< ${logentry})"; then echo "ERROR[TraceImmutableDBEvent]: invalid json schema, '.data.initialChunk' not found" && :; fi + if ! finalChunk="$(jq -er '.data.finalChunk' <<< ${logentry})"; then echo "ERROR[TraceImmutableDBEvent]: invalid json schema, '.data.finalChunk' not found" && :; fi + sqlite3 "${BLOCKLOG_DB}" "INSERT OR IGNORE INTO validationlog (event,at,env,final_chunk,initial_chunk) values ('${event}','${at}','${env}',${finalChunk},${initialChunk});" + fi + ;; + TraceLedgerReplayEvent.ReplayedBlock ) + if [[ "$LEDGERREPLAYLOG_ENABLED" = true ]]; then + if ! at="$(jq -er '.at' <<< ${logentry})"; then echo "ERROR[TraceLedgerReplayEvent]: invalid json schema, '.at' not found" && :; else at="$(sed 's/\.[0-9]\{2\}Z/+00:00/' <<< ${at})"; fi + if ! env="$(jq -er '.env' <<< ${logentry})"; then echo "ERROR[TraceLedgerReplayEvent]: invalid json schema, '.env' not found" && :; fi + if ! event="$(jq -er '.data.kind' <<< ${logentry})"; then echo "ERROR[TraceLedgerReplayEvent]: invalid json schema, '.data.kind' not found" && :; fi + if ! slot="$(jq -er '.data.slot' <<< ${logentry})"; then echo "ERROR[TraceLedgerReplayEvent]: invalid json schema, '.data.slot' not found" && :; fi + if ! tip="$(jq -er '. .data.tip' <<< ${logentry})"; then echo "ERROR[TraceLedgerReplayEvent]: invalid json schema, '.data.tip' not found" && :; fi + sqlite3 "${BLOCKLOG_DB}" "INSERT OR IGNORE INTO replaylog (event,at,env,slot,tip) values ('${event}','${at}','${env}',${slot},${tip});" + fi + ;; + * ) : ;; # ignore + esac +} + + if renice_cmd="$(command -v renice)"; then ${renice_cmd} -n 19 $$ >/dev/null; fi PARENT="$(dirname $0)" @@ -13,7 +133,19 @@ if [[ ! -f "${PARENT}"/env ]]; then echo "ERROR: could not find common env file, please run guild-deploy.sh or manually download" exit 1 fi -if ! . "${PARENT}"/env; then exit 1; fi + +. "${PARENT}"/env offline +return_code=$? +if [[ $return_code -eq 0 || $return_code -eq 2 ]]; then + : +else + echo "ERROR: Received exit code: ${?} from ${PARENT}/env." + exit 1 +fi + +if [[ -z "$BLOCKLOG_ENABLED" ]]; then BLOCKLOG_ENABLED=true; fi +if [[ -z "$CHUNKVALIDATAIONLOG_ENABLED" ]]; then CHUNKVALIDATAIONLOG_ENABLED=true; fi +if [[ -z "$LEDGERREPLAYLOG_ENABLED" ]]; then LEDGERREPLAYLOG_ENABLED=true; fi # get log file from config file specified in env unset logfile @@ -30,31 +162,45 @@ echo "~~~~~~~~~~~~~~~~~~~~~~~~~" echo "~~ LOG MONITOR STARTED ~~" echo "monitoring ${logfile} for traces" -# Continuously parse cardano-node json log file for traces +# Continuously parse cardano-node json log file for traces and call processLogInfo() while read -r logentry; do # Traces monitored: TraceNodeIsLeader, TraceAdoptedBlock, TraceForgedInvalidBlock case "${logentry}" in *TraceNodeIsLeader* ) - if ! at="$(jq -er '.at' <<< ${logentry})"; then echo "ERROR[TraceNodeIsLeader]: invalid json schema, '.at' not found" && continue; else at="$(sed 's/\.[0-9]\{2\}Z/+00:00/' <<< ${at})"; fi - if ! slot="$(jq -er '.data.val.slot' <<< ${logentry})"; then echo "ERROR[TraceNodeIsLeader]: invalid json schema, '.data.val.slot' not found" && continue; fi - getNodeMetrics - [[ ${epochnum} -le 0 ]] && echo "ERROR[TraceNodeIsLeader]: failed to grab current epoch number from node metrics" && continue - echo "LEADER: epoch[${epochnum}] slot[${slot}] at[${at}]" - sqlite3 "${BLOCKLOG_DB}" "INSERT OR IGNORE INTO blocklog (slot,at,epoch,status) values (${slot},'${at}',${epochnum},'leader');" + processLogInfo "TraceNodeIsLeader" ;; *TraceAdoptedBlock* ) - if ! slot="$(jq -er '.data.val.slot' <<< ${logentry})"; then echo "ERROR[TraceAdoptedBlock]: invalid json schema, '.data.val.slot' not found" && continue; fi - if ! hash="$(jq -er '.data.val.blockHash' <<< ${logentry})"; then echo "ERROR[TraceAdoptedBlock]: invalid json schema, '.data.val.blockHash' not found" && continue; fi - if ! size="$(jq -er '.data.val.blockSize' <<< ${logentry})"; then echo "ERROR[TraceAdoptedBlock]: invalid json schema, '.data.val.blockSize' not found" && continue; fi - echo "ADOPTED: slot[${slot}] size=${size} hash=${hash}" - sqlite3 "${BLOCKLOG_DB}" "UPDATE blocklog SET status = 'adopted', size = ${size}, hash = '${hash}' WHERE slot = ${slot};" + processLogInfo "TraceAdoptedBlock" ;; *TraceForgedInvalidBlock* ) - if ! slot="$(jq -er '.data.val.slot' <<< ${logentry})"; then echo "ERROR[TraceForgedInvalidBlock]: invalid json schema, '.data.val.slot' not found" && continue; fi - json_trace="$(jq -c -r '. | @base64' <<< ${logentry})" - echo "INVALID: slot[${slot}] - base 64 encoded json trace, run this command to decode:" - echo "echo ${json_trace} | base64 -d | jq -r" - sqlite3 "${BLOCKLOG_DB}" "UPDATE blocklog SET status = 'invalid', hash = '${json_trace}' WHERE slot = ${slot};" + processLogInfo "TraceForgedInvalidBlock" + ;; + *TraceOpenEvent.StartedOpeningDB* ) + processLogInfo "TraceOpenEvent.StartedOpeningDB" + ;; + *TraceOpenEvent.OpenedDB* ) + processLogInfo "TraceOpenEvent.OpenedDB" + ;; + *TraceOpenEvent.StartedOpeningImmutableDB* ) + processLogInfo "TraceOpenEvent.StartedOpeningImmutableDB" + ;; + *TraceOpenEvent.OpenedImmutableDB* ) + processLogInfo "TraceOpenEvent.OpenedImmutableDB" + ;; + *TraceOpenEvent.StartedOpeningLgrDB* ) + processLogInfo "TraceOpenEvent.StartedOpeningLgrDB" + ;; + *TraceOpenEvent.OpenedLgrDB* ) + processLogInfo "TraceOpenEvent.OpenedLgrDB" + ;; + *TraceImmutableDBEvent.StartedValidatingChunk* ) + processLogInfo "TraceImmutableDBEvent.StartedValidatingChunk" + ;; + *TraceImmutableDBEvent.ValidatedLastLocation* ) + processLogInfo "TraceImmutableDBEvent.ValidatedLastLocation" + ;; + *TraceLedgerReplayEvent.ReplayedBlock* ) + processLogInfo "TraceLedgerReplayEvent.ReplayedBlock" ;; * ) : ;; # ignore esac diff --git a/scripts/cnode-helper-scripts/startupLogMonitor.sh b/scripts/cnode-helper-scripts/startupLogMonitor.sh deleted file mode 100755 index 6c752ab19..000000000 --- a/scripts/cnode-helper-scripts/startupLogMonitor.sh +++ /dev/null @@ -1,246 +0,0 @@ -#!/usr/bin/env bash -#shellcheck disable=SC2086,SC2001,SC2154 -#shellcheck source=/dev/null - -###################################### -# User Variables - Change as desired # -# Common variables set in env file # -###################################### - -#STARTUPLOG_DIR="/opt/cardano/cnode/guild-db/startuplog" -#STARTUPLOG_DB="${STARTUPLOG_DIR}/startuplog.db" -#BATCH_AUTO_UPDATE="Y" - -###################################### -# Do NOT modify code below # -###################################### - - -##################### -# Functions # -##################### - -usage() { - cat <<-EOF >&2 - - Usage: $(basename "$0") [operation ] - Script to run startupLogMonitor, best launched through systemd deployed by 'deploy-as-systemd.sh' - - -d Deploy startupLogMonitor as a systemd service - -u Skip script update check overriding UPDATE_CHECK value in env (must be first argument to script) - EOF - exit 1 -} - -createStartuplogDB() { - if ! mkdir -p "${STARTUPLOG_DIR}" 2>/dev/null; then echo "ERROR: failed to create directory to store blocklog: ${STARTUPLOG_DIR}" && return 1; fi - if [[ ! -f ${STARTUPLOG_DB} ]]; then # create a fresh DB with latest schema - sqlite3 ${STARTUPLOG_DB} <<-EOF - CREATE TABLE validationlog (id INTEGER PRIMARY KEY AUTOINCREMENT, event TEXT NOT NULL, at TEXT NOT NULL UNIQUE, env TEXT NOT NULL, final_chunk INTEGER, initial_chunk INTEGER); - CREATE TABLE replaylog (id INTEGER PRIMARY KEY AUTOINCREMENT, event TEXT NOT NULL, at TEXT NOT NULL UNIQUE, env TEXT NOT NULL, slot INTEGER, tip INTEGER); - CREATE TABLE statistics (id INTEGER PRIMARY KEY AUTOINCREMENT, event TEXT NOT NULL, start INTEGER, end INTEGER, env TEXT NOT NULL); - EOF - echo "SQLite startuplog DB created: ${STARTUPLOG_DB}" - fi -} - -deploy_systemd() { - echo "Deploying ${CNODE_VNAME}-startup-logmonitor as systemd service.." - ${sudo} bash -c "cat <<-'EOF' > /etc/systemd/system/${CNODE_VNAME}-startup-logmonitor.service - [Unit] - Description=Cardano Node - Startup Log Monitor - Wants=network-online.target - After=network-online.target - - [Service] - Type=simple - Restart=on-failure - RestartSec=20 - User=${USER} - WorkingDirectory=${CNODE_HOME}/scripts - ExecStart=/bin/bash -l -c \"exec ${CNODE_HOME}/scripts/startupLogMonitor.sh -u\" - ExecStop=/bin/bash -l -c \"exec kill -2 \$(ps -ef | grep -m1 ${CNODE_HOME}/scripts/startupLogMonitor.sh | tr -s ' ' | cut -d ' ' -f2) &>/dev/null\" - KillSignal=SIGINT - SuccessExitStatus=143 - StandardOutput=syslog - StandardError=syslog - SyslogIdentifier=${CNODE_VNAME}-startup-logmonitor - TimeoutStopSec=5 - KillMode=mixed - - [Install] - WantedBy=multi-user.target - EOF" && echo "${CNODE_VNAME}-startup-logmonitor.service deployed successfully!!" && ${sudo} systemctl daemon-reload && ${sudo} systemctl enable ${CNODE_VNAME}-startup-logmonitor.service -} - -################### -# Execution # -################### - -while getopts du opt; do - case ${opt} in - d ) DEPLOY_SYSTEMD="Y" ;; - u ) SKIP_UPDATE=Y ;; - \? ) usage ;; - esac -done -shift $((OPTIND -1)) - -[[ -z "${SKIP_UPDATE}" ]] && SKIP_UPDATE=N - -PARENT="$(dirname $0)" - -[[ ! -f "$(dirname $0)"/env ]] && echo -e "\nCommon env file missing, please ensure latest guild-deploy.sh was run and this script is being run from ${CNODE_HOME}/scripts folder! \n" && exit 1 -. "$(dirname $0)"/env offline -case $? in - 1) echo -e "ERROR: Failed to load common env file\nPlease verify set values in 'User Variables' section in env file or log an issue on GitHub" && exit 1;; - 2) clear ;; -esac - -[[ -z "${BATCH_AUTO_UPDATE}" ]] && BATCH_AUTO_UPDATE=N -[[ -z "${STARTUPLOG_DIR}" ]] && export STARTUPLOG_DIR="${CNODE_HOME}/guild-db/startuplog" -[[ -z "${STARTUPLOG_DB}" ]] && export STARTUPLOG_DB="${STARTUPLOG_DIR}/startuplog.db" -[[ -z ${SUDO} ]] && SUDO='Y' -[[ "${SUDO}" = 'Y' ]] && sudo="sudo" || sudo="" - -####################################################### -# Version Check # -####################################################### -clear - -if [[ ${UPDATE_CHECK} = Y && ${SKIP_UPDATE} != Y ]]; then - - echo "Checking for script updates..." - - # Check availability of checkUpdate function - if [[ ! $(command -v checkUpdate) ]]; then - echo -e "\nCould not find checkUpdate function in env, make sure you're using official guild docs for installation!" - exit 1 - fi - - # check for env update - ENV_UPDATED=${BATCH_AUTO_UPDATE} - checkUpdate "${PARENT}"/env N N N - case $? in - 1) ENV_UPDATED=Y ;; - 2) exit 1 ;; - esac - - # check for startupLogMonitor.sh update - checkUpdate "${PARENT}"/startupLogMonitor.sh ${ENV_UPDATED} - case $? in - 1) $0 "-u" "$@"; exit 0 ;; # re-launch script with same args skipping update check - 2) exit 1 ;; - esac - - # source common env variables in case it was updated - . "${PARENT}"/env offline &>/dev/null - case $? in - 0) : ;; # ok - 2) echo "continuing with startup log monitoring..." ;; - *) exit 1 ;; - esac - -fi - -# get log file from config file specified in env -unset logfile -if [[ "${CONFIG##*.}" = "yaml" ]]; then - [[ $(grep "scName.*\.json" "${CONFIG}") =~ scName:.\"(.+\.json)\" ]] && logfile="${BASH_REMATCH[1]}" -elif [[ "${CONFIG##*.}" = "json" ]]; then - logfile=$(jq -r '.setupScribes[] | select (.scFormat == "ScJson") | .scName' "${CONFIG}") -fi -[[ -z "${logfile}" ]] && echo -e "${FG_RED}ERROR:${NC} failed to locate json logfile in node configuration file\na setupScribe of format ScJson with extension .json expected" && exit 1 - - -if renice_cmd="$(command -v renice)"; then ${renice_cmd} -n 19 $$ >/dev/null; fi - -#Deploy systemd if -d argument was specified -if [[ "${DEPLOY_SYSTEMD}" == "Y" ]]; then - deploy_systemd && exit 0 - exit 2 -fi - -[[ ! -f ${STARTUPLOG_DB} ]] && createStartuplogDB - -echo "~~~~~~~~~~~~~~~~~~~~~~~~~" -echo "~~ STARTUP LOG MONITOR STARTED ~~" -echo "monitoring ${logfile} for traces" - -# Continuously parse cardano-node json log file for traces -while read -r logentry; do - # Traces monitored: - # - TraceOpenEvent.StartedOpeningDB - # - TraceOpenEvent.OpenedDB - # - TraceOpenEvent.StartedOpeningImmutableDB - # - TraceOpenEvent.OpenedImmutableDB - # - TraceOpenEvent.StartedOpeningLgrDB - # - TraceOpenEvent.OpenedLgrDB - # - TraceImmutableDBEvent.StartedValidatingChunk - # - TraceImmutableDBEvent.ValidatedLastLocation - # - TraceLedgerReplayEvent - - case "${logentry}" in - *TraceOpenEvent.StartedOpeningDB* ) - if ! at="$(jq -er '.at' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.at' not found" && continue; else at="$(sed 's/\.[0-9]\{2\}Z/+00:00/' <<< ${at})"; fi - if ! env="$(jq -er '.env' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.env' not found" && continue; fi - if ! event="$(jq -er '.data.kind' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.data.kind' not found" && continue; fi - sqlite3 "${STARTUPLOG_DB}" "INSERT OR IGNORE INTO statistics (event,start,env) values ('OpenDB','${at}','${env}');" - ;; - *TraceOpenEvent.OpenedDB* ) - if ! at="$(jq -er '.at' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.at' not found" && continue; else at="$(sed 's/\.[0-9]\{2\}Z/+00:00/' <<< ${at})"; fi - if ! env="$(jq -er '.env' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.env' not found" && continue; fi - if ! event="$(jq -er '.data.kind' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.data.kind' not found" && continue; fi - sqlite3 "${STARTUPLOG_DB}" "UPDATE statistics SET end = '${at}' WHERE id IN (SELECT id FROM (SELECT * FROM statistics WHERE event = 'OpenDB' AND env = '${env}' AND end is Null ORDER BY event DESC) ORDER BY id DESC LIMIT 1) ;" - ;; - *TraceOpenEvent.StartedOpeningImmutableDB* ) - if ! at="$(jq -er '.at' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.at' not found" && continue; else at="$(sed 's/\.[0-9]\{2\}Z/+00:00/' <<< ${at})"; fi - if ! env="$(jq -er '.env' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.env' not found" && continue; fi - if ! event="$(jq -er '.data.kind' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.data.kind' not found" && continue; fi - sqlite3 "${STARTUPLOG_DB}" "INSERT OR IGNORE INTO statistics (event,start,env) values ('OpenImmutableDB','${at}','${env}');" - ;; - *TraceOpenEvent.OpenedImmutableDB* ) - if ! at="$(jq -er '.at' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.at' not found" && continue; else at="$(sed 's/\.[0-9]\{2\}Z/+00:00/' <<< ${at})"; fi - if ! env="$(jq -er '.env' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.env' not found" && continue; fi - if ! event="$(jq -er '.data.kind' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.data.kind' not found" && continue; fi - sqlite3 "${STARTUPLOG_DB}" "UPDATE statistics SET end = '${at}' WHERE id IN (SELECT id FROM (SELECT * FROM statistics WHERE event = 'OpenImmutableDB' AND env = '${env}' AND end is Null ORDER BY event DESC) ORDER BY id DESC LIMIT 1) ;" - ;; - *TraceOpenEvent.StartedOpeningLgrDB* ) - if ! at="$(jq -er '.at' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.at' not found" && continue; else at="$(sed 's/\.[0-9]\{2\}Z/+00:00/' <<< ${at})"; fi - if ! env="$(jq -er '.env' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.env' not found" && continue; fi - if ! event="$(jq -er '.data.kind' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.data.kind' not found" && continue; fi - sqlite3 "${STARTUPLOG_DB}" "INSERT OR IGNORE INTO statistics (event,start,env) values ('OpenLedgerDB','${at}','${env}');" - ;; - *TraceOpenEvent.OpenedLgrDB* ) - if ! at="$(jq -er '.at' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.at' not found" && continue; else at="$(sed 's/\.[0-9]\{2\}Z/+00:00/' <<< ${at})"; fi - if ! env="$(jq -er '.env' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.env' not found" && continue; fi - if ! event="$(jq -er '.data.kind' <<< ${logentry})"; then echo "ERROR[TraceOpenEvent]: invalid json schema, '.data.kind' not found" && continue; fi - sqlite3 "${STARTUPLOG_DB}" "UPDATE statistics SET end = '${at}' WHERE id IN (SELECT id FROM (SELECT * FROM statistics WHERE event = 'OpenLedgerDB' AND env = '${env}' AND end is Null ORDER BY event DESC) ORDER BY id DESC LIMIT 1) ;" - ;; - *TraceImmutableDBEvent.StartedValidatingChunk* ) - if ! at="$(jq -er '.at' <<< ${logentry})"; then echo "ERROR[TraceImmutableDBEvent]: invalid json schema, '.at' not found" && continue; else at="$(sed 's/\.[0-9]\{2\}Z/+00:00/' <<< ${at})"; fi - if ! env="$(jq -er '.env' <<< ${logentry})"; then echo "ERROR[TraceImmutableDBEvent]: invalid json schema, '.env' not found" && continue; fi - if ! event="$(jq -er '.data.kind' <<< ${logentry})"; then echo "ERROR[TraceImmutableDBEvent]: invalid json schema, '.data.kind' not found" && continue; fi - if ! initialChunk="$(jq -er '.data.initialChunk' <<< ${logentry})"; then echo "ERROR[TraceImmutableDBEvent]: invalid json schema, '.data.initialChunk' not found" && continue; fi - if ! finalChunk="$(jq -er '.data.finalChunk' <<< ${logentry})"; then echo "ERROR[TraceImmutableDBEvent]: invalid json schema, '.data.finalChunk' not found" && continue; fi - sqlite3 "${STARTUPLOG_DB}" "INSERT OR IGNORE INTO validationlog (event,at,env,final_chunk,initial_chunk) values ('${event}','${at}','${env}',${finalChunk},${initialChunk});" - ;; - *TraceImmutableDBEvent.ValidatedLastLocation* ) - if ! at="$(jq -er '.at' <<< ${logentry})"; then echo "ERROR[TraceImmutableDBEvent]: invalid json schema, '.at' not found" && continue; else at="$(sed 's/\.[0-9]\{2\}Z/+00:00/' <<< ${at})"; fi - if ! env="$(jq -er '.env' <<< ${logentry})"; then echo "ERROR[TraceImmutableDBEvent]: invalid json schema, '.env' not found" && continue; fi - if ! event="$(jq -er '.data.kind' <<< ${logentry})"; then echo "ERROR[TraceImmutableDBEvent]: invalid json schema, '.data.kind' not found" && continue; fi - if ! initialChunk="$(jq -er '.data.initialChunk' <<< ${logentry})"; then echo "ERROR[TraceImmutableDBEvent]: invalid json schema, '.data.initialChunk' not found" && continue; fi - if ! finalChunk="$(jq -er '.data.finalChunk' <<< ${logentry})"; then echo "ERROR[TraceImmutableDBEvent]: invalid json schema, '.data.finalChunk' not found" && continue; fi - sqlite3 "${STARTUPLOG_DB}" "INSERT OR IGNORE INTO validationlog (event,at,env,final_chunk,initial_chunk) values ('${event}','${at}','${env}',${finalChunk},${initialChunk});" - ;; - *TraceLedgerReplayEvent.ReplayedBlock* ) - if ! at="$(jq -er '.at' <<< ${logentry})"; then echo "ERROR[TraceLedgerReplayEvent]: invalid json schema, '.at' not found" && continue; else at="$(sed 's/\.[0-9]\{2\}Z/+00:00/' <<< ${at})"; fi - if ! env="$(jq -er '.env' <<< ${logentry})"; then echo "ERROR[TraceLedgerReplayEvent]: invalid json schema, '.env' not found" && continue; fi - if ! event="$(jq -er '.data.kind' <<< ${logentry})"; then echo "ERROR[TraceLedgerReplayEvent]: invalid json schema, '.data.kind' not found" && continue; fi - if ! slot="$(jq -er '.data.slot' <<< ${logentry})"; then echo "ERROR[TraceLedgerReplayEvent]: invalid json schema, '.data.slot' not found" && continue; fi - if ! tip="$(jq -er '. .data.tip' <<< ${logentry})"; then echo "ERROR[TraceLedgerReplayEvent]: invalid json schema, '.data.tip' not found" && continue; fi - sqlite3 "${STARTUPLOG_DB}" "INSERT OR IGNORE INTO replaylog (event,at,env,slot,tip) values ('${event}','${at}','${env}',${slot},${tip});" - ;; - * ) : ;; # ignore - esac -done < <(tail -F -n0 "${logfile}")