
Security News
OpenGrep Restores Fingerprinting in JSON and SARIF Outputs
OpenGrep has restored fingerprint and metavariable support in JSON and SARIF outputs, making static analysis more effective for CI/CD security automation.
A simple cloud services watchdog with digest notification support & no external dependencies
Warning: Currently in alpha, use at own risk
Barky is intended to run custom monitoring in a simple and effective way, using a tool with no external dependencies (like databases) in order to drive effective alert/outage notifications.
What problem does this solve?
Standard alerting systems can be noisy and overwhelming, sending hundreds of notifications when issues arise. With Barky’s digest feature, you get a single alert with all the key details and updates as the situation changes. You’ll be notified when the issue is resolved and receive updates at a frequency you control, keeping you informed without the overload.
What does this do?
It runs a custom set of evaluators (configured in simple markup using YAML) with (current) support for the following checks:
Evaluations supported:
The results of the evaluators are emitted via stdout
using a pipe delimited format (per monitor). This can be logged and ingested into any cloud log management tool for tracking.
In addition to this, the results are evaluated and alerts emitted in a digest format via the following supported channels:
So, the pipeline is Evaluate > Digest
where the evaluation emits status of things monitored and the digest step emits any alerts (triggered, ongoing or resolution) via the configured channels. The digest step is optional.
Commands:
run
- runs the evaluatorloop
- runs the evaluator in a loop (every 30s) until terminatedkillall
- kills all running barky processes based on lock files in the current directory# run the evaluator without digest
npx barky@latest run configs/my.yaml
# run the evaluator with specific evaluator (comma separate for more)
npx barky@latest run configs/my.yaml --eval=web
# run the evaluator and digest step
npx barky@latest run configs/my.yaml --digest=configs/digest/my-team.yaml --title="ACME Public"
# run in a loop (every 30s) until exit
npx barky@latest loop configs/my.yaml --digest=configs/digest/my-team.yaml --title="ACME Public"
# bypass prompt for package installation
npx --yes barky@latest run configs/my.yaml --eval=web --digest=configs/digest/my-team.yaml --title="ACME Public"
Evaluator configuration is managed via a YAML file which is referenced via a CLI command argument.
The config
keyword in the YAML file defines locale and timezone as per the example below:
config:
locale: en-ZA
timezone: Africa/Johannesburg
port: 3000 # the port to run the web UI on (defaults to 3000)
The following high level keys are supported (note, it is case sensitive):
For convenience, you can store rules in separate yaml files and include them as follows. Note that paths should be relative to parent.
import:
- rules/web.yaml
- rules/sumo.yaml
- rules/mysql.yaml
For each evaluator, a ping is emitted to stdout
with information about what type was evaluated, and how many were evaluated.
Example:
# date|type|label|identifer|success|result_text|result|time_taken
2023-08-25T13:17:24.471Z|web|monitor|ping|1|7 evaluated|7|132.78
In addition, each evaluator app supports the following properties:
name
- String - a friendly name for the relevant appquiet
- Any - if set to a truthy value, will suppress success outputtimeout
- Numeric - a value in milliseconds (example: 10000 for 10s)vary-by
- Array<string|string[]> - enables variations of fields like name, url or query (depending on the type of evaluator)every
- String value representing how often to evaluate the rule, defaults to "30s" - value must be multiples of 30s, examples: 60s, 90s, 10m, 1h (only applicable in loop mode)Examples of vary-by:
web:
www.codeo.co.$1:
vary-by: [za,us,gb]
url: https://www.codeo.co.za/en-$1
This will generate 3 apps to be evaluated with:
names: [www.codeo.co.za, www.codeo.co.us, www.codeo.co.gb]
urls: [https://www.codeo.co.za/en-za, https://www.codeo.co.za/en-us, https://www.codeo.co.za/en-gb]
A more complex example:
web:
www.codeo.co.$1:
vary-by:
- [za, zar]
- [com, usd]
url: https://www.codeo.co.$1/currency=$2
This would generate 2 apps to be evaluated with:
names: [www.codeo.co.za, www.codeo.com]
urls: [https://www.codeo.co.za/currency=zar, https://www.codeo.com/currency=usd]
Simple example:
config:
locale: en-ZA
timezone: Africa/Johannesburg
web:
www.codeo.co.za:
url: https://www.codeo.co.za
This will trigger a check against www.codeo.co.za, and will validate that a 200 status code is returned. It will automatically include a __barky={timestamp}
querystring parameter to bust any caching and will also submit with user agent barky
.
Example successful output:
# date|type|label|identifer|success|result_text|result|time_taken
2023-08-23T15:05:53.554Z|web|health check|www.codeo.co.za|1|OK|200|184.70
Example failure output:
# date|type|label|identifer|success|result_text|result|time_taken
2023-08-23T15:07:35.339Z|web|health check|www.codeo.co.za|0|Expected status:200,received 500|500|86.43
If there is a problem with global configuration, you can expect a monitor output like this:
2023-08-23T15:08:56.172Z|watchdog|monitor|configuration|0|invalid yaml definition in file 'configs/test.yaml' - Implicit keys need to be on a single line at line 4, column 1: timezone: Africa/Johannesburg @asd ^ }||0.00
If there is a problem with an evaluator, you can expect a monitor output like this:
# date|type|label|identifer|success|result_text|result|time_taken
2023-08-23T15:13:33.467Z|web|monitor|www.codeo.co.za|0|missing url for web app 'www.codeo.co.za'||0.00
You can get further information about any error by running the tool using the --debug
switch.
Additional values that can be configured:
method
defaults to get
status
defaults to 200max-redirects
defaults to 5 - set to 0 to disable redirectstimeout
defaults to 5000 (5 seconds)headers
- a custom set of headers (see example below) - these can include environment variables using $ prefixvary-by
- enables variations of a given url, an instance for each variation is monitoredtriggers
- a list of custom response triggers (expect values to be truthy to pass)
text
- a string to search for in the response body (case-insensitive)json
- a JavaScript expression to evaluate on the JSON data returned by the requestmatch
- a regular expression to match against the response bodyalert
- defines the alert rules, see belowtls
- if property is missing, the defaults below apply
verifiy
- defaults to true, set to false to disable all certificate verificationexpiry
- defaults to 7d
(7 days) - will alert if certificate expires within this periodAlerts
Fields:
channels
- an array of channels to use (example: [sms, slack]
)links
- optional array of links to include when alerts triggerrules
- an array of rules (the first matched rule will always be used) - if match
expression is used, only rules matching the expression will be evaluated
description
- not requiredcount|any
- count means trigger after defined consecutive count of errors, any means trigger after any
count of errors in the window period definedwindow
- not required, but useful to constrain any
operator to the given window, example: -30m
means last 30 minutes. Maximum window is 24h
. Defaults to 5 minutes if not specifiedmatch
- an optional match expression to match against the monitor identifier (see below for format)days
- array of days and only required if you want to constrain the trigger to specific days of week (see example)time
- array or single range value, only required if you want to constrain the trigger to specific times of the day (times are in the timezone specified in the config)exception-policy
- the name of the alert policy (defined in the digest configuration) to use for monitor failures (such as timeouts or exceptions), if not set then the same alert configuration rules defined above will be used when the monitor incurs an unhandled errorThe match expression is composed as follows: type|label|identifier
. For example: web|web-performance|www.codeo.co.za
. The regular expression for match
will thus be compared against this string value (case-insensitive).
Advanced example:
config:
locale: en-ZA
timezone: Africa/Johannesburg
web:
www.codeo.co.za.$1:
vary-by: [za,us,gb]
url: https://www.codeo.co.za/en-$1 # the vary-by instance value is captured into $1
status: 200
timeout: 10000
max-redirects: 0 # don't follow redirects
headers:
Authorization: $my-auth-token # uses environment variable my-auth-token
x-my-custom-value: "123"
triggers:
# either match, text or json can be used (in conjunction if necessary)
- text: ok # this checks the response contains the text "ok"
message: Expected to find text "ok" in response but didn't
- match: "\\d+" # this checks the response matches the regex
message: Expected to find a numeric valid in the response, but didn't
# for json responses, the response is parsed and the expression is evaluated against the parsed object
# for example a response of { "result": 123 } could be evaluated with `result > 100`
# if the key has non-alpha-numeric values, these are replaced with underscore, i.e.
# { "my-key": 123 } would be accessed with `my_key`; note that sub-properties can also
# be dereferenced using normal notation (i.e. `my_key.sub_key > 123`)
- json: someKey.result > 100 # this checks the response is valid json and matches the expression
# note that the message can include an evaluation of properties on the json result, see below
message: "Expected someKey.result to be greater than 100 but was {{someKey?.result ?? 'unknown'}}"
alert:
channels: [sms, slack]
links:
- label: Playbook
url: "https://notion.so"
rules:
- description: Weekdays
match: .*
count: 2 # any consecutive 2 failures trigger alert
days: [mon, tue, wed, thu, fri]
time: [00:00-04:00, 6:00- 17:00] # local time as per timezone
- description: Weekends
match: .*
window: -5m
any: 3
days: [sat, sun]
time: 04:00 - 17:00
The Sumo Logic evaluator supports two modes: logs, metrics (default mode is logs).
Log Evaluator
The example below will search Sumo Logic using the given query, and iterate over the result set. The time window
searched is specified by period
.
The triggers
define the set of rules that will trigger alerts. The first trigger that matches the value for the identifier will be selected and evaluated.
The trigger's rule expression is evaluated as JavaScript.
For Sumo Logic queries, the default domain is api.eu.sumologic.com
- however, this can be overridden using an
environment variable called sumo-domain
.
The example below does not have any alerts configured, see web example above to see what you can do with alerts.
sumo:
web-performance:
name: web-performance
quiet: true # successful evaluation is suppressed
token: sumo-token # the tool will expect an environment variable with the appropriate token using this key
period: -10m to -0m
query: >
# this query gets 90th percentile response time, and error rate for sites with traffic in the last 10 minutes
_sourceCategory=system/linux/nginx _collector=*mycollector* not(host=*test*)
| if(status matches "5*", 1, 0) as error
| if(status matches "5*", 0, 1) as ok
| where responsetime >= 0
| pct(responsetime, 90) as _90p, sum(error) as error, sum(ok) as ok, count by host
| where _count > 10
| error / _count * 100 as error_rate
| host as site
| _90p as response_time
| fields site, response_time, error_rate
| order by response_time desc, error_rate desc
identifier: site # this specifies what field in the result set is the identifier to iterate over
triggers:
- match: myslowsite\.(com|net) # special rules for myslowsite.com and myslowsite.net
rules:
- expression: response_time >= 2
message: "Response time is too high: {{response_time}}s"
- expression: error_rate > 1
message: "Error rate is too high: {{error_rate}}%"
- match: .* # catch all
rules:
- expression: response_time >= 0.5
message: "Response time is too high: {{response_time}}s"
- expression: error_rate > 1
message: "Error rate is too high: {{error_rate}}%"
The trigger.rule object has the following additional properties that can be set:
Example period formats
-10m to -5m
is 10 minutes ago to 5 minutes ago-1h to 0h
is hour ago to now-1d to -1h
is 24 hours ago to 1 hour agotoday
is from 00:00AM until nowyesterday
is from 00:00AM yesterday until 00:00AM todayMetric Evaluator
The metrics evaluator enables you to define rules across the high level metrics results:
For each aggregated metric result, it exposes these values and any other columns (_collector, _source, etc) in the result set.
sumo:
cpu-performance:
mode: metrics
token: sumo-token # the tool will expect an environment variable with the appropriate token using this key
period: -5m to -0m
query: >
metric=cpu_total _source=*yoursource*
| quantize to 1s
| avg by _sourcehost # any metric can be chosen here and it will return min, max avg, count in the period
identifier: _sourcehost # this specifies what field in the result set is the identifier to iterate over
emit: [avg] # only emit the average cpu in the output
triggers:
- match: .* # catch all
rules:
- expression: avg >= 50
message: "CPU is too high: {{avg}}%"
The example below will execute the given mysql query, and iterate over the result set.
The triggers
define the set of rules that will trigger alerts. The first trigger that matches the value for the identifier and relevant rules will be selected and evaluated.
The trigger's rule expression is evaluated as JavaScript.
The example below does not have any alerts configured, see web example above to see what you can do with alerts.
Note, the connection
value is used to lookup environment variables by convention (mysql-{your-key}-host|password...
).
mysql:
queue-processing:
name: queue-performance
quiet: true # successful evaluation is suppressed
connection: aws-aurora
timeout: 15000 # query will timeout after 15s
query: >
set transaction isolation level read uncommitted;
select queue, unprocessed, minutes_to_process from some_view;
identifier: queue # specifies what field to use to match against the validator and emit as identifier (if field not in result set, is set to value set here)
emit: [unprocessed, minutes_to_process] # optional, if not set, all fields are emitted in log
triggers:
- match: .* # catch all
rules:
- expression: minutes_to_process >= 10
message: "Queue is backlogged with {{ unprocessed }} msgs & will take {{ minutes_to_process }} minutes to catch up"
alert:
# see web evaluator for example alert configuration
The trigger.rule object has the following additional properties that can be set:
The example below demonstrates how to run a custom shell script.
shell:
my-script:
timeout: 5s # defaults to 10 seconds
name: my-script
path: ./my-script.sh # relative to current yaml file (or absolute path)
responseType: json
triggers:
- rules:
- expression: "exitCode !== 0"
message: "Script failed with exit code {{exitCode}}"
- expression: failed_requests > 0
message: "Failed requests was {{ failed_requests }}"
alert:
# see web evaluator for example alert configuration
Supported response types:
json
- Barky expects a json string response (the raw response will be emitted into a stdout
variable), note that if the result is a JSON array or JSONL, multiple results will be returnedstring
- Barky expects a string response and will emit this into a stdout
variableAll environment variables are injected into the script for use. All non-alphanumeric and underscore characters are replaced with _
in the variable name. Example my-var
becomes my_var
and can be accessed with $my_var
.
A more complex example:
shell:
validate-country-$1:
vary-by: [za,us,gb]
path: ./my-script.sh # each fanned out vary-by result will have the variation passed as an argument, i.e. ./my-script.sh za
responseType: json
identifier: id
triggers:
- match: .* # catch all (you can match on the identifier field in the result set, and if missing or not set, the vary-by value will be used here (identifier is pipe delimited if multipart))
rules:
- expression: "exitCode !== 0"
message: "Script failed with exit code {{exitCode}}"
- expression: my_field > 0
message: "Failed requests was {{ failed_requests }}"
alert:
# see web evaluator for example alert configuration
The digest is the second phase of the tool, and is optional. This controls the execution of alerts.
The digest execution requires configuration of channels and their output. The digest is run as part of the monitor execution, so will only have access to the monitors configured.
Supported channels:
When executed, the digester evaluates and compares last monitor snapshot to the current snapshot and makes decisions as to what to do based on configuration.
In addition to defining the channel configuration, the digest may also optionally configure alert policies that can be used in the alert configuration. You may want to define shared configuration for aspects such as exception policies, for when a monitor cannot evaluate due to an unhandled error.
Example configuration:
mute-windows: # alerts are silenced if generated in these window periods
- match: mysql.*performance # only for monitors matching this regex
time: 00:00 - 06:00
- date: 2023-08-27 # only matches for this specific date
time: 22:00 - 24:00 # 2PM to 4PM for a specific date
- time: 00:00 - 06:00 # every weekday midnight to 6AM
days: [mon, tue, wed, thu, fri]
alert-policies:
monitor-exception:
channels: [slack]
rules:
- description: More than 3 monitor errors in a 10 minute window
any: 3
window: 10m
channels:
sms:
type: sms
provider: clickatell_legacy # currently only supported provider
template:
prefix: '{{ title }}' # this is a global variable passed in via the CLI "title" param
postfix: Please see Ops Slack channel for any updates.
interval: 5m # how often to send alert updates
contacts: # list of people to contact
- name: Rohland
mobile: +2782...
slack:
type: slack
template:
prefix: '<!channel> {{ title }}' # <!channel> alerts everyone in the given channel
postfix:
summary: "<https://acme.com/dashboard|Please see dashboard here:>"
interval: 1h # how often to post updates as a new message
token: slack-token # we expect an environment variable with this name
channel: "#ops"
workspace: "acme" # this is your workspace name in slack (optional)
A note on templates: Slack prevents message updates from exceeding 4k characters, so if the Slack notification exceeds this, a summary message is posted instead with high level stats of what's going on. In this context, the summary template is included.
Mute Windows
Any number of windows can be defined where alerts will be silenced. This is useful for maintenence windows, or when you know that a monitor will be failing for a period of time.
Fields:
match
: regex to match monitor identifier(see below for format) (not required)date
: specific date to match (not required)time
: time range to match (required)The match
regular expression is used to match against the monitor identifier that is a string value composed of:
type
- the type of monitor (example: web, sumo, mysql)label
- the name of the monitor (example: web-performance)identifier
- the name of the failing identifier (example: www.codeo.co.za)The value is composed as follows: type::label::identifier
. For example: web::web-performance::www.codeo.co.za
. The regular expression for match
will thus be compared against this string value (case-insensitive).
For SMS, any initial change into a failure state for the relevant team, will trigger a single SMS, which will include a summary of what has gone wrong, and will indicate that further updates will be sent via Slack. An update is posted every 15 minutes.
Example outage alerts (prefix and postfix can be configured):
{prefix} Outage STARTED at 17:40:00. 1 health check affected. {postfix}
Example update configured at interval:
{prefix} Outage ONGOING for 15 minutes (since 17:40:00). 1 health check affected. {postfix}
Example resolution notification:
{prefix} Outage RESOLVED at 17:59:00. Duration was 19 minutes. {postfix}
Currently, the only supported provider is the Clickatell legacy SMS gateway at https://sms-gateway.clickatell.com/. The provider expects the following environment variables to be configured:
For Slack, more detail is posted about an outage, and everyone is notified upon the initial outage via @everyone
. Example:
@channel Outage 🔥: Started at: 11:15AM Duration: 5 minutes
There is an outage affecting 2 health checks:
- web:health check → www.codeo.co.za (expected 200, received 500)
- web:health check → www.codeo2.co.za (expected 200, received 500)
Last updated: 11:20AM
The above message will be updated at the interval the tool is updated, and at the notification_interval
a new message will be started (to assist with the notification scrolling offscreen in Slack).
Example resolution:
✅ @channel Previous outage resolved at 10:11:08. Duration was 1 minute. See above for more details about affected services.
A simple web interface is exposed on the configured port (defaults to 3000, editable using the global config) that presents a combined output of all alerts and their current status. It enables dynamic muting/un-muting of alerts, and provides a summary of active, resolved and muted alerts. The UI is updated every 10 seconds.
Security of this interface is left in the hands of the user.
Messages using {{ some_var }}
syntax have access to a few helper functions:
Configure a local .env.local
file with the configuration as outlined in the documentation above, and then execute
as follows. Note, Barky will prioritise .env.local
over entries in a .env
file.
# this expects that you have a subfolder called path and files called config.yml and digest.yml
npm start -- loop ./path/config --eval=web --digest=./path/digest --debug
FAQs
A simple cloud services watchdog with digest notification support & no external dependencies
The npm package barky receives a total of 157 weekly downloads. As such, barky popularity was classified as not popular.
We found that barky demonstrated a healthy version release cadence and project activity because the last version was released less than a year ago. It has 1 open source maintainer collaborating on the project.
Did you know?
Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.
Security News
OpenGrep has restored fingerprint and metavariable support in JSON and SARIF outputs, making static analysis more effective for CI/CD security automation.
Security News
Security experts warn that recent classification changes obscure the true scope of the NVD backlog as CVE volume hits all-time highs.
Security Fundamentals
Attackers use obfuscation to hide malware in open source packages. Learn how to spot these techniques across npm, PyPI, Maven, and more.