Falco can send alerts to one or more channels:
The channels are configured via the falco configuration file falco.yaml
. See the Falco Configuration page for more details. Here are details on each of those channels.
When configured to send alerts via standard output, a line is printed for each alert. Here's an example:
stdout_output:
enabled: true
10:20:05.408091526: Warning Sensitive file opened for reading by non-trusted program (user=root command=cat /etc/shadow file=/etc/shadow)
Standard output is useful when using Fluentd or Logstash to capture logs from containers. Alerts can then be stored in Elasticsearch, and dashboards can be created to visualize the alerts. For more information, read this blog post.
When run in the background via the -d/--daemon
command line option, standard output messages are discarded.
When configured to send alerts to a file, a message is written to the file for each alert. The format is very similar to the Standard Output format:
file_output:
enabled: true
keep_alive: false
filename: ./events.txt
When keep_alive
is false (the default), for each alert the file is opened for appending, the single alert is written, and the file is closed. The file is not rotated or truncated. If keep_alive
is set to true, the file is opened before the first alert and kept open for all subsequent alerts. Output is buffered and will be flushed only on close. (This can be changed with --unbuffered
).
If you'd like to use a program like logrotate to rotate the output file, an example logrotate config is available here.
As of Falco 0.10.0, falco will close and reopen its file output when signaled with SIGUSR1
. The logrotate example above depends on it.
When configured to send alerts to syslog, a syslog message is sent for each alert. The actual format depends on your syslog daemon, but here's an example:
syslog_output:
enabled: true
Jun 7 10:20:05 ubuntu falco: Sensitive file opened for reading by non-trusted program (user=root command=cat /etc/shadow file=/etc/shadow)
Syslog messages are sent with a facility of LOG_USER. The rule's priority is used as the priority of the syslog message.
When configured to send alerts to a program, Falco starts the program for each alert and writes its contents to the program's standard input. You can only configure a single program output (e.g. route alerts to a single program) at a time.
For example, given a falco.yaml
configuration of:
program_output:
enabled: true
keep_alive: false
program: mail -s "Falco Notification" someone@example.com
If the program cannot normally accept an input from standard input, xargs
can be used to pass the falco events with an argument. For example :
program_output:
enabled: true
keep_alive: false
program: "xargs -I {} aws --region ${region} sns publish --topic-arn ${falco_sns_arn} --message {}"
When keep_alive
is false (the default), for each alert falco will run the program mail -s ...
and write the alert to the program. The program is run via a shell, so it's possible to specify a command pipeline if you wish to add additional formatting.
If keep_alive
is set to true, before the first alert falco will spawn the program and write the alert. The program pipe will be kept open for subsequent alerts. Output is buffered and will be flushed only on close. (This can be changed with --unbuffered).
Note: the program spawned by falco is in the same process group as falco and will receive all signals that falco receives. If you want to, say, ignore SIGTERM to allow for a clean shutdown in the face of buffered outputs, you must override the signal handler yourself.
As of Falco 0.10.0, falco will close and reopen its file output when signaled with SIGUSR1
.
If you'd like to send falco notifications to a slack channel, here's the required configuration to massage the JSON output to a form required for the slack webhook endpoint:
# Whether to output events in json or text
json_output: true
…
program_output:
enabled: true
program: "jq '{text: .output}' | curl -d @- -X POST https://hooks.slack.com/services/XXX"
If you'd like to send a stream of alerts over a network connection, here's an example:
# Whether to output events in json or text
json_output: true
…
program_output:
enabled: true
keep_alive: true
program: "nc host.example.com 1234"
Note the use of keep_alive: true
to keep the network connection persistent.
If you'd like to send alerts to an HTTP[s] endpoint, you can use the http_output
option:
json_output: true
...
http_output:
enabled: true
url: http://some.url/some/path/
Currently only unencrypted HTTP endpoints or valid, secure HTTPs endpoints are supported (ie invalid or self signed certificates are not supported).
For all output channels, you can switch to JSON output either in the configuration file or on the command line. For each alert, falco will print a JSON object, on a single line, containing the following properties:
time
: the time of the alert, in ISO8601 format.rule
: the rule that resulted in the alert.priority
: the priority of the rule that generated the alert.output
: the formatted output string for the alert.output_fields
: for each templated value in the output expression, the value of that field from the event that triggered the alert.Here's an example:
{"output":"16:31:56.746609046: Error File below a known binary directory opened for writing (user=root command=touch /bin/hack file=/bin/hack)","priority":"Error","rule":"Write below binary dir","time":"2017-10-09T23:31:56.746609046Z", "output_fields": {"evt.t\
ime":1507591916746609046,"fd.name":"/bin/hack","proc.cmdline":"touch /bin/hack","user.name":"root"}}
Here's the same output, pretty-printed:
{
"output" : "16:31:56.746609046: Error File below a known binary directory opened for writing (user=root command=touch /bin/hack file=/bin/hack)"
"priority" : "Error",
"rule" : "Write below binary dir",
"time" : "2017-10-09T23:31:56.746609046Z",
"output_fields" : {
"user.name" : "root",
"evt.time" : 1507591916746609046,
"fd.name" : "/bin/hack",
"proc.cmdline" : "touch /bin/hack"
}
}
If you'd like to send alerts to an external program connected via gRPC API (for example, the falco-exporter), you need to enable both the grpc
and grpc_output
options as described under the gRPC Configuration section.
Was this page helpful?
Glad to hear it! Please tell us how we can improve.
Sorry to hear that. Please tell us how we can improve.