split_or_unseparated_processor
| Selects the processor by detecting \n . If unsure, this processor is recommended. | line_split_processor
| Split an object into logs at each \n character using Python splitlines . Optionally, the log can be split using: chunks , a boolean parameter which enables the remaining parameters.
indices , a Python integer array which selects lines from the object.
substrings , a Python string array which selects lines from the object containing the configured strings.
regex , a Python string array which selects lines from the object using python re.findall .
| text_file_processor
| Finds header information and adds it to each event. The line_separator parameter is required. The header can be identified using one of these parameters. header (boolean) and header_field_separator to get headers from the first line of data.
field_names for a manual header.
field_separator for numbered fields.
Additionally, if use_json is true, Python orjson will convert the data to JSON. | unseparated_json_processor
| Split a JSON array into individual Devo events with Python raw_decode . The key parameter is permitted. The parameter may be a string or array. The include parameter is permitted. It a dictionary which renames JSON keys. The rename parameter can rename keys selected by the key parameter. | single_json_object_processor
| For messages containing one JSON object. Uses Python orjson to process a single JSON object. Not for arrays. The key parameter is permitted. | separated_json_processor
| Similar to other separators. The default separator is \n . The separator parameter is permitted. | json_object_to_linesplit_processor
| Split by configured value. The key string parameter or keys array parameter are permitted. | unseparated_json_processor_extract_key
| The unseparated_json_processor with an additional extraction_key parameter permitted. Use this when filtering on two levels of JSON keys. If the log message has this format Code Block |
---|
{
"id": 1,
"timestamp": 2,
"logEvents": {
"another_id": 3,
"another_timestamp": 4,
"message": "send to devo"
}
} |
The configuration Code Block |
---|
"file_format": {
"type": "unseparated_json_processor_extract_key",
"config": {
"key": "logEvents",
"extraction_key": "message"
}
}, |
will send send to devo
to Devo. | json_array_processor
| Split a JSON array into individual Devo events with Python orjson . The key string parameter or keys array parameter are permitted. | json_line_arrays_processor
| Processes JSON separated by \n . Use separated_json_processor instead. | aws_access_logs_processor
| For AWS access logs and \n splits. | bluecoat_processor
| Bluecoat. | jamf_processor
| Jamf logs. | parquet_processor
| Parquet processing using Python pandas.read_parquet . The data is converted to JSON. | guardduty_processor
| For GuardDuty processors. | vpc_flow_processor
| AWS VPC. | alt_vpc_flow_processor
| Used for exception handling. | kolide_processor
| For Kolide. | json_array_vpc_processor
| AWS VPC. | rds_processor
| RDS processor for the RDS service unseparated_json_processor . Use this if the events come in one massive JSON object. | windows_security_processor
| Windows security logs. |
Custom Service Optionsfile_format
| type A string specifying which processor to use, from the list above.
| config A dictionary of the processor’s parameters.
| record_field_mapping
| A dictionary where each key defines a variable that can be parsed out from each record (which may be referenced later in filtering). To parse something and call it type by getting type from a certain key in the record (which may be multiple layers deep). Code Block |
---|
{"type": {"keys": ["file", "type"]}, "operations": [] } |
Suppose we have logs that look like this: Code Block |
---|
{“file”: {“type”: { “log_type” : 100}}} |
To get the log_type , we should list all the keys needed to parse through the JSON in order: Code Block |
---|
keys: [“file”, “type”, “log_type”] |
In many cases, you will probably only need one key, for example, in a flat JSON that isn’t nested: Code Block |
---|
{“log_type”: 100, “other_info”: “blah” ….} |
Here you would just specify keys: ["log_type"] . There are some operations that can be used to further alter the parsed information (like split and replace ). This snippet would grab whatever is located at log["file"]["type"] and name it as type. record_field_mapping defines variables by taking them from logs, and these variables can then be used for filtering. If you have a log in JSON format like this which will be set to Devo: Code Block |
---|
{“file”: {“value”: 0, “type”: “security_log”}} |
Specifying type in record_field_mapping will allow the collector to extract the security_log and save it as type . To change the tag using a field mapping, change the routing_template to something like my.app.datasource.[record-type] . In the case of the log above, it would be sent to my.app.datasource.security_log . To filter out (not send) any records which have the type security_log , write a line_filter_rule as follows: Code Block |
---|
{"source": "record", "key": "type", "type": "match", "value": "security_log" } |
We specified the source as record because we want to use a variable from the record_field_mapping . We specified the key as type because that is the name of the variable we defined. We specified type as match because any record matching this rule we want to filter out. And we specified the value as security_log because we specifically do not want to send any records with the type equalling security_log .
The split operation is the same as if you ran the Python split function on a string. Given a filename logs/account_id/folder_name/filename and you want to save the account_id as a variable to use for tag routing or filtering. You could write a file_field_definition like this: Code Block |
---|
"account_id": [{"operator": "split", "on": "/", "element": 1}] |
This would store a variable called account_id by taking the entire filename and splitting it into pieces based on where it finds backslashes, then take the element as position one. In Python, it would look like this: Code Block |
---|
filename.split(“/”)[1] |
|
TaggingTagging can be done in many different ways. One way tagging works is by using the file field definitions: Code Block | "file_field_definitions": {
"log_type": [
{
on a string. Given a filename logs/account_id/folder_name/filename and you want to save the account_id as a variable to use for tag routing or filtering. You could write a file_field_definition like this: Code Block |
---|
"account_id": [{"operator": "split", |
|
2
}
]
},These are the elements of the filename object: Image RemovedIf you look at the highlighted object filename, you can see that we are splitting and looking for the 2nd value. This starts at 0 like arrays. So: "routing_template": "my.app.test_cequence.[file-log_type]"
Our final tag is my.app.test_cequence.detector Here is another example: file_field_definitions
| Defined as a dictionary mapping of variable names (you decide) that lists parsing rules. Each parsing rule has an operator with its own keys. Parsing rules are applied in the order they are listed in the configuration. The split operator uses the on and element keys. The file name will split into pieces considering the character or character sequence specified in the on key, and will extract whatever it is at the specified element index, as in the example below. The replace operator uses the to_replace and replace_with keys.
For example, if your filename is server_logs/12409834/ff.gz , this configuration would store the log_type as serverlogs : Code Block |
---|
This would store a variable called account_id by taking the entire filename and splitting it into pieces based on where it finds backslashes, then take the element as position one. In Python, it would look like this: Code Block |
---|
filename.split(“/”)[1] |
|
Automatic TaggingTags can be generated using record field mapping or file field definitions. File name split Code Block |
---|
"file_field_definitions": {
"log_type": [
{
"operator": "split",
"on": "/",
"element": 2
}
]
}, |
If the filename field is foo/bar/baz/qux.json
then "routing_template": "my.app.test.[file-log_type]"
results in tag my.app.test.baz File name split and replaceIf the filename field is foo_bar/baz/qux.gz
it is helpful to remove the unwanted special character when creating the tag. For example, Code Block |
---|
"file_field_definitions": {
| { "log_type": [
{
"operator": "split",
"on": "/",
"element": 0
},
{
"operator": "replace",
"to_replace": "_ | ",",
"replace_with": ""
}
]
} |
combined with "routing_template A string defining how to build the tag to send each message, for example, ": "my.app. wow.[record-type]test.[file-log_type] If the type extracted during record_field_mapping was null , the record would be sent to the tag "
will result in my.app. wowtest. nullfoobar
Options for filteringLine-level filtersThese are a list of rules for filtering out single events. |