New Event Structure
Opened this issue · 37 comments
#2355 changed the primary user experience of Tracee to be event oriented (previously events were considered internal and hidden from the user). Therefore:
- The event schema needs to be formalized and stabilized. Since it's no longer internal.
- The event structure needs to be generalized. Since events are will now be used for detections, captures and more.
Following is the updated event schema based on the comments below:
timestamp
name
id
- machine readable id (integer). Note: current event id isn't good since it is architecture specific- //
version
- use semver where major is a breaking change in the event (e.g. one of the event's fields under data has been changed or removed), minor is a non breaking change (e.g. a new field was added to the event under data) and patch (e.g. a bug fix). Since this data is static, we may remove this or make optional - //
tags
- since this data is static, we may remove this or make optional labels
- doesn't exist. For future use.policies
matched
actions
- doesn't exist, for future use - list of actions taken (currently the only action we have is print).
workload
process
executable
path
name
- the binary name (basename of the path) - doesn't exist, consider adding (in another issue)
uniqueId
- unique id of the processpid
hostPid
executionTime
- time of last exec. Doesn't exist, consider adding (in another issue)realUser
id
name
- doesn't exist, consider adding (in another issue)
user
- effective user. Doesn't exist, consider adding (in another issue)id
name
ancestors
- process ancestors array. Only direct parent will be populated by default with the following fields:uniqueId
pid
hostPid
- Other ancestor fields may be populated by threat detection events
thread
startTime
name
(aka "comm")tid
hostTid
capabilities
- doesn't exist, consider adding (in another issue)syscall
- the syscall that triggered this eventcompat
- boolean. moved fromflags.compat
userStackTrace
- if enabled, will be here
container
id
name
image
id
repoDigest
name
isRunning
- boolean. moved fromflags
startTime
- Timestamp of container start time. Doesn’t exist. Will replacestarted
pid
- entrypoint's pid. Doesn’t exists, consider adding
k8s
pod
name
uid
labels
namespace
name
data
- Any relevant field (per-event schema)
returnValue
(if relevant will appear here)
threat
(if relevant will appear here) - static data about threats (can be omitted)description
mitre
tactic
name
technique
name
id
severity
triggeredBy
(will appear on threat detection events)
1.name
2.id
3.data
We also discussed versioning the event schema, but not including the version with each event, for efficiency.
while this represent the complete event schema, I also think we should try to emit minimal events. For example, I suppose most people won't need the full description with every event, or the full stack trace. We already allow conditionally including some elements (stack trace, syscall, exec-env), and I think we need to expand this control to all event fields. for discussing this I'll create another issue.
@itaysk Do you think we should maybe use the protocol.Event
feature for this (see types/protocol
)? We could add a ContentType
header to differentiate events. Or do we want to stick with trace.Event
with split up context fields?
I've had this thought as well in relation to the recent Metadata
field we've added for the finding events, it could've fit into either the protocol.EventHeader
or that Findings can have a different payload body, but be considered an event through the protocol.
Actually I'm not very familiar with how protocol is used now. I thought it's supposed to become obsolete in the unified binary approach. Can you elaborate on your proposal or perhaps add an example?
The protocol is more a way to differentiate different payloads going into the rules engine.
For example in CNDR it is used to define the ability to receive events sent directly from CNDR into the signature (for example to set data storage in the signature).
The same way the protocol can define a different payload for signature inputs, it can be used to define different event payloads with different shapes (for example, before we changed detections to be trace.Event
we could have defined it it to be a different payload in the protocol.Event
instead).
I am not sure if we want to currently support different payloads for events going into the engine (since the engine is now supposed to be internal).
May I suggest we move the event definition to a proto
file while we're at it?
Should make it easier to integrate tracee into gRPC systems later.
The protocol is more a way to differentiate different payloads going into the rules engine.
For example in CNDR it is used to define the ability to receive events sent directly from CNDR into the signature (for example to set data storage in the signature).
The same way the protocol can define a different payload for signature inputs, it can be used to define different event payloads with different shapes (for example, before we changed detections to be trace.Event we could have defined it it to be a different payload in the protocol.Event instead).
I am not sure if we want to currently support different payloads for events going into the engine (since the engine is now supposed to be internal).
I'm still not sure there's a use case for this, or I didn't fully understand it, but regardless seems like we reached the same conclusion that the "rule engine" is gone now.
+1 for proto
Has the event type serialization, using protobufs, ever been tested? I'm particularly concerned about:
Protocol buffers tend to assume that entire messages can be loaded into memory at once and are not larger than an object graph. For data that exceeds a few megabytes, consider a different solution; when working with larger data, you may effectively end up with several copies of the data due to serialized copies, which can cause surprising spikes in memory usage.
I mean, we will not serialize eBPF events, for sure, but if we're considering converting the Tracee type to protobuf, in a high rate, coming from an external source, then we should do some measurement before taking the decision, IMO.
Has the event type serialization, using protobufs, ever been tested? I'm particularly concerned about:
Protocol buffers tend to assume that entire messages can be loaded into memory at once and are not larger than an object graph. For data that exceeds a few megabytes, consider a different solution; when working with larger data, you may effectively end up with several copies of the data due to serialized copies, which can cause surprising spikes in memory usage.
I've actually written a PR (#2070) once where I added a ebpf -> rules serialization with protobuf. From my measurement in that PR, protobuf serialization was quicker than both json and gob ONLY if the struct was initially in the protobuf form already - conversion from trace.Event
to proto.Event
was the overhead in that printer format.
When taking conversion time into consideration it was quicker than both.
Good to know that.
I updated the struct to have process related context grouped together (pids, tids, comm, star time, namespaces, cgroup, uid).
Also moved processorId to be part of the context
Adding this here for future reference and consideration: https://www.elastic.co/blog/ecs-elastic-common-schema-otel-opentelemetry-faq
After talking with @yanivagman we thought about some slight changes:
- rename
args
tofields
- rename
kubernetes
topod
(and remove pod prefix from internal fields) - remove parent information from
process
intro: - new
ancestors
field which is an array ofprocess
(not all fields will be implemented so need toomitEmpty
- move
flags.containerStarted
tocontainer.started
- move
flags.isCompat
toprocess.compat
TBD:
severity
- location and nameversion
- maybe move to root
I've updated the description to reflect this
Let's also take the opportunity to only expose the fields mentioned here to the user. Any other (internal) fields should not be part of the trace.Event.
We can do this by embedding the trace.Event struct into an internal struct used by the pipeline only where we can also add some extra fields (e.g. matchedActions, matchedPolicies bitmap, cgroup id, etc.)
More detailed struct with some comments added:
timestamp
id
- better if this was an integer, but we also need to keep backwards compatibility with e.g. TRC-XXX ids. Maybe call this differently so we can add integer ID in the future?name
metadata
description
severity
(orpriority
for non signature events?)tags
version
(this version describes the event fields version, and not the schema of the event structure)misc
-map[string]interface{}
context
process
- all process related contextexecutionTime
- doesn't exist, consider addingname
id
namespaceId
userId
thread
startTime
id
namespaceId
mountNamespaceId
- consider removingpidNamespaceId
- consider removingutsName
- consider removingsyscall
- moved herestackTrace
- if enabled, will be herecompat
- moved fromflags.compat
ancestors
- array of all process ancestors, in a structure similar toprocess
. Element 0 is parent.container
- all container related contextid
name
image
imageDigest
started
- moved fromflags
- consider setting this as a timestamp of container start time and not boolean
pod
- all pod related context (from kubernetes)name
namespace
uid
sandbox
processorId
- moved here
fields
- renamed fromargs
(or should we call itdata
)?- every event attribute
returnValue
(if relevant will appear here)
matchedPolicies
id - better if this was an integer
why?
executionTime - doesn't exist, consider adding
agree, but let's discuss in a separate issue as it's about adding an entirely new feature? (and won't break the event structure)
thread
what's the motivation for adding this layer?
mountNamespaceId - consider removing
pidNamespaceId - consider removing
utsName - consider removing
+1
ancestors
since we're discussing the "event structure", having a field in the root implies it's an "event" field. In this case, I'm internally reading this as "event ancestors" but the meaning is "process ancestors". should we relocate it under process or prefix it with process to clarify?
started - moved from flags - consider setting this as a timestamp of container start time and not boolean
agree, but let's discuss in a separate issue as it's about adding an entirely new feature?
pod - all pod related context (from kubernetes)
if this is a kubernetes thing, should we add kubernetes to the name? if this is not, are we sure the structure will work for other pod implementation?
sandbox
What does this mean? from reading the code I think I can gather that it is looking at container lable io.kubernetes.docker.type==sandbox
but I don't understand what/why/how it's related to kubernetes pod. @NDStrahilevitz can you please explain?
matchedPolicies
This is the only field in the event that is Tracee-specific, and not meaningful by itself. For example, if I'm streaming all events to splunk and then someone else at another time sees this event there, all other field would make sense since they describe what happened, but to understand this field I need to understand trace and how it was configured started.
I won't turn this into a debate about the usefulness of this event, but at least I'd suggest to prefix with "tracee". or even better, if we have intentions to add more tracee-specific info int the future (#3153) then better to put it under a "tracee" level.
id - better if this was an integer
why?
Makes it easier and more efficient to process Tracee's events in later stages.
Currently, both id and name are strings, which is redundant (yet required for backwards compatibility with already written signatures, TRC-XXX)
executionTime - doesn't exist, consider adding
agree, but let's discuss in a separate issue as it's about adding an entirely new feature? (and won't break the event structure)
There was no intention to do it now, just adding some fields which can be useful to have in the struct. We can open a seperate discussion.
thread
what's the motivation for adding this layer?
Grouping fields which are related to the context of the thread and not specific to the process (can be different between different threads of the same process)
ancestors
since we're discussing the "event structure", having a field in the root implies it's an "event" field. In this case, I'm internally reading this as "event ancestors" but the meaning is "process ancestors". should we relocate it under process or prefix it with process to clarify?
Agree. Let's move under process then.
started - moved from flags - consider setting this as a timestamp of container start time and not boolean
agree, but let's discuss in a separate issue as it's about adding an entirely new feature?
Ditto
pod - all pod related context (from kubernetes)
if this is a kubernetes thing, should we add kubernetes to the name? if this is not, are we sure the structure will work for other pod implementation?
Better to make it generic and not just for k8s. For fields that are k8s specific we can add a prefix
matchedPolicies
This is the only field in the event that is Tracee-specific, and not meaningful by itself. For example, if I'm streaming all events to splunk and then someone else at another time sees this event there, all other field would make sense since they describe what happened, but to understand this field I need to understand trace and how it was configured started. I won't turn this into a debate about the usefulness of this event, but at least I'd suggest to prefix with "tracee". or even better, if we have intentions to add more tracee-specific info int the future (#3153) then better to put it under a "tracee" level.
Agree. Let's add "tracee" level then
So the modified struct will look like this (added a few future suggestions as well):
timestamp
name
id
- machine readable id (integer). Note: current event id isn't good since it is architecture specificmetadata
tags
dataVersion
(this version describes the event fields version, and not the schema of the event structure)detection
name
- keep backwards compatibility with e.g. TRC-XXX ids. Consider removing and use id instead.severity
mitreCategory
context
process
executionTime
- doesn't exist, consider adding (in another issue)binary
- binary path used to execute the process. Doesn't exist, consider adding (in another issue)pid
namespacePid
userId
userName
- doesn't exist, consider adding (in another issue)ancestors
- array of all process ancestors, where each entry has the above process fields. Currently only populate element 0, which is the parent process.thread
startTime
name
(aka "comm")tid
namespaceTid
capabilities
- doesn't exist, consider adding (in another issue)mountNamespaceId
- consider removingpidNamespaceId
- consider removingutsName
- consider removingsyscall
- moved herecompat
- moved fromflags.compat
userStackTrace
- if enabled, will be here
container
id
name
image
id
name
digest
started
- moved fromflags
- consider setting this as a timestamp of container start time and not boolean
pod
name
namespace
uid
sandbox
- consider renaming
data
- per-event schema of relevant fields (e.g. argument for system calls)
returnValue
(if relevant will appear here)
tracee
matchedPolicies
To avoid breaking clients we want to support the old event structure as a printer, so give users more time to migrate to the new structure. So we will have a json-v1printer which is the old format, and json-v2 (aka "json" printer) with the new format.
Plan for migrating ParsedArgs list to schema
As part of the issue an internal PipelineObject
will be added to tracee, whose fields will be (roughly) event
, and pipelineData
.
The following flow should run if parse-args
is enabled:
- The
ParsedArgs
list will be moved into thepipelineData
- When piping the event into the signature engine (if in the new binary) or in the sink stage (legacy mode), the
data
field will be swapped with theParsedArgs
.
CAVEAT: With the following logic, signatures supporting parsed arguments can not run alongside signatures not supporting parsed arguments. In practical terms that means the open source rego and golang signatures can not run at the same time.
@NDStrahilevitz parsing the args (meaning, translating from machine-readable into user-readable form) is a static operation no? it could even be done by the printer or the consumer. why do we need to include it in the pipeline?
@NDStrahilevitz parsing the args (meaning, translating from machine-readable into user-readable form) is a static operation no? it could even be done by the printer or the consumer. why do we need to include it in the pipeline?
Ideally that would be the case, and it was when tracee-ebpf was standalone. This allowed parsing arguments to be used as an implicit dependency by signature authors. If you take a look at all our open source signatures, you can see that they rely on the arguments parsed forms and not their originals. Furthermore, rego signatures make fixing this even harder, since while golang signatures can at least use the libbpfgo constants for readable logic, rego signatures don't have that option, making argument parsing practically required for them.
Following is a revised structure for Tracee event.
Changes:
- Removed metadata section
- Updated some fields names to match ECS/Opentelemetry schema
- threat field group was added under the data section in case of a threat detection event
timestamp
name
id
- machine readable id (integer). Note: current event id isn't good since it is architecture specific- //
version
- use semver where major is a breaking change in the event (e.g. one of the event's fields under data has been changed or removed), minor is a non breaking change (e.g. a new field was added to the event under data) and patch (e.g. a bug fix). Since this data is static, we may remove this or make optional - //
tags
- since this data is static, we may remove this or make optional labels
- doesn't exist. For future use.policies
matched
actions
- for future use (currently the only action we have is print)
context
(likeresource
in open telemetry, but structured - https://github.com/open-telemetry/opentelemetry-specification/blob/main/specification/logs/data-model.md#log-and-event-record-definition).process
(see also https://opentelemetry.io/docs/specs/semconv/resource/process/ and https://www.elastic.co/guide/en/ecs/8.9/ecs-process.html)executionTime
- time of last exec. Doesn't exist, consider adding (in another issue)executable
path
- binary path used to execute the process. Doesn't exist, consider adding (in another issue)name
pid
namespacePid
- should we name itvpid
as in the kernel?entityId
- unique id of the process, as defined at [process.entity_id](https://www.elastic.co/guide/en/ecs/8.9/ecs-process.html#field-process-entity-id)realUser
id
name
- doesn't exist, consider adding (in another issue)
user
- effective user. Doesn't exist, consider adding (in another issue)id
name
parent
- nested process ancestors. Only the following fields will be populated by default:pid
entityId
executable
start
we may want to include this filed only for direct parentparent
- nested process ancestors (won't be populated, unless required, e.g. on threat detection)
thread
start
name
(aka "comm")tid
namespaceTid
- should we name itvtid
(in the kernel this is called vpid)?capabilities
- doesn't exist, consider adding (in another issue)syscall
- moved herecompat
- boolean. moved fromflags.compat
userStackTrace
- if enabled, will be here
container
(see also https://opentelemetry.io/docs/specs/semconv/resource/container/)id
name
image
id
repoDigest
name
started
- boolean. moved fromflags
start
- Timestamp of container start time. Doesn’t exist. Will replacestarted
pid
- entrypoint's pid. Doesn’t exists, consider adding
k8s
(see also https://opentelemetry.io/docs/specs/semconv/resource/k8s/)pod
name
uid
labels
namespace
name
data
(likeattributes
in open telemetry, but structured - https://github.com/open-telemetry/opentelemetry-specification/blob/main/specification/logs/data-model.md#log-and-event-record-definition).- Any relevant field (per-event schema)
returnValue
(if relevant will appear here)triggerEvent
(will appear on thread detection events)name
id
data
threat
(if relevant will appear here) - static data about threats (can be omitted)description
mitreTactic
name
mitreTechnique
id
- the ID of the technique, as could be found in the [cti json]externalId
- the ID given in the mitre websitename
severity
This is what I think will match our needs for the threats
field:
threat
(if relevant will appear here) - (see also https://www.elastic.co/guide/en/ecs/8.9/ecs-threat.html)description
- detailed description of the threattactic
3.id
4.name
technique
6.id
- the ID of the technique, as could be found in the cti json
7.external_id
- the ID given in the mitre website
8.name
kubernetes_technique
- see Matrix ATT&CK for Kubernetes from Microsoft
10.name
indicator
(or another name) - a dictionary for threat related dataseverity
(note: in ecs, this field is not defined here but under event - but do we want it common to all events?)triggeringEvent
(can this be under enrichments? better name?)
14.name
15.id
16.data
I am also think that using the "related" field of the ECS can do good for keeping the IOCs in one place
Moved it to next release, the event structure is finished and merged, now it is missing integrating it on tracee internals.
Few comments:
- Threat is static information, and hence should be available outside of event (including grpc interface)
- The event header should also include the "text ID" of event (ART-X, TRC-X etc.)
- To be able to use numeric ID, consumer need to be able to use the definitions of ids without dependency on tracee ecosystem
3.1 The id definition must be managed in backward compatible manner ( i.e. only add, no edits, no deletes)
The data we need and is not in context currently:
- under process section missing full process commandline
- under k8s pod section missing deployment and type
- under process section missing full process commandline
This is available in the sched_process_exec
event and (I think) in the process tree and its data source.
This isn't to claim we shouldn't put it in the event context, rather that it is already available through alternative means.
- under k8s pod section missing deployment and type
@itaysk we should be able to add these through container labels like other kubernetes data we already get in container enrichment. I can open an issue for this if agreed.
This is available in the
sched_process_exec
event and (I think) in the process tree and its data source.
When process tree was being created there was a specific discussion about COMM versus BINARY PATH and where the info should go (https://github.com/aquasecurity/tracee/pull/3364/files#diff-773e2917cb050cc42ce31d36b08db6c9e3da89ab6dff75f8a9b3eba5171316d3R12).
Alon and I agreed that COMM would be part of the TaskInfo structure (for the Process or Thread) and the Binary Path would be part of the File Info structure (for the Binary and the Interpreter of the BInary).
The COMM (Process Name for the process tree) does not pick the args (its basically procfs comm field), that would come together with argv array (which is an argument for the exec event).
We can introduce COMM + ARGS in the process tree if needed (or something else), no problem.
Hey @mcherny, thank you for the questions:
Few comments:
- Threat is static information, and hence should be available outside of event (including grpc interface)
Do you have an example of how this will be used? Because the Threat information is part of specific even (behaviour events), I'm considering this should be a part of the event definition as optional, and not have its own API. Thoughts?
- The event header should also include the "text ID" of event (ART-X, TRC-X etc.)
We didn't add those because they are internal to Aqua, our plan was to actually remove all together for opensource signatures, and let the internal project handle the translation between those ids, and tracee ids.
- To be able to use numeric ID, consumer need to be able to use the definitions of ids without dependency on tracee ecosystem
3.1 The id definition must be managed in backward compatible manner ( i.e. only add, no edits, no deletes)
Agree! Right now the ids are stable for base events but dynamic for signatures, I need to look into how we can make it always stable.
- under process section missing full process commandline
This is available in the
sched_process_exec
event and (I think) in the process tree and its data source. This isn't to claim we shouldn't put it in the event context, rather that it is already available through alternative means.
Assuming we need this on each event that this may be relevant (e.x. file open), how would I consume by alternative means if I need it outside of tracee process, that is at gRPC client?
Assuming we need this on each event that this may be relevant (e.x. file open), how would I consume by alternative means if I need it outside of tracee process, that is at gRPC client?
In general, if someone needs info on something on each event of its kind, for a particular usecase, a signature/derived event is probably called for.
Few comments:
- Threat is static information, and hence should be available outside of event (including grpc interface)
That's why we wrote:
threat (if relevant will appear here) - static data about threats (can be omitted)
There is an advantage to include this information in the OSS project when a threat is detected (not so frequent) so user can get information about the threat without consulting the documentation
- The event header should also include the "text ID" of event (ART-X, TRC-X etc.)
In continuation to the first point, this is also static data and can be added to the threat
section by an internal mapping if required by some project
- To be able to use numeric ID, consumer need to be able to use the definitions of ids without dependency on tracee ecosystem
3.1 The id definition must be managed in backward compatible manner ( i.e. only add, no edits, no deletes)
Agree. We have an old issue opened for that #1098
- under process section missing full process commandline
This is available in the
sched_process_exec
event and (I think) in the process tree and its data source. This isn't to claim we shouldn't put it in the event context, rather that it is already available through alternative means.Assuming we need this on each event that this may be relevant (e.x. file open), how would I consume by alternative means if I need it outside of tracee process, that is at gRPC client?
Sounds like a reasonable addition to tracee that should be optional. We should discuss such additions in a separate issue and keep this issue for event structure to support the already existing features of tracee
I've been doing some work in the capture area. Considering we want to merge it into the "everything is an event" scheme at some point, shouldn't we also include an artifact
field in the structure (I recall @yanivagman making this sort of suggestion some time ago)? I assume that otherwise, each capture will be its own event.
I've been doing some work in the capture area. Considering we want to merge it into the "everything is an event" scheme at some point, shouldn't we also include an
artifact
field in the structure (I recall @yanivagman making this sort of suggestion some time ago)? I assume that otherwise, each capture will be its own event.
Eventually capture will be an action taken by some event. In that case, the data about it will be part of policies->actions
Following is the updated event schema based on the comments above:
timestamp
name
id
- machine readable id (integer). Note: current event id isn't good since it is architecture specific- //
version
- use semver where major is a breaking change in the event (e.g. one of the event's fields under data has been changed or removed), minor is a non breaking change (e.g. a new field was added to the event under data) and patch (e.g. a bug fix). Since this data is static, we may remove this or make optional - //
tags
- since this data is static, we may remove this or make optional labels
- doesn't exist. For future use.policies
matched
actions
- doesn't exist, for future use - list of actions taken (currently the only action we have is print).
workload
process
executable
path
name
- the binary name (basename of the path) - doesn't exist, consider adding (in another issue)
uniqueId
- unique id of the processpid
hostPid
executionTime
- time of last exec. Doesn't exist, consider adding (in another issue)realUser
id
name
- doesn't exist, consider adding (in another issue)
user
- effective user. Doesn't exist, consider adding (in another issue)id
name
ancestors
- process ancestors array. Only direct parent will be populated by default with the following fields:uniqueId
pid
hostPid
- Other ancestor fields may be populated by threat detection events
thread
startTime
name
(aka "comm")tid
hostTid
capabilities
- doesn't exist, consider adding (in another issue)syscall
- the syscall that triggered this eventcompat
- boolean. moved fromflags.compat
userStackTrace
- if enabled, will be here
container
id
name
image
id
repoDigest
name
isRunning
- boolean. moved fromflags
startTime
- Timestamp of container start time. Doesn’t exist. Will replacestarted
pid
- entrypoint's pid. Doesn’t exists, consider adding
k8s
pod
name
uid
labels
namespace
name
data
- Any relevant field (per-event schema)
returnValue
(if relevant will appear here)
threat
(if relevant will appear here) - static data about threats (can be omitted)description
mitre
tactic
name
technique
name
id
severity
triggeredBy
(will appear on threat detection events)
1.name
2.id
3.data
So we'll keep with triggeredBy outside of event data?