Table of Contents generated with DocToc
- Before getting started...
- Configuration and credential management
- Sync = Get, Transform, Load, Cleanup
- Error handling principles
- Schema
- Making tests
- Other
This doc contains guidelines on creating a Cartography intel module. If you want to add a new data type to Cartography, this is the guide for you! It is fairly straightforward to copy the structure of an existing intel module and test it, but we'll share some best practices in this doc to save you some time. We look forward to receiving your PR!
Read through and follow the setup steps in the Cartography developer guide. Learn the basics of running, testing, and linting your code there.
If you need to supply an API key or other credential to your Cartography module, we recommend adding a CLI argument. An example of this can be seen in our Okta module where we require the user to specify the name of an environment variable containing their Okta API key. This credential will then be bound to Cartography's Config object which is present in all modules. You can specify different arguments from the commandline for your module via the Config object.
Note that it is your module's responsibility to validate arguments that you introduce. For example with the Okta module, we validate that config.okta_api_key has been defined before attempting to continue.
A cartography intel module consists of one sync function. sync should call get, then load, and finally cleanup.
The get function retrieves necessary data
from a resource provider API, which is GCP in this particular example.
get should be "dumb" in the sense that it should not handle retry logic or data
manipulation. It should also raise an exception if it's not able to complete successfully.
The transform function manipulates data
to make it easier to ingest to the graph. We have some best practices on handling transforms:
We should directly access dicts in cases where not having the data should cause a sync to fail.
For example, if we are transforming AWS data, we definitely need an AWS object's ARN field because it uniquely
identifies the object. Therefore, we should access an object's ARN using data['arn'] as opposed to
using data.get('arn') (the former will raise a KeyError if arn does not exist and the latter will just return
None without an exception).
We want the sync to fail if an important field is not present in our data. The idea here is that it is better to fail a sync than to add malformed data.
On the other hand, we should use data.get('SomeField') if SomeField is something optional that can afford to be
None.
For the sake of consistency, if a field does not exist, set it to None and not "".
The load function ingests the processed data to Neo4j, as seen in this GCP VPC example.
There are many best practices to consider here.
cartography's global config object carries around an update_tag property
which is a tag/label associated with the current sync. Cartography's CLI code sets this to a Unix timestamp of when the CLI was run,
but in theory could be anything provided that it is unique per sync.
All cartography intel modules need to set the lastupdated property on all nodes and all relationships to this
update_tag. You can see a couple examples of this in our
AWS ingestion code and our
GCP ingestion code.
-
id- an ID should be a string that uniquely identifies the node. In AWS, this is usually an ARN. In GCP, this is usually a partial URI.If possible, we should use API-provided fields for IDs and not create our own. In some cases though this is unavoidable - see GCPNetworkTag.
When setting an
id, ensure that you also include the field name that it came from. For example, since we've decided to usepartial_uris as an id for a GCPVpc, we should include bothpartial_uriandidon the node. This way, a user can tell what fields were used to derive theid. This is accomplished here. -
lastupdated- See below on how to set this. -
firstseen- See below on how to set this.
Cartography currently does not create indexes on relationships, so we should keep relationships lightweight with only these two fields:
In this example of ingesting GCP VPCs, we connect VPCs with GCPProjects
based on GCPProject ids and GCPVpc ids.
ids are indexed, as seen here
and here.
All of these queries use indexes for faster lookup.
Be sure to update the indexes.cypher file with your new node type. Indexing on ID is required, and indexing on anything else that will be frequently queried is encouraged.
Set the lastupdated and firstseen fields on both nodes and relationships. Suppose we are creating
the following chain:
MERGE (n:NodeType)-[r:RELATIONSHIP]->(n2:NodeType2)-
To handle nodes in this case,
-
Every
MERGEquery that creates a new node should look like thisON CREATE SET n.firstseen = {UpdateTag} SET n.lastupdated = {UpdateTag}, node.field1 = {value1}, node.field2 = {value2}, ... node.fieldN = {valueN}
-
-
To handle relationships in this case,
-
Every
MERGEquery that creates a new relationship should look like thisON CREATE SET r.firstseen = {UpdateTag} SET r.lastupdated = {UpdateTag}
-
Node connections can be complex. In many cases we need to connect many different node types together, so we use an
_attach function to manage this.
The best way to explain _attach is through an example, like when we connect GCP instances to their VPCs.
In this case, we create a helper _attach function
that accepts the instance's id and connects the instance to the VPC using a MERGE query.
This pattern can also be seen when attaching AWS RDS instances to EC2 security groups.
We have just added new nodes and relationships to the graph, and we have also updated previously-added ones
by using MERGE. We now need to delete nodes and relationships that no longer exist, and we do this by simply removing
all nodes and relationships that have lastupdated NOT set to the update_tag of this current run.
-
Delete all old nodes
You can see this in our GCP VPCs example. We run
DETACH DELETEto delete an old node and disconnect it from all other nodes. -
Delete all old relationships
You can see this in the GCP VPC example here and here.
-
Q: We just
DETACH DELETE'd the node. Why do we need to delete the relationships too? -
A: There are cases where the node may continue to exist but the relationships between it and other nodes have changed. Explicitly deleting stale relationships accounts for this case. See this short discussion.
-
-
Don't catch the base Exception class when error handling because it makes problems difficult to trace.
-
Do catch the narrowest possible class of exception.
-
Only catch exceptions when your code can resolve the issue. Otherwise, allow exceptions to bubble up.
- Update the schema with every change!
-
Before making tests, read through and follow the setup steps in the Cartography developer guide.
-
Add fake data for testing at
tests/data. We can see the GCP VPC example here: https://github.com/lyft/cartography/blob/0652c2b6dede589e805156925353bffc72da6c2b/tests/data/gcp/compute.py#L2. -
Add unit tests to
tests/unit/cartography/intel. See this example. These tests ensure thattransform*manipulates the data in expected ways. -
Add integration tests to
tests/integration/cartography/intel. See this example. These tests assume that you have neo4j running at localhost:7687 with no password, and ensure that nodes loaded to the graph match your mock data.
-
We prefer and will accept PRs which incrementally add information from a particular data source. Incomplete representations are OK provided they are consistent over time. For example, we don't sync 100% of AWS resources but the resources that exist in the graph don't change across syncs.
-
Each intel module offers its own view of the graph
ℹ️ This best practice is a little more less precise, so if you've gotten to this point and you need clarification, just submit your PR and ask us.
As much as possible, each intel module should ingest data without assuming that a different module will ingest the same data. Explained another way, each module should "offer its own perspective" on the data. We believe doing this gives us a more complete graph. Below are some key guidelines clarifying and justifying this design choice.
-
Use
MERGEwhen connecting one node type to another node type. -
It is possible (and encouraged) for more than one intel module to modify the same node type.
For example, when we connect RDS instances to their associated EC2 security groups there are actually two different intel modules that retrieve EC2 security group data: the RDS module returns partial group data, and the EC2 module returns more complete data as it calls APIs specific for retrieving and loading security groups. Because both the RDS and EC2 modules
MERGEon a unique ID, we don't need to worry about creating duplicate nodes in the graph.Another less obvious benefit of using
MERGEacross more than one intel module to connect nodes in this way is that in many cases, we've seen an intel module discover nodes that another module was not aware of!
-