Outgoing feed - Amazon S3 push#
Note
This article describes how to configure outgoing feeds for a particular feed source. To see how to configure outgoing feeds in general, see Create and configure outgoing feeds.
Specifications |
|
---|---|
Transport type |
Amazon S3 push |
Content type |
For more information, see Configure content types. |
Description |
The feed publishes entities and observables in the selected content type to the specified destination location on the designated Amazon S3 bucket. Each time the outgoing feed task runs, it generates a data package containing zero or more entities, depending on the outgoing feed update strategy, and on the feed data source containing data that match the feed configuration. |
Configure the outgoing feed#
Create or edit an outgoing feed.
From the Transport type drop-down menu, select Amazon S3 push.
From the Content type drop-down menu, select the appropriate content type for the data you want to publish through the outgoing feed.
The selected content type for the feed should match the actual format of the source data you are publishing through the feed.
This can vary, depending on the dataset sources you retrieve the data for publication from.
The selected content type for the feed should match the source data format.
This can vary, depending on the dataset source(s) you retrieve the data from.From the Datasets drop-down menu, select one or more existing datasets to use as sources to populate the outgoing feed content.
For the feed not to be empty, at least one selected dataset should contain entities and observables in the same format as the configured content type for the feed.From the Update strategy drop-down menu, select the preferred method to populate the outgoing feed with data before publishing it.
Update strategies help define how content is aggregated and packaged for publication when an outgoing feed task runs:Append: every time the outgoing feed task runs, it fetches only new, unpublished data to generate the content for publication.
The published packages contain only new entities and observables ingested in the platform after the previous execution of the outgoing feed.Replace: every time the outgoing feed task runs, it fetches new and existing data to generate the content for publication.
The published packages contain new and existing entities and observables included also in the previous execution of the outgoing feed.Diff: this option is available only for the EclecticIQ Entities CSVand EclecticIQ Observables CSV content types. Every time the outgoing feed task runs, new data is compared against existing data to identify any differences between the two datasets:
At entity level: any entities added to or removed from the set, if EclecticIQ Entities CSV is the designated content type for the feed.
At observable level: any observable added to or removed from the entities in the set, if EclecticIQ Observables CSV is the designated content type for the feed.
Depending on the selected CSV content option, each row in the CSV output contains information about one entity being added or removed, or one observable being added or removed.
An extra diff column is added to the output CSV to indicate if a row, and therefore either an entity or an observable, has been added to or removed from the set.
This option enables identifying changes in a feed between two executions without downloading the whole feed every time.Note
Update strategies help define how content is aggregated and packaged for publication when an outgoing feed task runs:
Update strategies rely on the
last_updated_at
database field to identify entities whosetimestamp
value was updated since the previous execution of the outgoing feed.
Entities with a more recenttimestamp
value compared to the previous execution of the outgoing feed are packaged and included in the published content of the outgoing feed.Changes to the
data
section of an entity create a new version of the entity.
They also add a new log entry to the entity history to record the changes.Changes to the
meta
section of an entity do not create a new version of the entity.
However, they do update the timestamp value of thelast_update_at
database field.
In the Secret access key field, enter your Amazon Web Services secret access key.
Sign up to Amazon Web Services, and then create one or more accounts, as necessary, to use their S3 data storage service.
The secret key is part of your authentication credentials to log in to and to access Amazon S3 services.In the Access key field, enter your access key.
Along with your secret key, the access key enables you to authenticate to access Amazon S3 services.In the Bucket field, enter the name of the Amazon S3 bucket to use as a target location for the outgoing feed published content.
Buckets are data containers in the S3 environment.
Buckets are region-specific, and their names must comply with standard DNS naming conventions.
The default format of the URL to access a bucket ishttps://${bucket_name}.s3-${aws-region}.amazonaws.com
.In the Path field, enter the path to the target directory where the content published through the outgoing feed is stored, relative to the bucket root.
Example:/intel/actors/hacktivists
.To store your changes, click Save; to discard them, click Cancel.
View and retrieve outgoing feed content#
In the top navigation bar click Data configuration > Outgoing feeds.
In the Outgoing feeds view, click anywhere in the row corresponding to the outgoing feed whose content you want to view or retrieve.
In the selected outgoing feed detail pane, click the Created packages tab.
In the Created packages tab, under the Download column header, click the name of a package to download it, and to save it to a target location.