Einstein Analytics
Supported pipeline types:
|
When you configure the destination, you define connection information, including the API version that the destination uses to connect to Einstein Analytics.
You specify the edgemart alias or name of the dataset to upload data to. You can also optionally define the name of the edgemart container or app that contains the dataset.
The destination can upload external data to a new dataset or to an existing dataset using an append, delete, overwrite, or upsert operation. Based on the operation type, you define the metadata of the data to be uploaded in JSON format.
You can optionally use an HTTP proxy to connect to Salesforce Einstein Analytics. When enabled in Salesforce, you can configure the destination to use mutual authentication to connect to Salesforce.
Changing the API Version
Data Collector ships with version 43.0 of the Salesforce Web Services Connector libraries. You can use a different Salesforce API version if you need to access functionality not present in version 43.0.
Define the Operation
- Append - Appends data to the dataset, creating the dataset if it doesn’t exist.
- Delete - Deletes rows from the dataset. The rows to delete must contain a single field with a unique identifier.
- Overwrite - Replaces data in the dataset, creating the dataset if it doesn't exist.
- Upsert - Inserts or updates rows in the dataset, creating the dataset if it doesn’t exist. The rows to upsert must contain a single field with a unique identifier.
For more information about unique identifiers, see the Salesforce Developer documentation.
Metadata JSON
- Data file that contains the external data.
- Optional metadata file that describes the schema of the data in JSON format.
The Einstein Analytics destination creates the data file based on the incoming record. You define the metadata in JSON format when you configure the destination.
You must define metadata for the append, upsert, and delete operations. For append and upsert, the metadata must match the metadata of the dataset being uploaded to. For delete, the metadata must be a subset of the dataset columns.
You can optionally define metadata for the overwrite operation so that Einstein Analytics can correctly interpret the data type of the data. If you do not enter metadata, then Einstein Analytics treats every field as text.
For more information about how Einstein Analytics handles JSON metadata for uploaded external data, see the Salesforce Developer documentation.
Dataflow (Deprecated)
In previous releases, you could configure the destination to use an Einstein Analytics dataflow to combine multiple datasets together. However, using dataflows is now deprecated and will be removed in a future release. We recommend configuring the destination to use the append operation to combine data into a single dataset.
An Einstein Analytics dataflow includes instructions and transformations to combine datasets. Create the dataflow in Einstein Analytics. Then when you configure the Einstein Analytics destination, specify the name of the existing dataflow. The dataflow should not contain any content, as the Einstein Analytics destination overwrites any existing content.
By default, a dataflow runs every 24 hours. However, you can configure the dataflow to run each time the destination closes and uploads a dataset to Einstein Analytics. In Einstein Analytics, you can run a dataflow a maximum of 24 times in a 24 hour period. So if you choose to run the dataflow after each dataset upload, make sure that the configured dataset wait time is more than an hour.
For more information about creating dataflows, see the Salesforce Einstein Analytics documentation.