Pull from Optimizely

Optimizely is an experimentation platform for testing, learning, and deploying positive digital experiences.

Optimizely can send enriched exports events data to Amperity via Amazon S3. Enriched events include details such as event timestamps, event IDs, event tags, event names, visitor IDs, session IDs, experiment IDs, and variation IDs.

This topic describes the steps that are required to pull interactions records to Amperity from Optimizely:

  1. Get details

  2. Review Amazon S3 requirements for Optimizely

  3. Add data source and feed

Get details

Amperity can be configured to pull data from Optimizely using Amazon S3. This requires the following configuration details:

  1. Access to an Optimizely data service hosted in Amazon S3.

  2. The Amazon Resource Name (ARN) for a role with cross-account access.

  3. The name of the Amazon S3 bucket.

  4. A list of objects (by filename and file type) in the Amazon S3 bucket to be pulled to Amperity.

  5. A sample for each file to simplify feed creation.

Note

Amperity supports using cross-account role assumption with Amazon S3 buckets when Optimizely supports the use of cross-account roles and your tenant uses the Amazon S3 data source.

Amazon S3 requirements

Amazon S3 requires the following:

Amazon S3 credentials

Amperity requires the ability to connect to, and then read data from the Amazon S3 bucket used by Optimizely. The credentials that allow that connection and the permissions to read data are entered into the Amperity user interface while configuring a courier. These credentials are created and managed by the owner of the Amazon S3 bucket. Use SnapPass to share credentials with your Amperity representative, if necessary.

SnapPass allows sharing secrets in a secure, ephemeral way. Input a single or multi-line secret, define how long someone has to view the secret, and then generate a single-use URL. Share the URL to share the secret.

Optimizely S3 partitions

Enriched events exports are exported to a bucket named optimizely-events-data that contains two partitions: decisions and conversions.

The paths to these partitions are similar to:

s3://optimizely-events-data/v1/account_id=<account_id>/
type=decisions/date={YYYY-MM-DD}/experiment=<experiment_id>

or

s3://optimizely-events-data/v1/account_id=<account_id>/
type=events/date={YYYY-MM-DD}/event=<event_name>

where:

  • optimizely-events-data is the name of the Amazon S3 bucket

  • account_id is your unique account identifier

  • date is the creation date for the data

  • experiment_id is the unique experiment identifier used for the decisions partition

  • event_name is the event or entity identifier used for the events partition

The daily partition files are ready when _SUCCESS is appended to the partition path.

Note

Optimizely uses AWS Key Management Service for encryption. Amperity must be able to decrypt these files to pull them to the Amazon S3 or Azure Blob Storage location used by your tenant.

Add data source and feed

Add a data source that pulls data from Optimizely.

Configure Amperity to pull one or more files, and then for each file review the settings, define the schema, activate the courier, and then run a manual workflow. Review the data that is added to the domain table.

To add a data source for Optimizely

Step 1.

Open the Sources page to configure Optimizely.

Click the Add courier button to open the Add courier dialog box.

Select Optimizely. Do one of the following:

  1. Click the row in which Optimizely is located. Sources are listed alphabetically.

  2. Search for Optimizely. Start typing “ama”. The list will filter to show only matching sources.

Step 2.

Credentials allow Amperity to connect to Optimizely and must exist before a courier can be configured to pull data from Optimizely. Select an existing credential from the Credential dropdown, and then click Continue.

Tip

A courier that has credentials that are configured correctly will show a “Connection successful” status, similar to:

Add
Step 3.

Select the file that will be pulled to Amperity, either directly (by going into the SFTP site and selecting it) or by providing a filename pattern.

Add

Click Browse to open the File browser. Select the file that will be pulled to Amperity, and then click Accept.

Use a filename pattern to define files that will be loaded on a recurring basis, but will have small changes to the filename over time, such as having a datestamp appended to the filename.

Note

For a new feed, this file is also used as the sample file that is used to define the schema. For an existing feed, this file must match the schema that has already been defined.

Add

Use the PGP credential setting to specify the credentials to use for an encrypted file.

Add
Step 4.

Review the file.

Add

The contents of the file may be viewed as a table and in the raw format. Switch between these views using the Table and Raw buttons, and then click Refresh to view the file in that format.

Note

PGP encrypted files can be previewed. Apache Parquet PGP encrypted files must be less than 500 MB to be previewed.

Amperity will infer formatting details, and then adds these details to a series of settings located along the left side of the file view. File settings include:

  • Delimiter

  • Compression

  • Escape character

  • Quote character

  • Header row

Review the file, and then update these settings, if necessary.

Step 5.

A feed defines the schema for a file that is loaded to Amperity, after which that data is loaded into a domain table and ready for use with workflows within Amperity.

New feed

To use a new feed, choose the Create new feed option, select an existing source from the Source dropdown or type the name of a new data source, and then enter the name of the feed.

Add

After you choose a load type and save the courier configuration, you will configure the feed using the data within the sample file.

Existing feed

To use an existing feed, choose the Use existing feed option to use an existing schema.

Add

This option requires this file to match all of the feed-specific settings, such as incoming field names, field types, and primary keys. The data within the file may be different

Pull data

Define how Amperity will pull data from Optimizely and how it is loaded to a domain table.

Add

Use the Upsert option to use the selected file update existing records and insert records that do not exist.

Use the Truncate and upsert option to delete all records in the existing table, and then insert records.

Note

When a file is loaded to a domain table using an existing file, the file that is loaded must have the same schema as the existing feed. The data in the file may be new.

Step 6.

Use the feed editor to do all of the following:

  • Set the primary key

  • Choose the field that best presents when the data in the table was last updated; if there is not an obvious choice, use the “Generate an updated field” option.

  • For each field in the incoming data, validate the field name and semantic tag columns in the feed. Make any necessary adjustments.

  • For tables that contain customer records, enable the “Make available to Stitch” to ensure the values in this data source are used for identity resolution.

When finished, click Activate.

Step 7.

Find the courier related to the feed that was just activated, and then run it manually.

On the Sources page, under Couriers, find the courier you want to run and then select Run from the actions menu.

Add

Select a date from the calendar picker that is before today, but after the date on which the file was added to the Optimizely bucket.

Add

Leave the load options in the Run courier dialog box unselected, and then click Run.

After the courier has run successfully, inspect the domain table that contains the data that was loaded to Amperity. After you have verified that the data is correct, you may do any of the following:

  • If the data contains customer records, edit the feed and make that data available to Stitch.

  • If the data should be loaded to Amperity on a regular basis, add the courier to a courier group that runs on the desired schedule.

  • If the data will be a foundation for custom domain tables, use Spark SQL to build out that customization.