Skip to main content

Running LiveRamp’s Local Encoder In a Docker-Run Environment

Local Encoder enables you to generate securely-encoded RampIDs for your consumer data files within your own cloud environment and then utilize that data for onboarding or addressability use cases, depending on your needs. In this way your consumer data is never exposed to an external network, while still enabling full use of the LiveRamp solutions. The encoded RampIDs produced by the application cannot be decoded back to the original consumer identifiers.

Local Encoder can be run on any infrastructure that supports running Docker images. The Local Encoder Docker image is currently distributed via the Amazon Elastic Container Registry (ECR).

For more information on Local Encoder, including information on security, use cases, data you can send, and output options, see "LiveRamp Local Encoder".

For information on running Local Encoder using Docker run, see the sections below.

Overall Steps

Running the Local Encoder in a Docker-run environment involves the following overall steps:

  1. You provide LiveRamp with your PGP public key or Keybase username.

  2. LiveRamp provides you with credentials.

  3. You decrypt the AWS IAM Secret Access Key.

  4. You use the credentials to configure AWS to access LiveRamp’s Docker image in ECR.

  5. You use Docker to access ECR and pull the Docker image for Local Encoder.

  6. You test the installation.

  7. You complete other configuration activities, as necessary:

    • If you want to save processed file names permanently, you configure how file names are stored.

    • If you want to use a different input or output location than the defaults, you mount folders.

    • You create and mount a configuration file that contains information used by the run command.

  8. You run the Local Encoder.

  9. You format your data files and then upload them to the appropriate input location.

  10. You run the encoding operation within your VPC:

    1. The data is normalized and hygiene is performed.

    2. The identifiers in the data are converted into derived RampIDs.

    3. If appropriate, the derived RampIDs for each record are encoded into secure RampID packets or identity envelopes.

    4. The input identifiers are removed and replaced with the appropriate RampID output type (RampIDs, RampID packets, or identity envelopes).

  11. For Onboarding use cases the following steps are performed:

    1. The output containing RampID packets is delivered to LiveRamp.

    2. LiveRamp decodes the RampID packets into their individual derived RampIDs.

    3. LiveRamp matches those derived RampIDs to their associated maintained RampIDs.

    4. LiveRamp creates the appropriate fields and segments from the segment data in your LiveRamp platform (such as Connect or Safe Haven).

  12. For Addressability use cases the following steps are performed:

    1. The output containing RampIDs or identity envelopes is output to the destination of your choice.

    2. You leverage the output to build a mapping table of your customer IDs to RampIDs or identity envelopes.

Prerequisites

Running the Local Encoder in a Docker-run environment requires that you have the following prerequisites:

Note

You will use the AWS command line tool to issue commands at your system's command line to perform Amazon ECR and other AWS tasks. We recommend that you have the latest version of the AWS CLI installed. For information about installing the AWS CLI or upgrading it to the latest version, see Installing the AWS Command Line Interface.

Decrypt the AWS IAM Secret Access Key

In addition to providing the Local Encoder account ID and the AWS IAM Access Key ID, LiveRamp will provide the AWS IAM Secret Access Key in an encrypted format. The AWS IAM Secret Access Key must be decrypted for use in subsequent steps.

To decrypt the AWS IAM Secret Access Key:

  1. Export your public and private key to files, name them “public.key” and “private.key”, and save them in the “/tmp” directory.

  2. Save the encrypted AWS IAM Secret Access Key to a file named “secret.txt” and save it in the “/tmp” directory.

    Note

    If the “/tmp” directory has not already been created, you will need to create it.

  3. From Docker, run the following command:

    docker run -it --rm alpine apk add --no-cache wget gnupg && export GPG_TTY=$(tty) && gpg --import /tmp/public.key && gpg --import /tmp/private.key && base64 -d -i /tmp/secret.txt | gpg --decrypt > /tmp/output.txt
  4. Open the “output.txt” file to access the decrypted AWS IAM Secret Access Key.

Configure AWS to Access LiveRamp’s Docker Image

To configure AWS to be able to access LiveRamp’s Local Encoder Docker image in ECR, set your AWS IAM access key and secret access key in one of the following methods.

Run AWS Configure

From the command line, run the command “aws configure” with the following values:

AWS Access Key Id: [LR_VAULT_LR_AWS_ACCESS_KEY_ID]
AWS Secret Access Key: [LR_VAULT_LR_AWS_SECRET_ACCESS_KEY]
Default region name: eu-central-1
Default output format: json

Edit Files Manually

To set your keys by editing the files manually (MAC OS example):

  1. From Finder, navigate to the Home directory.

    Local_Encoder-Docker_Run-Finder_Home_Directory.png
  2. Press (command + shift + .) to reveal the hidden folders within the Home directory Note: If you don’t see these files, you might not have the latest version of AWS CLI installed. Follow the instructions in this Amazon article to install the latest version.

  3. From the .aws folder, open the “config” and “credentials” files and then enter your aws_access_key_id and aws_secret_access_key (2.2) values as shown below:

    Local_Encoder-Docker_Run-key_values.png

Pull the Docker Image

The Docker image you need to pull from ECR is “461694764112.dkr.ecr.eu-central-1.amazonaws.com/vault-app:latest”.

Once you’ve configured AWS to access LiveRamp’s Local Encoder Docker image in ECR, perform the following steps to pull the Docker image:

  1. Make sure the Docker Engine is running

  2. Execute the following commands in the CLI:

    aws ecr get-login-password --region eu-central-1 | docker login --username AWS --password-stdin 461694764112.dkr.ecr.eu-central-1.amazonaws.com

    Note

    Make sure to run the command exactly as shown.

    The command output should declare “Login Succeeded”.

    Note

    If you get an error code, check that you have configured AWS correctly.

  3. Run the following command to retrieve the image:

    docker pull 461694764112.dkr.ecr.eu-central-1.amazonaws.com/vault-app:latest

    The docker image should be available on your local machine or VM to use.

  4. To confirm access to the image, run the following command and make sure that the image shows up in the list of images:

    docker images

Test the Installation

To confirm that the installation has been done properly, we recommend that you test the installation. The Local Encoder offers a test mode, which can let you run the app without the data being sent to LiveRamp. To run the app in test mode, run the following command in Docker:

docker run --cap-add IPC_LOCK \
--env LR_VAULT_ACCOUNT_ID=[test] \
--env LR_VAULT_ACCOUNT_TYPE=awsiam \
--env LR_VAULT_DRY_RUN=true \
--env LR_VAULT_INPUT=/tmp/input \
--env LR_VAULT_OUTPUT=/tmp/output \
--env LR_VAULT_LOCALE=au \
--name vault-app vault-app
  • LR_VAULT_ACCOUNT_ID: You can use any value for dry run mode

  • LR_VAULT_ACCOUNT_TYPE: You can use either “awsiam” or ”aws”

  • LR_VAULT_DRY_RUN: Use “true” to enable test mode

  • LR_VAULT_INPUT: The location of the local folder containing input files to process

  • LR_VAULT_OUTPUT: The location of the local folder for output files

  • LR_VAULT_LOCALE: The two-digit country code

If you would like a file with test data, contact your LiveRamp representative. The output of that test file can be run against a truth set to ensure that the application has run as expected.

Complete Configuration Activities

See the sections below for information on completing any desired additional configuration activities. For more information on configuration parameters, see the "Configuration Parameters" section below.

Configure the Output Type

The Local Encoder application offers multiple output formats:

  • RampID packets: Used for Onboarding by brands and marketers, these RampID packets can be delivered to LiveRamp, where they can be transformed into RampIDs and used to generate fields and segments in your LiveRamp application. This is the default output type.

  • RampIDs: Used for addressability by publishers and platforms who want to create a RampID mapping.

  • Identity envelopes: Used for addressability by publishers and platforms who want to get RampIDs into the bidstream to safely engage with the programmatic advertising ecosystem.

Note

  • If you plan to receive RampIDs or identity envelopes, contact your LiveRamp representative for approval and account configuration changes.

  • Identity envelope output is only available with Version 1.6 of Local Encoder.

After you are approved for RampID or identity envelope output, update your configuration to add an additional variable using one of the methods listed below:

Note

To receive RampID packets, you do not need to make any changes to the configuration.

  • Environmental Variable (when running the Local Encoder):

    • For RampID output, add the following line to the Environmental Variable:

      --env LR_VAULT_PACKET_TYPE: unencoded
    • For identity envelope output, add the following line to the Environmental Variable:

      --env LR_VAULT_ENVELOPES_FLOW=true
  • Config File (when implementing a configuration file):

    • For RampID output, add the following line to the Config File:

      packet_type: unencoded
    • For identity envelope output, add the following line to the Config File:

      envelopes_flow: true

Note

When editing the yml file, any formatting issues will prevent the configuration file from working properly. We recommend that you run the file through a YAML validator.

Utilize Encryption

Optional encryption is available with Local Encoder (version 1.5 and greater). This functionality encrypts each row of data before it is sent to LiveRamp for processing.

Note

Adding encryption increases the processing time approximately 20% depending on the size of the file and number of records. LiveRamp recommends limiting file size to 15GB.

To utilize encryption, use one of the following methods, depending on your implementation:

  • Add the following variable to the code sample, just before the name variable: --env LR_VAULT_PUBLIC_KEY_ENCRYPTION=true (for more information, see the "Run the Local Encoder" section)

  • Add the following line to the configuration file, just before the country_code_column parameter: public_key_encryption: true (for more information, see the "Implement a Configuration File" section)

Enable the Error Log File

You can configure Local Encoder to return an error log to your desired location that provides details on any errors that were encountered during processing, including the row number for the row where the error occurred. You can then correct the errors and re-process the corrected data.

The error log includes the following information (see the example below):

I-Local_Encoder-error_log_file_example.png
  • The date and time of the error

  • The type of error

  • The row number where the error occurred

  • The name of the file

To enable the error log file and set the file delivery location, see the sections below, depending on the method used.

The Local Configuration Method

When using the local configuration method, include the following parameters to enable the error log file and set the delivery location for the file:

  • “-v <localFolder>:/var/logs/vault-app”

  • “customer_logging_enabled=true”

See below for an example:

docker run --rm --cap-add IPC_LOCK \
-v ~/configurationFileFolder:/config \
-v <localFolder>:/var/logs/vault-app \
vault-app \
-config-file test-configuration.yml

The Command Line or configmaps.yaml Method

When using the command line or configmaps.yaml method, include the following lines to enable the error log file and set the delivery location for the file: :

  • “-v <localFolder>:/var/logs/vault-app”

  • “--env LR_VAULT_CUSTOMER_LOGGING_ENABLED=true”

See below for an example that sets the delivery location to “:/Documents/localencoder/log”:

docker run --rm --cap-add IPC_LOCK \
-v ~/Documents/localencoder/input:/tmp/input \
-v ~/Documents/localencoder/output:/tmp/output \
-v ~/Documents/localencoder/log:/var/logs/vault-app \
--env LR_VAULT_ACCOUNT_ID=[LR_VAULT_ACCOUNT_ID] \
--env LR_VAULT_ACCOUNT_TYPE=[LR_VAULT_ACCOUNT_TYPE] \
--env LR_VAULT_LR_AWS_ACCESS_KEY_ID=[AWS_ACCESS_KEY_ID] \
--env LR_VAULT_LR_AWS_SECRET_ACCESS_KEY=[AWS_SECRET_ACCESS_KEY] \
--env LR_VAULT_INPUT=/tmp/input \
--env LR_VAULT_OUTPUT=/tmp/output \
--env LR_VAULT_LOCALE=eu-central-1 \
--env LR_VAULT_CUSTOMER_LOGGING_ENABLED=true
--name vault-app [Image Name]

Store File Names

To save processed files names permanently (after application shutdown), there are two options:

Save to the H2 Database

To save file names to the H2 database connected to the local file (default), provide the following environment variables (either in the command line or in the configuration file):

- LR_VAULT_METASTORE_DB_URL - datasource url, local file example: *h2:~/db/LOCAL_FILE_NAME*
- LR_VAULT_METASTORE_DB_DRIVER - driver class name, for H2: *org.h2.Driver*
- LR_VAULT_METASTORE_DB_USERNAME - username for H2 database
- LR_VAULT_METASTORE_DB_PASSWORD - password for H2 database
- LR_VAULT_METASTORE_DB_PLATFORM - SQL dialect of the underlying database, for H2: *org.hibernate.dialect.H2Dialect*

When using H2 for our database, you have to mount a folder on the host machine in order to permanently store the database file after the docker container is shut down.

docker run --rm --cap-add IPC_LOCK \
-v ~/configurationFileFolder:/config \
-v ~/localencoderdb:/PATH_TO_DB vault-app \
-config-file [test-configuration].yml

In this example, we are mounting the "localencoderdb" folder from the host machine root directory to "/PATH_TO_DB" directory inside the LocalEncoder docker container. PATH_TO_DB is the path to the database folder we provided in the H2 LR_VAULT_METASTORE_DB_URL setting. e.g. "h2:~/db/vault-app-metadatastore" in this case PATH_TO_DB would be "db" and this is also the default case for H2. The folder name on the host machine is optional, but it always has to be mounted to "/PATH_TO_DB".

Connect to an Existing Database

To connect to an existing database, provide the following environment variables:

- LR_VAULT_METASTORE_DB_URL - database url, local MySQL database example: *mysql://host.docker.internal:3306/DATABASE_NAME*; local PostgreSQL database example *postgresql://host.docker.internal:5432/DATABASE_NAME*
- LR_VAULT_METASTORE_DB_DRIVER - driver class name, for MySQL: *com.mysql.cj.jdbc.Driver*; for PostgreSQL: *org.postgresql.Driver*
- LR_VAULT_METASTORE_DB_USERNAME - username for database
- LR_VAULT_METASTORE_DB_PASSWORD - password for database
- LR_VAULT_METASTORE_DB_PALTFORM - SQL dialect of the underlying database, for MySQL: *org.hibernate.dialect.MySQL5InnoDBDialect*; for PostgreSQL: *org.hibernate.dialect.PostgreSQLDialect*

Use Local Input and Output Sources

When using local folders for input and output sources, you have to mount the folder from our host machine:

docker run --rm --cap-add IPC_LOCK \
-v ~/configurationFileFolder:/config \
-v ~/inputFolderOnHostMachine:/LR_VAULT_INPUT \
-v ~/outputFolderOnHostMachine:/LR_VAULT_OUTPUT
vault-app \
-config-file test-configuration.yml

In this example, we are mounting input and output folders from the host machine root directory to LR_VAULT_INPUT and LR_VAULT_OUTPUT directories inside LocalEncoder docker containers. LR_VAULT_INPUT and LR_VAULT_OUTPUT are local input and output values that we provided either in a configuration file or by passing env var LR_VAULT_INPUT and LR_VAULT_OUTPUT. Folder names on the host machine are optional, they always have to be mounted to LR_VAULT_INPUT and LR_VAULT_OUTPUT.

By default, the application will output to the LiveRamp S3. If you would like to point output to a different location, set the LR_Local Encoder_OUTPUT variable:

The application outputs the data to a destination of your choice. Use the LR_VAULT_OUTPUT variable to set your destination, as shown below.  To onboard data to LiveRamp, set the output to the LiveRamp S3 bucket.

--env LR_VAULT_OUTPUT=[Your local directory, or S3 bucket]

Implement a Configuration File

A configuration file contains information that is used by the run command and allows for the use of run commands that are more streamlined.

When using a configuration file, you need to mount the folder of the configuration file on the host machine to the "/config" folder inside the docker container.

docker run --rm --cap-add IPC_LOCK \
-v ~/configurationFileFolder:/config \
vault-app \
-config-file test-configuration.yml

In this example, the "test-configuration.yml" file is located in "~/configurationFileFolder" on the host machine, and we are mounting it to the "/config" folder inside the docker container. Then passing "-config-file test-configuration.yml" as a command line argument to LocalEncoder app. Folder name on the host machine is optional, it always has to be mounted to "/config".

In order to use the configuration file, you need to mount a directory where the file is located to /config directory. Example: -v ~/Documents/localencoder:/config - the YAML config file should be located in the Local encoder folder. Then pass the file name as a command line argument. Docker Run example:

docker run --rm --cap-add IPC_LOCK \
-v ~/configurationFileFolder:/config \
[docker_image] \
-config-file [your_filename].yml

Create a yaml configuration file that follows this example:

Note

Any formatting issues will prevent the configuration file from working properly. We recommend that you run the file through a YAML validator.

liveramp:
  local_encoder:
    account_type: awsiam
    account_id: [LR_VAULT_ACCOUNT_ID]
    profile: PROD
    locale: [COUNTRY_CODE]
    output: [your_directory]
    input: [your_directory]
    filename_pattern: pattern
    header_mapping: custom_header=email1,custom_header2=email2
    country_code_column: [COUNTRY_CODE]
    public_key_encryption: true
    mode: default
    dry_run: false
    lr_aws:
      lr_id: 461694764112
      lr_access_key_id: [AWS_ACCESS_KEY_ID]
      lr_secret_access_key: [AWS_SECRET_ACCESS_KEY]
      lr_region: eu-central-1
    metastore:
      db:
        url:
        driver:
        username: 
        password: 
        platform: 

Include the following information in the file parameters:

  • account_id: Use the Local Encoder account ID you received from LiveRamp in place of the “[LR_VAULT_ACCOUNT_ID]” variable.

  • locale: Use the two-digit country code for the country the data originated in.

  • lr_access_key_id: Use the AWS Access Key ID you received from LiveRamp in place of the “[AWS_ACCESS_KEY_ID]” variable.

  • lr_secret_access_key: Use the AWS Secret Access Key you received from LiveRamp and decrypted in place of the “[AWS_SECRET_ACCESS_KEY]” variable.

  • Header_mapping: Include this line if you are using headers for the identifier columns that are different from the standard headers. Map each custom header name to the corresponding standard header name in the format shown, separating each mapping with a comma. If you’re using the standard headers, you can leave this line out.

  • country_code_column: Include this line if you’ll be providing data from multiple countries. If you’re not going to be providing data from multiple countries, you can leave this line out.

  • public_key_encryption: Include this line and set it to “true” if you want to have each row encrypted before being sent to LiveRamp for processing. For more information, see the “Utilize Encryption” section below.

Run the Local Encoder

To run the Local Encoder as Docker image --cap-add IPC_LOCK flag needs to be passed. To start a Docker container run the following command.

docker run --rm --cap-add IPC_LOCK \
--env LR_VAULT_ACCOUNT_ID=[LR_VAULT_ACCOUNT_ID] \
--env LR_VAULT_ACCOUNT_TYPE=[LR_VAULT_ACCOUNT_TYPE] \
--env LR_VAULT_LR_AWS_ACCESS_KEY_ID=[AWS_ACCESS_KEY_ID] \
--env LR_VAULT_LR_AWS_SECRET_ACCESS_KEY=[AWS_SECRET_ACCESS_KEY] \
--env LR_VAULT_INPUT=/tmp/input \
--env LR_VAULT_OUTPUT=/tmp/output \
--env LR_VAULT_LOCALE=eu-central-1 \
--env LR_VAULT_PUBLIC_KEY_ENCRYPTION=true
--name vault-app [Image Name]

Note

Only include the row --env LR_VAULT_PUBLIC_KEY_ENCRYPTION=true if you want to have each row encrypted before being sent to LiveRamp for processing. For more information, see the “Utilize Encryption” section above.

Format the File

Input files must include identifier fields and (for Onboarding use cases where you're receiving RampID packets) can also include segment data fields if desired.

Before uploading a file to the input location, make sure to format the data according to these guidelines:

  • Include a header row in the first line of every file consistent with the contents of the file. Files cannot be processed without headers.

  • If you want to maintain the ability to sort the output file, you must include a column containing row IDs (“RID”) as the first column of the file.

    Note

    The row identifier column is only required to maintain sort order and should not contain any customer personally-identifiable data.

  • Make sure that the only identifiers included are the allowed identifier touchpoints listed below.

  • If you’re sending data for consumers in multiple countries or if you’re including phone numbers, you must include the appropriate country code column (depending on the method used) to identify the country of each record. For more information, see the “Optional Configuration Parameters” section below.

  • Include a maximum of 500 segment data fields in a single file (for Onboarding use cases where you're receiving RampID packets).

  • Segment data field types can be in the form of a string, numeral, enum, etc.

  • The application supports three file formats: CSV, PSV, and TSV.

  • Files must be rectangular (have the same number of columns for every row).

  • If any values contain the file’s delimiter character (for example, a comma in a .csv file), make sure that your values are contained within quotes.

  • The recommended maximum file size is 20GB.

Allowed Identifier Touchpoints

You can include any of the following allowed identifier touchpoints for translation to RampIDs in both Onboarding and Addressability use cases:

  • Plaintext email address (maximum of three per record)

  • SHA-256 hashed email address (maximum of three per record)

  • Plaintext mobile phone number (maximum of two per record)

  • SHA-256 hashed mobile phone number (maximum of two per record)

  • Plaintext landline phone number (maximum of one per record)

  • SHA-256 hashed landline phone number (maximum of one per record)

Additional Allowed Touchpoints for Onboarding Use Cases

For Onboarding use cases (where you're receiving RampID packets, the following additional identifier touchpoints are also allowed for translation to RampIDs:

  • Name and postcode, which consists of first name, last name, and postcode (maximum of one per record)

  • AAID (maximum of one per record)

  • IDFA (maximum of one per record)

  • IMEI (maximum of one per record)

Example Header

See the header shown below for an example of what the header might look like when sending data in a pipe-separated file (psv) for an Onboarding use case, where segment data fields are included:

RID|EMAIL1|EMAIL2|EMAIL3|SHA256EMAIL1|SHA256EMAIL2|SHA256EMAIL3|MOBILE1|MOBILE2|SHA256MOBILE1|SHA256MOBILE2|LANDLINE1|SHA256LANDLINE1|FIRSTNAME|LASTNAME|POSTCODE|AAID|IDFA|IMEI|ATTRIBUTE_1|...|ATTRIBUTE_N

Replace ATTRIBUTE_1 … N in the example header with the name of your CRM attributes.

Example Output Files

For more information on the format of the output file, see "Output Examples".

Upload the File to the Input Bucket

Uploading a file to your Local Encoder services input bucket kicks off the encoding operation. To upload your file, run a command similar to the example below (this example shows using an AWS S3, but this could be any local directory):

aws s3 cp [your_file].csv s3://com-liveramp-vault-[your-vpc-id]-input

Caution

To successfully process a file, the input bucket cannot contain more than 10 files. Before uploading a new file to the input bucket, check that the bucket will not have more than 10 files once the new file has been uploaded.

Once the file has been processed, you’ll get a confirmation message that includes the number of records processed.

  • For RampID packet output, all consumer identifier data in a row is transformed into derived RampIDs, packaged into one data structure and encrypted again, yielding a RampID packet.

  • For RampID output, all consumer identifier data in a row is transformed into derived RampIDs in the form of a JSON string in a “RampID” column.

  • For identity envelope output, all consumer identifier data is transformed into derived RampIDs. A selection logic is applied, then the RampID is additionally obfuscated and encrypted into an identity envelope. Only one identity envelope is returned per row of data. A timestamp column is appended to the end of each row. This column gives the expiration date and time for the identity envelope in Unix format (timezone UTC).

For more information on the format of the output file, see "Output Examples".

Configuration Parameters

See the sections below for information on the required and optional parameters to use, depending on the deployment method being used.

Required Configuration Parameters

Parameter Name

Parameter for Local Configuration

Parameter for Command Line or Configmaps.yaml

Example Value(s)

Notes

AWS user ID

account_id

LR_VAULT_ACCOUNT_ID

AID.….

Provided by LiveRamp

Account type

account_type

LR_VAULT_ACCOUNT_TYPE

awsiam

LiveRamp AWS account ID

AWS_LR_ACCOUNT_ID

461694764112

Provided by LiveRamp

AWS IAM access key ID

lr_access_key_id

LR_VAULT_LR_AWS_ACCESS_KEY_ID

AKI......

Provided by LiveRamp

AWS IAM secret access key

lr_secret_access_key

LR_VAULT_LR_AWS_SECRET_ACCESS_KEY

LiveRamp provides the secret encrypted with customer key

Input File Location

input

LR_VAULT_INPUT

  • s3://input-bucket

  • gs://input-bucket

  • /tmp/input-folder

  • For AWS S3 bucket, prefix is “s3://”.

  • For GCS buckets, prefix is “gs://”.

Output file Location

output

LR_VAULT_OUTPUT

  • s3lr://bucket-name

  • gs://bucket-name

  • s3://bucket-name

  • /tmp/output-folder

  • For an S3 bucket that belongs to a LiveRamp account, prefix is “‘s3lr://”’.

  • For AWS S3 bucket, prefix is “s3://’.

  • For GCS buckets, prefix is “gs://”.

AWS region for LR resources

lr_region

LR_VAULT_LR_AWS_REGION

eu-central-1

LiveRamp’s AWS Region

Origin of the data being processed

locale

LR_VAULT_LOCALE

us

Two letter country code representing the origin of the data being processed (for example, Australia = “au”, Great Britain = “GB”). Not case sensitive.

Optional Configuration Parameters

Parameter Name

Parameter for Local Configuration

Parameter for Command Line or Configmaps.yaml

Example Value(s)

Notes

Customer Profile

profile

LR_VAULT_PROFILE

  • prod

  • dev

Default is "prod"

Filename

filename_pattern

LR_VAULT_FILENAME_PATTERN

  • ^test.*

The regex to use to determine which files in the input folder or bucket should be processed (for example, entering “^test.*” would include a file named “test.csv”). The app will process files from the folder/bucket with filenames that match the regex.

Country Header

country_code_column

LR_VAULT_COUNTRY_CODE_COLUMN

  • COUNTRY_CODE (default value)

  • {{Customer provided}} (optional)

  • The header name for the column containing country codes for each row.

  • The values in the country code column determine the hygiene applied to phone numbers in the file.

  • Must match the value in input file.

Public Key Encryption

public_key_encryption

LR_VAULT_PUBLIC_KEY_ENCRYPTION

true

Include this parameter and set to “true” to encrypt each row of data in the files before they’re sent to LiveRamp for processing.

Header Mapping

header_mapping

LR_VAULT_HEADER_MAPPING

{{customer determined}}

newvalue=defaultvalue,

A list of key=value pairs which can be used to replace the default headers for the identifier columns in the file. For example, if the email columns have the headers “primary_email” and “alt_email”, then the header mapping should be set to “primary_email=email1,alt_email=email2”.

Error Log

customer_logging_enabled

LR_VAULT_CUSTOMER_LOGGING_ENABLED

Include this parameter and set to “true” to enable the generation of an error log file showing any processing errors (such as illegal characters or incorrect header values) and the row number for where the error occurred.

Error Log Location

-v <localFolder>:/var/logs/vault-app

-v <localFolder>:/var/logs/vault-app

Include this parameter to set the delivery location for the error log file.

Mode

mode

LR_VAULT_MODE

  • default

  • task

The default value is "default" for long-running file processing, set to "task" to enable single file processing (the application will shut down after processing a single file, not available for Kubernetes setup).

Packet Type

packet_type

LR_VAULT_PACKET_TYPE

unencoded

Include this parameter and set to "unencoded" to receive RampIDs rather than RampID packets. Include this parameter only when you are receiving RampIDs.

Envelope Output

envelopes_flow

LR_VAULT_ENVELOPES_FLOW

true

Include this parameter and set to "true" to have the output RampIDs packaged into identity envelopes. Include this parameter only when you are receiving identity envelopes.

Test Mode

dry_run

LR_VAULT_DRY_RUN

  • true

  • false

The default value is “'false”'. Set to “true” to run the app in dry run mode. Only outputs encrypted packets in dry run mode.

Customer AWS Access Key

N/A

AWS_ACCESS_KEY

Only for customers using S3 bucket as input source. Access key for client’s AWS.

Customer AWS Secret Access Key

N/A

AWS_SECRET_ACCESS_KEY

Only for customers using S3 bucket as input source. Secret access key for client’s AWS.

Customer AWS Region

N/A

AWS_REGION

Only for customers using S3 bucket as input source. AWS region in which the bucket is residing.

Customer AWS Region

N/A

AWS_DEFAULT_REGION

Only for customers using S3 bucket as input source. AWS region in which the bucket is residing.

Customer GCS Bucket Credentials

N/A

GOOGLE_APPLICATION_CREDENTIALS

Only for customers using GCS bucket as input/output source. Path to your Google Credentials JSON file.

Customer GCS Project

gcp_project_name

LR_VAULT_GCP_PROJECT_NAME

Only for customers using GCS bucket as input/output source. The name of your GCP project. Added if the default profile name can't be found.