Export Results to GCS
You can set up an export of analytics and list results to GCS, which has the overall steps shown below.
Note
Partners invited to a clean room must have their export destination connections (grouped under "Destinations") approved by clean room owners. Contact your Customer Success representative to facilitate the approval.
The IAM role from LiveRamp Clean Room needs to have write/delete and read access on the customer bucket/folder.
Configuring exports outside of a clean room (i.e., at the organization level) is still supported, but will be deprecated. Setting up clean room question exports within a clean room is recommended.
Overall Steps
Perform the following overall steps in Google Cloud Platform to set up an export of analytics and list results to BigQuery:
Once the above steps have been performed in Google Cloud Platform, perform the following overall steps in LiveRamp Clean Room:
For information on performing these steps and for export details, see the sections below.
Perform Steps in Google Cloud Platform
Perform the steps in the sections below in Google Cloud Platform to set up an export of analytics and list results to BigQuery.
Create a Google Service Account
To create a Google service account in Google Cloud Platform:
From the Google Cloud Platform main menu, select IAM & Admin → Service Accounts.
Click
. Save the service account email because you will need it in later steps.Enter a name for the service account.
Click
.Configure roles and additional user access as needed.
Click
.
Create a Google Service Account Key
After you've created a Google service account, create a Google service account key:
From the Google Cloud Platform main menu, select IAM & Admin → Service Accounts.
Select the check box for the service account you created in the previous procedure.
From the More Options menu for the service account's row, select Manage keys.
Click
and then click .For the key type, select JSON and then click CREATE.
The private key will be stored in your Download folder (it will look similar to the following example). Save this for use in the “Add the Credentials” section below.
![]() |
Grant the Service Account Permissions to Access Bucket Objects
To grant bucket objects permissions to the service account:
From the Google Cloud Platform console, log in as a Project Editor.
From the home dashboard, select IAM & Admin → Roles.
Select CREATE ROLE.
Enter a title, description, and ID for the custom role.
Select Add permissions.
Enter "Storage Admin" in the filter.
Add the following permissions:
storage.managedFolders.create
storage.managedFolders.delete
storage.managedFolders.get
storage.managedFolders.list
storage.multipartUploads.abort
storage.multipartUploads.create
storage.multipartUploads.list
storage.multipartUploads.listParts
storage.objects.create
storage.objects.delete
storage.objects.get
storage.objects.getIamPolicy
storage.objects.list
storage.objects.overrideUnlockedRetention
storage.objects.restore
storage.objects.setIamPolicy
storage.objects.setRetention
Storage.objects.update
Click
.
Note
If you don’t want to set up a custom role with the above permissions you can also assign the service account with the Storage Object Admin role. Instructions on assigning a role to the service account can be found below.
Assign the Custom Role to the Cloud Storage Service Account
To assign the custom role to the cloud storage service account:
From the Google Cloud Platform console, log in as a Project Editor.
Search for "Cloud Storage".
From the home dashboard, select Storage → Browser.
Click the bucket name for the bucket you want to configure for access.
Select SHOW INFO PANEL in the upper-right corner. The information panel for the bucket displays.
From the information panel, click
.In the Add members field, add the service account.
From the Select a role dropdown, select Custom → role (where "role" is the custom Cloud Storage role you created in the previous procedure).
Click
.Confirm the custom role and service account are now associated with the bucket.
Perform Steps in LiveRamp Clean Room
Once the above steps have been performed in Google Cloud Platform, perform the overall steps in the sections below in LiveRamp Clean Room.
Enable the Clean Room for Exports
Before setting up an export, the clean room owner must enable export for the selected source clean room:
From the LiveRamp Clean Room navigation pane, select Clean Rooms → Clean Rooms (or click Go to Clean Rooms from the Clean Rooms tile).
In the row for the clean room you would like to export from, click the More Options menu (the three dots), and then select Edit.
From the Configuration step, click
.From the Parameters step, adjust any data control parameters as needed and then slide the Enable Export toggle to the right.
Click
.Verify that your data control parameters are correct and then click
.
Add the Credentials
To set up an export to a cloud location, the clean room owner must first add either their own credentials or those of their partner:
From the LiveRamp Clean Room navigation pane, select Data Management → Credentials.
Click
.Enter a descriptive name for the credential.
For the Credentials Type, select "Google Service Account".
For the Project ID, enter the project ID.
Enter your Credential JSON (credential JSON is hidden by default, viewable here for demonstration purposes).
Click
.Verify that your credentials have been added to LiveRamp Clean Room:
Add an Export Destination Connection
To add an export destination connection:
From the LiveRamp Clean Room navigation pane, select Destinations & Integrations → Destinations.
Click
.Select GCS Export.
Enter a name and select the GCS Credential created in the "Add the Credentials" section above.
Note
Leave the Public Key field blank.
Click
.Confirm that the new export has been added to your list of GCS export destination accounts.
Note
The status of the destination connection will be "Pending" initially, but you can continue to export data. Once the first successful export has been processed the status changes to "Complete".
Set Up a Data Export Within the Clean Room
To set up a data export within the clean room:
From the LiveRamp Clean Room navigation pane, select Clean Rooms → Clean Rooms.
From the tile for the desired clean room, click
.From the Clean Room navigation pane, select Destinations. The Destinations screen shows all destination connections provisioned to the clean room.
Check the check box for the desired destination connection and then click
(AWS S3 example shown).Verify that your destination connection has been added (S3 IAM example shown).
Create a New Export
To create a new export:
From the LiveRamp Clean Room navigation pane, select Clean Rooms → Clean Rooms.
From the tile for the desired clean room, click
.From the Clean Room navigation pane, select Exports.
Click
to open the wizard to create a new export.Select the question that you want to export outputs for and then click
.Check the radio button for the specific GCS destination account you want to send run outputs to.
Enter the GCS export path to save the results to and then click
.Note
Provide only the bucket name for the analytical or user list outputs. Do not include the
gs://
prefix before the bucket name.Verify that the job has been created. Exports are added to the page. You may view the details of an export by clicking on the name.
Note
Exports can be paused, which will stop them from sending data upon the completion of each run.
Exports cannot be edited or deleted. Changes should be made by pausing the export and creating a new export.
Export Details
When a question runs, the results will be written to the defined bucket. Each row will have an associated "Run ID" column. There will be a second metadata table created. Users can join on Run ID to get run metadata (Run Name, Runtime Parameters, etc.).
GCS Export File Structure
gs://<data-source-location>/<custom-prefix>
/date=<yyyy-mm-dd>
/cleanroom=<clean-room-id>
/question=<question-id>
/run=<run-id>
/data
For Analytical Questions, this contains 1 .csv file. For List Questions, this contains 1+ .parquet files with List data.
Note
List files do not contain column headers.
metadata.json
This contains each column header name, data type, and whether it is encrypted. If encryption is used, the dek and kek are also provided.