Skip to main content

Google Cloud Storage

Overview

The Google Cloud Storage (GCS) connector provides a convenient way of importing event data directly from a GCS bucket into the Decentriq DMP.

Prerequisites

  • Have a Google Cloud account.
  • Have an existing Google Cloud Storage bucket containing the data to import.
  • Have existing service account credentials.

Step-by-step guide

Step 1: Select the source

Follow the steps to create a new import and choose Google Cloud Storage from the list of available data sources.

Select Google Cloud Storage

Step 2: Configure connection

Input the required information:

  • Bucket: Name of the GCS bucket from which the data should be imported.
  • Service Account JSON: The service account credentials associated with the Google Cloud account (see the Google documentation for more details). This takes the form of a JSON file generated when setting up the service account and contains the following information:
    • type: Identifies the type of credentials (this will be set to service_account).
    • project_id: The ID of your Google Cloud project.
    • private_key_id: The identifier for the private key.
    • private_key: The actual private key used for authentication - in PEM format.
    • client_email: The email address of the service account.
    • client_id: A unique identifier for the service account.
    • auth_uri: The URL to initiate OAuth2 authentication requests.
    • token_uri: The URL to retrieve OAuth2 tokens.
    • auth_provider_x509_cert_url: The URL to get Google's public certificates for verifying signatures.
    • client_x509_cert_url: The URL to the public certificate for the service account.
    • universe_domain: The domain of the API endpoint.
  • Path prefix: An optional prefix used to limit the scope of retrieved objects to a specific folder or directory within the bucket.

GCS connection details

Click Test connection to verify that the credentials and configuration are valid.

Step 3: Select files

Click Continue to proceed to file selection. Configure which files should be imported:

  • File names pattern: Define a path and file name pattern to match files (for example data/*.csv or data/*-{yyyy}-{mm}-{dd}.csv). Wildcards (*) are supported.
  • File names include a date (optional): Enable this if the pattern contains date tokens ({yyyy}, {mm}, {dd}), then select an absolute or relative time range.
  • Click Apply to update the list of matching files. The table displays all files that match the configured pattern and time range.

File selection

Step 4: Configure parsing

Click Continue to proceed to data parsing. Define how the selected files should be parsed, including format-specific options such as delimiters, headers, or timestamp handling.

Data parsing

Step 5: Map fields

Click Continue to proceed to field mapping. Map the parsed fields to DMP event attributes. This step defines how incoming data is transformed into DMP events and user profile updates.

Field mapping

Step 6: Schedule and name

Click Continue to proceed to scheduling. Name the import and define how often it should run.

Scheduling

Step 7: Save the import

Click Create to create the import. The import then appears in the imports table with the name provided in the previous step.

Imports table