With the integration of Blueshift and Google BigQuery, you can import customer data into your Blueshift account very easily and build a 360 degree view of all your customers.

For information about data types, user attributes, and data formats, see Import customer data.

Prerequisites

Before you can import data from Google BigQuery, you must set up integration of Blueshift with Google BigQuery and configure at least one adapter.

Set up an import task

You can import catalogs, events, and customer data using Google BigQuery. The starting point for your import tasks depends on the type of data you are importing.

To set up a task to import customer data from Google BigQuery, complete the following steps:

  1. Go to the Customer Data > Profiles in the left navigation. Click +CUSTOMERS > Import Customers.
  2. Select Google BigQuery as the Source.
  3. Add a Name for the task. The import task form opens.
  4. In the Destination section, you can see the type of data being imported as Customers.
  5. Set up Notification Preferences to send a notification email to the specified email addresses when there is a change in the status of the task or when a certain percentage of records fail during import.
  6. In the Source section, select the adapter that you want to use for the import task.
    Screenshot 2024-07-16 at 7.28.32 PM.png

     

  7. For Import From, select either Table or View. If your data is spread across multiple tables, it is recommended that you provide a View.
  8. Select the Table or the View from which the data is to be imported.
  9. Sample data consisting of 10 records is fetched from the table or view specified in the Source section. This data is displayed in the Configuration section.
     

    databricks_customer_fetchdata.png

  10. Map the fields from the imported data to the fields in Blueshift and specify the data type for the field.
    • For customer data, one column must be mapped to either of the following customer identifiers in Blueshift: email, retailer_customer_id, customer_id, device_id, or cookie.
    • The Source Attribute Name is the attribute in Google BigQuery and the Destination Attribute Name is the attribute in Blueshift.
      Note: You cannot change the data type for a custom attribute once it is imported.
       

      databricks_customer_mapfields.png

  11. In case the sample data does not contain all the available fields, Add more fields to the data mapping.
  12. Click Test Run to test the mapping. A maximum of 10 records are fetched during this test run.
     

    databricks_customer_testrun.png

  13. Verify that the data mapping is done correctly. Edit the data mapping if required. Click Test Run again after you make the changes.
  14. For Additional Configurations, select the Type of Import.
    • Select Full Import if you are importing bulk data. For a Full import, the entire data from the selected table or view is imported from Google BigQuery every time you run the import task.
    • Select Incremental Import to set up an incremental import task. Select the Diff Identification and the Diff Identifier.
       

      databricks_typeofimport.png

  15. In the Schedule section, set the Start Time for the import task.
     

    databricks_schedule.png

  16. To set up a recurring task, under Schedule select the Is it a recurring data import? option.
  17. Set the frequency using the Task executes every field. You can set the frequency in minutes, hours, days, weeks, or months. You cannot set an import frequency shorter than a day for a Full import.
  18. Click Save to save the task.
  19. Click Launch to run the import task.

Import task status

The index page for Customer imports shows the status for each Google BigQuery task with an overall status that identifies which state the task is in. For more information, see Customer Import Tasks.

Was this article helpful?
0 out of 0 found this helpful

Comments

0 comments

Please sign in to leave a comment.