You can import customer data in CSV format from either the Amazon S3 bucket provided by Blueshift or your custom Amazon S3 bucket. You can set up a one time or recurring import task to import the data.
For information about data types, user attributes, and data formats, see Import customer data.
You can upload the customers data CSV file to the S3 bucket provided by Blueshift (using the default adapter Blueshift Import) or to your custom S3 bucket (using the Custom Import adapter).
If you are using the Blueshift Import option:
- Add data files to the folder bsft-customers/<site-name>/customer_attributes.
- You can obtain the S3 credentials from the Account Settings > API Keys tab. The credentials are visible only to users with an Admin role.
If you are using the Custom Import option:
- Ensure that your S3 bucket has a CORS configuration.
- Provide your own S3 credentials and S3 file path. For example,
Complete the following steps to set up an import task:
- To import customer data, go to the Customer Data > Profiles in the left navigation. Click +CUSTOMERS > Import Customers.
- Select Upload via S3 Bucket as the Source.
- On the Customer CSV Upload page, specify the Customer Attribute Feed Name. This field is used for identification purposes only.
- Specify an email address in the Email Upload Summary section to receive a notification when the upload process is completed. If there are errors, an error report is also sent. In the CC field, specify the email addresses of othet people who should receive the notification.
- For the Adapter, select Blueshift Import if you are using the S3 bucket provided by Blueshift. Select Custom Import to use your own S3 location.
- If you select Blueshift Import, specify the S3 Path where you upload the customers data files. For example, you can create a location on the S3 bucket called /customer-imports/blueshiftreads, and upload files to this path.
- If you select Custom import, provide the access credentials for the S3 bucket. For more information, see Working with Amazon S3 buckets in the AWS documentation.
- Access Key: The Access Key ID that we use to get files that contain your customers' data from your S3 bucket.
Secret Access Key: The Secret Access Key that we use to get the files from your S3 bucket.
For more information on the Access Key ID and Secret Access Key, see Programmatic access on AWS documentation.
- Bucket: Provide the URL to your S3 bucket where you store the files.
- AWS Region: Specify the AWS region where the S3 bucket resides.
- Path: Specify the path on the S3 bucket where you upload the files with your customers' data.
- Add the Update Frequency. You can select from Only once, Weekly, Daily, or Hourly.
- Specify the Update Time. You can upload a file Immediately or set a date and time for the upload.
- For recurring uploads, specify an End date and time.
- Click advanced to specify the character encoding of the CSV file you upload. You can select Unicode (UTF-8), Unicode (UTF-16), or Western Europe (ISO-8859-1).
- Set whether to Update All Matching Profiles and to Allow Blank User Attributes.
Note: Attributes with data type as Boolean cannot be set to a NULL or blank value and must be set as either true or false.
- Click Fetch Sample S3 File to fetch a sample CSV file from the location that you specified in the Adapter field.
Map the fields from the imported data to the fields in Blueshift and specify the data type for the field.
- For customer data, one column must be mapped to either of the following customer identifiers in Blueshift: email, retailer_customer_id, customer_id, or device_id.
- The Column is the attribute in the imported data and the Customer Attribute is the attribute in Blueshift.
Note: You cannot change the data type for a custom attribute once it is imported.
- Click Import Customers to run the import task.
You can view the task status from the View Uploads page.