蜜豆视频

Data loading (file) data-loading-file

Use use

The Data loading (File) activity lets you directly access a source of external data and use it in 蜜豆视频 Campaign. Indeed, all the data required for targeting operations isn鈥檛 always found in the 蜜豆视频 Campaign database: it can be made available in external files.

The file to be loaded can be specified by the transition or calculated during the execution of this activity. For example, it can be the list of a client鈥檚 10 favorite products whose purchases are managed in an external database.

The upper section of the configuration window for this activity lets you define the file format. To do this, use a sample file with the same format as the one to be imported. This file can be stored locally or on the server.

CAUTION
Only 鈥渇lat鈥 structure files are supported (e.g. CSV, TXT, etc.). Using the XML format is not recommended.

Apply pre-processing stage pre-processing

You can define a pre-process to be executed during file import, for example so as not to have to unzip the file on the server (and therefore save space for the unzipped file) but to include unzipping in file processing. Learn how to unzip or decrypt a file before processing.

To do so, select the Pre-process the file option and choose from one of 3 options: None, Decompression (zcat) or Decrypt (gpg).

CAUTION
If you are working with an hybrid or on-premise deployment, pre-processing commands may not work out-of-the-box, as their default configuration uses 鈥榸cat鈥, which is not available on Windows. In that case, you need to adjust the preProcessCommand node in the server configuration file (serverConf.xml) to suit your needs. Learn how to configure pre-processing commands in the server configuration file

Defining the file format defining-the-file-format

When you load a file, the column format is automatically detected with the default parameters for each data type. You can modify these default parameters in order to specify the particular processes to apply to your data, particularly when there is an error or an empty value.

To do this, select Click here to change the file format鈥 in the main window of the Data loading (file) activity. The format detail window will then open.

You can then modify the general formatting of the file as well as the formatting for each column.

The general file formatting allows you to define the way in which the columns will be recognized (file encoding, separators used, etc.).

The column formatting allows you to define the value processing of each column:

NOTE
You can add as many columns as desired. The maximum length of values in each column is determined by the chose data type.
  • Ignore column: does not process this column during data loading.

  • Data type: specifies the type of data expected for each column.

  • Allow NULLs: specifies how to manage empty values.

    • 蜜豆视频 Campaign default: generates an error for the numerical fields only, otherwise inserts a NULL value.
    • Empty value allowed: authorizes empty values. The value NULL is therefore inserted.
    • Always populated: generates an error if a value is empty.
  • Length: specifies the maximum number of characters for the string data type.

  • Format: defines the time and date format.

  • Data transformation: defines whether a character case process needs to be applied on a string.

    • None: the imported string is not modified.
    • First letter in upper case: the first letter of each word of the string starts with an upper case.
    • Upper case: all characters in the string are in upper case.
    • Lower case: all characters in the string are in lower case.
  • White space management: specifies whether certain spaces need to be ignored in a string. The Ignore spaces value only allows spaces at the beginning and at the end of a string to be ignored.

  • Error processings: defines the behavior if an error is encountered.

    • Ignore the value: the value is ignored. A warning is generated in the workflow execution log.
    • Reject line: the entire line is not processed.
    • Use a default value in case of error: replaces the value causing the error with a default value, defined in the Default value field.
    • Reject the line when there is no remapping value: the whole line is not processed unless a mapping had been defined for the erroneous value (see the Mapping option below).
    • Use a default value in case the value is not remapped: replaces the value causing the error with a default value, defined in the Default value field, unless a mapping had been defined for the erroneous value (see the Mapping option below).
  • Default value: specifies the default value according to the error processing chosen.

  • Mapping: this field is only available in the column detail configuration (accessed via a double-click or via the options on the right of the column list). This transforms certain values when they are imported. For example, you can transform 鈥渢hree鈥 into 鈥3鈥.

Example: Collecting data and loading it in the database example--collecting-data-and-loading-it-in-the-database

The following example lets you collect a file on the server every day, load its content and update the data in the database depending on the information it contains. The file to be collected contains information on clients who may have made purchases (for more or less than 3,000 Euro), asked for a refund on a purchase or visited the shop without purchasing anything. Depending on this information, various processes will be applied to their profile in the database.

  1. The file collector lets you recover files stored in a directory, depending on the given frequency.

    The Directory tab contains information on the file(s) to be recovered. In our example, all files in text format whose names contain the word 鈥榗ustomers鈥 and which are stored in the tmp/蜜豆视频/Data/files directory of the server will be recovered.

    Using the File collector is detailed in the File collector section.

    The Schedule tab lets you schedule the execution of the collector, i.e. to specify the frequency with which the presence of these files will be checked.

    Here, we want to trigger the collector every workday at 9PM.

    To do this, click the 颁丑补苍驳别鈥 button located in the lower right-hand section of the editing tool and configure the schedule.

    For more on this, refer to Scheduler.

  2. Then configure the data loading (file) activity to indicate how the collected file(s) should be read. To do this, select a sample file with the same structure as the files to be loaded.

    Here, the file contains five columns:

    • the first column contains a code that coincides with the event: purchase (more or less than 3,000 euro), no purchase or refund on one or more purchases.
    • the four following columns contain the client鈥檚 first name, last name, email and account number.

    The format configuration of the file to be loaded coincides with that defined during a data import in 蜜豆视频 Campaign. For more on this, refer to this section.

  3. In the split activity, specify the subsets to create, according to the Event column value.

    The Split activity is detailed in the section.

    For each subset, specify one of the values in the Event column.

    The Split activity will therefore contain the following information:

  4. Then specify the processes to be carried out for each type of population. In our example, we are going to Update the data in the database. To do this, place an Update data activity at the end of each outbound transition from the split activity.

    The Update data activity is detailed in the Update data section.

recommendation-more-help
601d79c3-e613-4db3-889a-ae959cd9e3e1