Table of Contents

Search

  1. Preface
  2. Informatica Developer
  3. The Model Repository
  4. Searches in Informatica Developer
  5. Connections
  6. Physical Data Objects
  7. Flat File Data Objects
  8. Logical View of Data
  9. Viewing Data
  10. Application Deployment
  11. Application Patch Deployment
  12. Application Patch Deployment Example
  13. Continuous Integration and Continuous Delivery (CI/CD)
  14. Object Import and Export
  15. Appendix A: Data Type Reference
  16. Appendix B: Keyboard Shortcuts
  17. Appendix C: Connection Properties

Developer Tool Guide

Developer Tool Guide

Run-time: Read Properties

Run-time: Read Properties

The Developer tool displays run-time properties for flat file sources in the
Run-time: Read section
in the
Advanced
view. The Data Integration Service uses this information when it reads data from a flat file.
The following table describes the read properties that you configure for flat file data objects:
Property
Description
Input type
Type of source input. You can choose the following types of source input:
  • File. For flat file sources.
  • Command. For source data or a file list generated by a shell command.
Source type
Indicates source type of files with the same file properties. You can choose one of the following source types:
  • Direct. A source file that contains the source data.
  • Indirect. A source file that contains a list of files. The Data Integration Service reads the file list and reads the files in sequential order.
  • Directory. Source files that are in a directory. You must specify the directory location in the source file directory property. The Data Integration Service reads the files in ascending alphabetic order. The Data Integration Service does not read files in the subdirectories.
Source file name
File name of the flat file source.
Source file directory
Directory where the flat file sources exist. The machine that hosts Informatica services must be able to access this directory.
Default is the SourceDir system parameter.
Concurrent Read Partitioning
Order in which multiple partitions read input rows from a source file. If the Data Integration Service does not create partitions for the mapping, it ignores this value.
Select one of the following options:
  • Optimize throughput. The Data Integration Service does not preserve input row order.
  • Keep relative order. The Data Integration Service preserves the input row order for the rows read by each partition.
  • Keep absolute order. The Data Integration Service preserves the input row order for all rows read by all partitions.
Connection Type
The type of connection. Select from the following options:
  • None. The source file does not require a connection.
  • Hadoop File System. The source file resides in HDFS.
Default is None.
Command
Command used to generate the source file data. Use a command to generate or transform flat file data and send the standard output of the command to the flat file reader when the mapping runs. The flat file reader reads the standard output as the flat file source data. Generating source data with a command eliminates the need to stage a flat file source.
Use a command or script to send source data directly to the Data Integration Service instead of using a pre-mapping command to generate a flat file source. You can also use a command to generate a file list.
For example, to use a directory listing as a file list, use the following command:
cd MySourceFiles; ls sales-records-Sep-*-2005.dat
Truncate string null
Strips the first null character and all characters after the first null character from string values.
Enable this option for delimited flat files that contain null characters in strings. If you do not enable this option, the Data Integration Service generates a row error for any row that contains null characters in a string.
Default is disabled.
Line sequential buffer length
Number of bytes that the Data Integration Service reads for each line.
This property, together with the total row size, determines whether the Data Integration Service drops a row. If the row exceeds the larger of the line sequential buffer length or the total row size, the Data Integration Service drops the row and writes it to the mapping log file. To determine the total row size, add the column precision and the delimiters, and then multiply the total by the maximum bytes for each character.
Default is 1024.
Generate Run-time Column Names
Determines how to generate the column metadata at run time.
Select one of the following options:
  • Automatically (Field1, Field 2…). The Data Integration Service includes column metadata based on the position of the column in the data. The column metadata for the flat file is constructed at run time from the row that the user specifies in the
    start from import line
    attribute in the format properties of the flat file data object.
  • From data file header (first line). The first row of the flat file contains a header row that the Data Integration Service uses to determine the column metadata.
  • From control file. The Data Integration Service constructs column metadata based on the data in a control file, such as column name, precision, data type, scale, and number of bytes to process.
Control file name
Name of the control file. Required if you generate run-time column names from control file.
Control file directory
Directory where the control file exist. Required if you generate run-time column names from control file.
Default Field Type
Data type of the additional ports generated at run time.
Default Precision
Precision of the additional ports generated at run time.
Default Scale
Scale of the additional ports generated at run time.
Constraints
Conditional expression that the values on a data row must satisfy. Use the Expression editor to enter an expression that evaluates to TRUE. When the Data Integration Service reads constraints, it drops the rows that do not evaluate to TRUE.
For example, a source flat file has an AGE column. You can set a constraint with AGE < 70 on the flat file data object. The Data Integration Service reads rows from the source flat file with the constraint AGE < 70. If the Data Integration Service reads rows with AGE >= 70, it drops those rows.

0 COMMENTS

We’d like to hear from you!