🖌️
CaO_Data_Platform (KODDA)
  • Introduction
  • Getting Started
    • Login Screen
    • Home Screen
    • Manage Users
    • Manage Connections
    • User Interface
  • User Guide
    • Generate Mapping
      • Data Type Mappings
      • Generate Results
    • Edit Mapping
      • Mapping Details
      • Mapping Elements
      • Mapping Activities
    • Design Workflow
    • Scheduler
    • Browse Data
    • Map History
    • Workflow History
    • Upload Files
    • Migrate Objects
    • Google Analytics
    • Data Dictionary
  • Architecture Guide
  • Release Notes
  • Examples
    • Incremental Change Tracking
    • Incremental Loads Using Row Version
  • Product Availability Matrix
  • API Docs
Powered by GitBook
On this page
  • Manage Connections
  • Example: Oracle
  • Additional Connection Types
  • File
  • Google Analytics
  • Hive
  • Informix
  • MySQL
  • Oracle
  • PostgreSQL
  • Redshift
  • Salesforce
  • Snowflake
  • SQL Server
  1. Getting Started

Manage Connections

PreviousManage UsersNextUser Interface

Last updated 4 years ago

Manage Connections

Use the Manage Connections screen to create and modify data source and/or target connections.

  • To create a new connection, simply click the and enter the required information.

  • Once a connection is created use the icon to test your new connection.

Connection Names should use alphanumeric characters only.

Different connection types have their own unique connection parameters. Consult your database documentation or database administrator for your connection details required.

Example: Oracle

Parameter

Description

Connection Name

A unique name identifying the connection.

Connection Role

Indicates whether a connection is to be used as a source (reading only), target (writing only) or both.

Host

The host name or IP address of the database server

Port

The port number of the database server

Service Name

Oracle Service Name

SID

Oracle SID

User Name

The user name used to authenticate with the connection.

Password

The password used to authenticate the user.

Use TNS

Create a connection using the TNS file on the application Server (on-prem only)

Additional Connection Types

See the following tables for more detail on specific connection parameter.

File

Parameter

Description

File Type

Example: Delimited File

File System Type

Local, HDFS, S3

Connection Name

A unique name identifying the connection

Connection Role

Indicates whether a connection is to be used as a source (reading only), target (writing only) or both.

Directory

The path to the directory for reading or writing files.

File Extension

The extension to use for files that are to be read or written. When reading files only files with this extension are read. If multiple extensions need to be supported from the same directory for multiple files, create a seperate connection. Default: csv

Delimiter

Defines column separator. A separator longer than one character is permitted. Default: Comma

Quote Character

Defines quote character. Column values surrounded with the quote character are parsed with the quote characters removed. This is useful when values contain the separator or line breaks. No more than one character is allowed. An empty value disables quoting. Default: Double quote

Column Header

Used to specify that the files contain a column header with column names. If False and column names are not set, then columns are named sequentially COLUMN1, COLUMN2 when used as a source. If True, the column header is read from the first line of the file. Default: True

Connection Type: File (Advanced Settings)

Parameter

Description

Group Files

When True all files with a file name matching the File Tail Pattern and the File Extension are read as if they were a single file.

Trim Values

If True leading and trailing white space is trimmed from each column value in the file. Column values inside quotes are not trimmed.

Lines to Skip

After opening a file, skip this many lines before starting to interpret the contents.

Column Names

Used when the Column Header property is set to False in order to provide a set of column names to be applied to the files that are read. Provide a list of column names for a table separated by the delimiter. When multiple file formats exist you can specify the columns for each file by using the following format {"<filename without extension or tail pattern>":"<column list>","<table name2>":"<column list2>"}

File Tail Pattern

Regular expression for matching filenames when property Group Files is True. If the regular expression contains groups (surrounded by parentheses) then the value of each group in matching filenames is added as an extra column when reading the file. Example: For files named test-001-20081112.csv and test-002-20081113.csv with a file tail pattern of -(\d+)-(\d+). The column values 001 and 20081112 are added to each line read from the first file and 002 and 20081113 are added to each line read from the second file.

File Tail Column Names

Comma-separated list of column names for the additional columns generated by regular expression groups in the property File Tail Pattern.

Character Set

Defines the character set name of the files being read, such as WIN-1252, UTF-8, UTF-16 etc...

Connection Type: File (HDFS Settings)

Parameter

Description

HDFS Host

The host name or IP address of the HDFS name node.

HDFS Port

The port number of the HDFS name node. For high availability clusters use HA

Additional HDFS Parms

Specific parameters to pass to the HDFS connection. These are parameters that are typically found in the config files and should be specified in the following format {'parameter':'value','parameter2':'value2' }

HDFS Config Location (on-prem)

The location of the HDFS config files.

HDFS User

The HDFS user used to connect.

Use HDFS Autoconfig (on-prem)

Autoconfiguration uses the hdfs-site.xml and core-site.xml found on the node. You can specify the location of this file by setting the HDFS Config location parameter.

Connection Type: File (S3 Settings)

Parameter

Description

AWS Access Key

AWS access key ID

AWS Secret Access Key

AWS access key secret

AWS Region Name

Region in which the S3 bucket resides.

Bucket Name

AWS S3 bucket name

Google Analytics

Parameter

Description

Connection Name

A unique name identifying the connection Kodda.

Service Account Credentials (JSON)

Queries Per 100s

The maximum number of queries to allow per 100 seconds. The maximum value allowed in this field is 1000.

Hive

Parameter

Description

Connection Name

A unique name identifying the connection.

Host

The host name for HS2. For Impala, this can be any of the impalad's.

Port

The port number for HS2. The Impala default is 21050. The Hive default is 10000.

Authorization Mechanism

Specify the authentication mechanism. 'NOSASL' for unsecured Impala.'PLAIN' for unsecured Hive (because Hive requires the SASL transport). 'GSSAPI' for Kerberos and 'LDAP' for Kerberos with LDAP.

Temp File Location

The HDFS directory location to use to stream data being delivered to Hive. Data is loaded here first prior to being imported into the target Hive tables.

User Name

The user name used to authenticate with the connection.

Password

The password used to authenticate the user.

Connection Type: Hive (Advanced Settings)

Parameter

Description

Enable SSL

Enable SSL.

Default Database

The default database.

CA Certificate

Local path to the the third-party CA certificate. If SSL is enabled but the certificate is not specified, the server certificate will not be validated.

Kerberos Service Name

Authenticate to a particular 'impalad' service principal.

Connection Timeout

The connection timeout in seconds.

Hive (HDFS)

Parameter

Description

HDFS Host

The host name of the HDFS name node.

HDFS Port

The port number of the HDFS name node. For high availability clusters use HA

Additional HDFS Parms

Specific parameters to pass to the HDFS connection. These are parameters that are typically found in the config files and should be specified in the following format {'parameter':'value','parameter2':'value2' }

HDFS Config Location (on-prem)

The locat ion of the HDFS config files.

HDFS User

The HDFS user used to connect.

Use HDFS Autoconfig (on-prem)

Auto configuration uses the hdfs-site.xml and core-site.xml found on the node. You can specify the location of this file by setting the HDFS Config location parameter.

Informix

Parameter

Description

Connection Name

A unique name identifying the connection.

Host

The host name or IP address of the Informix server.

Server

The Informix database server/intance name.

Database

The name of the database being connected to

Protocol

The communication protocol used with the server.

Service

The database server process that runs on the host.

User Name

The user name used to authenticate with the connection.

Password

The password used to authenticate the user.

Database Locale

The language locale of the database. Default: en_US.819

Client Locale

The language locale of the client. Default: en_US.CP1252

MySQL

Parameter

Description

Connection Name

A unique name identifying the connection.

Host

The host name or IP address of the database server

Port

The port number of the database server

Database Name

The name of the database being connected to

User Name

The user name used to authenticate with the connection.

Password

The password used to authenticate the user.

MySQL (Advanced Settings)

Parameter

Description

SSL KEY

The path name of the client private key file.">SSL KEY</label>

SSL CERT

The path name of the client public key certificate file.

CA Cert

The path name of the C ertificate Authority (CA) certificate file. This option, if used, must specify the same certificate used by the server.

SSL CA Path

The path name of the directory that contains trusted SSL CA certificate files.

SSL CIPHER

The list of permitted ciphers fo r SSL encryption.

Oracle

Parameter

Description

Connection Name

A unique name identifying the connection.

Host

The host name or IP address of the database server

Port

The port number of the database server

Service Name

Oracle Service Name

SID

Oracle SID

User Name

The user name used to authenticate with the connection.

Password

The password used to authenticate the user.

Use TNS

Create a connection using the TNS file on the application Server (on-prem only)

PostgreSQL

Parameter

Description

Connection Name

A unique name identifying the connection.

Host

The host name or IP address of the database server

Port

The port number of the database server

Database Name

The name of the database being connected to

User Name

The user name used to authenticate with the connection.

Password

The password used to authenticate the user.

Redshift

Parameter

Description

Connection Name

A unique name identifying the connection.

Host

The host name or IP address of the database server

Port

The port number of the database server

Database Name

Th e name of the database being connected to

User Name

The user name used to authenticate with the connection.

Password

The pass word used to authenticate the user.

Redshift (S3)

Parameter

Description

AWS Access Key

AWS access key ID

AWS Secret Access Key

AWS access key secret

AWS Region Name

Region in which the S3 bucket resides.

Bucket Name

AWS S3 bucket name

Salesforce

Parameter

Description

Connection Name

A unique name identifying the connection.

Connection Role

Source only

Instance

The hostname of the salesforce instance.

Security Token

The security token used to authenticate the username.

Organization ID

The organization ID to use for the Salesforce connection

User Name

The user name to use when authenticating with Salesforce

Password

The password to use when authenticating with Salesforce.

Salesforce (Advanved Settings)

Parameter

Description

Domain

The domain to use when connecting to sales force.(i.e. test) defaults to login.

Proxies

Optional map to proxy server.

Api Version

The salesforce api version to use

Max Bulk API Limit %

A threshold in percentages that is used to determine if bulk queries can run using the bulk api.

Max Rest API Limit %

A threshold in percentages that is used to determine if queries can run using the rest api.

Snowflake

Parameter

Description

Connection Name

A unique name identifying the connection.

Account Name

The host name or IP address of the database server

Warehouse Name

The por t number of the database server

Database Name

The name of the database being connected to

User Name

The user name used to aut henticate with the connection.

Password

The password used to authenticate the user.

Snowflake (Advanced Settings)

Parameter

Description

Authenticator

For OKTA https://<okta_acct_name>.okta.com (i.e. the URL endpoint for Okta) to authenticate through native OKTA.

Snowflake Region

Region ID of the Snowflake Region where your account is located. By default, the region is US West. If your account is located in a region other than US West, the region ID must be explicitly specified.

Role Name

Name of the default role to use.

Snowflake (S3)

Parameter

Description

AWS Access Key

AWS access key ID

AWS Secret Access Key

AWS access key secret

AWS Region Name

Region in which the S3 bucket resides.

Bucket Name

AWS S3 bucket name

SQL Server

Parameter

Description

Connection Name

A unique name identifying the connection.

Host

The host name or IP address of the database server

Port

The port number o f the database server

Instance Name

The database instance name to connect to

Database Name

The name of the database being connected to

User Name

The user name used to authenticate with the connection.

Password

The password used to authenticate the user.

The provided from in JSON format. Paste full contents of JSON file.

Google
service account credentials
Connections
add-button
run-button