skip to main content
Using Hybrid Data Pipeline : Creating data sources with the Web UI : Amazon Redshift parameters
  

Try Now

Amazon Redshift parameters

The following tables describe parameters available on the tabs of an Amazon Redshift Data Source dialog:
*General tab
*Security tab
*OData tab
*Advanced tab

General tab

Click the thumbnail to view the screen. Required fields are marked with an asterisk.
General tab of the Amazon Redshift data source setup dialogGeneral tab of the Amazon Redshift data source setup dialog
Table 2. General tab connection parameters for Amazon Redshift
Field
Description
Data Source Name*
A unique name for the data source. Data source names can contain only alphanumeric characters, underscores, and dashes.
Description
A general description of the data source.
User Id
The login credentials for your Amazon Redshift server.
Hybrid Data Pipeline uses this information to connect to the data store. The administrator of the server must grant permission to a user with these credentials to access the data store and the target data.
Note: You can save the Data Source definition without specifying the login credentials. In that case, when you test the Data source connection, you will be prompted to specify these details. Applications using the connectivity service will have to supply the data store credentials (if they are not saved in the Data Source) in addition to the Data Source name and the credentials for the Hybrid Data Pipeline account.
Password
A case-sensitive password that is used to connect to your Amazon Redshift database. A password is required if user ID/password authentication is enabled on your database. Contact your system administrator to obtain your password.
Note: By default, the password is encrypted.
By default, the characters in the Password field you type are not shown. If you want the password to be displayed in clear text, click the eye Watchful eye password icon icon. Click the icon again to conceal the password.
Server Name*
Specifies either the IP address in IPv4 or IPv6 format, or a combination of the two, or the server name (if your network supports named servers) of the primary database server, for example, RedshiftServer or 122.23.15.12
Valid Values:
server_name | IP_address
where:
server_name
is the name of the server to which you want to connect.
IP_address
is the IP address of the server to which you want to connect.
Port Number
The port number of the Amazon Redshift server.
Database*
The name of the database that is running on the database server.

Security tab

Click the thumbnail to view the screen. Required fields are marked with an asterisk.
Security tab of the Amazon Redshift data source setup dialogSecurity tab of the Amazon Redshift data source setup dialog
Table 3. Security tab Connection Parameters for Amazon Redshift
Field
Description
Encryption Method
Determines whether data is encrypted and decrypted when transmitted over the network between the Hybrid Data Pipeline connectivity service and the database server.
Valid Values:
noEncryption | SSL | requestSSL
If set to noEncryption, data is not encrypted or decrypted.
If set to SSL, data is encrypted using SSL. If the database server does not support SSL, the connection fails and the Hybrid Data Pipeline connectivity service throws an exception.
If set to requestSSL, the login request and data is encrypted using SSL. If the database server does not support SSL, the connectivity service establishes an unencrypted connection.
*When SSL is enabled, the following properties also apply:
Host Name In Certificate
ValidateServerCertificate
Crypto Protocol Version
Default: SSL
Crypto Protocol Version
Specifies a protocol version or a comma-separated list of the protocol versions that can be used in creating an SSL connection to the data source. If the protocol (or none of the protocols) is not supported by the database server, the connection fails and the connectivity service returns an error.
Valid Values:
cryptographic_protocol [[, cryptographic_protocol ]...]
where:
cryptographic_protocol
is one of the following cryptographic protocols:
TLSv1 | TLSv1.1 | TLSv1.2
The client must send the highest version that it supports in the client hello.
Note: Good security practices recommend using TLSv1.2 if your data source supports that protocol version, due to known vulnerabilities in the earlier protocols.
Example
Your security environment specifies that you can use TLSv1.1 and TLSv1.2. When you enter the following values, the connectivity service sends TLSv1.2 to the server first.
TLSv1.1,TLSv1.2
Default: TLSv1, TLSv1.1, TLSv1.2
Host Name In Certificate
Specifies a host name for certificate validation when validation is enabled (Validate Server Certificate=ON). This optional parameter provides additional security against man-in-the-middle (MITM) attacks by ensuring that the server that the Hybrid Data Pipeline connectivity service is connecting to is the server that was requested.
Valid Values:
host_name | #SERVERNAME#
where host_name is a valid host name.
If host_name is specified, the Hybrid Data Pipeline connectivity service compares the specified host name to the DNSName value of the SubjectAlternativeName in the certificate. If a DNSName value does not exist in the SubjectAlternativeName or if the certificate does not have a SubjectAlternativeName, the Hybrid Data Pipeline connectivity service compares the host name with the Common Name (CN) part of the certificate’s Subject name. If the values do not match, the connection fails and the connectivity service throws an exception.
If #SERVERNAME# is specified, the Hybrid Data Pipeline connectivity service compares the server name that is specified in the connection URL or data source of the connection to the DNSName value of the SubjectAlternativeName in the certificate. If a DNSName value does not exist in the SubjectAlternativeName or if the certificate does not have a SubjectAlternativeName, the Hybrid Data Pipeline connectivity service compares the host name to the CN part of the certificate’s Subject name. If the values do not match, the connection fails and the connectivity service throws an exception. If multiple CN parts are present, the connectivity service validates the host name against each CN part. If any one validation succeeds, a connection is established.
Default: Empty string
Validate Server Certificate
Determines whether the Hybrid Data Pipeline connectivity service validates the certificate that is sent by the database server when SSL encryption is enabled (Encryption Method=SSL). When using SSL server authentication, any certificate that is sent by the server must be issued by a trusted Certificate Authority (CA). Allowing the Hybrid Data Pipeline connectivity service to trust any certificate that is returned from the server even if the issuer is not a trusted CA is useful in test environments because it eliminates the need to specify truststore information on each client in the test environment.
Valid Values:
ON | OFF
If set to ON, the Hybrid Data Pipeline connectivity service validates the certificate that is sent by the database server. Any certificate from the server must be issued by a trusted CA in the truststore file. If the Host Name In Certificate parameter is specified, the Hybrid Data Pipeline connectivity service also validates the certificate using a host name. The Host Name In Certificate parameter is optional and provides additional security against man-in-the-middle (MITM) attacks by ensuring that the server the connectivity service is connecting to is the server that was requested.
If set to OFF, the Hybrid Data Pipeline connectivity service does not validate the certificate that is sent by the database server. The connectivity service ignores any truststore information that is specified by the Java system properties. Truststore information is specified using Java system properties.
Default: ON

OData tab

Click the thumbnail to view the screen. Required fields are marked with an asterisk.
OData tab of the Amazon Redshift data source setup dialogOData tab of the Amazon Redshift data source setup dialog
Table 4. OData tab connection parameters for Amazon Redshift
Field
Description
OData Version
Enables you to choose from the supported OData versions. OData configuration made with one OData version will not work if you switch to a different OData version. If you want to maintain the data source with different OData versions, you must create different data sources for each of them.
OData Access URI
Specifies the base URI for the OData feed to access the data source, for example, https://example.com:8443/api/odata4/<datasourcename>. You can copy the URI and paste it into your application's OData configuration.
The URI contains the case-insensitive name of the data source to connect to, and the query that you want to execute. This URI is the OData Service Root URI for the OData feed. The Service Document for the data source is returned by issuing a GET request to the data source's service root.
The OData Service Document returns the names of the entities exposed by the Data Source OData service. To get details such as the properties of the entities exposed, the data types for those properties and the relationships between entities, the Service Metadata Document can be fetched by adding /$metadata to the service root URI.
Schema Map
Enables OData support. If a schema map is not defined, the OData API cannot be used to access the data store using this data source definition. Use the Configure Schema editor to select the tables/columns to expose through OData.
Page Size
Determines the number of entities returned on each page for paging controlled on the server side. On the client side, requests can use the $top and $skip parameters to control paging. In most cases, server side paging works well for large data sets. Client side pagination works best with a smaller data sets where it is not as expensive to fetch subsequent pages.
Valid Values: 0 | n
where n is an integer from 1 to 10000.
When set to 0, the server default of 2000 is used.
Default: 0
Refresh Result
Controls what happens when you fetch the first page of a cached result when using Client Side Paging. Skip must be omitted or set to 0. You can use the cached copy of that first page, or you can re-execute the query to get a new result, discarding the previously cached result. Re-executing the query is useful when the data being fetched may change between two requests for the first page. Using the cached result is useful if you are paging back and forth through results that are not expected to change.
Valid Values:
When set to 0, the OData service caches the first page of results.
When set to 1, the OData service re-executes the query.
Default: 1
Inline Count Mode
Specifies how the connectivity service satisfies requests that include the $count parameter when it is set to true (for OData version 4) or the $inlinecount parameter when it is set to allpages (for OData version 2). These requests require the connectivity service to include the total number of entities that are defined by the OData query request. The count must be included in the first page in server-driven paging and must be included in every page when using client-driven paging.
The optimal setting depends on the data store and the size of results. The OData service can run a separate query using the count(*) aggregate to get the count, before running the query used to generate the entities. In very large results, this approach can often lead to the first page being returned faster. Alternatively, the OData service can fetch the entire result before returning the first page. This approach works well for small results and for data stores that cannot optimize the count(*) aggregate; however, it may have a longer initial response time for the first page if the result is large.
Valid Values:
When set to 1, the connectivity service runs a separate count(*) aggregate query to get the count of entities before executing the query to return results. In very large results, this approach can often lead to the first page being returned faster.
When set to 2, the connectivity service fetches all entities before returning the first page. For small results, this approach is always faster. However, the initial response time for the first page may be longer if the result is large.
Default: 1
Top Mode
Indicates how requests typically use $top and $skip for client side pagination, allowing the service to better anticipate how to process queries.
Valid Values:
Set to 0 when the application generally uses $top to limit the size of the result and rarely attempts to get additional entities by combining $top and $skip.
Set to 1 when the application uses $top as part of client-driven paging and generally combines $top and $skip to page through the result.
Default: 0
OData Read Only
Controls whether write operations can be performed on the OData service. Write operations generate a 405 Method Not Allowed response if this option is enabled.
Valid Values:
ON | OFF
When ON is selected, OData access is restricted to read-only mode.
When OFF is selected, write operations can be performed on the OData service.
Default: OFF

Advanced tab

Click the thumbnail to view the screen. Required fields are marked with an asterisk.
Advanced tab of the Amazon Redshift data source setup dialogAdvanced tab of the Amazon Redshift data source setup dialog
Table 5. Advanced tab connection parameters for Amazon Redshift
Field
Description
Catalog Options
Determines which type of metadata information is included in result sets when an application calls DatabaseMetaData methods. To include multiple types of metatdata information, add the sum of the values that you want to include. In this case, specify 6 to query database catalogs for column information and to emulate getColumns() calls.
Valid Values:
2 | 4
If set to 2, the Hybrid Data Pipeline connectivity service queries database catalogs for column information.
If set to 4, a hint is provided to the connectivity service to emulate getColumns() calls using the ResultSetMetaData object instead of querying database catalogs for column information. Using emulation can improve performance because the SQL statement that is formulated by the emulation is less complex than the SQL statement that is formulated using getColumns(). The argument to getColumns() must evaluate to a single table. If it does not, because of a wildcard or null value, for example, the Hybrid Data Pipeline connectivity service reverts to the default behavior for getColumns() calls.
Default: 2
Extended Options
Specifies a semi-colon separated list of connection options and their values. Use this configuration option to set the value of undocumented connection options that are provided by Progress DataDirect technical support. You can include any valid connection option in the Extended Options string, for example:
Database=Server1;UndocumentedOption1=value[;UndocumentedOption2=value;]
If the Extended Options string contains option values that are also set in the setup dialog, the values of the options specified in the Extended Options string take precedence.
Valid Values:
string
Default:none
Initialization String
A semicolon delimited set of commands to be executed on the data store after Hybrid Data Pipeline has established and performed all initialization for the connection. If the execution of a SQL command fails, the connection attempt also fails and Hybrid Data Pipeline returns an error indicating which SQL commands failed.
Syntax:
command[[; command]...]
Where:
command
is a SQL command. Multiple commands must be separated by semicolons. In addition, if this property is specified in a connection URL, the entire value must be enclosed in parentheses when multiple commands are specified. For example, assuming a schema name of SFORCE:
InitializationString=(REFRESH SCHEMA SFORCE)
The default is an empty string.
Login Timeout
The amount of time, in seconds, that the Hybrid Data Pipeline connectivity service waits for a connection to be established before timing out the connection request.
Valid Values:
0 | x
where x is a positive integer that represents a number of seconds.
If set to 0, the connectivity service does not time out a connection request.
If set to x, the connectivity service waits for the specified number of seconds before returning control to the application and throwing a timeout exception.
Default: 30
Max Pooled Statements
The maximum number of prepared statements to cache for this connection. If the value of this property is set to 20, the connectivity service caches the last 20 prepared statements that are created by the application.
Query Timeout
Sets the default query timeout (in seconds) for all statements created by a connection.
Valid Values:
-1 | 0 | x
If set to -1, the query timeout functionality is disabled. The Hybrid Data Pipeline connectivity service silently ignores calls to the Statement.setQueryTimeout() method.
If set to 0, the default query timeout is infinite (the query does not time out).
If set to x, the Hybrid Data Pipeline connectivity service uses the value as the default timeout for any statement that is created by the connection. To override the default timeout value set by this connection option, call the Statement.setQueryTimeout() method to set a timeout value for a particular statement.
Default: 0
Resultset Meta Data Options
Determines whether the Hybrid Data Pipeline connectivity service returns table name information in the ResultSet metadata for Select statements.
Valid Values:
0 | 1
If set to 0 and the ResultSetMetaData.getTableName() method is called, the Hybrid Data Pipeline connectivity service does not perform additional processing to determine the correct table name for each column in the result set. The getTableName() method may return an empty string for each column in the result set.
If set to 1 and the ResultSetMetaData.getTableName() method is called, the Hybrid Data Pipeline connectivity service performs additional processing to determine the correct table name for each column in the result set. The connectivity service returns schema name and catalog name information when the ResultSetMetaData.getSchemaName() and ResultSetMetaData.getCatalogName() methods are called if the Hybrid Data Pipeline connectivity service can determine that information.
Default: 0
Metadata Exposed Schemas
Restricts the metadata exposed by Hybrid Data Pipeline to a single schema. The metadata exposed in the SQL Editor, the Configure Schema Editor, and third party applications will be limited to the specified schema. JDBC, OData, and ODBC metadata calls will also be restricted. In addition, calls made with the Schema API will be limited to the specified schema.
Warning: This functionality should not be regarded as a security measure. While the Metadata Exposed Schemas option restricts the metadata exposed by Hybrid Data Pipeline to a single schema, it does not prevent queries against other schemas on the backend data store. As a matter of best practice, permissions should be set on the backend data store to control the ability of users to query data.
Valid Values
<schema>
Where:
<schema>
is the name of a valid schema on the backend data store.
Default: No schema is specified. Therefore, all schemas are exposed.
See the steps for:
How to create a data source in the Web UI