README
     Progress(R) DataDirect(R) Hybrid Data Pipeline
     Hybrid Data Pipeline Server
     Release 4.0
     October 2016

***********************************************************************
Copyright (c) 1990-2016 Progress Software Corporation and/or its 
subsidiaries or affiliates. All Rights Reserved.

***********************************************************************

CONTENTS

Release 4.0 Features
Product Components
Server Requirements
Installation Directory
Notes, Known Problems, and Restrictions
Documentation
Installed Files
Third-Party Acknowledgments


     Release 4.0 Features

Progress DataDirect Hybrid Data Pipeline is a data access platform that provides
simple, secure access to cloud and on-premises data sources, such as RDBMS, Big
Data, and NoSQL. Hybrid Data Pipeline allows business intelligence tools and
applications to use ODBC, JDBC, or OData to access data from supported
datasources. Hybrid Data Pipeline can be installed in the cloud or behind a
firewall. Hybrid Data Pipeline can then be configured to work with applications
and data sources in nearly any business environment.

* Supports access to over 20 data sources through a single, unified interface.

* Supports secure access to data on-premises or in the cloud.

* Can be hosted in the cloud or on premises.

* Supports SaaS, SQL, NoSQL, and Big Data data sources.

* Supports ODBC, JDBC, and OData APIs.


     Product Components

Progress DataDirect Hybrid Data Pipeline consists of four primary, separately
installed components.

* The Hybrid Data Pipeline server provides access to multiple data sources
  through a single, unified interface. The server can be hosted on premises or
  in the cloud.

* The On-Premises Connector enables the Hybrid Data Pipeline to establish a
  secure connection from the cloud to an on-premises data source.

* The ODBC driver enables ODBC applications to communicate to a data source
  through the Hybrid Data Pipeline server.

* The JDBC driver enables JDBC applications to communicate to a data source
  through the Hybrid Data Pipeline server.

In addition to these four primary components, Progress DataDirect also provides
a customized version of OpenAccess server. The OpenAccess server is a
connectivity layer required for Eloqua, Marketo, and HubSpot data sources in a
Hybrid Data Pipeline environment.


     Server Requirements

Hybrid Data Pipeline must be installed on a 64-bit Linux machine (2 core, 8 GB
RAM minimum) running one of the following operating systems:

* CentOS Linux x64, version 4.0 and higher

* Oracle Linux x64, version 4.0 and higher

* Red Hat Enterprise Linux x64, version 4.0 and higher

* SUSE Linux Enterprise Server, Linux x64, version 10.x, 11, 12, and 13


     Installation Directory

The default installation directory for the Hybrid Data Pipeline server is:

  /opt/Progress/DataDirect/Hybrid_Data_Pipeline/Hybrid_Server

  Note: If you do not have access to "/opt", your user's home directory will
  take the place of this directory.


     Notes, Known Problems, and Restrictions

All Data Sources
----------------
* It is recommended that Login Timeout not be disabled (set to 0) for a Data
  Source.

* Using setByte to set parameter values with the Hybrid Data Pipeline JDBC
  driver fails when the data source does not support the TINYINT SQL type. Use
  setShort or setInt to set the parameter value instead of setByte.

* When an administrator tries to add new users using the Add Users window, the
  the Password and Confirm Password fields occasionally do not appear properly
  in the popup window.

Google Analytics
----------------
* Validation message is not showing when a user enters a Start Date value less
  than the End Date value in Create/Update Google Analytics page.

* Once a Google Analytics OAuth profile is created for a specific Google
  account, changing the Google Account associated with the profile results in
  "the configuration options used to open the database do not match the options
  used to create the database" error being returned for any existing data
  sources.

HubSpot
-------
* Hubspot is not supported for the GA release of Hybrid Data Pipeline.

* HubSpot introduced a new limit on their API in July 2016, limiting API calls
  to 10 per second. However, you can try the request again after waiting 1
  second.

* We are investigating adding a connection option to automatically throttle API
  calls to deal with this HubSpot change. For more information, see the HubSpot
  forum post at
  http://integrate.hubspot.com/t/error-you-have-reached-your-secondly-limit/300.

OData
-----
* $expand only supports one level deep.
  For example, with the entity hierarchy:
  Customers
  |-- Orders
  | |-- OrderItems
  |-- Contacts

  The following queries are supported:
  Customers?$expand=Orders
  Customers?$expand=Contacts
  Customers?$expand=Orders,Contacts

  However, this query is not supported:
  Customers?$expand=Orders,OrderItems

  OrderItems is a second level entity with respect to Customers. To query Orders
  and OrderItems, the query must be rooted at Orders. For example:
  Orders?$expand=OrderItems
  Orders(id)?$expand=OrderItems

* When manually editing the ODataSchemaMap value, the table names and column
  names specified in the value are case-sensitive. The case of the table and
  column names must match the case of the tables and column names reported by
  the data source.
  Note: It is highly recommended that you use the OData Schema Editor to
  generate the value for the ODataSchemaMap data source option. The Schema
  Editor takes care of table and column name casing and other syntactic details.

* When using the substring function on properties that map to a CHAR column in
  the data source, it is data source dependent as to whether the substring
  function treats the trailing spaces as significant. When going against Oracle,
  the trailing spaces are preserved. When going against other data sources, the
  trailing spaces are discarded.

* The $expand clause is not supported with OpenEdge data sources.

* The day scalar function is not working when specified in a $filter clause when
  querying a DB2 data source.

Oracle Sales Cloud
------------------
* Create Mapping is not fully supported for the Oracle Sales Cloud data source.
  Typically, when editing a data source from the Data Sources page, a user would
  need to select "Force New" for Create Mapping under the Mapping tab to refresh
  a schema. However, this currently results in an input/output error. As a
  workaround, create a new data source with the desired configuration.

* External storage for processing large results is not currently supported for
  Oracle Sales Cloud. All processing currently takes place in memory. This
  primarily impacts queries with post processing options and limits the size of
  the query that can be successfully processed to the system resources available
  to the Hybrid Data Pipeline connectivity service.

* The drivers currently report ATTACHMENT type fields in the metadata but do not
  support retrieving data for these fields. These fields are set to NULL.

* Join queries between parent and child tables are not supported.

* Queries on child tables whose parent has a composite primary key are not
  supported. For example, the children of ACTIVITIES_ACTIVITYCONTACT and
  LEADS_PRODUCTS are not accessible.

* Queries on the children of relationship objects are not supported. For
  example, the children of ACCOUNTS_RELATIONSHIP, CONTACTS_RELATIONSHIP, and
  HOUSEHOLDS_RELATIONSHIP are not accessible.

* Queries on grandchildren with multiple sets of Parent IDs and Grand Parent IDs
  used in an OR clause are not supported. For example, the following query is
  not supported.
     select * from ACCOUNTS_ADDRESS_ADDRESSPURPOSE
        where (ACCOUNTS_PARTYNUMBER = 'OSC_12343' AND
               ACCOUNTS_ADDRESS_ADDRESSNUMBER = 'AUNA-2XZKGH')
           or (ACCOUNTS_PARTYNUMBER = 'OSC_12344' AND
               ACCOUNTS_ADDRESS_ADDRESSNUMBER = 'AUNA-2YZKGH')

* When querying documented objects like "CATALOGPRODUCTITEMS" and
  "CATEGORYPRODUCTITEMS", no more than 500 records are returned, even when
  more records may be present. This behavior is also seen with some custom
  objects. We are currently working with Oracle support to resolve this issue.

* A query on OPPORTUNITIES_CHILDREVENUE_PRODUCTS or LEADS_PRODUCTGROUPS with
  a filter on the primary key column returns 0 records even when more records
  are present. We are currently working with Oracle support to resolve this
  issue.

* Queries that contain subqueries returning more than 100 records are not
  supported. For example, the following query is not supported.
     select * from ACCOUNTS_ADDRESS
        where ACCOUNTS_PARTYNUMBER
           in (select top 101 PARTYNUMBER from ACCOUNTS)

* When you create custom objects, your Oracle Sales Cloud administrator must
  enable these objects for REST API access through Application Composer.
  Otherwise, you will not be able to query against these custom objects.

Oracle Service Cloud
--------------------
* When you create a custom object, your Oracle Service Cloud administrator must
  enable all four columns of the Object Fields tab of the Object Designer, or
  you cannot query against the custom objects.

* The initial connection when the relational map is created can take some time.
  It is even possible to receive an error "504: Gateway Timeout". When this
  happens, Hybrid Data Pipeline  continues to build the map in the background
  such that subsequent connection attempts are successful and have full access
  to the relational map.

Marketo
-------
* Marketo is not supported for the GA release of Hybrid Data Pipeline.

Microsoft Dynamics CRM
----------------------
* Testing has shown the following two errors from Microsoft Dynamics CRM Online
  when executing queries against the ImportData and TeamTemplate tables:
  - Attribute errortype on Entity ImportData is of type picklist but has Child
    Attributes Count 0
  - Attribute issystem on Entity TeamTemplate is of type bit but has Child
    Attributes Count 0
  Note: We have filed a case with Microsoft and are waiting to hear back about
  the cause of the issue.

* The initial on-premise connection when the relational map is created can take
  some time. It is even possible to receive an error "504: Gateway Timeout".
  When this happens, Hybrid Data Pipeline continues to build the map in the
  background such that subsequent connection attempts are successful and have
  full access to the relational map.

OpenEdge 10.2b
--------------
* Setting the MaxPooledStatements data source option in an OpenEdge data store
  to a value other than zero can cause statement not prepared errors to be
  returned in some situations.

SugarCRM
--------
* Data sources that are using the deprecated enableExportMode option will still
  see a problem until they are migrated to the new data source configuration.

* Data source connections by default now use Export Mode to communicate with the
  Sugar CRM server, providing increased performance when querying large sets of
  data. Bulk export mode causes NULL values for currency columns to be returned
  as the value 0. Because of this, there is no way to differentiate between a
  NULL value and 0, when operating in export mode. This can be a problem when
  using currency columns in the SQL statements, because Hybrid Data Pipeline
  must satisfy some filter conditions on queries, such as with operations like
  =, <>, >, >=, <, <=, IS NULL and IS NOT NULL. For example, suppose a currency
  column in a table in SugarCRM has 3 null values and 5 values that are 0. When
  a query is executed to return all NULL values (SELECT * FROM <table> WHERE
  <>currency column> IS NULL), then 3 rows are returned. However, if a query is
  executed to return all rows where the column performs an arithmetic operation
  (SELECT * FROM <table> WHERE <currency column> + 1 = 1), then all 8 records
  are returned because the 3 NULL values are seen as 0.

On-Premises Connector
---------------------
* See the hdpopcreadme.txt file for the latest notes when using the On-Premises
  Connector. This file is located in the installation directory for the
  On-Premises Connector.

JDBC Driver
-----------
* See the hdpjdbcreadme.txt file for the latest notes when accessing Hybrid Data
  Pipeline through the JDBC driver. This file is located in the installation
  directory for the JDBC driver.

ODBC Driver
-----------
* See the hdpodbcreadme.txt file for the latest notes when accessing Hybrid Data
  Pipeline through the ODBC driver. This file is located in the installation
  directory for the ODBC driver.


     Documentation

Hybrid Data Pipeline documentation consists of the following guides and readmes.

* PROGRESS DATADIRECT HYBRID DATA PIPELINE INSTALLATION GUIDE
  Available online at
  https://documentation.progress.com/output/DataDirect/hybridpipeinstall

* PROGRESS DATADIRECT HYBRID DATA PIPELINE QUICK START
  Available online at
  https://documentation.progress.com/output/DataDirect/hybridpipestart

* PROGRESS DATADIRECT HYBRID DATA PIPELINE USER'S GUIDE
  Available online at
  https://documentation.progress.com/output/DataDirect/hybridpipeline

* The Hybrid Data Pipeline readme file: hdpreadme.txt
  Installed file and available online at
  https://documentation.progress.com/output/DataDirect/hdpreadmes/hdpreadme.htm

* The On-Premise Connector readme file: hdpopcreadme.txt
  Installed file and available online at
  https://documentation.progress.com/output/DataDirect/hdpreadmes/hdpopcreadme.htm

* The JDBC Driver readme file: hdpjdbcreadme.txt
  Installed file and available online at
  https://documentation.progress.com/output/DataDirect/hdpreadmes/hdpjdbcreadme.htm

* The ODBC Driver readme file: hdpodbcreadme.txt
  Installed file and available online at
  https://documentation.progress.com/output/DataDirect/hdpreadmes/hdpodbcreadme.htm

* The OpenAccess Server readme file: hdpoaserverreadme.txt
  Installed file and available online at
  https://documentation.progress.com/output/DataDirect/hdpreadmes/hdpoaserverreadme.htm


     Installed Files

When you install the Hybrid Data Pipeline server, the installer creates logs and
scripts. These files can be used to evaluate and troubleshoot issues. These
files can be found in the following locations where INSTALL_DIR is the
installation directory for the Hybrid Data Pipeline server.

INSTALL_DIR/ddcloud/: 
---------------------
deploy.log                 Log file that provides deployment details

deploy.sh                  Shell script to update existing deployment or echo
                           all output to stdout

error.log                  Log file that provides list of server errors

getlogs.sh                 Shell script that creates a compressed tar file with
                           all server logs

setlogginglevel.sh         Shell script to specify level of detail written to 
                           log files

In addition, the installation process creates four configuration files that are
needed to integrate Hybrid Data Pipeline components.

INSTALL_DIR/redist/: 
--------------------
config.properties          File that contains branding information

ddcloud.pem                File for the self-signed certificate

ddcloudTrustStore.jks      The Java keystore

OnPremise.properties       File with server and port information


     Third-Party Acknowledgments

Third-party acknowledgments are listed on the following Web page:

https://www.progress.com/legal/hybridpipe-third-party


October 2016
~~~~~~~~~~~~
End of README