4.5.0 archive

Note: This version of Hybrid Data Pipeline has reached end of life. These release notes are for reference purposes.

Resolved Issues

The following issues have been resolved. An asterisk (*) indicates an issue that was resolved in a software patch subsequent to the GA release.

Issue HDP-3974 Installation fails when choosing a unicode external database*
When a unicode external database was selected during the installation process, the Hybrid Data Pipeline server failed to install. This fix is available in build

Issue HDP-3989 Validate Server Certificate persistence in the Web UI*
The Web UI was not persisting the value of the Validate Server Certificate parameter after it had been set to OFF. After exiting the data source and returning back, this resulted in the test connection failing. This fix is available in build

Issue HDP-3785 Data source password replacing plus sign (+) with space*
When creating a password for a MySQL CE data source in the Web UI, the plus sign (+) was incorrectly being replaced with a space. This fix is available in build

Issue HDP-3878 OData model creation failure*
OData model creation was failing when the connectivity service was building an OData model from a very large database. Additionally, if unable to read metadata from unique or unusual tables, the creation of the OData model would result in either no rows returned or only partial rows returned. Hybrid Data Pipeline now builds the OData model from the tables selected to be in the model, as opposed to all the tables in the database. This enhancement is available in build



Hybrid Data Pipeline now supports multitenancy. Multitenancy allows a system administrator to isolate groups of users, such as organizations or departments, that are being hosted through the Hybrid Data Pipeline service. The provider maintains a physical instance of Hybrid Data Pipeline, while each tenant (group of users) is provided with its own logical instance of the service. In a multitenant environment, the default system tenant contains multiple child tenants. The user accounts that reside in one tenant are isolated from those in other tenants.

Data source sharing

Hybrid Data Pipeline now supports data source sharing via the Data Sources API. Data source owners can now share data sources with other users. Standard users can share data sources with individual user accounts. Administrators can share data sources with tenants and individual user accounts. Data source sharing allows administrators to provision users for limited or query-only access to Hybrid Data Pipeline resources.

Third-party JDBC support and validation tool

Hybrid Data Pipeline support for third-party JDBC drivers is now GA. Administrators can use a command line validation tool to determine whether a third-party JDBC driver will work with the Hybrid Data Pipeline server and On-Premises Connector. If validated, a third-party driver can be used to support OData, JDBC, and ODBC connectivity in the Hybrid Data Pipeline environment. Once the driver is integrated with the Hybrid Data Pipeline environment, users can create Hybrid Data Pipeline data sources for the backend data store supported by the third-party JDBC driver.

IP address whitelists

Administrators can now restrict access to Hybrid Data Pipeline by creating an IP address whitelist to determine which IP addresses (either individual IP addresses or a range of IP addresses) can access resources such as the Data Sources API, the Users API, and the Web UI. IP address whitelists can be implemented at system, tenant, and user levels.

Web UI
  • The Web UI has been refreshed with modern look and feel to provide an improved user experience. As part of the refresh, the Web UI URL has been changed to http(s)://<servername>:<portnumber>/hdpui.
  • The OData Configure Schema editor has been enhanced and now provides a better way to configure an OData schema map.
  • The process for creating Google Analytics data sources has also been improved. 
SQL Server data store

Hybrid Data Pipeline now supports the following features.

  • Transparent connectivity to Microsoft Azure Synapse Analytics (formerly Microsoft Azure SQL Data Warehouse) and Microsoft Analytics Platform System data sources
  • Always On Availability Groups via the Multi-Subnet Failover, Application Intent, and Server Name options
  • Azure Active Directory authentication (Azure AD authentication) via the Authentication Method, User, Password, Server Name, and Port Number options
Exporting non-relational data source files

The Data Source API now supports operations to export the relational map files for non-relational data sources. When a data source is created for a web service such as Salesforce, Hybrid Data Pipeline generates files to map the object model to a relational model. These files may be used to resolve issues that can arise when performing queries against data sources such as these.

Evaluation period

The evaluation period for Hybrid Data Pipeline has been changed from 90 to 30 days. 


Connect any application to any data source anywhere

Explore all DataDirect Connectors

A product specialist will be glad to get in touch with you

Contact Us