Data source connectivity

Using Pentaho Data Storage Optimizer

Version
10.1.x
Audience
anonymous
Part Number
MK-95PDSO000-02

The following table contains the supported data sources and respective requirements to connect with Data Storage Optimizer.

Data source Requirements
AWS S3
  • AWS region where the S3 bucket was created
  • Access key and secret access key
  • Read-only permissions to the S3 bucket
Azure Blob Storage
  • Account Fully Qualified Domain Name (FQDN)
  • Client ID and client key
  • authTokenEndpoint
HCP
  • AWS region where the S3 bucket was created
  • Access key and secret access key
  • Read-only permissions to the S3 bucket
HDFS
  • Hadoop version 2.7.2 and later
  • URI should provide a hostname and share folder details
  • Path of the directory that needs to be scanned
  • Read-only access to the directory
SharePoint and OneDrive
  • Application (client) ID, Directory (tenant ID), and clientSecret from a registered app on the Azure portal
  • Delegated permissions and Application permissions in the registered app
  • Read-only permissions to the SharePoint and OneDrive sites
SMB/CIFS
  • URI should provide host name and share folder details
  • Username and password to access the SMB/CIFS share directory
  • Path of directory that needs to be scanned
  • Read-only access is required