Skip to main content

Pentaho+ documentation has moved!

The new product documentation portal is here. Check it out now at docs.hitachivantara.com

 

Hitachi Vantara Lumada and Pentaho Documentation

Install

This article covers the installation of Lumada Data Catalog onto a Kubernetes cluster, using Helm and Hitachi Vantara owned Docker images.

Before installing, you must have:

  • Reviewed the system requirements for Data Catalog.
  • Knowledge of your organization’s networking environment.
  • The root permissions for your designated server.
  • The ability to connect to your organization’s data sources.

Based on the deployment pattern your organization will use, you can also set up standalone remote agents. A remote agent serves the same purpose as a local agent, and is set up on an edge node of a separate Hadoop/Spark cluster.

The installation process includes:

By the end of these steps, the following Data Catalog components will be set up on your designated Kubernetes cluster:

Component Description
LDC Application ServerRuns the Data Catalog application
Agent (Local agent)Runs jobs executed by the application server
MongoDBDatabase backing the application
KeycloakOpen Source Identity Provider (IDP)
MinIOObject storage (used for debug purposes only)
Spark History ServerHelps store spark job logs on an MinIO or S3 bucket (by default this will be in the MinIO component included in set up)
REST ServerProvides several API calls that customers can use to interact with Data Catalog, and documentation covering these supported API calls
Data Catalog upgrade paths

You can upgrade Data Catalog from previous releases as shown in the table below.

From releaseTo releaseProcedure
7.27.3Upgrade with Helm
7.1.17.3Upgrade with Helm
6.17.3Contact the Hitachi Vantara Lumada and Pentaho Support Portal