Skip to main content
Hitachi Vantara Lumada and Pentaho Documentation

Table Input

The Table Input step reads information from a connected database using SQL statements. Basic SQL statements can be generated automatically by clicking the Get SQL select statement button.

AEL considerations

When using the Table Input step with Adaptive Execution Layer, consider the following factors for Hive and Impala database connections, which can affect performance and results.

  • The following options in the step are not supported:
    • Enable Lazy conversion.
    • Execute for each row.
    • Pooling and Clustering database connection options.
  • Structures, arrays, and user-defined data types are not supported.
  • Because of limitations for CDS Powered by Apache Spark in CDH 6.x, AEL does not support Hive or Impala in YARN mode. If you would like specific information, see the Cloudera documentation:
    • When communicating with a Hive metastore, Spark SQL does not respect Sentry ACLs.
    • Using the JDBC Datasource API to access Hive or Impala is not supported.
    • Using the ORC file format is not supported.
    • Reading Hive tables containing data files in the ORC format from Spark applications is not supported. As an alternative, you can use the Parquet data format for columnar data using Impala.
  • You can use Hortonworks to access Hive managed tables in ORC format, as described in Configuring the AEL daemon for the Hive Warehouse Connector on your Hortonworks cluster.
  • You can use AEL with managed tables in Hive to execute joined Table input to Table output transformation steps. However, you must follow a specific workflow for correct data processing. For details, see Using Table input to Table output steps with AEL for managed tables in Hive.
  • If you are using AEL with external tables in Hive, then use the Hadoop input and Hadoop output steps to process tabular data.

Connect to a Hive database

When using the Table input and Table output steps, you can connect to Hive in one of two ways to achieve the best processing rate for small and large tables within the same cluster:

Connect to an Impala database

We support Impala versions 2.2.x. Before you can use the Table Input step to connect to an Impala database through AEL, you must download and install the Cloudera Impala driver.

Perform the following steps to download and install the Cloudera Impala driver:


  1. Stop the AEL daemon.

  2. Go to, select Downloads and click Impala JDBC Driver Downloads.

  3. Select Impala JDBC Connector 2.5.42 from the menu and follow the site's instructions for downloading.

    A ZIP file containing the Impala_jdbc_2.5.42 driver is downloaded.
  4. Unzip the file to a local folder.

    The contents of the ZIP file are extracted to the folder. The unpacked contents include a documentation folder and two ZIP files. You only need the
  5. Open the ClouderaImpalaJDBC-2.5.42 folder and unzip the file to a local folder.

    The associated JAR files are extracted from the ZIP file.
  6. Copy all the JAR files, except log4j-1.2.14.jar, to the pentaho/design-tools/data-integration/adaptive-execution/extra folder.

    CautionThe log4j-1.2.14.jar file should not be copied as it is already present and will cause conflicts.
  7. Save and close the file.

    The Cloudera Impala driver is installed.
  8. Restart the AEL daemon.


You can now use the Table Input step to connect to an Impala database through AEL


The following fields are general to this transformation step:

Step nameSpecify the unique name of the Table Input step on the canvas. You can customize the name or leave it as the default.
ConnectionUse the list to select the name of an existing database connection. Select the connection according to the database you are accessing.
  • If you are using the following database types with Spark on AEL, select one of the following:
    • Cloudera Impala: Select if you are using an Impala database.
    • Hadoop Hive 2/3: Select if you are using small unmanaged tables in Hive on a secure HDP cluster, or if you are using Hive tables on an Amazon EMR cluster.
    • Hive Warehouse Connector: Select if you are using large unmanaged tables in Hive, or if you are using managed Hive tables on a secure HDP cluster.
  • For all others, select the connection according to the relational database type.

If you do not have an existing connection, click New or Wizard. If you need to modify an existing connection, click Edit. See Define Data Connections for instructions.


The Table Input step has the following options:

SQLSpecify a SQL statement to read information from the connected database. You can also click the Get SQL select statement button to browse tables and automatically generate a basic select statement.
Store column info in step metaSelect this option to use the cached metadata stored in the KTR without making a database connection to query the table.
NoteIf you are using Spark as your processing engine, select this option. (Required)
Enable lazy conversionSelect this option to enable the lazy conversion algorithm. When selected, lazy conversion avoids unnecessary data type conversions when possible, which can significantly improve performance.
Replace variables in script?Select this option to replace variables in the script. This feature provides testing capabilities with or without variable substitutions.
Insert data from stepSpecify the input step name where PDI can expect information to come from. This information can then be inserted into the SQL statement. The locator where PDI inserts information is indicated with a question mark: ?.
Execute for each row?Select this option to execute the query for each individual row.
Limit sizeSpecify the number of lines to read from the database. A value of zero (0) indicates to read all lines.
Preview (button)Click Preview to open a new window and view an execution log derived from a temporary transformation with two steps: the Table Input step and the Dummy step. To see the log, click Logs in the Preview window that opens.


Below is an SQL statement:

SELECT * FROM customers WHERE changed_date BETWEEN ? AND ?

This SQL statement requests two calendar dates, to create a range, that are read from the Insert data from step option. The target date range can be provided using the Get System Info step. For example, if you want to read all customers that have had their data changed yesterday, you can get a target range for yesterday and read the customer data.

Metadata injection support

You can use the Metadata Injection supported fields with the ETL Metadata Injection step to pass metadata to your transformation at runtime. The following fields of the Table Input step support metadata injection:

  • SQL
  • Limit size
  • Execute for each row
  • Replace variables in script
  • Enable lazy conversion
  • Cached row meta
  • Connection