SAP HANA data point is used to configure connectivity to SAP HANA database. For each instance of SAP HANA database, a separate data point will have to be created. Multiple schemas within the same SAP HANA instance can be added in single SAP HANA data point. The SAP HANA data point will be associated to any SAP HANA data object created and data flows defining SAP HANA as the native processing platform. 

To work with SAP HANA Data Point, follow below steps:

Step I: Create a New Data Point

Step II: Provide connection details

1. To connect to SAP HANA database, following details need to be provided in the Properties tab.

  • Host: Specify the Hostname or the IP address of the SAP HANA system.
  • Port: Specify the port associated to the SAP HANA system.
  • App User: Specify the user id that will be used to connect to the SAP HANA system and make sure the User have necessary privileges to access the data in the schemas.
  • Password: Specify the password for the App User. To use the project parameter for the password, check the Use Project Parameters option, and you can view and select the required Project Parameter from the Password drop-down. 
  • Jdbc Options: Specify the options that should be used along with JDBC URL used to connect to SAP HANA server.
    For example, following details are provided in JDBC Options to connect to SAP HANA: user=diyotta, password=****, db=TEST_DB
  • Database: Collection of information that is organized so that it can be easily accessed, managed and updated.

Mandatory field names are suffixed with *. To establish the connection, provide all the mandatory property field values.
  • If Default agent is assigned to the Project then automatically, the Default agent will be associated with the new Data point created.
  • If Default agent is not assigned to the Project then, no agent will be assigned automatically and appropriate agent needs to be assigned to the data point.
  • When connecting to the Agent server then, the agent installation user should have appropriate privilege to access the path where file will be placed.
  • When connecting to the remote server then, the firewall needs to be opened from the Agent server to it and user specified to connect should have appropriate privilege to access the path where file will be placed.

2. Assign Agent: To assign or change the associated agent click Change. The Change Agent window appears and displays the list of available Agents. From the list, click on required Agent Name.

Note: To search for a specific agent, enter the keyword in the search bar, and the window displays the search result list. Select the required agent and click Ok.

Step III: Test the data point connection

  • To validate that the data point is able to connect to the SAP HANA data point database using the details provided, refer Test Data Point Connection.

Step IV: Enter the Schema details

Manage the required schemas to be accessed through the Schemas tab. Multiple schemas can be added here.

Note: To work with the application its mandatory to have at least one schema connection, by default the Default schema is associated with the new data point created.

1. To add a new schema, click Add(+).

New entry for schema is generated and success message is displayed at the bottom right corner of the screen.

  • The "Name" field is a friendly name that can be assigned to the schema for easy identification. This friendly name is displayed when a schema needs to be chosen from the data point and when the schema association with other components is displayed.
  • Provide the physical name of the database in the "Database" field. When clicked on the entry in the "Database" field a drop-down appears with the list of databases in the system. As you enter the database keyword, the drop-down shows the specific database. The database name can either be selected from this drop-down list or you can manually enter.
  • To view the database drop-down, it is a prerequisite to test the connection. For more information, refer Test Data Point Connection.
  • Provide the physical name of the schema in the "Schema" field. The  SAP HANA schema name can manually entered. When clicked on the entry in the "Schema" field a drop-down appears with the list of schemas in the system. As you enter the schema keyword, the drop-down shows the specific schema. The schema name can either be selected from this drop-down list or you can manually enter.
To search for a specific schema, enter the keyword in the search bar, and the page displays the related schemas.


  • Following operations are allowed on the schema entries: Add, Cut, Copy, Paste, Up, Down, Delete, and Search.
  • From the list of schemas, multiple schemas can be selected and we can perform/apply these operations.

Step V: Save the data point 


Step VI: Modify the configured Extract properties 

When moving data from one system to another the data is extracted from the source system, moved over the network and loaded into the target system. The SQLs and commands generated during execution of the jobs to extract the data are generated based on the properties defined for these. The properties associated with the extraction of data should depend on the format, performance and variety of the data being moved. These properties vary based on the environment and the type of the system. Diyotta comes with default properties that covers most of the known scenarios.

To modify these properties, refer Editing Extract Properties in SAP HANA Data Point.

  • The default values for extract properties can be configured in the Admin module and these properties reflect in the Studio module.
  • The extract properties set in data point are by default used in the source and target instance of the data flow.
  • It is a good practice to set the extract properties as per the company standards in the data point.
  • However, if needed any specific property can be overridden in the data flow.