This is the multi-page printable view of this section. Click here to print.

Return to the regular view of this page.

OpenText Core Analytics Database

This topic provides an overview of OTCAD.

OpenText™ Core Analytics Database (OTCAD) is a managed database service that delivers a unified, high performance advanced analytics and machine learning platform in the cloud. The database gets automatically provisioned in an OpenText AWS account. OpenText manages the database hardware and environment. OTCAD is optimized for efficiency and query performance, ensuring you meet your service level agreements within your budget.

OTCAD enhances the speed and scalability of a database without the need to manage hardware. OTCAD is a data warehouse and data lake platform that enables organizations to deploy databases securely. To get access to OTCAD, contact your sales representative or contact us.

OTCAD UI tour

You can view and manage the provisioned database. After logging in to the OTCAD application, the home page appears. This page displays the three-dot menu and other SQL editor options as shown in this image:

OTCAD user interface

Option Description
a More options: Displays the More options button with options to access the SQL editor, Settings, Data pipelines, and the Integrations page.
b Stop: Displays the option to stop the database.
c History view: Displays the queries run previously.
d Favorite view: Displays the bookmarked SQL queries.
e Help: Displays OpenText™ Analytics Database user documentation.
f View profile: Displays the user profile and option to sign out of the application.

More options

When you log in to the application, the More options button is displayed with these options:

  1. SQL editor: Select this option to work with the SQL editor.
  2. Settings: Select this option to configure the database settings.
  3. Data pipeline: Select this option to work with data pipelines.
  4. Integrations: Select this option to display the Integrations page.

Database connection options

The Home page displays the option to stop an active database.

  • To stop OTCAD, select Stop. <--Or-->
  • To re-start an inactive database, select Start.

SQL Editor

Use OTCAD SQL editor to write, edit, and execute SQL statements in a query view. You can write and execute analytical queries without using any external tools.

You can use the SQL editor to:

  • gain valuable insights from your data without needing expertise to set up database connections.
  • write, execute, and save queries that need to be repeated.
  • view sample data returned from your queries.
  • validate the execution of SQL queries.
  • view data about the execution of SQL queries.
  • export queries into a spread sheet.
  • execute SQL commands for OTCAD operations, such as creation of schemas, tables, and so on.

Working with SQL editor

After logging in to the OTCAD application, the home page appears.

To launch the SQL editor, select SQL Editor in the three-dots menu. The Schema view displays with a list of available schemas and tables. You can either select a table or create an SQL query.

Create an SQL query

You create SQL queries either in the Home page or the Schema view.

Do one of the following:

  1. Select a table:

    • Select Create query in the Schema view. --Or-->
    • Select Create new query in the Home page.
  2. Enter your query in the SQL editor.

  3. Select Run.

  4. Enter the name of the query at the top.

  5. To view the schema details of a table, select View Schema. The table columns display along with the datatypes.

  6. To filter the SQL query results, select the filter icon. You can add a filter. search the results by the column name and the match type, and select Apply.

  7. To save the SQL query results, do the following: a. Select Export CSV. b. Enter the file name and select Export.

    You can export upto a maximum of 10 thousand records. To import data that exceeds 10 thousand records, modify the SQL query and export again.

History view

The History view tab stores the queries run previously. When you run a query, it appears in the Query History list. Queries of the last 2 weeks only are displayed.

To run a SQL query:

  1. Select .
  2. Select Run.

To remove a SQL query:

  1. Select .
  2. Select Remove.

Favorite view

You can bookmark SQL queries and run them again later in the Favorite view tab. Saved SQL queries display as a list of favorites next to the History view tab.

To bookmark a SQL query as a favorite:

  1. Select the SQL query.
  2. Select Run.
  3. Select Favorite.
  4. Enter a name and select Add.

To run a SQL query from the Favorites view:

  1. Select .
  2. Select Run. After successful execution of the query, the results are displayed in the SQL editor.

To rename a favorite SQL query:

  1. Select the SQL query.
  2. Select .
  3. Select Rename.
  4. Enter the new name and select Update.

To remove a favorite SQL query:

  1. Select the query.
  2. Select .
  3. Select Remove.
  4. When prompted for confirmation, select Remove.

1 - OpenText Core Analytics Database Settings Overview

This topic provides an overview of database configuration settings in OTCAD.

You can configure the default database configuration settings in OpenText™ Core Analytics Database (OTCAD).

  1. Log in to OTCAD.
  2. Select Settings from More options. The Settings Overview page opens. In this page, you can configure the following OTCAD settings:
    • Connection data: You can configure the default connection data settings in OTCAD.
    • CIDR: CIDR is fundamental to subnetting, the process of dividing a large network into smaller, more manageable subnetworks. By using different prefix lengths, you can create subnets with varying numbers of host addresses, enabling efficient use of IP address space.
    • Scaling: Database scaling is the ability of a database to obtain or reduce resources depending on the system's load, establishing a direct correlation between the allocation of resources and their demand. Scaling ensures consistency in the database server quality while maximizing resource usage.
  3. Select Connection Data.
  4. Copy this information to connect your database with other applications or clients:
    • Host DNS: The DNS host provides the authoritative nameservers that answer or resolve queries for your website. This is how your site is found on the web.
    • Host Port: The host port is a network port on a computer or device that is used to establish a connection with another device or service, often for communication or data transfer.
    • Database Name: The unique name of the database.
    • Refresh Token: The refresh token is a special type of token used in authentication flows, particularly in OAuth 2.0, to obtain new access tokens when the current ones expire.
    • Access Token: An access token is a security credential that allows an application or user to access protected resources on a server. It acts as a digital key, verifying the user's or application's identity and authorization level, and is used to grant access without requiring the user to repeatedly provide their credentials.
  5. Select Close to close the Connection Details page.

1.1 - Connection Data Settings

This topic provides an overview of connection data settings in OTCAD.

You can configure the default connection data settings in OTCAD. You can manage IP addresses with Classless Inter-Domain Routing (CIDR) and scale your plan to suit your requirements.

1.2 - CIDR

This topic provides an overview of managing IP addresses with Classless Inter-Domain Routing (CIDR) in OTCAD.

CIDR

Cloud infrastructure engineers use CIDR notation to manage IP address ranges for cloud-based networks. IP addresses enable resources in your Virtual Private Cloud (VPC) to communicate with each other and with resources over the Internet. CIDR notation is a way to represent an IP address and its network mask.

OpenText supports ranges of IP addresses using Classless Inter-Domain Routing (CIDR) notation. For example, 192.168.1.0/24 represents all IPv4 addresses in the range of 192.168.1.0 to 192.168.1.255.

The format of these addresses is as follows:

  • An individual IPv4 address is 32 bits, with 4 groups of up to 3 decimal digits. For example, 10.0.1.0.
  • An individual IPv6 address is 128 bits, with 8 groups of 4 hexadecimal digits. For example, 2001:0db8:85a3:0000:0000:8a2e:0370:7334.

Public IPv4 and IPv6 addresses

All subnets have an attribute that determines whether a network interface created in the subnet automatically receives a public IPv4 address. Therefore, when you launch an instance into a subnet that has this attribute enabled, a public IP address is assigned to the primary network interface that's created for the instance. You can control whether your instance receives a public IP address by doing the following:

  • Modifying the public IP addressing attribute of your subnet.
  • Enabling or disabling the public IP addressing feature during instance launch, which overrides the subnet's public IP addressing attribute.
  • Unassigning a public IP address from your instance after launch by managing the IP addresses associated with a network interface.

Public IPv6 addresses are IPv6 addresses that can be configured to remain private or configured to be reachable over the Internet.

Private IPv4 and IPv6 addresses

Private IPv4 addresses are not reachable over the Internet and can be used for communication between the instances in your VPC. When you launch an instance into a VPC, a primary private IP address from the IPv4 address range of the subnet is assigned to the primary network interface (for example, eth0) of the instance. Each instance is also given a private (internal) DNS hostname that resolves to the private IP address of the instance.

You cannot share private IPv6 addresses on the Internet. You can use a private IPv6 address if you want your private networks to support IPv6 and you have no intention of routing traffic from these addresses to the Internet. You can connect to the internet from a resource that has a private IPv6 address. But you must route traffic through a resource in another subnet with a public IPv6 address.

Configure CIDR settings

To configure CIDR settings:

  1. In the Settings Overview page, select View CIDR. The CIDR page displays.
  2. Select Private.
  3. To add an IP address, select Add IP.
  4. From the Value drop-down list, select either IPv4 or IPv6 depending on the IP address you want to add. Do one of the following: a. Enter a valid IPv4 or IPv6 address. b. Copy a valid IPv4 or IPv6 address from another source and select Paste.
  5. Select .

1.3 - Connection Data Settings

This topic provides an overview of scaling your database in OTCAD.

Scaling

Workload scaling enhances application performance and resource efficiency in dynamic environments. Scaling ensures that applications can handle varying workloads without performance degradation. OpenText provides the ability to automatically scale resources up or down based on real-time metrics, allowing organizations to respond quickly to changes in traffic. This elasticity not only improves user experience but also optimizes resource utilization, helping to minimize costs associated with underused or overprovisioned resources.

You cam manage compute to your database to improve the performance of complex analytic queries that have been running for a long time. Adding compute helps these queries run faster.

Configure the compute of your database to suit your query requirements and workload performance needs.

To manage compute for your database:

  1. In the Settings Overview page, select Scaling.
  2. Select View Scaling. The horizontal database scaling options are displayed. You can scale the database either manually or automatically. Do one of the following: a. To manually scale the database, select Manual. b. Slide across the slider horizontally to scale the database depending on the size - small, small-medium, medium, medium-large, or large. c. Select Apply changes. d. In the View scaling log area, click the Audit log and Scaling log tabs to view the log details. <-Or-> a. To automatically scale the database, select Auto. Select this option only if you need to automatically scale the database based on defined parameters. b. From the Scale up to drop-down list, select either Large or Medium depending on the size of your database. c. Use the CPU usage toggle button to scale out the database when the CPU usage exceeds a value in the range 1% to 100%. d. Use the Memory usage toggle button to scale out the database when the memory usage exceeeds a value in the range 1% to 100%. e. Select Apply changes. f. In the View scaling log area, click the Audit log and Scaling log tabs to view the log details. The Audit log provides information about the following:
    • Date & time
    • Event
    • Performed by The Scaling log provides information about the following:
    • Date & time
    • Scale event
    • Compute (units)
    • Triggered
    • Performed by

1.4 - Maintenance schedule

This topic provides an overview of the database maintenance schedule settings in OTCAD.

Maintenance schedule

The maintenance schedule enables the database administrator to specify a specific day of the week and time to schedule periodic maintenance tasks such as upgrades, patches, and so on.

During a maintenance activity, a banner appears at the top of the page after you sign in to the database. You can still sign in and continue working during this time. Only database administrators can configure the maintenance schedule. The database will be in read-only mode during a maintenance activity. By default, maintenance activities are scheduled every Sunday at 6 AM UTC.

To modify the maintenance schedule:

  1. Select Settings from More options.

  2. Select Maintenance schedule.

    The Maintenance schedule page appears, displaying the current database version and the date and time of the previous upgrade.

  3. In the Day of event list, select the day of the week to schedule the database maintenance activity.

  4. In the From and To lists, select the time interval to schedule the database maintenance activity.

  5. Select Save.

    The database is scheduled for maintenance on the designated day of the week and time.

    Changes to the maintenance schedule made at least one week in advance will only be considered. If you need to postpone the current maintenance schedule, contact the technical support team.

2 - Overview of Data pipelines

This topic provides an overview of data pipelines in OTCAD.

Data pipelines allow you to define where, when, and how you need to load data into OTCAD with minimal manual intervention.

2.1 - Data pipelines

This topic provides an overview of data pipelines in OTCAD.

Data pipeline

A data pipeline also known as a data loader is a declarative, automated way to process data continuously from external sources like Amazon S3 or Kafka. They allow you to define where, when, and how you need to load data into OTCAD with minimal manual intervention.

Data pipelines UI tour

You can view and manage data pipelines. After logging in to the OTCAD application, you land on the home page. Select the More options button and Data Pipeline. The Data pipeline page appears as shown in this image:

OTCAD user interface

Option Description
a Create a pipeline: Select this option to create a data pipeline.
b Filter by: Select this option to filter data by different criteria.
c Search pipeline name or schema: Search for a data pipeline based on the pipeline name or schema.
d Actions: Select in the Actions column to view pipeline details, edit, clone, pause or resume schedule, and delete a data pipeline.

This page displays the following information in the Overview card:

  • Total number of pipelines - The Total Pipelines card displays the total number of pipelines configured in the system.
  • Pipelines that failed execution - Select the Failed Execution list and the period for which you need to view the pipelines that failed to execute.
  • Active pipelines - The Active Pipelines card displays the total number of pipelines that are in the Active status.
  • From the Duration list, select the duration for which you need to view the data pipelines. You can choose from the following:
    • Last 24 hours
    • Last week
    • Last month
    • Last 6 months
    • Last year
    • Anytime

The Pipelines area displays the following information:

  • Pipeline name - The name of the data pipeline.
  • Created by - The user ID of the person who created the data pipeline.
  • Data source - The source location of the files that contain the data to be loaded.
  • Schema - The tables, columns (fields), data types, and relationships among different tables in the database.
  • Destination table - The database table where data is written after it has been processed or transformed from a source table or other data source.
  • Last run on - The timestamp at which the data pipeline was last run.
  • Last run state - Indicates the state of the data pipeline when it was last run. The values are:
    • Executed - The data pipelines that are executed successfully.
    • Failed - The data pipelines that have failed to execute.
    • Partially executed - The data pipelines that are executed partially.
    • Scheduled - The data pipelines that are scheduled to run at a specific duration.
  • Pipeline status - Indicates the present status of the data pipeline. The values are:
    • Active - The data pipelines that are active.
    • Inactive - The data pipelines that are inactive.
    • Paused - The data pipelines that are paused for execution.
  • Actions - Options to view pipeline details, edit, clone, pause or resume schedule, and delete a data pipeline.

Create a data pipeline

You can create a data pipeline using either the AWS object store or Kafka data source.

Create a data pipeline using AWS object store

AWS object store, primarily Amazon Simple Storage Service (S3) is a highly scalable, durable, and cost-effective cloud storage service for unstructured data. It stores data as objects in flat containers called buckets.

To create a data pipeline using AWS object store, do the following:

  1. In the Data pipelines page, select +Create a pipeline.

    The Create a pipeline page is displayed.

  2. In the Pipeline name field, enter the name of the pipeline.

  3. Select AWS object store.

  4. In the Access key ID field, enter your AWS account access key id.

  5. In the Secret access key field, enter your AWS account secret access key.

  6. From the Region list, select the region (geography) of the S3 bucket where the files are present.

  7. In the S3 Bucket/File/Folder path field, enter the name or the folder path where the files are present.

  8. Select the Data is encrypted option to specify the following parameters:

    • Select either AWS Key Management Service Key or Customer managed keys if you wish to encrypt and load data into the S3 bucket.
    • Select Encryption key ID.
  9. Select Next.

  10. In the Retry limit field, specify the number of times the system should attempt to retry a failed file load.

  11. In the Parameters field, specify the copy parameters. For more information, see Parameters.

  12. Select Next.

  13. From the Destination table list, select the destination table to which you need to load the data.

  14. Select Next.

  15. Specify the schedule at which the data pipeline needs to run. Do one of the following:

    • Select Schedule.

      • From the date pickers, select the Start date and End date.
      • In the Repeat every field, specify the duration at which the data pipeline needs to run.
      • From the Unit list, select the minute, hour, day, week, or month at which the data pipeline needs to run.
      • Select the On day option and specify the day on which the data pipeline needs to run.
      • Select the On option and specify the exact day and month on which the data pipeline needs to run.
      • Select the option Trigger when something is added to run the data pipeline when a file is added to the S3 bucket.
      • Enter the SQS credentials in the Access key ID, Secret access key, and Resource URL fields.

      <--Or-->

    • Select Execute once.

  16. Select Finish.

    The data pipeline is created and displayed in the Data pipelines page.

Create a data pipeline using Kafka data source

Define data pipelines to ingest real-time data through Apache Kafka, an open-source distributed real-time streaming platform by leveraging Kafka topics for efficient streaming and processing. A Kafka topic is a category or feed name for streams of data, similar to a table in a database. Consumers read data from Kafka topics, with topics being organized into partitions for parallel processing. Each message in a topic is a record with a key, value, and timestamp. Kafka topics are logs where messages are ordered by an offset within each partition.

To create a data pipeline using Kafka data source, do the following:

  1. In the Data pipelines page, select +Create a pipeline. The Create a pipeline page is displayed.

  2. In the Pipeline name field, enter the name of the pipeline.

  3. Select Kafka.

  4. In the Bootstrap servers field, enter the initial list of Kafka broker addresses that a Kafka client uses to connect to the Kafka cluster.

  5. Do one of the following:

    • Select SSL if your Kafka broker is configured with SSL.

      SSL is short for Secure Sockets Layer. It is a protocol that creates an encrypted link between a web server and a web browser to ensure that all data transmitted between them is confidential.

    • In the Client key field, enter the encrypted client key generated for the SSL certificate.

    • In the Password field, enter the password associated with the encrypted client key.

    • In the Client certificate field, enter the certificate used to authenticate the client with the Kafka broker.

    • In the CA certificate field, enter the certificate authority (CA) certificate used to validate the Kafka broker.

    <--Or-->

    • Select SASL if your Kafka broker is authenticated with Simple Authentication and Security Layer (SASL).

      SASL is a Kafka framework for authenticating clients and brokers, which can be used with or without TLS/SSL encryption. It allows Kafka to support different authentication mechanisms.

    • In the SASL mechanism list, select one of the following:

      • Plain - A simple username/password authentication mechanism used with TLS for encryption to implement secure authentication.
      • SCRAM - SHA - 256 - A secure password-based authentication method that uses a challenge-response protocol and the SHA-256 hashing algorithm to verify user credentials without sending the password in plain text over the network.
      • SCRAM - SHA - 512 - A secure authentication mechanism that uses the SHA-512 cryptographic hash function to verify a user's credentials in a "challenge-response" format, which prevents the password from being sent directly over the network.
      • In the Username field, enter a valid username.
      • In the Password field, enter the password for SASL authentication.
  6. Select Proceed.

  7. In the Define configuration area, specify the configuration settings for the data source.

  • In the Topic field, enter the Kafka topic. A Kafka topic is a logical grouping of messages, split into multiple partitions for parallelism.
  • In the Partition box, type or select the number of partitions for the Kafka topic. A partition is a sequence of records within a topic, stored on brokers and consumed independently.
  • In the Start offset box, type or select the incremental ID for each partition. The offset is a unique identifier for each record within a partition, used to track consumer progress.
  • Select +Add topic to add more topics and partitions.
  • Select one of the available parser options depending on the message type in the topic:
    • AVRO - An Avro schema registry is a crucial component in systems that utilize Apache Kafka and Avro for data serialization. Its primary purpose is to centralize the management and evolution of Avro schemas, providing a robust mechanism for ensuring data compatibility and governance in streaming data pipelines.

    • In the URL field, enter the schema registry URL.

    • In the Subject field, enter the subject information from the schema registry.

    • In the Version fieled, enter the version of the schema registry.

      <--Or-->

    • In the External Schema field, enter the schema of the AVRO message. Ensure that the schema is in JSON format.

    • JSON - JSON schema registry for Kafka provides a mechanism to define, manage, and enforce the structure of JSON data being produced to and consumed from Kafka topics. This ensures data consistency and compatibility, especially in distributed systems where multiple applications interact with the same data streams.

    • Kafka - A Kafka schema registry is an external service that acts as a central repository for managing and validating schemas for data in a Kafka cluster. It allows producers to register schemas and consumers to retrieve them, ensuring data consistency and compatibility as schemas evolve over time.

  1. Select Proceed.

  2. From the Destination table list, select the destination table to which you need to load the data.

  3. Select Proceed.

  4. Specify the schedule at which the data pipeline needs to run. Do one of the following:

  • Select Schedule.

    • From the date pickers, select the Start date and End date.
    • In the Repeat every field, specify the duration at which the data pipeline needs to run.
    • From the Unit list, select the minute, hour, day, week, or month at which the data pipeline needs to run.
    • Select the On day option and specify the day on which the data pipeline needs to run.
    • Select the On option and specify the exact day of the week on which the data pipeline needs to run.

    <--Or-->

  • Select Execute once.

  1. Select Finish.

    The data pipeline is created and displayed in the Data pipelines page.

Filter a data pipeline

You can filter data pipelines based on certain criteria. Filter data pipelines in one of these ways:

  1. Schema

    • Select the Filter icon in the Data pipelines page.
    • Expand the Schema list.
    • Select the required schema.
  2. Data source

    • Select the Filter icon in the Data pipelines page.
    • Expand the Data source list.
    • Select the data source or data sources for which you wish to view the data pipeline.
  3. Pipeline status

    • Select the Filter icon in the Data pipelines page.
    • Expand the Pipeline status list.
    • Select Active to view the data pipelines that are active. A data pipeline is in the Active status when the schedule end date is a future date and there are one or more ingestions that are yet to complete.
    • Select Inactive to view the data pipelines that are inactive. A data pipeline is in the Inactive status either when the end date is past or there are no ingestions that are yet to complete.
  4. Id of the person who created the pipeline

    • Select the Filter icon in the Data pipelines page.
    • Expand the Created by list.
    • Select the user Id of the person who created the data pipeline.

Search a data pipeline

All data pipelines are displayed by default. You can search data pipelines using specific search criteria. To search data pipelines, do the following:

  1. In the Data pipelines page, select +Search pipeline name or schema.
  2. Enter either the pipeline name or the schema name.
  3. You can sort the data pipeline by the pipeline name or the date on which it was last run. To sort the data pipelines in the ascending or descending order, select the Sort icon for the Pipeline name or Last run on column in the Data pipelines page.

View pipeline details

You can view pipeline details in the Data pipelines page. To view the details of a data pipeline, do the following:

  1. In the Data pipeline page, hover over the Pipeline name column and select the +View details icon for a data pipeline. You can view the following details of the data pipeline in this page.

    • Instance started at - Displays the timestamp of the data pipeline execution.

    • Transaction ID - Displays the transaction ID of the data pipeline execution.

    • Total file size - Displays the size of the file processed by the data pipeline.

    • File loaded or Rows loaded - Displays the number of files (for AWS object store) or the total number of rows loaded into the target table (if the data source is Kafka).

    • Status - Indicates the state of the data pipeline when it was last run. The values are:

      • Executed - All rows are successfully loaded into the target table and there are no entries in the rejects table.
      • Failed - For data pipelines that use AWS object data store, no rows are loaded into the target table. For data pipelines that use Kafka data source, some rows are loaded into the target table and the rejects table contains one or more entries.
      • Partially executed - For data pipelines that use AWS object data store, some rows are loaded into the target table and the rejects table contains one or more entries that failed. For data pipelines that use Kafka data source, no rows are loaded into the target table and the rejects table contains one or more entries.
    • For data pipelines that use the AWS object store and are either in Failed or Partially executed status, select the i icon.

      The Files tab appears with the following information:

      • Transaction ID
      • Timestamp of the transaction
      • Batch ID
      • File name
      • File size
      • Status
      • Rows loaded
      • Actions column () with options to view the details and synchronize the file.
        • Select View details.

          The error log appears with the file name, row number, failure reason, and rejected data. Resolve the error and reload the file.

        • Select Sync file to synchronize the data between the source and destination files. A message "File sync in progress" appears at the top of the page.

  2. Select Execute pipeline to execute the selected data pipeline. Executing a data pipeline loads all files that have not already been loaded and that have not reached the retry limit. Executing the data pipeline commits the transaction.

Edit a data pipeline using AWS object store

After creating a data pipeline, you can edit the details to suit your requirements. To edit a data pipeline, do the following:

  1. In the Data pipelines page, mouse over the Pipeline name column and select the +Edit pipeline icon for a data pipeline.

  2. Select AWS object store.

    Information is pre-populated in all fields.

  3. In the Access key ID field, enter your AWS account access key id.

  4. In the Secret access key field, enter your AWS account secret access key. All the details about the data pipeline are populated, except the Access key ID and Secret access key.

    For more information about editing a data pipeline using AWS object store, see Create a data pipeline.

Edit a data pipeline using Kafka data source

After creating a data pipeline, you can edit the details to suit your requirements. To edit a data pipeline, do the following:

  1. In the Data pipelines page, hover over the Pipeline name column and select the +Edit pipeline icon for a data pipeline.

  2. Select Kafka.

    Information is pre-populated in all fields.

  3. Do one of the following:

  • Select SSL if your Kafka broker is configured with SSL.

    • In the Client key field, enter the encrypted client key generated for the SSL certificate.
    • In the Password field, enter the password associated with the encrypted client key.
    • In the Client certificate field, enter the certificate used to authenticate the client with the Kafka broker.
    • In the CA certificate field, enter the certificate authority (CA) certificate used to validate the Kafka broker.

    <--Or-->

  • Select SASL if your Kafka broker is authenticated with Simple Authentication and Security Layer (SASL).

    • In the SASL mechanism list, select one of the following:
    • Plain - A simple username/password authentication mechanism used with TLS for encryption to implement secure authentication.
    • SCRAM - SHA - 256 - A secure, password-based authentication method that uses a challenge-response protocol and the SHA-256 hashing algorithm to verify user credentials without sending the password in plain text over the network.
    • SCRAM - SHA - 512 - A secure authentication mechanism that uses the SHA-512 cryptographic hash function to verify a user's credentials in a "challenge-response" format, which prevents the password from being sent directly over the network.
    • In the Username field, enter a valid username.
    • In the Password field, enter the password for SASL authentication.

For more information about editing a data pipeline using Kafka data source, see Create a data pipeline using Kafka data source.

Clone a data pipeline using AWS object store

You can create a clone or replica of an existing data pipeline. The configurations of the existing data pipeline are copied to the cloned data pipeline. You can edit these configuration settings in the cloned data pipeline.

  1. In the Data pipelines page, select in the Actions column.
  2. Select Clone. The Create a pipeline page is displayed.
  3. Select AWS object store.
  4. In the Pipeline name field, enter the name of the pipeline.
  5. In the Access key ID field, enter your AWS account access key id.
  6. In the Secret access key field, enter your AWS account secret access key. Information in all other fields is pre-populated. You can edit this information. For more information about cloning a data pipeline using AWS object store, see Create a data pipeline using AWS object store.

Clone a data pipeline using Kafka data source

You can create a clone or replica of an existing data pipeline. The configurations of the existing data pipeline are copied to the cloned data pipeline. You can edit these configuration settings in the cloned data pipeline.

  1. In the Data pipelines page, select in the Actions column.

  2. Select Clone. The Create a pipeline page is displayed.

  3. Select Kafka.

  4. In the Pipeline name field, enter the name of the pipeline.

  5. Do one of the following:

    • Select SSL if your Kafka broker is configured with SSL.

      • In the Client key field, enter the encrypted client key generated for the SSL certificate.
      • In the Password field, enter the password associated with the encrypted client key.
      • In the Client certificate field, enter the certificate used to authenticate the client with the Kafka broker.
      • In the CA certificate field, enter the certificate authority (CA) certificate used to validate the Kafka broker.

    <--Or-->

    • Select SASL if your Kafka broker is authenticated with Simple Authentication and Security Layer (SASL).

    • In the SASL mechanism list, select one of the following:

      • Plain - A simple username/password authentication mechanism used with TLS for encryption to implement secure authentication.
      • SCRAM - SHA - 256 - A secure, password-based authentication method that uses a challenge-response protocol and the SHA-256 hashing algorithm to verify user credentials without sending the password in plain text over the network.
      • SCRAM - SHA - 512 - A secure authentication mechanism that uses the SHA-512 cryptographic hash function to verify a user's credentials in a "challenge-response" format, which prevents the password from being sent directly over the network.
      • In the Username field, enter a valid username.
      • In the Password field, enter the password for SASL authentication. Information in all other fields is pre-populated. You can edit this information. For more information about cloning a data pipeline using Kafka data source, see Create a data pipeline using Kafka data source.

Pause and resume schedule

You can pause data ingestion anytime to suit your requirement.

  1. In the Data pipelines page, select in the Actions column.

  2. Select Pause schedule.

  3. In the Confirmation dialog, select Pause.

    This will pause all data ingestions until the state is changed. Data that is already ingested will not be affected. A message "Data ingestion paused successfully" is displayed.

You can resume data ingestion anytime to suit your requirement.

  1. In the Data pipelines page, select the data pipeline that is paused and click in the Actions column.

  2. Select Resume schedule.

  3. In the Confirmation dialog, select Resume.

    This will resume all data ingestions based on the defined schedule. A message "Data ingestion resumed successfully" is displayed.

Delete a data pipeline

You can delete a data pipeline that is no longer in use or required.

  1. In the Data pipelines page, select a data pipeline and click in the Actions column.
  2. Select Resume schedule.
  3. In the Confirmation dialog, select Remove.

3 - Overview of UDx

This topic provides an overview of UDx in OTCAD.

A user-defined extension (UDx) is a component that expands OTCAD functionality—for example, new types of data analysis and the ability to parse and load new types of data.

3.1 - Developing user-defined extensions (UDx)

The primary strengths of UDxs are:.

The primary strengths of UDx are:

  • You can use them wherever you use internal functions.

  • They take full advantage of the distributed computing feature of the system. The extensions usually execute in parallel on each node in the cluster.

  • The complicated aspects of developing a distributed piece of analytic code are handled by the system. Your main programming task is to read in data, process it, and then write it out using the SDK APIs.

Fenced mode

Fenced mode runs the UDx code outside of the main system process. Fenced UDx crashes do not impact the core system process. There is a small performance impact when running UDx code in fenced mode. On average, using fenced mode adds about 10% more time to execution.

Fenced mode is currently available for all UDx with the exception of user-defined aggregates. All UDx run in fenced mode, since the Python runtimes cannot run directly within the system process. Using fenced mode does not affect the development of your UDx. Fenced mode is enabled by default for all UDx.

OTCAD does not support the unfenced mode.

Scalar functions (UDSFs)

A user-defined scalar function (UDSF) returns a single value for each row of data it reads. You can use a UDSF anywhere you can use a built-in database function. For more information about UDSF, see Scalar functions (UDSFs).

3.2 - UDx

This topic provides an overview of UDx in OTCAD.

OTCAD enables you to create, list, and delete custom UDx, so that you can integrate them with existing database tables. UDx allow you to reuse and share code that extends built-in functionality. Use UDx to perform specific tasks such as complex calculations, transformations, or custom data manipulations.

You can reuse a UDx several times. A function always returns a value explicitly by specifying an expression. It is a good choice to calculate and return a value. You can use UDFs to extend built-in functions or to encapsulate calculations that are standard for your organization. UDFs you create can be called in a way similar to built-in functions.

In the present OTCAD release, you write a UDF’s logic (it's handler) – in python.

About UDx

You can view and manage UDx. After logging in to OTCAD, the home page appears. Select the More options button and UDx. The User-defined extensions page appears with the Functions and UDx jobs tabs.

The Functions tab lists the functions that are available in the system. The following columns are displayed:

  • Schema name - The name of the schema.
  • Function name - The name of the function.
  • Language - The language in which the function is developed. For e.g, Python.
  • Library - The name of the library to which the UDx is associated.
  • Created by - The user ID of the person who created the function.
  • Actions - Option to delete the function.

The UDx jobs tab lists the UDx along with the status that are available in the system. The following columns are displayed:

  • Library - The name of the library containing the UDx.
  • Language - The language in which the library is developed. For e.g, Python.
  • Functions - The number of functions associated with the UDx.
  • Created by -The user ID of the person who created the library.
  • Creation date - The date and time of creation of the library.
  • Status - The present status of the UDx job.
  • Details - Details about the UDx in the system.

Create a library

To create a library, do the following in the User-defined extensions page:

  1. Select +Add new.

    The Create a library page appears.

  2. In the Library name field, enter the name of the library.

  3. In the Schema list, choose the required schema.

  4. In the Language list, choose the required language. For e.g, Python.

  5. In the File name field, enter the name of the python file.

  6. Select Browse and choose the zip file that contains the python file you selected in the previous step.

  7. In the Function area, for the Name field, enter the name of the function.

    For more information about UDSF, see Scalar functions (UDSFs).

  8. In the Function area, for the Define class field, enter the name of the class that is defined in the python file selected in step 5. For e.g, MultiplyAndSumFactory.

  9. Click +Add new function to add another function. The newly added functions appear in the UDx jobs tab.

  10. Click Create. The library is created and displayed in the UDx Jobs tab of the User-defined extensions page.

Delete a function

You can delete functions that are no longer used or needed.

  1. In the User defined extensions page, select in the Action column.
  2. Select Delete.
  3. In the Confirmation dialog, select Delete.

4 - Overview of Integrations

This topic provides an overview of third party integrations in OTCAD.

OTCAD integrates with third party applications. This topic provides information about OTCAD integration with third party applications.

4.1 - Integrations

This topic provides an overview of OTCAD integration with third party applications.

OTCAD is a cloud solution that enhances the speed and scalability of an "OpenText™ Analytics Database without the need to manage hardware. Use OTCAD to build databases securely and scale them to large amounts of data in the cloud.

Integrations

When you log in to the application, the More options button is displayed.

  1. Expand More options.
  2. Select Integrations. The Integrations page displays. OTCAD supports machine learning with VerticaPy for data science and AI-driven data science applications. It also supports the following third-party applications:
  3. For each third-party application, select View Docs to view the Connection guide.

5 - Overview of Dashboards

This topic provides an overview of dashboards in OTCAD.

Dashboards identify system bottlenecks on specific clusters and improve the performance of the system.

5.1 - Dashboards

This topic provides an overview of dashboards in OTCAD.

Dashboards

Dashboards enable you to view health metrics for your database.

Dashboards UI tour

You can view and manage dashboards. After signing in to the OTCAD application, the home page appears. Select the More options button and Dashboard. The Dashboard page appears as shown in this image:

Dashboard

The dashboard displays key metrics and events, including details about long-running queries, average query execution time, memory-intensive queries, and scaling activities.

This page displays the following information:

Option Description
a More options: Select this option to view the More menu options.
b Dashboard: View the dashboards in OTCAD by different criteria.
c Active users: This card displays the total number of users who are active over the last 24 hours. Select > to view the list of active users. For more information about active users, see Active users.
d Data stored: This card displays the aggregate data storage in the database from the first day of the current month till date.
e Query overview: This card displays the total count of queries that completed and the queries that failed to execute over the last 24 hours.
f

Cluster performance: This card displays information about the overall cluster performance. It provides information about average CPU usage, average memory usage, and disk I/O usage over the last 24 hours.

  • Select Average CPU usage to view the CPU percentage used by the database server across the primary cluster over the last 24 hours. Select > to view the average CPU usage and maximum CPU usage over a duration of time. For more information about CPU usage, see CPU usage.
  • Select Average Memory usage to view the memory percentage usage information over the last 24 hours. Select > to view the average memory usage and maximum usage over a duration of time. For more information about average memory usage, see Memory usage.
  • Select Disk I/O usage to view the percentage of time a disk is actively engaged in processing I/O requests over the last 24 hours. Select > to view the average disk I/O usage and maximum usage over a duration of time. For more information about average disk I/O usage, see Disk I/O usage.
  • Select Scaling activity to view the consumption of compute in units over the last 24 hours. For more information about scaling activity, see Scaling activity.
g

Query insights: This card displays information about open sessions, queued queries, memory intensive queries, and long running queries.

  • Select Open sessions to view information about the top 5 sessions that are active over the last 1 hour. Select > to view all the sessions that are active over the last 1 hour. For more information about open sessions, see Sessions.
  • Select Queued queries to view information about the top 5 queries in the queue over the last 1 hour. Select > to view all the queries in the queue over the last 1 hour. For more information about queued queries, see Queued queries.
  • Select Memory intensive queries to view information about the top 5 queries that consume more than 25% memory over the last 24 hours. Select > to view information about all queries that consume more than 25% memory over the last 24 hours. For more information about memory intensive queries, see Memory intensive queries.
  • Select Long running queries to view information about the top 5 queries that took more than 10 minutes to execute over the last 24 hours. Select > to view information about all the queries that took more than 10 minutes to execute over the last 24 hours. For more information about long running queries, see Queries by duration.

Cluster performance

Expand Cluster performance in the Dashboard page. The following dashboards are displayed:

CPU usage

Identify CPU usage and address system bottlenecks proactively.

The Average CPU Usage card displays information about the average cluster CPU usage trends in the system over a time duration. The Maximum Usage card displays information about the maximum CPU usage over a duration of time.

In the Visualization tab, do the following:

  1. In the Duration list, do one of the following:
  • Choose the time duration in minutes, hours, or days.

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date. Ensure that the start date does not exceed the end date by more than 30 days.
  1. In the View list, choose one of the values: Average, Minimum, or Maximum. Choose Average to view the average CPU usage for every minute over the selected duration of time. Choose Minimum to view the minimum CPU usage for every minute over the selected duration of time. Choose Maximum to view the maximum CPU usage for every minute over the selected duration of time.

  2. In the Axis Interval list, choose the duration for which you need to view the data in the graph. Choose to view the data for 1 minute, 5 minutes, 15 minutes, 1 hour, or 1 day. By default, line charts are plotted for minute-wise data. Every minute has one data point. For example, if you select 1 day, a total of 1440 data points are plotted on the graph with an axis interval of 1 minute.

In the List view tab, information about the minute-wise time, average CPU usage, maximum CPU usage, and minimum CPU usage is displayed over the selected duration. Choose the Download button to export this information in a spreadsheet.

Memory usage

This helps you compare memory that OTCAD uses versus memory in use by the entire system.

The Average Memory Usage card displays information about the average memory usage trends in the system over the selected duration. The Maximum Usage card displays information about the maximum memory usage over the selected duration of time.

In the Visualization tab, do the following:

  1. In the Duration list, do one of the following:
  • Choose the time duration in minutes, hours, or days.

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date for which you need to visualize the memory usage. Ensure that the start date does not exceed the end date by more than 30 days.
  1. In the View list, choose one of the values: Average, Minimum, or Maximum. Choose Average to view the average memory usage for every minute over the selected duration. Choose Minimum to view the minimum memory usage for every minute over the selected duration of time. Choose Maximum to view the maximum memory usage for every minute over the selected duration of time.

  2. In the Axis Interval list, choose the duration for which you need to view the data in the graph. Choose to view the data for 1 minute, 5 minutes, 15 minutes, 1 hour, or 1 day. By default, line charts are plotted for minute-wise data. Every minute has one data point. For example, if you select 1 day, a total of 1440 data points are plotted on the graph with an axis interval of 1 minute.

In the List view tab, information about the minute-wise time, average memory usage, maximum memory usage, and minimum memory usage is displayed over the selected duration. Choose the Download button to export this information in a spreadsheet.

Disk I/O usage

Identify and address disk I/O bottlenecks for resource optimization, capacity planning, and so on.

The Disk I/O Usage card displays information about the disk I/O usage in the system over the selected duration. The Maximum Usage card displays information about the maximum disk I/O usage over the selected duration.

In the Visualization tab, do the following:

  1. In the Duration list, do one of the following:
  • Choose the time duration in minutes, hours, or days.

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date for which you need to visualize the disk I/O usage. Ensure that the start date does not exceed the end date by more than 30 days.
  1. In the View list, choose one of the values: Average, Minimum, or Maximum. Choose Average to view the average disk I/O usage for every minute over the selected duration. Choose Minimum to view the minimum disk I/O usage for every minute over the selected duration of time. Choose Maximum to view the maximum disk I/O usage for every minute over the selected duration.
  2. In the Axis Interval list, choose the duration for which you need to view the data in the graph. Choose to view the data for 1 minute, 5 minutes, 15 minutes, 1 hour, or 1 day. By default, line charts are plotted for minute-wise data. Every minute has one data point. For example, if you select 1 day, a total of 1440 data points are plotted on the graph with an axis interval of 1 minute.

In the List view tab, information about the minute-wise time, average disk I/O usage, maximum disk I/O usage, and minimum disk I/O usage is displayed over the selected duration of time. Choose the Download button to export this information in a spreadsheet.

Scaling activity

Identify and analyze the potential cost impact due to scaling. View the utilization of additional compute units over the last 24 hours.

The Current Scale Instance card displays information about the current compute units in the system over the selected duration of time. The Maximum Scale card displays information about the maximum compute units over the selected duration of time.

In the Visualization tab, the Duration list is displayed. Do one of the following:

  • Choose the time duration in minutes, hours, or days.

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date for which you need to visualize the scaling activity. You can select a start date that is more than one month prior to the end date.

In the List view tab, information about the date and time, scale event, compute (units), trigger, and user ID of the person who performed the scaling is displayed. Choose Download to export this information in a spreadsheet.

Query insights

Expand Query insights in the Dashboards page. The following dashboards are displayed:

Queries by volume

Identify the query volumes over a duration to understand database performance, identify bottlenecks, and optimize resource utilization. Identify slow or frequently used queries that require optimization. If query volumes increase quickly, optimize the database infrastructure to handle the increased workload.

The Successful queries card displays information about queries that are successfully executed over the last 24 hours. The Queued queries card displays information about SQL queries that are queued for execution over the last 24 hours. The Query status card displays the number of SQL queries that completed successfully and failed over the last 24 hours.

In the Visualization tab, do the following.

  1. In the Duration list, do one of the following:
  • Choose the time duration in minutes, hours, or days.

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date for which you need to visualize the SQL queries by volume. Ensure that the start date does not exceed the end date by more than 30 days.
  1. In the View list, choose one of the values: Volume, Success rate, or Failure rate. Choose Volume to view the number of SQL queries executed over the selected duration. Choose Success rate to view the percentage of SQL queries that completed over the selected duration. Choose Failure rate to view the percentage of SQL queries that failed over the selected duration.

  2. In the Axis Interval list, choose the duration for which you need to view the data in the graph. Choose to view the data for 1 minute, 5 minutes, 15 minutes, 1 hour, or 1 day. By default, line charts are plotted for minute-wise data at an interval of 5 minutes. Every minute has one data point. For example, if you select 1 day at an interval of 1 minute, a total of 1440 data points are plotted on the graph.

In the List view tab, the Duration list is displayed. Do one of the following:

  • Choose the time duration in minutes, hours, or days.

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date for which you need to view the SQL queries by volume. Ensure that the start date does not exceed the end date by more than 30 days.

Information about the time, user name, request status, query, query label, query duration, memory acquired, rows processed, user type, client type, session ID, statement ID, and transaction ID is displayed. Choose Download to export this information in a spreadsheet. Select the Copy icon to copy the query to the clipboard.

Filter criteria

You can filter the dashboards in the Query insights folder based on certain criteria. Choose the Filter icon in each dashboard and filter data using the following criteria:

  1. Request status
    • Expand the Request status list.
    • Choose Completed to view the SQL queries that executed succesfully.
    • Choose Failed to view the SQL queries that failed to execute.
  2. Duration
    • Choose the required duration from the options available to view the SQL queries for the selected duration.
    • Choose Custom and enter the minimum and maximum time in seconds or minutes to view the SQL queries for the selected duration.
  3. Client type - The client type is categorized by the client application. Different client types are displayed depending on the client application. A few examples of client types are given below:
    • JDBC driver - Select this option to view SQL queries executed through the client type JDBC driver.
    • ODBC driver - Select this option to view SQL queries executed through the client type ODBC driver.
    • ADO.NET Driver - Select this option to view SQL queries for the client type ADO.NET driver.
    • vsql - Select this option to view SQL queries for the client type vsql.
  4. Rows processed
    • Choose the required options to view SQL queries by the number of rows selected.
  5. Memory acquired
    • Choose the required options to view the memory acquired for executing the SQL queries.
  6. Currently running
    • Choose Yes to view the SQL queries that are currently running.
  7. System v/s user
    • Choose System to view the SQL queries executed by the system.
    • Choose User to view the SQL queries executed by specific users.

Queries by duration

Analyze the count of queries that executed over the last 24 hours. Slow-running queries hinder overall database efficiency. Analyze the count of queries in the given interval by execution time. This dashboard displays only those queries that took more than 10 minutes to execute over the last 24 hours.

The Average query duration card displays the mean time taken per query for a selected duration.

In the Visualization tab, the Duration list is displayed.

  1. Do one of the following:
  • Choose the time duration in minutes, hours, or days.

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date for which you need to visualize the SQL queries by duration. Ensure that the start date does not exceed the end date by more than 30 days.
  1. In the View list, choose one of the values: Query Id, Query phase, Query duration, or Query type. Choose Query Id to view the SQL queries based on the query ID over the selected duration. Choose Query phase to view the SQL queries based on the query phase over the selected duration. Choose Query duration to view the SQL queries over specific durations of time. Choose Query type to view the SQL queries by type over the selected duration.

In the List view tab, the Duration list is displayed. Do one of the following:

  • Choose the time duration in minutes, hours, or days.

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date for which you need to view the queries by duration. Ensure that the start date does not exceed the end date by more than 30 days.

Information about the time, user name, request status, query, query label, query duration, memory acquired, rows processed, user type, client type, session ID, statement ID, and transaction ID is displayed. Choose Download to export this information in a spreadsheet. Select the Copy icon to copy the query to the clipboard.

Queries by memory usage

Analyze the count of queries over the last 24 hours by memory consumption. When a query uses a large share of memory, it can hinder the execution of other simultaneous queries by restricting access to shared resources.

The Average memory usage card displays information about the average memory that is used by all the queries over the selected duration. The Average memory acquired card displays information about memory acquired for executing the SQL queries over the selected duration.

In the Visualization tab, the Duration list is displayed.

  1. Do one of the following:
  • Choose the time duration in minutes, hours, or days.

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date for which you need to visualize the SQL queries by duration. Ensure that the start date does not exceed the end date by more than 30 days.
  1. In the View list, choose either Memory usage or Acquired memory. Choose Memory usage to view the average memory that is used by all the queries over the selected duration. Choose Acquired memory to view the memory acquired for executing the SQL queries over the selected duration.
  2. In the Axis Interval list, choose the duration for which you need to view the data in the graph. Choose to view the data for 1 minute, 5 minutes, 15 minutes, 1 hour, or 1 day. By default, line charts are plotted for minute-wise data at an interval of 5 minutes. Every minute has one data point. For example, if you select 1 day at an interval of 1 minute, a total of 1440 data points are plotted on the graph.

In the List view tab, the Duration list is displayed. Do one of the following:

  • Choose the time duration in minutes, hours, or days.

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date for which you need to view the queries by memory usage. Ensure that the start date does not exceed the end date by more than 30 days.

Information about the time, user name, request status, query, query label, query duration, memory acquired, rows processed, user type, client type, session ID, statement ID, and transaction ID is displayed. Choose Download to export this information in a spreadsheet. Select the Copy icon to copy the query to the clipboard.

Queries by user

Identify users who generate a high volume of queries over the last 24 hours. See how users work with the database and use these insights to plan resources and manage capacity effectively.

The Average query count card displays information about the total number of queries executed by different users over the selected duration.

In the Visualization tab, the Duration list is displayed.

  1. Do one of the following:
  • Choose the time duration in minutes, hours, or days.

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date for which you need to visualize the SQL queries by duration. Ensure that the start date does not exceed the end date by more than 30 days.
  1. In the View list, choose one of the values: All users, Completed v/s failed, or System v/s user. Choose All users to view the total queries executed by each user over the selected duration. Choose Completed v/s failed to view a comparison of total queries completed v/s total queries failed over the selected duration. Choose System v/s user to view a comparison of total queries completed by all users v/s total queries executed by the system over the selected duration.

In the List view tab, the Duration list is displayed. Do one of the following:

  • Choose the time duration in minutes, hours, or days.

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date for which you need to view the queries by user. Ensure that the start date does not exceed the end date by more than 30 days.

Information about the time, user name, request status, query, query label, query duration, memory acquired, rows processed, user type, client type, session ID, statement ID, and transaction ID is displayed. Choose Download to export this information in a spreadsheet. Select the Copy icon to copy the query to the clipboard.

Queries in queue

Identify queries awaiting allocation of system resources in the last 1 hour. Analyze query queue patterns and trends to plan future capacity needs. This ensures the system scales appropriately to handle anticipated growth in query volume.

The Average query duration card displays the mean time taken per query during the selected time range.

In the Visualization tab, the Duration list is displayed.

  1. Do one of the following:
  • Choose the time duration in minutes, hours, or days.

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date for which you need to visualize the SQL queries in queue. Ensure that the start date does not exceed the end date by more than 30 days.
  1. In the Axis Interval list, choose the duration for which you need to view the data in the graph. Choose to view the data for 1 minute, 5 minutes, 15 minutes, 1 hour, or 1 day. By default, line charts are plotted for minute-wise data at an interval of 5 minutes. Every minute has one data point. For example, if you select 1 day at an interval of 1 minute, a total of 1440 data points are plotted on the graph.

In the List view tab, the Duration list is displayed. Do one of the following:

  • Choose the time duration in minutes, hours, or days.

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date for which you need to view the queries in queue. Ensure that the start date does not exceed the end date by more than 30 days.

Information about the query entry time, user name, query, query label, time in queue, priority, memory requested, pool name, client type, statement ID, and transaction ID is displayed. Choose Download to export this information in a spreadsheet. Select the Copy icon to copy the query to the clipboard.

Queries in queue by duration

Identify the queries that are queued for resource allocation. Use query queue patterns and trends to plan capacity. Scale the system as needed to manage expected growth in query volume.

The Average query duration card displays the mean time taken per query during the selected time range.

In the Visualization tab, the Duration list is displayed. Do one of the following:

  • Choose the time duration in minutes, hours, or days.

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date for which you need to visualize the SQL queries in queue by duration. Ensure that the start date does not exceed the end date by more than 30 days.

In the List view tab, the Duration list is displayed. Do one of the following:

  • Choose the time duration in minutes, hours, or days.

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date for which you need to view the queries in queue by duration. Ensure that the start date does not exceed the end date by more than 30 days.

Information about the query entry time, user name, query, query label, time in queue, priority, memory requested, pool name, client type, statement ID, and transaction ID is displayed. Choose Download to export this information in a spreadsheet. Select the Copy icon to copy the query to the clipboard.

Requests by query type

Analyze request volumes for a specific duration to assess database performance, identify bottlenecks, and optimize resource use. Review request types to manage resources effectively and ensure the database handles varying loads without slowing down.

The Average query duration card displays the mean time taken per query during the selected time range over the last 24 hours.

In the Visualization tab, the Duration list is displayed.

  1. Do one of the following:
  • Choose the time duration in minutes, hours, or days.

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date for which you need to visualize the requests by query type. Ensure that the start date does not exceed the end date by more than 30 days.
  1. In the Axis Interval list, choose the duration for which you need to view the data in the graph. Choose to view the data for 1 minute, 5 minutes, 15 minutes, 1 hour, or 1 day. By default, line charts are plotted for minute-wise data at an interval of 5 minutes. Every minute has one data point. For example, if you select 1 day at an interval of 1 minute, a total of 1440 data points are plotted on the graph.

In the List view tab, the Duration list is displayed. Do one of the following:

  • Choose the time duration in minutes, hours, or days.

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date for which you need to view the queries in queue. Ensure that the start date does not exceed the end date by more than 30 days.

Information about the time, user name, request status, query, query label, query duration, memory acquired, rows processed, user type, client type, session ID, statement ID, and transaction ID is displayed. Choose Download to export this information in a spreadsheet. Select the Copy icon to copy the query to the clipboard.

System metrics

Expand System metrics in the Dashboards page. The following dashboards are displayed:

Connections

Identify the total database connections over the last 24 hours to understand the database performance, identify bottlenecks, and optimize resource utilization. Find sessions that are stuck or failing to troubleshoot quickly. Detect deadlocked sessions to fix them and enhance database performance.

The Average connection count card displays the average number of active database connections recorded over the selected duration. It provides a quick overview of connection trends, helping users monitor database usage and identify potential performance issues or unusual activity.

In the Visualization tab, the Duration list is displayed. Do one of the following:

  • Choose the time duration in minutes, hours, or days.

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date for which you need to visualize the SQL queries in the connections. Ensure that the start date does not exceed the end date by more than 30 days.

  • In the View list, choose Choose Open vs closed to view connections that are open vs connections that are closed. Choose Internal vs user to view connections that are internal to the system vs connections specific to users.

In the List view tab, the Duration list is displayed. Do one of the following:

  • Choose the time duration in minutes, hours, or days.

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date for which you need to view the queries in queue by duration. Ensure that the start date does not exceed the end date by more than 30 days.

Information about the connection time, connection duration, user name, client host name, session status, connection type, request, is currently running, request duration, transaction ID, statement ID, client type, client label, client OS, and client PID is displayed. Choose Download to export this information in a spreadsheet. Select the Copy icon to copy the query to the clipboard.

Filter criteria

You can filter the Connections and Connections by duration dashboards in the System metrics folder based on certain criteria. Choose the Filter icon and filter data using the following criteria:

  1. connectionDuration
    • Choose the required duration from the options available to view the SQL queries for the selected duration.
    • Choose Custom and enter the minimum and maximum time in seconds or minutes to view the SQL queries for the selected duration.
  2. Connection status - Choose Open to view the connections that are open. Choose Closed to view the connections that are closed.
  3. Connection type
    • Choose Internal to view the SQL queries executed by the system.
    • Choose User to view the SQL queries executed by specific users.
  4. Client type - The client type is categorized by the client application. Different client types are displayed depending on the client application. A few examples of client types are given below:
    • JDBC driver - Select this option to view SQL queries executed through the client type JDBC driver.
    • ODBC driver - Select this option to view SQL queries executed through the client type ODBC driver.
    • ADO.NET Driver - Select this option to view SQL queries for the client type ADO.NET driver.
    • vsql - Select this option to view SQL queries for the client type vsql.

Connections by duration

Find the total number of database connections for the selected duration to assess database performance, locate bottlenecks, and optimize resource use.

The Average connection count card displays the average number of active database connections recorded over the selected duration. It provides a quick overview of connection trends, helping users monitor database usage and identify potential performance issues or unusual activity.

In the Visualization tab, the Duration list is displayed. Do one of the following:

  • Choose the time duration in minutes, hours, or days.

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date for which you need to visualize the SQL queries in the connections. Ensure that the start date does not exceed the end date by more than 30 days.

In the List view tab, the Duration list is displayed. Do one of the following:

  • Choose the time duration in minutes, hours, or days.

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date for which you need to view the queries in queue by duration. Ensure that the start date does not exceed the end date by more than 30 days.

Information about the connection time, connection duration, user name, client host name, session status, connection type, request, is currently running, request duration, transaction ID, statement ID, client type, client label, client OS, and client PID is displayed. Choose Download to export this information in a spreadsheet. Select the Copy icon to copy the query to the clipboard.

Data storage

Detect trends in the amount of data stored within the database. Prevent over-provisioning and lower costs from unused storage. Optimize storage allocation to minimize waste and cost.

The Data storage card displays the average data stored for the selected duration over the last month.

In the Visualization tab, the Duration list is displayed.

  1. Do one of the following:
  • Choose the time duration. Choose one of the following:

    • Month to date (MTD)- the period from the beginning of the current calendar month up to but not including the current date. For example, if today is 10 November, MTD covers data from November 1 to November 9.
    • Last month
    • Quarter to date (QTD) - the period from the first day of the calendar quarter up to the current date. For example, if the current date is November 22, QTD covers data from October 1 to November 22.
    • Last quarter

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date for which you need to visualize the requests by query type. Ensure that the start date does not exceed the end date by more than 30 days.
  1. In the Axis Interval list, choose the duration for which you need to view the data in the graph. Choose to view the data by days, months or quarters. By default, line charts are plotted for day-wise data.

In the List view tab, the date of data storage in the database and data storage (in GB and TB) is displayed. Choose Download to export this information in a spreadsheet. Select the Copy icon to copy the query to the clipboard.

Data transfer

Monitor outbound data transfer trends from the database to detect unusual activity. Identify and block unauthorized transfers to prevent data leakage. Timely investigation and mitigation helps you identify and fix irregular transfer patterns that indicate security breaches or unauthorized access.

The Average data transfer card displays the average data that is transferred for the selected duration over the last month.

In the Visualization tab, the Duration list is displayed.

  1. Do one of the following:
  • Choose the time duration. Choose one of the following:

  • Month to date (MTD)- the period from the beginning of the current calendar month up to but not including the current date. For example, if today is 10 November, MTD covers data from November 1 to November 9.

  • Last month

  • Quarter to date (QTD) - the period from the first day of the calendar quarter up to the current date. For example, if the current date is November 22, QTD covers data from October 1 to November 22.

  • Last quarter

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date for which you need to visualize the requests by query type. Ensure that the start date does not exceed the end date by more than 30 days.
  1. In the Axis Interval list, choose the duration for which you need to view the data in the graph. Choose to view the data by days, months or quarters. By default, line charts are plotted for day-wise data.

In the List view tab, the Duration list is displayed. Do one of the following:

  • Choose the time duration in minutes, hours, or days.

<--or-->

  • Choose Custom. From the date pickers, select the start date and end date for which you need to view the queries in queue. Ensure that the start date does not exceed the end date by more than 30 days.

In the List view tab, the date of outbound data transfer from the database and data storage (in GB and TB) is displayed. Choose Download to export this information in a spreadsheet. Select the Copy icon to copy the query to the clipboard.

Active users

Identify active users logged into the database over the last 24 hours. Keep the system responsive and maintain a smooth user experience during peak loads. Identify abnormal patterns that may signal security threats. Confirm active sessions belong to authorized users to safeguard the database.

In the List view tab, the user name, super user, roles, and last login time is displayed. Choose Download to export this information in a spreadsheet.