Known issues

Known issues release notes for 11.0.x.

Vertica makes every attempt to provide you with an up-to-date list of significant known issues in each release. We will update this list as we resolve issues and as we learn of new issues.

11.0.2-0

No Known issues as of hotfix 11.0.2-3.

11.0.1-0

Updated 10/25/21

Issue Key Component Description
VER‑78310 Client Drivers - JDBC

JDBC complex types return NULL arrays as empty arrays. For example, when executing this SQL statement:

SELECT ARRAY[null,ARRAY[1,2,3],null,ARRAY[4,5],null] as array;

The array column the server returns will be:

[[],[1,2,3],[],[4,5],[]]

Because of the null values in the string literal, it should return this value:

[null,[1,2,3],null,[4,5],null]

This is a work around to a limitation in Simba.

VER‑78074 Procedural Languages, UDX Stored procedures that contain DML queries on tables with key constraints do not return a value.
VER‑69803 Hadoop The infer_external_table_ddl meta-function will hang if you do not set authentication information in the current session when analyzing a file on GCP.
VER‑64916 Kafka Integration When Vertica exports data collector information to Kafka via notifier, the serialization logic does not properly handle Vertica internal data type translation. When importing or copying that data back into the Management Console, Vertica uses the raw data in Kafka and inserts garbled data into the extended monitoring database.
VER‑64352 SDK

Under special circumstances, a sequence of statements to create and drop Python libraries fails. For example, the following session statements attempt to create Python libraries lib1 and lib2:

CREATE LIBRARY lib1 AS :lib1_path
DEPENDS :numpy_path LANGUAGE 'Python';  DROP LIBRARY lib1;  CREATE
LIBRARY lib2 AS :lib2_path DEPENDS :sklearn_path LANGUAGE 'Python';

Here, lib1 is a Python library that imports module numpy, while lib2 imports module sklearn, which relies on numpy. After lib1 is dropped, the CREATE LIBRARY statement for lib2 fails.

VER‑62983 Hadoop When hcatalogconnector schemas are created with custom_partitions enabled, poor performance has been observed when there are many (500+) partitions. By default, custom_partitions are disabled on hcatalogconnector schemas.
VER‑61420 Data Removal - Delete, Purge, Partitioning Partition operations, such as move_partitions_to_table, must split storage containers that have both partitions that match the operations and partitions that do not. Version 9.1 introduced an inefficiency whereby such a split may separate a storage container into one more storage container than necessary. Note that any extraneous containers created this way will eventually be merged by the TM.
VER‑61069 Execution Engine In very rare circumstances, if a Vertica process crashes during shutdown, the remaining processes might hang indefinitely.
VER‑48020 Hadoop Canceling a query that involves loading data from ORC or Parquet files can be slow if there are too many files involved in the load.

11.0.0-0

Updated 08/11/2021

Issue Key Component Description
VER‑48020 Hadoop Canceling a query that involves loading data from ORC or Parquet files can be slow if there are too many files involved in the load.
VER‑61069 Execution Engine In very rare circumstances, if a Vertica process crashes during shutdown, the remaining processes might hang indefinitely.
VER‑61420 Data Removal - Delete, Purge, Partitioning Partition operations, such as move_partitions_to_table, must split storage containers that have both partitions that match the operations and partitions that do not. Version 9.1 introduced an inefficiency whereby such a split may separate a storage container into one more storage container than necessary. Note that any extraneous containers created this way will eventually be merged by the TM.
VER‑62983 Hadoop When hcatalogconnector schemas are created with custom_partitions enabled, poor performance has been observed when there are many (500+) partitions. By default, custom_partitions are disabled on hcatalogconnector schemas.
VER‑64352 SDK

Under special circumstances, a sequence of statements to create and drop Python libraries fails. For example, the following session statements attempt to create Python libraries lib1 and lib2:

CREATE LIBRARY lib1 AS :lib1_path DEPENDS :numpy_path LANGUAGE 'Python';
DROP LIBRARY lib1;                
CREATE LIBRARY lib2 AS :lib2_path DEPENDS :sklearn_path LANGUAGE 'Python';

Here, lib1 is a Python library that imports module numpy, while lib2 imports module sklearn, which relies on numpy. After lib1 is dropped, the CREATE LIBRARY statement for lib2 fails.

VER‑69803 Hadoop The infer_external_table_ddl meta-function will hang if you do not set authentication information in the current session when analyzing a file on GCP.
VER‑78074 Procedural Languages, UDX Stored procedures that contain DML queries on tables with key constraints do not return a value.
VER‑78310 Client Drivers - JDBC

JDBC complex types return NULL arrays as empty arrays. For example, when executing this SQL statement:

SELECT ARRAY[null,ARRAY[1,2,3],null,ARRAY[4,5],null] as array;
The array column the server returns will be:
[[],[1,2,3],[],[4,5],[]]

Because of the null values in the string literal, it should return this value:

[null,[1,2,3],null,[4,5],null]

This is a work around due to a limitation in Simba.