From cbf1fd15a5ca51a749ed71ab2d0d9a6638fc8de6 Mon Sep 17 00:00:00 2001 From: DB Tsai Date: Mon, 23 Mar 2026 13:56:27 -0700 Subject: [PATCH 1/3] sqlState --- .../resources/error/error-conditions.json | 198 ++++++++++-------- .../SupportsAtomicPartitionManagement.java | 2 +- .../catalog/SupportsPartitionManagement.java | 4 +- .../read/PartitionReaderFactory.java | 2 +- .../apache/spark/sql/connector/read/Scan.java | 6 +- .../ContinuousPartitionReaderFactory.java | 2 +- .../util/V2ExpressionSQLBuilder.java | 6 +- .../spark/sql/connector/write/DeltaWrite.java | 2 +- .../connector/write/DeltaWriteBuilder.java | 2 +- .../sql/connector/write/LogicalWriteInfo.java | 4 +- .../spark/sql/connector/write/Write.java | 4 +- .../sql/connector/write/WriteBuilder.java | 4 +- 12 files changed, 127 insertions(+), 109 deletions(-) diff --git a/common/utils/src/main/resources/error/error-conditions.json b/common/utils/src/main/resources/error/error-conditions.json index 3c210ca7d985b..406e83911a999 100644 --- a/common/utils/src/main/resources/error/error-conditions.json +++ b/common/utils/src/main/resources/error/error-conditions.json @@ -1781,6 +1781,78 @@ ], "sqlState" : "42K03" }, + "DATASOURCE_BATCH_SCAN_NOT_SUPPORTED" : { + "message" : [ + ": Batch scan is not supported." + ], + "sqlState" : "0A000" + }, + "DATASOURCE_BATCH_WRITE_BUILD_NOT_SUPPORTED" : { + "message" : [ + " does not support batch write." + ], + "sqlState" : "0A000" + }, + "DATASOURCE_BATCH_WRITE_NOT_SUPPORTED" : { + "message" : [ + ": Batch write is not supported." + ], + "sqlState" : "0A000" + }, + "DATASOURCE_COLUMNAR_READER_NOT_SUPPORTED" : { + "message" : [ + "Cannot create columnar reader." + ], + "sqlState" : "0A000" + }, + "DATASOURCE_CONTINUOUS_SCAN_NOT_SUPPORTED" : { + "message" : [ + ": Continuous scan is not supported." + ], + "sqlState" : "0A000" + }, + "DATASOURCE_DELTA_BATCH_WRITE_NOT_SUPPORTED" : { + "message" : [ + ": Delta batch write is not supported." + ], + "sqlState" : "0A000" + }, + "DATASOURCE_DELTA_WRITE_BUILD_NOT_IMPLEMENTED" : { + "message" : [ + " does not implement build." + ], + "sqlState" : "0A000" + }, + "DATASOURCE_METADATA_SCHEMA_NOT_IMPLEMENTED" : { + "message" : [ + " does not implement metadataSchema." + ], + "sqlState" : "0A000" + }, + "DATASOURCE_MICRO_BATCH_SCAN_NOT_SUPPORTED" : { + "message" : [ + ": Micro-batch scan is not supported." + ], + "sqlState" : "0A000" + }, + "DATASOURCE_ROW_ID_SCHEMA_NOT_IMPLEMENTED" : { + "message" : [ + " does not implement rowIdSchema." + ], + "sqlState" : "0A000" + }, + "DATASOURCE_STREAMING_WRITE_BUILD_NOT_SUPPORTED" : { + "message" : [ + " does not support streaming write." + ], + "sqlState" : "0A000" + }, + "DATASOURCE_STREAMING_WRITE_NOT_SUPPORTED" : { + "message" : [ + ": Streaming write is not supported." + ], + "sqlState" : "0A000" + }, "DATETIME_FIELD_OUT_OF_BOUNDS" : { "message" : [ "." @@ -5537,6 +5609,12 @@ ], "sqlState" : "428FT" }, + "PARTITIONS_TRUNCATE_NOT_SUPPORTED" : { + "message" : [ + "Partitions truncate is not supported." + ], + "sqlState" : "0A000" + }, "PARTITION_BY_NOT_ALLOWED_WITH_INSERT_INTO" : { "message" : [ "partitionBy() cannot be used with insertInto(). Partition columns for table are determined by the table definition." @@ -5549,12 +5627,24 @@ ], "sqlState" : "42000" }, + "PARTITION_RENAME_NOT_SUPPORTED" : { + "message" : [ + "Partition renaming is not supported." + ], + "sqlState" : "0A000" + }, "PARTITION_TRANSFORM_EXPRESSION_NOT_IN_PARTITIONED_BY" : { "message" : [ "The expression must be inside 'partitionedBy'." ], "sqlState" : "42S23" }, + "PARTITION_TRUNCATE_NOT_SUPPORTED" : { + "message" : [ + "Partition truncate is not supported." + ], + "sqlState" : "0A000" + }, "PATH_ALREADY_EXISTS" : { "message" : [ "Path already exists. Set mode as \"overwrite\" to overwrite the existing path." @@ -6982,6 +7072,12 @@ ], "sqlState" : "XX000" }, + "UNEXPECTED_V2_EXPRESSION" : { + "message" : [ + "Unexpected V2 expression: ." + ], + "sqlState" : "42000" + }, "UNION_NOT_SUPPORTED_IN_RECURSIVE_CTE" : { "message" : [ "The UNION operator is not yet supported within recursive common table expressions (WITH clauses that refer to themselves, directly or indirectly). Please use UNION ALL instead." @@ -8203,6 +8299,18 @@ ], "sqlState" : "42000" }, + "V2_EXPRESSION_SQL_BUILDER_UDAF_NOT_SUPPORTED" : { + "message" : [ + " does not support user defined aggregate function: ." + ], + "sqlState" : "0A000" + }, + "V2_EXPRESSION_SQL_BUILDER_UDF_NOT_SUPPORTED" : { + "message" : [ + " does not support user defined function: ." + ], + "sqlState" : "0A000" + }, "VARIABLE_ALREADY_EXISTS" : { "message" : [ "Cannot create the variable because it already exists.", @@ -10696,91 +10804,6 @@ "CaseInsensitiveStringMap is read-only." ] }, - "_LEGACY_ERROR_TEMP_3133" : { - "message" : [ - " does not implement rowIdSchema" - ] - }, - "_LEGACY_ERROR_TEMP_3134" : { - "message" : [ - " does not implement metadataSchema" - ] - }, - "_LEGACY_ERROR_TEMP_3135" : { - "message" : [ - " does not support batch write" - ] - }, - "_LEGACY_ERROR_TEMP_3136" : { - "message" : [ - " does not support streaming write" - ] - }, - "_LEGACY_ERROR_TEMP_3137" : { - "message" : [ - ": Batch write is not supported" - ] - }, - "_LEGACY_ERROR_TEMP_3138" : { - "message" : [ - ": Streaming write is not supported" - ] - }, - "_LEGACY_ERROR_TEMP_3139" : { - "message" : [ - ": Delta batch write is not supported" - ] - }, - "_LEGACY_ERROR_TEMP_3140" : { - "message" : [ - " does not implement build" - ] - }, - "_LEGACY_ERROR_TEMP_3141" : { - "message" : [ - " does not support user defined function: " - ] - }, - "_LEGACY_ERROR_TEMP_3142" : { - "message" : [ - " does not support user defined aggregate function: " - ] - }, - "_LEGACY_ERROR_TEMP_3143" : { - "message" : [ - "Partition renaming is not supported" - ] - }, - "_LEGACY_ERROR_TEMP_3144" : { - "message" : [ - "Partition truncate is not supported" - ] - }, - "_LEGACY_ERROR_TEMP_3145" : { - "message" : [ - "Partitions truncate is not supported" - ] - }, - "_LEGACY_ERROR_TEMP_3147" : { - "message" : [ - ": Batch scan are not supported" - ] - }, - "_LEGACY_ERROR_TEMP_3148" : { - "message" : [ - ": Micro-batch scan are not supported" - ] - }, - "_LEGACY_ERROR_TEMP_3149" : { - "message" : [ - ": Continuous scan are not supported" - ] - }, - "_LEGACY_ERROR_TEMP_3150" : { - "message" : [ - "Cannot create columnar reader." - ] - }, "_LEGACY_ERROR_TEMP_3152" : { "message" : [ "Datatype not supported " @@ -10986,11 +11009,6 @@ " is not a boolean string." ] }, - "_LEGACY_ERROR_TEMP_3207" : { - "message" : [ - "Unexpected V2 expression: " - ] - }, "_LEGACY_ERROR_TEMP_3208" : { "message" : [ "The number of fields () in the partition identifier is not equal to the partition schema length (). The identifier might not refer to one partition." diff --git a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/catalog/SupportsAtomicPartitionManagement.java b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/catalog/SupportsAtomicPartitionManagement.java index 969a6a2874cf2..8c95e0060a654 100644 --- a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/catalog/SupportsAtomicPartitionManagement.java +++ b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/catalog/SupportsAtomicPartitionManagement.java @@ -127,6 +127,6 @@ default boolean purgePartitions(InternalRow[] idents) */ default boolean truncatePartitions(InternalRow[] idents) throws NoSuchPartitionException, SparkUnsupportedOperationException { - throw new SparkUnsupportedOperationException("_LEGACY_ERROR_TEMP_3145"); + throw new SparkUnsupportedOperationException("PARTITIONS_TRUNCATE_NOT_SUPPORTED"); } } diff --git a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/catalog/SupportsPartitionManagement.java b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/catalog/SupportsPartitionManagement.java index 3c130a0aa87e1..a0f6bcd261373 100644 --- a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/catalog/SupportsPartitionManagement.java +++ b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/catalog/SupportsPartitionManagement.java @@ -157,7 +157,7 @@ default boolean renamePartition(InternalRow from, InternalRow to) throws SparkUnsupportedOperationException, PartitionsAlreadyExistException, NoSuchPartitionException { - throw new SparkUnsupportedOperationException("_LEGACY_ERROR_TEMP_3143"); + throw new SparkUnsupportedOperationException("PARTITION_RENAME_NOT_SUPPORTED"); } /** @@ -172,6 +172,6 @@ default boolean renamePartition(InternalRow from, InternalRow to) */ default boolean truncatePartition(InternalRow ident) throws NoSuchPartitionException, SparkUnsupportedOperationException { - throw new SparkUnsupportedOperationException("_LEGACY_ERROR_TEMP_3144"); + throw new SparkUnsupportedOperationException("PARTITION_TRUNCATE_NOT_SUPPORTED"); } } diff --git a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/PartitionReaderFactory.java b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/PartitionReaderFactory.java index 01175f0513db0..5a73906d93816 100644 --- a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/PartitionReaderFactory.java +++ b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/PartitionReaderFactory.java @@ -51,7 +51,7 @@ public interface PartitionReaderFactory extends Serializable { * {@link InputPartition} class defined for the data source. */ default PartitionReader createColumnarReader(InputPartition partition) { - throw new SparkUnsupportedOperationException("_LEGACY_ERROR_TEMP_3150"); + throw new SparkUnsupportedOperationException("DATASOURCE_COLUMNAR_READER_NOT_SUPPORTED"); } /** diff --git a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/Scan.java b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/Scan.java index 81b89e5750d83..05a89fa1e5ab5 100644 --- a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/Scan.java +++ b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/Scan.java @@ -78,7 +78,7 @@ default String description() { */ default Batch toBatch() { throw new SparkUnsupportedOperationException( - "_LEGACY_ERROR_TEMP_3147", Map.of("description", description())); + "DATASOURCE_BATCH_SCAN_NOT_SUPPORTED", Map.of("description", description())); } /** @@ -95,7 +95,7 @@ default Batch toBatch() { */ default MicroBatchStream toMicroBatchStream(String checkpointLocation) { throw new SparkUnsupportedOperationException( - "_LEGACY_ERROR_TEMP_3148", Map.of("description", description())); + "DATASOURCE_MICRO_BATCH_SCAN_NOT_SUPPORTED", Map.of("description", description())); } /** @@ -112,7 +112,7 @@ default MicroBatchStream toMicroBatchStream(String checkpointLocation) { */ default ContinuousStream toContinuousStream(String checkpointLocation) { throw new SparkUnsupportedOperationException( - "_LEGACY_ERROR_TEMP_3149", Map.of("description", description())); + "DATASOURCE_CONTINUOUS_SCAN_NOT_SUPPORTED", Map.of("description", description())); } /** diff --git a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/streaming/ContinuousPartitionReaderFactory.java b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/streaming/ContinuousPartitionReaderFactory.java index 8c2da03372934..d99ee92432f0e 100644 --- a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/streaming/ContinuousPartitionReaderFactory.java +++ b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/streaming/ContinuousPartitionReaderFactory.java @@ -38,6 +38,6 @@ public interface ContinuousPartitionReaderFactory extends PartitionReaderFactory @Override default ContinuousPartitionReader createColumnarReader(InputPartition partition) { - throw new SparkUnsupportedOperationException("_LEGACY_ERROR_TEMP_3150"); + throw new SparkUnsupportedOperationException("DATASOURCE_COLUMNAR_READER_NOT_SUPPORTED"); } } diff --git a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/util/V2ExpressionSQLBuilder.java b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/util/V2ExpressionSQLBuilder.java index 343221f91a18d..8e3e2a359b69b 100644 --- a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/util/V2ExpressionSQLBuilder.java +++ b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/util/V2ExpressionSQLBuilder.java @@ -319,20 +319,20 @@ protected String visitInverseDistributionFunction( protected String visitUserDefinedScalarFunction( String funcName, String canonicalName, String[] inputs) { throw new SparkUnsupportedOperationException( - "_LEGACY_ERROR_TEMP_3141", + "V2_EXPRESSION_SQL_BUILDER_UDF_NOT_SUPPORTED", Map.of("class", this.getClass().getSimpleName(), "funcName", funcName)); } protected String visitUserDefinedAggregateFunction( String funcName, String canonicalName, boolean isDistinct, String[] inputs) { throw new SparkUnsupportedOperationException( - "_LEGACY_ERROR_TEMP_3142", + "V2_EXPRESSION_SQL_BUILDER_UDAF_NOT_SUPPORTED", Map.of("class", this.getClass().getSimpleName(), "funcName", funcName)); } protected String visitUnexpectedExpr(Expression expr) throws IllegalArgumentException { throw new SparkIllegalArgumentException( - "_LEGACY_ERROR_TEMP_3207", Map.of("expr", String.valueOf(expr))); + "UNEXPECTED_V2_EXPRESSION", Map.of("expr", String.valueOf(expr))); } protected String visitPartitionPredicate(PartitionPredicate partitionPredicate) { diff --git a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/DeltaWrite.java b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/DeltaWrite.java index 7c1903d7ac63b..4803d67be1092 100644 --- a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/DeltaWrite.java +++ b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/DeltaWrite.java @@ -32,6 +32,6 @@ public interface DeltaWrite extends Write { @Override default DeltaBatchWrite toBatch() { throw new SparkUnsupportedOperationException( - "_LEGACY_ERROR_TEMP_3139", Map.of("description", description())); + "DATASOURCE_DELTA_BATCH_WRITE_NOT_SUPPORTED", Map.of("description", description())); } } diff --git a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/DeltaWriteBuilder.java b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/DeltaWriteBuilder.java index d27fd7fb1f9c4..c437e401de1d0 100644 --- a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/DeltaWriteBuilder.java +++ b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/DeltaWriteBuilder.java @@ -32,6 +32,6 @@ public interface DeltaWriteBuilder extends WriteBuilder { @Override default DeltaWrite build() { throw new SparkUnsupportedOperationException( - "_LEGACY_ERROR_TEMP_3140", Map.of("class", getClass().getName())); + "DATASOURCE_DELTA_WRITE_BUILD_NOT_IMPLEMENTED", Map.of("class", getClass().getName())); } } diff --git a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/LogicalWriteInfo.java b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/LogicalWriteInfo.java index 63817aa0e3a80..f2f8ae0be4caa 100644 --- a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/LogicalWriteInfo.java +++ b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/LogicalWriteInfo.java @@ -55,7 +55,7 @@ public interface LogicalWriteInfo { */ default Optional rowIdSchema() { throw new SparkUnsupportedOperationException( - "_LEGACY_ERROR_TEMP_3133", Map.of("class", getClass().getName())); + "DATASOURCE_ROW_ID_SCHEMA_NOT_IMPLEMENTED", Map.of("class", getClass().getName())); } /** @@ -63,6 +63,6 @@ default Optional rowIdSchema() { */ default Optional metadataSchema() { throw new SparkUnsupportedOperationException( - "_LEGACY_ERROR_TEMP_3134", Map.of("class", getClass().getName())); + "DATASOURCE_METADATA_SCHEMA_NOT_IMPLEMENTED", Map.of("class", getClass().getName())); } } diff --git a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/Write.java b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/Write.java index f62d194fa7f9f..77ce6461192c2 100644 --- a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/Write.java +++ b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/Write.java @@ -56,7 +56,7 @@ default String description() { */ default BatchWrite toBatch() { throw new SparkUnsupportedOperationException( - "_LEGACY_ERROR_TEMP_3137", Map.of("description", description())); + "DATASOURCE_BATCH_WRITE_NOT_SUPPORTED", Map.of("description", description())); } /** @@ -67,7 +67,7 @@ default BatchWrite toBatch() { */ default StreamingWrite toStreaming() { throw new SparkUnsupportedOperationException( - "_LEGACY_ERROR_TEMP_3138", Map.of("description", description())); + "DATASOURCE_STREAMING_WRITE_NOT_SUPPORTED", Map.of("description", description())); } /** diff --git a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/WriteBuilder.java b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/WriteBuilder.java index fefc06c12df74..5462a3f28e548 100644 --- a/sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/WriteBuilder.java +++ b/sql/catalyst/src/main/java/org/apache/spark/sql/connector/write/WriteBuilder.java @@ -62,7 +62,7 @@ public StreamingWrite toStreaming() { @Deprecated(since = "3.2.0") default BatchWrite buildForBatch() { throw new SparkUnsupportedOperationException( - "_LEGACY_ERROR_TEMP_3135", Map.of("class", getClass().getName())); + "DATASOURCE_BATCH_WRITE_BUILD_NOT_SUPPORTED", Map.of("class", getClass().getName())); } /** @@ -73,6 +73,6 @@ default BatchWrite buildForBatch() { @Deprecated(since = "3.2.0") default StreamingWrite buildForStreaming() { throw new SparkUnsupportedOperationException( - "_LEGACY_ERROR_TEMP_3136", Map.of("class", getClass().getName())); + "DATASOURCE_STREAMING_WRITE_BUILD_NOT_SUPPORTED", Map.of("class", getClass().getName())); } } From fc9dc5593aa10af779708637538e8e27be18a6e5 Mon Sep 17 00:00:00 2001 From: DB Tsai Date: Tue, 24 Mar 2026 10:02:47 -0700 Subject: [PATCH 2/3] fix test failure --- .../resources/error/error-conditions.json | 144 +++++++++--------- 1 file changed, 72 insertions(+), 72 deletions(-) diff --git a/common/utils/src/main/resources/error/error-conditions.json b/common/utils/src/main/resources/error/error-conditions.json index 406e83911a999..8135f09067f92 100644 --- a/common/utils/src/main/resources/error/error-conditions.json +++ b/common/utils/src/main/resources/error/error-conditions.json @@ -1439,6 +1439,78 @@ ], "sqlState" : "KD011" }, + "DATASOURCE_BATCH_SCAN_NOT_SUPPORTED" : { + "message" : [ + ": Batch scan is not supported." + ], + "sqlState" : "0A000" + }, + "DATASOURCE_BATCH_WRITE_BUILD_NOT_SUPPORTED" : { + "message" : [ + " does not support batch write." + ], + "sqlState" : "0A000" + }, + "DATASOURCE_BATCH_WRITE_NOT_SUPPORTED" : { + "message" : [ + ": Batch write is not supported." + ], + "sqlState" : "0A000" + }, + "DATASOURCE_COLUMNAR_READER_NOT_SUPPORTED" : { + "message" : [ + "Cannot create columnar reader." + ], + "sqlState" : "0A000" + }, + "DATASOURCE_CONTINUOUS_SCAN_NOT_SUPPORTED" : { + "message" : [ + ": Continuous scan is not supported." + ], + "sqlState" : "0A000" + }, + "DATASOURCE_DELTA_BATCH_WRITE_NOT_SUPPORTED" : { + "message" : [ + ": Delta batch write is not supported." + ], + "sqlState" : "0A000" + }, + "DATASOURCE_DELTA_WRITE_BUILD_NOT_IMPLEMENTED" : { + "message" : [ + " does not implement build." + ], + "sqlState" : "0A000" + }, + "DATASOURCE_METADATA_SCHEMA_NOT_IMPLEMENTED" : { + "message" : [ + " does not implement metadataSchema." + ], + "sqlState" : "0A000" + }, + "DATASOURCE_MICRO_BATCH_SCAN_NOT_SUPPORTED" : { + "message" : [ + ": Micro-batch scan is not supported." + ], + "sqlState" : "0A000" + }, + "DATASOURCE_ROW_ID_SCHEMA_NOT_IMPLEMENTED" : { + "message" : [ + " does not implement rowIdSchema." + ], + "sqlState" : "0A000" + }, + "DATASOURCE_STREAMING_WRITE_BUILD_NOT_SUPPORTED" : { + "message" : [ + " does not support streaming write." + ], + "sqlState" : "0A000" + }, + "DATASOURCE_STREAMING_WRITE_NOT_SUPPORTED" : { + "message" : [ + ": Streaming write is not supported." + ], + "sqlState" : "0A000" + }, "DATATYPE_CANNOT_ORDER" : { "message" : [ "Type does not support ordered operations." @@ -1781,78 +1853,6 @@ ], "sqlState" : "42K03" }, - "DATASOURCE_BATCH_SCAN_NOT_SUPPORTED" : { - "message" : [ - ": Batch scan is not supported." - ], - "sqlState" : "0A000" - }, - "DATASOURCE_BATCH_WRITE_BUILD_NOT_SUPPORTED" : { - "message" : [ - " does not support batch write." - ], - "sqlState" : "0A000" - }, - "DATASOURCE_BATCH_WRITE_NOT_SUPPORTED" : { - "message" : [ - ": Batch write is not supported." - ], - "sqlState" : "0A000" - }, - "DATASOURCE_COLUMNAR_READER_NOT_SUPPORTED" : { - "message" : [ - "Cannot create columnar reader." - ], - "sqlState" : "0A000" - }, - "DATASOURCE_CONTINUOUS_SCAN_NOT_SUPPORTED" : { - "message" : [ - ": Continuous scan is not supported." - ], - "sqlState" : "0A000" - }, - "DATASOURCE_DELTA_BATCH_WRITE_NOT_SUPPORTED" : { - "message" : [ - ": Delta batch write is not supported." - ], - "sqlState" : "0A000" - }, - "DATASOURCE_DELTA_WRITE_BUILD_NOT_IMPLEMENTED" : { - "message" : [ - " does not implement build." - ], - "sqlState" : "0A000" - }, - "DATASOURCE_METADATA_SCHEMA_NOT_IMPLEMENTED" : { - "message" : [ - " does not implement metadataSchema." - ], - "sqlState" : "0A000" - }, - "DATASOURCE_MICRO_BATCH_SCAN_NOT_SUPPORTED" : { - "message" : [ - ": Micro-batch scan is not supported." - ], - "sqlState" : "0A000" - }, - "DATASOURCE_ROW_ID_SCHEMA_NOT_IMPLEMENTED" : { - "message" : [ - " does not implement rowIdSchema." - ], - "sqlState" : "0A000" - }, - "DATASOURCE_STREAMING_WRITE_BUILD_NOT_SUPPORTED" : { - "message" : [ - " does not support streaming write." - ], - "sqlState" : "0A000" - }, - "DATASOURCE_STREAMING_WRITE_NOT_SUPPORTED" : { - "message" : [ - ": Streaming write is not supported." - ], - "sqlState" : "0A000" - }, "DATETIME_FIELD_OUT_OF_BOUNDS" : { "message" : [ "." From 0dee74ec069862165abb7bce325db347a6ee1065 Mon Sep 17 00:00:00 2001 From: DB Tsai Date: Tue, 24 Mar 2026 15:02:48 -0700 Subject: [PATCH 3/3] trigger ci