Skip to content

query zstd parquet file error #53028

@mofei920

Description

@mofei920

clickhouse-local version: ClickHouse local version 23.7.1.2470 (official build)
mysql: 8.0.28

CREATE TABLE archive_test (
  id bigint unsigned NOT NULL,
  c1 bigint NOT NULL,
  PRIMARY KEY (id)
) ENGINE=InnoDB DEFAULT CHARSET=utf8mb4 COLLATE=utf8mb4_general_ci ROW_FORMAT=DYNAMIC;

INSERT INTO `archive_test` (`id`, `c1`)
VALUES
	(1, 11),
	(2, 12),
	(3, 13);

export data into parquet file in s3 successfully.

# clickhouse-local -q "SET output_format_parquet_row_group_size = 10240, output_format_parquet_compression_method='zstd'; INSERT INTO FUNCTION s3('http://bucket_name.oss_addr/mysql_archive.parquet', 'ak', 'sk') select id, c1 from mysql('mysql_hostname:3306','dbtest','archive_test','username','password')"

query from the parquet file。

# clickhouse-local -q "select count(*) FROM s3('http://bucket_name.oss_addr/mysql_archive.parquet', 'ak', 'sk')"

I get the error as below. but it's no problem when use output_format_parquet_compression_method=lz4. is it a bug?

Code: 33. DB::ParsingException: Error while reading Parquet data: IOError: ZSTD decompression failed: Restored data doesn't match checksum: While executing ParquetBlockInputFormat: While executing OSS. (CANNOT_READ_ALL_DATA)

Metadata

Metadata

Assignees

Labels

potential bugTo be reviewed by developers and confirmed/rejected.st-need-infoWe need extra data to continue (waiting for response). Either some details or a repro of the issue.

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions