You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When I try writing a simple csv file (initially, I tried parquet to the same effect), the file written by this library ends up not being readable by clickhouse. Clickhouse uses libhdfs3 library, so I assume the issue might be there as well.
What I tried so far:
the attached program that writes directly to hdfs: cannot be read by clickhouse
the attached program that writes to a local file, then upload using gohdfs put: cannot be read by clickhouse
the attached program that writes to a local file, then upload using hdfs dfs -put (hadoop tools): can be read by clickhouse
the attached program that writes directly to hdfs, use hdfs dfs -get followed by hdfs dfs -put: can be read by clickhouse, nothing missing
this is the response from running clickhouse-local:
server.internal :) select * from hdfs('hdfs://nameservice1/random/yet/existing/path/flat.csv', 'CSV')
SELECT *
FROM hdfs('hdfs://nameservice1/random/yet/existing/path/flat.csv', 'CSV')
Query id: e33d9bf7-41b0-4025-a5fc-8dc6ebb65c0f
0 rows in set. Elapsed: 60.292 sec.
Received exception:
Code: 210. DB::Exception: Fail to read from HDFS: hdfs://nameservice1, file path: /random/yet/existing/path/flat.csv. Error:
HdfsIOException: InputStreamImpl: cannot read file: /random/yet/existing/path/flat.csv, from position 0, size: 1048576.
Caused by: HdfsIOException: InputStreamImpl: all nodes have been tried and no valid replica can be read for Block: [block pool
ID: BP-2134387385-192.168.12.6-1648216715170 block ID 1367614010_294283603].: Cannot extract table structure from CSV
format file. You can specify the structure manually. (NETWORK_ERROR)
I'm using Hadoop 2.7.3.2.6.5.0-292.
The attached program doesn't produce any error.
The text was updated successfully, but these errors were encountered:
When I try writing a simple csv file (initially, I tried parquet to the same effect), the file written by this library ends up not being readable by clickhouse. Clickhouse uses libhdfs3 library, so I assume the issue might be there as well.
What I tried so far:
gohdfs put
: cannot be read by clickhousehdfs dfs -put
(hadoop tools): can be read by clickhousehdfs dfs -get
followed byhdfs dfs -put
: can be read by clickhouse, nothing missingThis is the same code:
this is the response from running clickhouse-local:
I'm using
Hadoop 2.7.3.2.6.5.0-292
.The attached program doesn't produce any error.
The text was updated successfully, but these errors were encountered: