How can i use terminal to read data from minio? #1933
Replies: 6 comments 30 replies
-
Hi, Amoro is a Lakehouse management system based on open data lake formats like Apache Iceberg.
Then you can create Iceberg tables through Terminal now, you can start your quickstart journey followed by:https://amoro.netease.com/quick-demo/ I paste a screenshot to help you to configue the catalog, wish it can help. |
Beta Was this translation helpful? Give feedback.
-
The profile The |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
Since you have chosen Hive MetaStore as the metadata service, you need to configure the implementation of Iceberg FileIO in catalog properties. Specifically, you can try adding the following configuration in catalog properties:
|
Beta Was this translation helpful? Give feedback.
-
i have many small file in MinIO while streaming data by kafka. even update
data will have a new file created. i wanna compact and merge all file into
one file
![image](https://github.com/NetEase/amoro/assets/87402004/1df37611-319c-43a5-9e20-8485df2708a8)
i use local optimize :
![image](https://github.com/NetEase/amoro/assets/87402004/5eefd91c-8469-4044-8e58-e913ea07b6a5)
![image](https://github.com/NetEase/amoro/assets/87402004/f240be2f-d68c-4eab-9c07-38af917c4f94)
tables committing but it error:
![image](https://github.com/NetEase/amoro/assets/87402004/d065d5fe-b1c6-47b5-9f44-44942fd24ed6)
this's error logs:
com.netease.arctic.server.exception.OptimizingCommitException: unexpected
commit error at
com.netease.arctic.server.optimizing.UnKeyedTableCommit.replaceFiles(UnKeyedTableCommit.java:214)
at
com.netease.arctic.server.optimizing.UnKeyedTableCommit.commit(UnKeyedTableCommit.java:174)
at
com.netease.arctic.server.optimizing.OptimizingQueue$TableOptimizingProcess.commit(OptimizingQueue.java:529)
at
com.netease.arctic.server.table.executor.OptimizingCommitExecutor.execute(OptimizingCommitExecutor.java:32)
at
com.netease.arctic.server.table.executor.BaseTableExecutor.executeTask(BaseTableExecutor.java:53)
at
com.netease.arctic.server.table.executor.BaseTableExecutor.lambda$scheduleIfNecessary$4(BaseTableExecutor.java:65)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266) at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750) Caused by:
org.apache.iceberg.exceptions.RuntimeIOException: Failed to create file:
s3a://datalake/lakehouse/postgres_ducdn__public_test_user/metadata/9dddf437-89d1-4c99-b95e-5a940f228609-m9.avro
at
org.apache.iceberg.hadoop.HadoopOutputFile.createOrOverwrite(HadoopOutputFile.java:87)
at
org.apache.iceberg.avro.AvroFileAppender.<init>(AvroFileAppender.java:56)
at org.apache.iceberg.avro.Avro$WriteBuilder.build(Avro.java:191) at
org.apache.iceberg.ManifestWriter$V2Writer.newAppender(ManifestWriter.java:243)
at org.apache.iceberg.ManifestWriter.<init>(ManifestWriter.java:58) at
org.apache.iceberg.ManifestWriter.<init>(ManifestWriter.java:34) at
org.apache.iceberg.ManifestWriter$V2Writer.<init>(ManifestWriter.java:220)
at org.apache.iceberg.ManifestFiles.write(ManifestFiles.java:163) at
org.apache.iceberg.SnapshotProducer.newManifestWriter(SnapshotProducer.java:488)
at
org.apache.iceberg.MergingSnapshotProducer$DataFileFilterManager.newManifestWriter(MergingSnapshotProducer.java:1057)
at
org.apache.iceberg.ManifestFilterManager.filterManifestWithDeletedFiles(ManifestFilterManager.java:414)
at
org.apache.iceberg.ManifestFilterManager.filterManifest(ManifestFilterManager.java:319)
at
org.apache.iceberg.ManifestFilterManager.lambda$filterManifests$0(ManifestFilterManager.java:195)
at org.apache.iceberg.util.Tasks$Builder.runTaskWithRetry(Tasks.java:413)
at org.apache.iceberg.util.Tasks$Builder.access$300(Tasks.java:69) at
org.apache.iceberg.util.Tasks$Builder$1.run(Tasks.java:315) at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at
java.util.concurrent.FutureTask.run(FutureTask.java:266) ... 3 more Caused
by: java.io.InterruptedIOException: getFileStatus on
s3a://datalake/lakehouse/postgres_ducdn__public_test_user/metadata/9dddf437-89d1-4c99-b95e-5a940f228609-m9.avro:
com.amazonaws.SdkClientException: Unable to execute HTTP request: Timeout
waiting for connection from pool at
org.apache.hadoop.fs.s3a.S3AUtils.translateInterruptedException(S3AUtils.java:352)
at org.apache.hadoop.fs.s3a.S3AUtils.translateException(S3AUtils.java:177)
at org.apache.hadoop.fs.s3a.S3AUtils.translateException(S3AUtils.java:151)
at
org.apache.hadoop.fs.s3a.S3AFileSystem.s3GetFileStatus(S3AFileSystem.java:2278)
at
org.apache.hadoop.fs.s3a.S3AFileSystem.innerGetFileStatus(S3AFileSystem.java:2226)
at
org.apache.hadoop.fs.s3a.S3AFileSystem.getFileStatus(S3AFileSystem.java:2160)
at org.apache.hadoop.fs.s3a.S3AFileSystem.create(S3AFileSystem.java:804) at
org.apache.hadoop.fs.FileSystem.create(FileSystem.java:1125) at
org.apache.hadoop.fs.FileSystem.create(FileSystem.java:1105) at
org.apache.hadoop.fs.FileSystem.create(
Vào Th 3, 12 thg 9, 2023 vào lúc 18:23 ZhouJinsong <
***@***.***> đã viết:
… Since you have chosen Hive MetaStore as the metadata service, you need to
configure the implementation of Iceberg FileIO in catalog properties.
Specifically, you can try adding the following configuration in catalog
properties:
io-impl=org.apache.iceberg.aws.s3.S3FileIO
—
Reply to this email directly, view it on GitHub
<#1933 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AU22MFFX7OOIM4DRU5ETTG3X2BA2HANCNFSM6AAAAAA4OTDCGQ>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
Beta Was this translation helpful? Give feedback.
-
you can try this.
This is a solution that other users in the community have implemented. |
Beta Was this translation helpful? Give feedback.
-
Hi, i have already minio running, now i wanna read data from minio storage. im newbie. Thanks!
Beta Was this translation helpful? Give feedback.
All reactions