Java Apache Beam - save file “LOCALY” by using DataflowRunner

邮差的信 提交于 2019-12-13 05:23:52

问题


Can send the java code but currently, it's not necessary.

I have an issue as when I run the job as (DirectRunner - using Google VM Instance) it is working fine, as it saves information to the local file and carries on...

The problem appears when trying to use (DataflowRunner), and the error which I receive:

java.nio.file.NoSuchFileExtension: XXXX.csv
.....
.....
XXXX.csv could not be delete.

It could be deleted as it not even created.

Problem - how to write the file locally when running through DataflowRunner??

P.S. Using Apache Beam

Pipeline (part of the code) - Reading from BigQuery and store data to Google storage (Special Character issue)


回答1:


AFAIK when it is ran as a dataflow instance, you have to write file to GCS service (aka storage bucket) rather than local disk.

Did you try that already? to create storage bucket: https://cloud.google.com/storage/docs/creating-buckets



来源:https://stackoverflow.com/questions/56156510/java-apache-beam-save-file-localy-by-using-dataflowrunner

易学教程内所有资源均来自网络或用户发布的内容,如有违反法律规定的内容欢迎反馈
该文章没有解决你所遇到的问题?点击提问,说说你的问题,让更多的人一起探讨吧!