Hi,
I am trying to use S3 as an artifact dir, so that I don't need to worry
about the shared volume in the flink job and task manager.
Currently I am passing the --artifacts_dir as the pipeline options, for
example:
```
ARGS = [
"--streaming",
"--runner=portableRunner",
"--environment_type=PROCESS",
'--environment_config={"command":"/opt/apache/beam/boot"}',
'--artifacts_dir=s3a://my-bucket/artifacts',
]
with beam.Pipeline(options=PipelineOptions(pipeline_args)) as pipeline:
# running my pipeline.
```
However, the job is able to run, but I couldn't find the artifacts in my s3
bucket, and looks like the artifacts are still stored in the default
`/tmp/beam-artifact-staging` . Wondering how I can configure to use the s3
as artifact directory? Thanks!
Sincerely,
Lydian Lee