See
<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/4726/display/redirect>
Changes:
------------------------------------------
[...truncated 1.08 MB...]
capabilities: "beam:coder:double:v1"
capabilities: "beam:coder:row:v1"
capabilities: "beam:coder:param_windowed_value:v1"
capabilities: "beam:coder:state_backed_iterable:v1"
capabilities: "beam:coder:sharded_key:v1"
capabilities: "beam:coder:nullable:v1"
capabilities: "beam:protocol:multi_core_bundle_processing:v1"
capabilities: "beam:protocol:progress_reporting:v1"
capabilities: "beam:protocol:harness_monitoring_infos:v1"
capabilities: "beam:protocol:control_request_elements_embedding:v1"
capabilities: "beam:protocol:state_caching:v1"
capabilities: "beam:version:sdk_base:apache/beam_java8_sdk:2.46.0.dev"
capabilities: "beam:transform:sdf_truncate_sized_restrictions:v1"
capabilities: "beam:transform:to_string:v1"
dependencies: <
type_urn: "beam:artifact:type:file:v1"
type_payload:
"\nL/tmp/artifacts/icedtea-sound-5xFkOD7aUmX8Jiw7lPqhn75zkNGCyUOfOuqi2rClWKM.jar\022@e71164383eda5265fc262c3b94faa19fbe7390d182c9439f3aeaa2dab0a558a3"
role_urn: "beam:artifact:role:staging_to:v1"
role_payload:
"\n=icedtea-sound-5xFkOD7aUmX8Jiw7lPqhn75zkNGCyUOfOuqi2rClWKM.jar"
>
dependencies: <
type_urn: "beam:artifact:type:file:v1"
type_payload:
"\nF/tmp/artifacts/jaccess-Rlp7GfuSFOBCBGdO8af2XxrTz5LqaQqwSMAXumdbHN0.jar\022@465a7b19fb9214e04204674ef1a7f65f1ad3cf92ea690ab048c017ba675b1cdd"
role_urn: "beam:artifact:role:staging_to:v1"
role_payload:
"\n7jaccess-Rlp7GfuSFOBCBGdO8af2XxrTz5LqaQqwSMAXumdbHN0.jar"
>
dependencies: <
type_urn: "beam:artifact:type:file:v1"
type_payload:
"\nI/tmp/artifacts/localedata-Nuzln2shadVmT-PppqaNc7GHBbUyMUb9pXsJTjRiAzs.jar\022@36ece59f6b2169d5664fe3e9a6a68d73b18705b5323146fda57b094e3462033b"
role_urn: "beam:artifact:role:staging_to:v1"
role_payload:
"\n:localedata-Nuzln2shadVmT-PppqaNc7GHBbUyMUb9pXsJTjRiAzs.jar"
>
dependencies: <
type_urn: "beam:artifact:type:file:v1"
type_payload:
"\nF/tmp/artifacts/nashorn-1iQRhJvU2nRCSB4ucg6f4b69TGiwE1yq0_LGG2MU5N0.jar\022@d62411849bd4da7442481e2e720e9fe1bebd4c68b0135caad3f2c61b6314e4dd"
role_urn: "beam:artifact:role:staging_to:v1"
role_payload:
"\n7nashorn-1iQRhJvU2nRCSB4ucg6f4b69TGiwE1yq0_LGG2MU5N0.jar"
>
dependencies: <
type_urn: "beam:artifact:type:file:v1"
type_payload:
"\nG/tmp/artifacts/cldrdata-MTK3mMuUzlWYlYkweGQNzI7dFpuLG0_8F8oxT0khQDc.jar\022@3132b798cb94ce559895893078640dcc8edd169b8b1b4ffc17ca314f49214037"
role_urn: "beam:artifact:role:staging_to:v1"
role_payload:
"\n8cldrdata-MTK3mMuUzlWYlYkweGQNzI7dFpuLG0_8F8oxT0khQDc.jar"
>
dependencies: <
type_urn: "beam:artifact:type:file:v1"
type_payload:
"\nD/tmp/artifacts/dnsns-YTF3f6B-OG4FmXLRf2ckHBi4VJnspBZTyIQCz-6Spwk.jar\022@6131777fa07e386e059972d17f67241c18b85499eca41653c88402cfee92a709"
role_urn: "beam:artifact:role:staging_to:v1"
role_payload: "\n5dnsns-YTF3f6B-OG4FmXLRf2ckHBi4VJnspBZTyIQCz-6Spwk.jar"
>
dependencies: <
type_urn: "beam:artifact:type:file:v1"
type_payload:
"\nr/tmp/artifacts/beam-sdks-java-io-expansion-service-2.46.0-SNAPSHOT-QrbUgaay_6e6lF0fiy5bRvZfNxhZB6msVxWyudt_YBM.jar\022@42b6d481a6b2ffa7ba945d1f8b2e5b46f65f37185907a9ac5715b2b9db7f6013"
role_urn: "beam:artifact:role:staging_to:v1"
role_payload:
"\ncbeam-sdks-java-io-expansion-service-2.46.0-SNAPSHOT-QrbUgaay_6e6lF0fiy5bRvZfNxhZB6msVxWyudt_YBM.jar"
>
>
>
>
root_transform_ids: "e4"
root_transform_ids: "e5"
root_transform_ids: "s1"
root_transform_ids: "s3"
requirements: "beam:requirement:pardo:splittable_dofn:v1"
2023/01/23 00:17:37 Cross-compiling
<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/go/test/integration/io/xlang/kafka/kafka_test.go>
as /tmp/worker-2-1674433057769559631
2023/01/23 00:17:39 Prepared job with id:
go-testkafkaio_basicreadwrite-929_6fd2267f-9df6-410d-a940-8c6c861fed6f and
staging token:
go-testkafkaio_basicreadwrite-929_6fd2267f-9df6-410d-a940-8c6c861fed6f
2023/01/23 00:17:39 Staged binary artifact with token:
2023/01/23 00:17:39 Submitted job:
go0testkafkaio0basicreadwrite0929-jenkins-0123001739-498aa2f5_b110975d-1390-4c2f-b807-8d3f9080a426
2023/01/23 00:17:39 Job state: STOPPED
2023/01/23 00:17:39 Job state: STARTING
2023/01/23 00:17:39 Job state: RUNNING
2023/01/23 00:18:35 Job state: DONE
2023/01/23 00:18:35 Warning: 6 errors during metrics processing: [failed to
deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1"
type:"beam:metrics:sum_int64:v1" payload:"\x01" labels:{key:"PCOLLECTION"
value:"qSkLrmJiCxExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/PairWithRestriction0"}
failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1"
type:"beam:metrics:sum_int64:v1" payload:"\x01" labels:{key:"PCOLLECTION"
value:"qSkLrmJiCxExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"}
failed to deduce Step from MonitoringInfo:
urn:"beam:metric:sampled_byte_size:v1"
type:"beam:metrics:distribution_int64:v1" payload:"\x01\xf0\"\xf0\"\xf0\""
labels:{key:"PCOLLECTION"
value:"qSkLrmJiCxExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/PairWithRestriction0"}
failed to deduce Step from MonitoringInfo:
urn:"beam:metric:sampled_byte_size:v1"
type:"beam:metrics:distribution_int64:v1" payload:"\x01\xf8\"\xf8\"\xf8\""
labels:{key:"PCOLLECTION"
value:"qSkLrmJiCxExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"}
failed to deduce Step from MonitoringInfo: urn:"beam:metric:element_count:v1"
type:"beam:metrics:sum_int64:v1" payload:"\x01" labels:{key:"PCOLLECTION"
value:"qSkLrmJiCxExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"}
failed to deduce Step from MonitoringInfo:
urn:"beam:metric:sampled_byte_size:v1"
type:"beam:metrics:distribution_int64:v1" payload:"\x01\xfd\"\xfd\"\xfd\""
labels:{key:"PCOLLECTION"
value:"qSkLrmJiCxExternal/KafkaIO.Read/KafkaIO.Read.ReadFromKafkaViaUnbounded/Read(KafkaUnboundedSource)/Create/Read(CreateSource)/ParDo(OutputSingleSource)/ParMultiDo(OutputSingleSource).output/SplitAndSize0"}]
--- PASS: TestKafkaIO_BasicReadWrite (103.66s)
PASS
ok github.com/apache/beam/sdks/v2/go/test/integration/io/xlang/kafka
109.695s
$ cd ../..
$ exit
$ exit
> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerJavaUsingPython
INFO:apache_beam.runners.portability.stager:Executing command:
['/home/jenkins/.apache_beam/cache/venvs/py-3.8-beam-2.44.0-743bdccef67087512a7a190820e727aab7bdcb14/bin/python',
'-m', 'pip', 'download', '--dest', '/tmp/dataflow-requirements-cache', '-r',
'/tmp/tmpownfz4rh/tmp_requirements.txt', '--exists-action', 'i', '--no-deps',
'--implementation', 'cp', '--abi', 'cp38', '--platform', 'manylinux2014_x86_64']
INFO:root:Default Python SDK image for environment is
apache/beam_python3.8_sdk:2.44.0
INFO:root:No image given, using default Python SDK image
INFO:root:Default Python SDK image for environment is
apache/beam_python3.8_sdk:2.44.0
INFO:root:Python SDK container image set to "apache/beam_python3.8_sdk:2.44.0"
for Docker environment
INFO:__main__:Listening for expansion requests at 41057
INFO:root:Missing pipeline option (runner). Executing pipeline using the
default runner: DirectRunner.
INFO:__main__:Shutting down expansion service.
org.apache.beam.sdk.extensions.python.transforms.RunInferenceTransformTest >
testRunInference FAILED
java.lang.RuntimeException at PortableRunner.java:205
Caused by: java.util.concurrent.ExecutionException at
CompletableFuture.java:357
Caused by:
org.apache.beam.vendor.grpc.v1p48p1.io.grpc.StatusRuntimeException at
Status.java:535
15 tests completed, 1 failed
> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerJavaUsingPython
> FAILED
> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingJava
>>> RUNNING integration tests with pipeline options: --runner=PortableRunner
>>> --job_endpoint=localhost:43797 --environment_cache_millis=10000
>>> --experiments=beam_fn_api
>>> pytest options:
>>> collect markers: -m=uses_java_expansion_service
ImportError while loading conftest
'<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/conftest.py'.>
conftest.py:22: in <module>
from apache_beam.options import pipeline_options
apache_beam/__init__.py:93: in <module>
from apache_beam import io
apache_beam/io/__init__.py:21: in <module>
from apache_beam.io.avroio import *
apache_beam/io/avroio.py:53: in <module>
from apache_beam.io import filebasedsink
apache_beam/io/filebasedsink.py:29: in <module>
from apache_beam.io import iobase
apache_beam/io/iobase.py:54: in <module>
from apache_beam.transforms import Impulse
apache_beam/transforms/__init__.py:24: in <module>
from apache_beam.transforms.external import *
apache_beam/transforms/external.py:44: in <module>
from apache_beam.runners import pipeline_context
apache_beam/runners/__init__.py:24: in <module>
from apache_beam.runners.direct.test_direct_runner import TestDirectRunner
/usr/lib/python3.7/os.py:223: in makedirs
mkdir(name, mode)
E OSError: [Errno 28] No space left on device:
'<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/apache_beam/runners/direct/__pycache__'>
> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingJava
> FAILED
> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingSql
> FAILED
>>> RUNNING integration tests with pipeline options: --runner=PortableRunner
>>> --job_endpoint=localhost:43797 --environment_cache_millis=10000
>>> --experiments=beam_fn_api
>>> pytest options:
>>> collect markers: -m=xlang_sql_expansion_service
Traceback (most recent call last):
File
"<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/bin/pytest",>
line 8, in <module>
sys.exit(console_main())
File
"<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.7/site-packages/_pytest/config/__init__.py",>
line 190, in console_main
code = main()
File
"<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.7/site-packages/_pytest/config/__init__.py",>
line 148, in main
config = _prepareconfig(args, plugins)
File
"<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.7/site-packages/_pytest/config/__init__.py",>
line 330, in _prepareconfig
pluginmanager=pluginmanager, args=args
File
"<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.7/site-packages/pluggy/_hooks.py",>
line 265, in __call__
return self._hookexec(self.name, self.get_hookimpls(), kwargs, firstresult)
File
"<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.7/site-packages/pluggy/_manager.py",>
line 80, in _hookexec
return self._inner_hookexec(hook_name, methods, kwargs, firstresult)
File
"<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.7/site-packages/pluggy/_callers.py",>
line 55, in _multicall
gen.send(outcome)
File
"<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.7/site-packages/_pytest/helpconfig.py",>
line 103, in pytest_cmdline_parse
config: Config = outcome.get_result()
File
"<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.7/site-packages/pluggy/_result.py",>
line 60, in get_result
raise ex[1].with_traceback(ex[2])
File
"<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.7/site-packages/pluggy/_callers.py",>
line 39, in _multicall
res = hook_impl.function(*args)
File
"<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.7/site-packages/_pytest/config/__init__.py",>
line 1058, in pytest_cmdline_parse
self.parse(args)
File
"<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.7/site-packages/_pytest/config/__init__.py",>
line 1346, in parse
self._preparse(args, addopts=addopts)
File
"<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.7/site-packages/_pytest/config/__init__.py",>
line 1249, in _preparse
early_config=self, args=args, parser=self._parser
File
"<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.7/site-packages/pluggy/_hooks.py",>
line 265, in __call__
return self._hookexec(self.name, self.get_hookimpls(), kwargs, firstresult)
File
"<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.7/site-packages/pluggy/_manager.py",>
line 80, in _hookexec
return self._inner_hookexec(hook_name, methods, kwargs, firstresult)
File
"<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.7/site-packages/pluggy/_callers.py",>
line 60, in _multicall
return outcome.get_result()
File
"<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.7/site-packages/pluggy/_result.py",>
line 60, in get_result
raise ex[1].with_traceback(ex[2])
File
"<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.7/site-packages/pluggy/_callers.py",>
line 34, in _multicall
next(gen) # first yield
File
"<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.7/site-packages/_pytest/capture.py",>
line 141, in pytest_load_initial_conftests
capman.start_global_capturing()
File
"<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.7/site-packages/_pytest/capture.py",>
line 688, in start_global_capturing
self._global_capturing = _get_multicapture(self._method)
File
"<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.7/site-packages/_pytest/capture.py",>
line 630, in _get_multicapture
return MultiCapture(in_=FDCapture(0), out=FDCapture(1), err=FDCapture(2))
File
"<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/build/gradleenv/1922375555/lib/python3.7/site-packages/_pytest/capture.py",>
line 388, in __init__
TemporaryFile(buffering=0),
File "/usr/lib/python3.7/tempfile.py", line 725, in TemporaryFile
prefix, suffix, dir, output_type = _sanitize_params(prefix, suffix, dir)
File "/usr/lib/python3.7/tempfile.py", line 265, in _sanitize_params
dir = gettempdir()
File "/usr/lib/python3.7/tempfile.py", line 433, in gettempdir
tempdir = _get_default_tempdir()
File "/usr/lib/python3.7/tempfile.py", line 368, in _get_default_tempdir
dirlist)
FileNotFoundError: [Errno 2] No usable temporary directory found in ['/tmp',
'/var/tmp', '/usr/tmp',
'<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python']>
> Task :runners:spark:2:job-server:sparkJobServerCleanup
<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/sdks/python/scripts/run_job_server.sh>:
line 19: cannot create temp file for here-document: No space left on device
Stopping job server pid: 3505391.
> Task :runners:spark:2:job-server:validatesCrossLanguageRunnerCleanup
Stopping expansion service pid: 3532735.
Stopping expansion service pid: 3532736.
FAILURE: Build completed with 3 failures.
1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task
':runners:spark:2:job-server:validatesCrossLanguageRunnerJavaUsingPython'.
> There were failing tests. See the report at:
> file://<https://ci-beam.apache.org/job/beam_PostCommit_XVR_Spark/ws/src/runners/spark/2/job-server/build/reports/tests/validatesCrossLanguageRunnerJavaUsingPython/index.html>
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
==============================================================================
2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task
':runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingJava'.
> Process 'command 'sh'' finished with non-zero exit value 4
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
==============================================================================
3: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task
':runners:spark:2:job-server:validatesCrossLanguageRunnerPythonUsingSql'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
==============================================================================
* Get more help at https://help.gradle.org
Deprecated Gradle features were used in this build, making it incompatible with
Gradle 8.0.
You can use '--warning-mode all' to show the individual deprecation warnings
and determine if they come from your own scripts or plugins.
See
https://docs.gradle.org/7.5.1/userguide/command_line_interface.html#sec:command_line_warnings
Execution optimizations have been disabled for 1 invalid unit(s) of work during
this build to ensure correctness.
Please consult deprecation warnings for more details.
BUILD FAILED in 48m 17s
268 actionable tasks: 178 executed, 78 from cache, 12 up-to-date
Publishing build scan...
https://gradle.com/s/sjxz2aqogy2we
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]