Back to Multiple platform build/check report for BioC 3.19: simplified long |
|
This page was generated on 2024-03-28 11:41:13 -0400 (Thu, 28 Mar 2024).
Hostname | OS | Arch (*) | R version | Installed pkgs |
---|---|---|---|---|
nebbiolo1 | Linux (Ubuntu 22.04.3 LTS) | x86_64 | R Under development (unstable) (2024-03-18 r86148) -- "Unsuffered Consequences" | 4708 |
palomino3 | Windows Server 2022 Datacenter | x64 | R Under development (unstable) (2024-03-16 r86144 ucrt) -- "Unsuffered Consequences" | 4446 |
lconway | macOS 12.7.1 Monterey | x86_64 | R Under development (unstable) (2024-03-18 r86148) -- "Unsuffered Consequences" | 4471 |
kunpeng2 | Linux (openEuler 22.03 LTS-SP1) | aarch64 | R Under development (unstable) (2024-03-19 r86153) -- "Unsuffered Consequences" | 4426 |
Click on any hostname to see more info about the system (e.g. compilers) (*) as reported by 'uname -p', except on Windows and Mac OS X |
Package 1730/2270 | Hostname | OS / Arch | INSTALL | BUILD | CHECK | BUILD BIN | ||||||||
RGMQL 1.23.1 (landing page) Simone Pallotta
| nebbiolo1 | Linux (Ubuntu 22.04.3 LTS) / x86_64 | OK | OK | WARNINGS | |||||||||
palomino3 | Windows Server 2022 Datacenter / x64 | OK | OK | ERROR | OK | |||||||||
lconway | macOS 12.7.1 Monterey / x86_64 | ... NOT SUPPORTED ... | ||||||||||||
kunpeng2 | Linux (openEuler 22.03 LTS-SP1) / aarch64 | OK | ERROR | skipped | ||||||||||
To the developers/maintainers of the RGMQL package: - Allow up to 24 hours (and sometimes 48 hours) for your latest push to git@git.bioconductor.org:packages/RGMQL.git to reflect on this report. See Troubleshooting Build Report for more information. - Use the following Renviron settings to reproduce errors and warnings. - If 'R CMD check' started to fail recently on the Linux builder(s) over a missing dependency, add the missing dependency to 'Suggests:' in your DESCRIPTION file. See Renviron.bioc for more information. - See Martin Grigorov's blog post for how to debug Linux ARM64 related issues on a x86_64 host. |
Package: RGMQL |
Version: 1.23.1 |
Command: /home/biocbuild/R/R-4.4-devel-2024.03.20/bin/R CMD build --keep-empty-dirs --no-resave-data RGMQL |
StartedAt: 2024-03-28 01:02:14 -0000 (Thu, 28 Mar 2024) |
EndedAt: 2024-03-28 01:03:18 -0000 (Thu, 28 Mar 2024) |
EllapsedTime: 64.1 seconds |
RetCode: 1 |
Status: ERROR |
PackageFile: None |
PackageFileSize: NA |
############################################################################## ############################################################################## ### ### Running command: ### ### /home/biocbuild/R/R-4.4-devel-2024.03.20/bin/R CMD build --keep-empty-dirs --no-resave-data RGMQL ### ############################################################################## ############################################################################## * checking for file ‘RGMQL/DESCRIPTION’ ... OK * preparing ‘RGMQL’: * checking DESCRIPTION meta-information ... OK * installing the package to build vignettes * creating vignettes ... ERROR --- re-building ‘RGMQL-vignette.Rmd’ using rmarkdown Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties 24/03/28 01:03:16 INFO SparkContext: Running Spark version 2.2.0 24/03/28 01:03:17 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 24/03/28 01:03:17 INFO SparkContext: Submitted application: GMQL-R 24/03/28 01:03:17 INFO SecurityManager: Changing view acls to: biocbuild 24/03/28 01:03:17 INFO SecurityManager: Changing modify acls to: biocbuild 24/03/28 01:03:17 INFO SecurityManager: Changing view acls groups to: 24/03/28 01:03:17 INFO SecurityManager: Changing modify acls groups to: 24/03/28 01:03:17 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(biocbuild); groups with view permissions: Set(); users with modify permissions: Set(biocbuild); groups with modify permissions: Set() 24/03/28 01:03:17 INFO PlatformDependent: Your platform does not provide complete low-level API for accessing direct buffers reliably. Unless explicitly requested, heap buffer will always be preferred to avoid potential system unstability. 24/03/28 01:03:17 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address. 24/03/28 01:03:17 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address. 24/03/28 01:03:17 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address. 24/03/28 01:03:17 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address. 24/03/28 01:03:17 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address. 24/03/28 01:03:17 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address. 24/03/28 01:03:17 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address. 24/03/28 01:03:17 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address. 24/03/28 01:03:17 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address. 24/03/28 01:03:17 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address. 24/03/28 01:03:17 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address. 24/03/28 01:03:17 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address. 24/03/28 01:03:17 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address. 24/03/28 01:03:17 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address. 24/03/28 01:03:17 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address. 24/03/28 01:03:17 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address. 24/03/28 01:03:17 ERROR SparkContext: Error initializing SparkContext. java.net.BindException: Cannot assign requested address: Service 'sparkDriver' failed after 16 retries (on a random free port)! Consider explicitly setting the appropriate binding address for the service 'sparkDriver' (for example spark.driver.bindAddress for SparkDriver) to the correct binding address. at java.base/sun.nio.ch.Net.bind0(Native Method) at java.base/sun.nio.ch.Net.bind(Net.java:555) at java.base/sun.nio.ch.ServerSocketChannelImpl.netBind(ServerSocketChannelImpl.java:337) at java.base/sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:294) at io.netty.channel.socket.nio.NioServerSocketChannel.doBind(NioServerSocketChannel.java:127) at io.netty.channel.AbstractChannel$AbstractUnsafe.bind(AbstractChannel.java:501) at io.netty.channel.DefaultChannelPipeline$HeadContext.bind(DefaultChannelPipeline.java:1218) at io.netty.channel.AbstractChannelHandlerContext.invokeBind(AbstractChannelHandlerContext.java:496) at io.netty.channel.AbstractChannelHandlerContext.bind(AbstractChannelHandlerContext.java:481) at io.netty.channel.DefaultChannelPipeline.bind(DefaultChannelPipeline.java:965) at io.netty.channel.AbstractChannel.bind(AbstractChannel.java:210) at io.netty.bootstrap.AbstractBootstrap$2.run(AbstractBootstrap.java:353) at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:399) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:446) at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:131) at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:144) at java.base/java.lang.Thread.run(Thread.java:833) 24/03/28 01:03:17 INFO SparkContext: Successfully stopped SparkContext Quitting from lines 250-251 [init] (RGMQL-vignette.Rmd) Error: processing vignette 'RGMQL-vignette.Rmd' failed with diagnostics: java.net.BindException: Cannot assign requested address: Service 'sparkDriver' failed after 16 retries (on a random free port)! Consider explicitly setting the appropriate binding address for the service 'sparkDriver' (for example spark.driver.bindAddress for SparkDriver) to the correct binding address. --- failed re-building ‘RGMQL-vignette.Rmd’ SUMMARY: processing the following file failed: ‘RGMQL-vignette.Rmd’ Error: Vignette re-building failed. Execution halted