Wow, can't believe I didn't know about /threadDump before, thanks!
Anyways I mostly see lots (on the order of 100s) of:
"Computer.threadPoolForRemoting [#85811]" Id=380625 Group=main
TIMED_WAITING on
java.util.concurrent.SynchronousQueue$TransferStack@687826ae
at sun.misc.Unsafe.park(Native Method)
- waiting on java.util.concurrent.SynchronousQueue$TransferStack@687826ae
at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:215)
at
java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(SynchronousQueue.java:460)
at
java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:362)
at java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:941)
at
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1066)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1127)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
I did see some (~20) CPS executors that looked suspicious (they're also
hanging forever when i go to the job):
"Running CpsFlowExecution[Owner[workflow_test/619:workflow_test#619]]"
Id=7310 Group=main RUNNABLE
at
org.kohsuke.stapler.framework.io.LargeText$BufferSession.skip(LargeText.java:532)
at org.kohsuke.stapler.framework.io.LargeText.writeLogTo(LargeText.java:211)
at
hudson.console.AnnotatedLargeText.writeRawLogTo(AnnotatedLargeText.java:162)
at
org.jenkinsci.plugins.workflow.job.WorkflowRun.copyLogs(WorkflowRun.java:359)
at
org.jenkinsci.plugins.workflow.job.WorkflowRun.access$600(WorkflowRun.java:107)
at
org.jenkinsci.plugins.workflow.job.WorkflowRun$GraphL.onNewHead(WorkflowRun.java:752)
- locked java.util.concurrent.atomic.AtomicBoolean@7eb2df49
at
org.jenkinsci.plugins.workflow.cps.CpsFlowExecution.notifyListeners(CpsFlowExecution.java:799)
at
org.jenkinsci.plugins.workflow.cps.CpsThreadGroup$4.run(CpsThreadGroup.java:320)
at
org.jenkinsci.plugins.workflow.cps.CpsVmExecutorService$1.run(CpsVmExecutorService.java:32)
at
hudson.remoting.SingleLaneExecutorService$1.run(SingleLaneExecutorService.java:112)
at
jenkins.util.ContextResettingExecutorService$1.run(ContextResettingExecutorService.java:28)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Number of locked synchronizers = 1
- java.util.concurrent.ThreadPoolExecutor$Worker@d3b1e0f
I noticed them before but I assumed they weren't related to this, issue, I
think they're waiting for slaves nodes that have been removed from our
jenkins (not just disconnected but deleted), I'll clean these up with the
/kill url on workflows and see if that improves things since the threaddump
makes me wonder if these are related.
Thanks again
On Thursday, January 14, 2016 at 4:14:16 PM UTC-5, Christopher Orr wrote:
>
> On 14/01/16 21:30, Kevin wrote:
> > Hi all, I've got some custom cloud providers that don't seem to get
> > triggered, I've traced it back to NodeProvisionerInvoker not being
> > called (even after 30 minutes of waiting), I can call the
> > nodeProvisioners by hand (using the update function) in the same way
> > that the NodeProvisionerInvoker does in the groovy script console and
> > things seem to work. Actually it seems like a number of bits related to
> > PeriodicWork classes aren't working since I assume git polling and the
> > periodic build setting are a PeriodicWork things, but they're also not
> > triggering.
> >
> > I did also look at PeriodicWork.all().collect {it.class.name}
> > and println jenkins.util.Timer.get() and got:
> >
> > hudson.diagnosis.HudsonHomeDiskUsageChecker,
> > hudson.diagnosis.MemoryUsageMonitor,
> > hudson.model.FingerprintCleanupThread,
> > hudson.model.LoadStatistics$LoadStatisticsUpdater,
> > hudson.model.WorkspaceCleanupThread,
> > hudson.slaves.ComputerRetentionWork,
> > hudson.slaves.ConnectionActivityMonitor,
> > hudson.slaves.NodeProvisioner$NodeProvisionerInvoker,
> > hudson.triggers.Trigger$Cron,
> > jenkins.model.DownloadSettings$DailyCheck,
> > org.jenkinsci.plugins.periodicreincarnation.PeriodicReincarnation]
> >
> >
> > and
> >
> >
> jenkins.util.ErrorLoggingScheduledThreadPoolExecutor@6d72bb0e[Running,
> > pool size = 10, active threads = 10, queued tasks = 2521, completed
> > tasks = 1134830]
>
> If the number of active threads remains equal to the pool size, then I
> would guess that some tasks are getting stuck.
>
> Can you see a bunch of timer-related or otherwise suspicious-looking
> tasks running if you go to /threadDump on your Jenkins instance?
>
> Regards,
> Chris
>
--
You received this message because you are subscribed to the Google Groups
"Jenkins Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email
to [email protected].
To view this discussion on the web visit
https://groups.google.com/d/msgid/jenkinsci-users/2b6e0c8f-1ad7-4a48-9b3f-c2f78d05a8ea%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.