See 
<https://builds.apache.org/job/Phoenix-master/1729/display/redirect?page=changes>

Changes:

[jamestaylor] PHOENIX-4070 Delete row should mask upserts at same timestamp

[jamestaylor] PHOENIX-4071 PDataType.compareTo(Object,Object,PDataType) does not

[jamestaylor] PHOENIX-4072 Prevent NPE for PreparedStatement.setObject of null

[jamestaylor] PHOENIX-4073 Disallow operations on a closed Connection

[jamestaylor] PHOENIX-3525 Cap automatic index rebuilding to inactive timestamp

[jamestaylor] PHOENIX-4074 Race condition in LazyValueGetter (Samarth Jain)

------------------------------------------
[...truncated 120.14 KB...]
        at 
org.apache.phoenix.hbase.index.write.ParallelWriterIndexCommitter$1.call(ParallelWriterIndexCommitter.java:129)
        at java.util.concurrent.FutureTask.run(FutureTask.java:262)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: 
Failed 1 action: org.apache.hadoop.hbase.DoNotRetryIOException
        at 
org.apache.phoenix.end2end.index.MutableIndexFailureIT$FailingRegionObserver.preBatchMutate(MutableIndexFailureIT.java:496)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$35.call(RegionCoprocessorHost.java:1007)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$RegionOperation.call(RegionCoprocessorHost.java:1674)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1750)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1706)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.preBatchMutate(RegionCoprocessorHost.java:1003)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutation(HRegion.java:3217)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3004)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2946)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.doBatchOp(RSRpcServices.java:843)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicRegionMutation(RSRpcServices.java:805)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.multi(RSRpcServices.java:2299)
        at 
org.apache.hadoop.hbase.client.MultiServerCallable.call(MultiServerCallable.java:136)
        at 
org.apache.hadoop.hbase.client.MultiServerCallable.call(MultiServerCallable.java:52)
        at 
org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithoutRetries(RpcRetryingCaller.java:212)
        at 
org.apache.hadoop.hbase.client.AsyncProcess$AsyncRequestFutureImpl$SingleServerRequestRunnable.run(AsyncProcess.java:771)
        at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
        at java.util.concurrent.FutureTask.run(FutureTask.java:262)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:745)
: 1 time, servers with issues: qnode1.quenda.co,41243,1502174038361, 
        at 
org.apache.hadoop.hbase.client.AsyncProcess$BatchErrors.makeException(AsyncProcess.java:260)
        at 
org.apache.hadoop.hbase.client.AsyncProcess$BatchErrors.access$2400(AsyncProcess.java:240)
        at 
org.apache.hadoop.hbase.client.AsyncProcess$AsyncRequestFutureImpl.getErrors(AsyncProcess.java:1733)
        at org.apache.hadoop.hbase.client.HTable.batch(HTable.java:912)
        at org.apache.hadoop.hbase.client.HTable.batch(HTable.java:926)
        at 
org.apache.phoenix.hbase.index.write.ParallelWriterIndexCommitter$1.call(ParallelWriterIndexCommitter.java:166)
        ... 5 more
: 1 time, servers with issues: qnode1.quenda.co,41243,1502174038361, 
        at 
org.apache.phoenix.end2end.index.MutableIndexFailureIT.updateTableAgain(MutableIndexFailureIT.java:450)
        at 
org.apache.phoenix.end2end.index.MutableIndexFailureIT.testIndexWriteFailure(MutableIndexFailureIT.java:259)
Caused by: org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: 
Failed 1 action: org.apache.hadoop.hbase.DoNotRetryIOException: Unable to 
update the following indexes: 
[T000018.A_FAIL_IDX],serverTimestamp=1502174368952,
        at 
org.apache.phoenix.util.ServerUtil.wrapInDoNotRetryIOException(ServerUtil.java:213)
        at 
org.apache.phoenix.index.PhoenixIndexFailurePolicy.handleFailure(PhoenixIndexFailurePolicy.java:148)
        at 
org.apache.phoenix.hbase.index.write.IndexWriter.writeAndKillYourselfOnFailure(IndexWriter.java:155)
        at 
org.apache.phoenix.hbase.index.write.IndexWriter.writeAndKillYourselfOnFailure(IndexWriter.java:139)
        at 
org.apache.phoenix.hbase.index.Indexer.doPostWithExceptions(Indexer.java:608)
        at org.apache.phoenix.hbase.index.Indexer.doPost(Indexer.java:556)
        at 
org.apache.phoenix.hbase.index.Indexer.postBatchMutateIndispensably(Indexer.java:540)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$37.call(RegionCoprocessorHost.java:1034)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$RegionOperation.call(RegionCoprocessorHost.java:1674)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1750)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1706)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.postBatchMutateIndispensably(RegionCoprocessorHost.java:1030)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutation(HRegion.java:3475)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3004)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2946)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.doBatchOp(RSRpcServices.java:843)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicRegionMutation(RSRpcServices.java:805)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.multi(RSRpcServices.java:2299)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:34958)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2347)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:123)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:188)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:168)
Caused by: 
org.apache.phoenix.hbase.index.exception.SingleIndexWriteFailureException: 
Failed 1 action: org.apache.hadoop.hbase.DoNotRetryIOException
        at 
org.apache.phoenix.end2end.index.MutableIndexFailureIT$FailingRegionObserver.preBatchMutate(MutableIndexFailureIT.java:496)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$35.call(RegionCoprocessorHost.java:1007)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$RegionOperation.call(RegionCoprocessorHost.java:1674)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1750)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1706)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.preBatchMutate(RegionCoprocessorHost.java:1003)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutation(HRegion.java:3217)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3004)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2946)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.doBatchOp(RSRpcServices.java:843)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicRegionMutation(RSRpcServices.java:805)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.multi(RSRpcServices.java:2299)
        at 
org.apache.hadoop.hbase.client.MultiServerCallable.call(MultiServerCallable.java:136)
        at 
org.apache.hadoop.hbase.client.MultiServerCallable.call(MultiServerCallable.java:52)
        at 
org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithoutRetries(RpcRetryingCaller.java:212)
        at 
org.apache.hadoop.hbase.client.AsyncProcess$AsyncRequestFutureImpl$SingleServerRequestRunnable.run(AsyncProcess.java:771)
        at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
        at java.util.concurrent.FutureTask.run(FutureTask.java:262)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:745)
: 1 time, servers with issues: qnode1.quenda.co,41243,1502174038361, 
        at 
org.apache.phoenix.hbase.index.write.ParallelWriterIndexCommitter$1.call(ParallelWriterIndexCommitter.java:170)
        at 
org.apache.phoenix.hbase.index.write.ParallelWriterIndexCommitter$1.call(ParallelWriterIndexCommitter.java:129)
        at java.util.concurrent.FutureTask.run(FutureTask.java:262)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.client.RetriesExhaustedWithDetailsException: 
Failed 1 action: org.apache.hadoop.hbase.DoNotRetryIOException
        at 
org.apache.phoenix.end2end.index.MutableIndexFailureIT$FailingRegionObserver.preBatchMutate(MutableIndexFailureIT.java:496)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$35.call(RegionCoprocessorHost.java:1007)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$RegionOperation.call(RegionCoprocessorHost.java:1674)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1750)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1706)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.preBatchMutate(RegionCoprocessorHost.java:1003)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.doMiniBatchMutation(HRegion.java:3217)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:3004)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.batchMutate(HRegion.java:2946)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.doBatchOp(RSRpcServices.java:843)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.doNonAtomicRegionMutation(RSRpcServices.java:805)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.multi(RSRpcServices.java:2299)
        at 
org.apache.hadoop.hbase.client.MultiServerCallable.call(MultiServerCallable.java:136)
        at 
org.apache.hadoop.hbase.client.MultiServerCallable.call(MultiServerCallable.java:52)
        at 
org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithoutRetries(RpcRetryingCaller.java:212)
        at 
org.apache.hadoop.hbase.client.AsyncProcess$AsyncRequestFutureImpl$SingleServerRequestRunnable.run(AsyncProcess.java:771)
        at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
        at java.util.concurrent.FutureTask.run(FutureTask.java:262)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:745)
: 1 time, servers with issues: qnode1.quenda.co,41243,1502174038361, 
        at 
org.apache.hadoop.hbase.client.AsyncProcess$BatchErrors.makeException(AsyncProcess.java:260)
        at 
org.apache.hadoop.hbase.client.AsyncProcess$BatchErrors.access$2400(AsyncProcess.java:240)
        at 
org.apache.hadoop.hbase.client.AsyncProcess$AsyncRequestFutureImpl.getErrors(AsyncProcess.java:1733)
        at org.apache.hadoop.hbase.client.HTable.batch(HTable.java:912)
        at org.apache.hadoop.hbase.client.HTable.batch(HTable.java:926)
        at 
org.apache.phoenix.hbase.index.write.ParallelWriterIndexCommitter$1.call(ParallelWriterIndexCommitter.java:166)
        ... 5 more
: 1 time, servers with issues: qnode1.quenda.co,41243,1502174038361, 
        at 
org.apache.phoenix.end2end.index.MutableIndexFailureIT.updateTableAgain(MutableIndexFailureIT.java:450)
        at 
org.apache.phoenix.end2end.index.MutableIndexFailureIT.testIndexWriteFailure(MutableIndexFailureIT.java:259)

[INFO] Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 22.522 s 
- in org.apache.phoenix.rpc.PhoenixServerRpcIT
[INFO] Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 163.382 
s - in org.apache.phoenix.iterate.ScannerLeaseRenewalIT
[ERROR] Tests run: 11, Failures: 8, Errors: 0, Skipped: 0, Time elapsed: 
533.807 s <<< FAILURE! - in 
org.apache.phoenix.end2end.index.PartialIndexRebuilderIT
[ERROR] 
testMultiValuesAtSameTS(org.apache.phoenix.end2end.index.PartialIndexRebuilderIT)
  Time elapsed: 64.03 s  <<< FAILURE!
java.lang.AssertionError
        at 
org.apache.phoenix.end2end.index.PartialIndexRebuilderIT.testMultiValuesAtSameTS(PartialIndexRebuilderIT.java:297)

[ERROR] 
testDeleteAndUpsertValuesAtSameTS1(org.apache.phoenix.end2end.index.PartialIndexRebuilderIT)
  Time elapsed: 63.388 s  <<< FAILURE!
java.lang.AssertionError
        at 
org.apache.phoenix.end2end.index.PartialIndexRebuilderIT.testDeleteAndUpsertValuesAtSameTS1(PartialIndexRebuilderIT.java:328)

[ERROR] 
testDeleteAndUpsertValuesAtSameTS2(org.apache.phoenix.end2end.index.PartialIndexRebuilderIT)
  Time elapsed: 65.364 s  <<< FAILURE!
java.lang.AssertionError
        at 
org.apache.phoenix.end2end.index.PartialIndexRebuilderIT.testDeleteAndUpsertValuesAtSameTS2(PartialIndexRebuilderIT.java:359)

[ERROR] 
testUpsertNullAfterFailure(org.apache.phoenix.end2end.index.PartialIndexRebuilderIT)
  Time elapsed: 64.396 s  <<< FAILURE!
java.lang.AssertionError
        at 
org.apache.phoenix.end2end.index.PartialIndexRebuilderIT.testUpsertNullAfterFailure(PartialIndexRebuilderIT.java:183)

[ERROR] 
testMultiVersionsAfterFailure(org.apache.phoenix.end2end.index.PartialIndexRebuilderIT)
  Time elapsed: 65.359 s  <<< FAILURE!
java.lang.AssertionError
        at 
org.apache.phoenix.end2end.index.PartialIndexRebuilderIT.testMultiVersionsAfterFailure(PartialIndexRebuilderIT.java:154)

[ERROR] 
testUpsertNullTwiceAfterFailure(org.apache.phoenix.end2end.index.PartialIndexRebuilderIT)
  Time elapsed: 64.456 s  <<< FAILURE!
java.lang.AssertionError
        at 
org.apache.phoenix.end2end.index.PartialIndexRebuilderIT.testUpsertNullTwiceAfterFailure(PartialIndexRebuilderIT.java:212)

[ERROR] 
testDeleteAfterFailure(org.apache.phoenix.end2end.index.PartialIndexRebuilderIT)
  Time elapsed: 63.312 s  <<< FAILURE!
java.lang.AssertionError
        at 
org.apache.phoenix.end2end.index.PartialIndexRebuilderIT.testDeleteAfterFailure(PartialIndexRebuilderIT.java:239)

[ERROR] 
testDeleteBeforeFailure(org.apache.phoenix.end2end.index.PartialIndexRebuilderIT)
  Time elapsed: 64.134 s  <<< FAILURE!
java.lang.AssertionError
        at 
org.apache.phoenix.end2end.index.PartialIndexRebuilderIT.testDeleteBeforeFailure(PartialIndexRebuilderIT.java:266)

[INFO] Tests run: 140, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 
843.503 s - in org.apache.phoenix.end2end.IndexExtendedIT
[INFO] 
[INFO] Results:
[INFO] 
[ERROR] Failures: 
[ERROR]   PartialIndexRebuilderIT.testDeleteAfterFailure:239
[ERROR]   PartialIndexRebuilderIT.testDeleteAndUpsertValuesAtSameTS1:328
[ERROR]   PartialIndexRebuilderIT.testDeleteAndUpsertValuesAtSameTS2:359
[ERROR]   PartialIndexRebuilderIT.testDeleteBeforeFailure:266
[ERROR]   PartialIndexRebuilderIT.testMultiValuesAtSameTS:297
[ERROR]   PartialIndexRebuilderIT.testMultiVersionsAfterFailure:154
[ERROR]   PartialIndexRebuilderIT.testUpsertNullAfterFailure:183
[ERROR]   PartialIndexRebuilderIT.testUpsertNullTwiceAfterFailure:212
[ERROR] Errors: 
[ERROR]   MutableIndexFailureIT.testIndexWriteFailure:259->updateTableAgain:450 
» Commit
[INFO] 
[ERROR] Tests run: 640, Failures: 8, Errors: 1, Skipped: 56
[INFO] 
[INFO] 
[INFO] --- maven-failsafe-plugin:2.20:verify (ParallelStatsEnabledTest) @ 
phoenix-core ---
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] Apache Phoenix ..................................... SUCCESS [  4.382 s]
[INFO] Phoenix Core ....................................... FAILURE [  01:35 h]
[INFO] Phoenix - Flume .................................... SKIPPED
[INFO] Phoenix - Kafka .................................... SKIPPED
[INFO] Phoenix - Pig ...................................... SKIPPED
[INFO] Phoenix Query Server Client ........................ SKIPPED
[INFO] Phoenix Query Server ............................... SKIPPED
[INFO] Phoenix - Pherf .................................... SKIPPED
[INFO] Phoenix - Spark .................................... SKIPPED
[INFO] Phoenix - Hive ..................................... SKIPPED
[INFO] Phoenix Client ..................................... SKIPPED
[INFO] Phoenix Server ..................................... SKIPPED
[INFO] Phoenix Assembly ................................... SKIPPED
[INFO] Phoenix - Tracing Web Application .................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 01:36 h
[INFO] Finished at: 2017-08-08T06:45:06Z
[INFO] Final Memory: 59M/657M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal 
org.apache.maven.plugins:maven-failsafe-plugin:2.20:verify 
(ParallelStatsEnabledTest) on project phoenix-core: There are test failures.
[ERROR] 
[ERROR] Please refer to 
<https://builds.apache.org/job/Phoenix-master/ws/phoenix-core/target/failsafe-reports>
 for the individual test results.
[ERROR] Please refer to dump files (if any exist) [date]-jvmRun[N].dump, 
[date].dumpstream and [date]-jvmRun[N].dumpstream.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please 
read the following articles:
[ERROR] [Help 1] 
http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :phoenix-core
Build step 'Invoke top-level Maven targets' marked build as failure
Archiving artifacts
[Fast Archiver] Compressed -874161186 B of artifacts by 182.3% relative to #1700
ERROR: Step ‘JIRA: Update relevant issues’ aborted due to exception: 
java.lang.ClassNotFoundException: com.atlassian.fugue.Effect
        at 
jenkins.util.AntClassLoader.findClassInComponents(AntClassLoader.java:1374)
        at jenkins.util.AntClassLoader.findClass(AntClassLoader.java:1327)
        at jenkins.util.AntClassLoader.loadClass(AntClassLoader.java:1080)
        at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
Caused: java.lang.NoClassDefFoundError: com/atlassian/fugue/Effect
        at 
com.atlassian.httpclient.apache.httpcomponents.DefaultHttpClientFactory.doCreate(DefaultHttpClientFactory.java:68)
        at 
com.atlassian.httpclient.apache.httpcomponents.DefaultHttpClientFactory.create(DefaultHttpClientFactory.java:35)
        at 
com.atlassian.jira.rest.client.internal.async.AsynchronousHttpClientFactory.createClient(AsynchronousHttpClientFactory.java:63)
        at 
com.atlassian.jira.rest.client.internal.async.AsynchronousJiraRestClientFactory.create(AsynchronousJiraRestClientFactory.java:35)
        at 
com.atlassian.jira.rest.client.internal.async.AsynchronousJiraRestClientFactory.createWithBasicHttpAuthentication(AsynchronousJiraRestClientFactory.java:42)
        at hudson.plugins.jira.JiraSite.createSession(JiraSite.java:280)
        at hudson.plugins.jira.JiraSite.getSession(JiraSite.java:255)
        at hudson.plugins.jira.Updater.perform(Updater.java:95)
        at 
hudson.plugins.jira.JiraIssueUpdater.perform(JiraIssueUpdater.java:64)
        at 
hudson.tasks.BuildStepCompatibilityLayer.perform(BuildStepCompatibilityLayer.java:81)
        at hudson.tasks.BuildStepMonitor$3.perform(BuildStepMonitor.java:45)
        at 
hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:735)
        at 
hudson.model.AbstractBuild$AbstractBuildExecution.performAllBuildSteps(AbstractBuild.java:676)
        at hudson.model.Build$BuildExecution.post2(Build.java:186)
        at 
hudson.model.AbstractBuild$AbstractBuildExecution.post(AbstractBuild.java:621)
        at hudson.model.Run.execute(Run.java:1760)
        at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
        at hudson.model.ResourceController.execute(ResourceController.java:97)
        at hudson.model.Executor.run(Executor.java:405)
Recording test results

Reply via email to