Hi, Sorry this may be a little off topic but I tried searching for docs on history server but couldn't really find much. Can someone point me to a doc or give me a point of reference for the use and intent of a history server?
-- Ankur > On 7 May 2015, at 12:06, Koert Kuipers <ko...@tresata.com> wrote: > > got it. thanks! > > On Thu, May 7, 2015 at 2:52 PM, Marcelo Vanzin <van...@cloudera.com> wrote: > Ah, sorry, that's definitely what Shixiong mentioned. The patch I mentioned > did not make it into 1.3... > > On Thu, May 7, 2015 at 11:48 AM, Koert Kuipers <ko...@tresata.com> wrote: > seems i got one thread spinning 100% for a while now, in > FsHistoryProvider.initialize(). maybe something wrong with my logs on hdfs > that its reading? or could it simply really take 30 mins to read all the > history on dhfs? > > jstack: > > Deadlock Detection: > > No deadlocks found. > > Thread 2272: (state = BLOCKED) > - sun.misc.Unsafe.park(boolean, long) @bci=0 (Compiled frame; information > may be imprecise) > - java.util.concurrent.locks.LockSupport.parkNanos(java.lang.Object, long) > @bci=20, line=226 (Compiled frame) > - > java.util.concurrent.SynchronousQueue$TransferStack.awaitFulfill(java.util.concurrent.SynchronousQueue$TransferStack$SNode, > boolean, long) @bci=174, line=460 (Compiled frame) > - > java.util.concurrent.SynchronousQueue$TransferStack.transfer(java.lang.Object, > boolean, long) @bci=102, line=359 (Interpreted frame) > - java.util.concurrent.SynchronousQueue.poll(long, > java.util.concurrent.TimeUnit) @bci=11, line=942 (Interpreted frame) > - java.util.concurrent.ThreadPoolExecutor.getTask() @bci=141, line=1068 > (Interpreted frame) > - > java.util.concurrent.ThreadPoolExecutor.runWorker(java.util.concurrent.ThreadPoolExecutor$Worker) > @bci=26, line=1130 (Interpreted frame) > - java.util.concurrent.ThreadPoolExecutor$Worker.run() @bci=5, line=615 > (Interpreted frame) > - java.lang.Thread.run() @bci=11, line=745 (Interpreted frame) > > > Thread 1986: (state = BLOCKED) > - java.lang.Thread.sleep(long) @bci=0 (Interpreted frame) > - org.apache.hadoop.hdfs.PeerCache.run() @bci=41, line=250 (Interpreted > frame) > - > org.apache.hadoop.hdfs.PeerCache.access$000(org.apache.hadoop.hdfs.PeerCache) > @bci=1, line=41 (Interpreted frame) > - org.apache.hadoop.hdfs.PeerCache$1.run() @bci=4, line=119 (Interpreted > frame) > - java.lang.Thread.run() @bci=11, line=745 (Interpreted frame) > > > Thread 1970: (state = BLOCKED) > > > Thread 1969: (state = BLOCKED) > - java.lang.Object.wait(long) @bci=0 (Interpreted frame) > - java.lang.ref.ReferenceQueue.remove(long) @bci=44, line=135 (Interpreted > frame) > - java.lang.ref.ReferenceQueue.remove() @bci=2, line=151 (Interpreted frame) > - java.lang.ref.Finalizer$FinalizerThread.run() @bci=36, line=209 > (Interpreted frame) > > > Thread 1968: (state = BLOCKED) > - java.lang.Object.wait(long) @bci=0 (Interpreted frame) > - java.lang.Object.wait() @bci=2, line=503 (Interpreted frame) > - java.lang.ref.Reference$ReferenceHandler.run() @bci=46, line=133 > (Interpreted frame) > > > Thread 1958: (state = IN_VM) > - java.lang.Throwable.fillInStackTrace(int) @bci=0 (Compiled frame; > information may be imprecise) > - java.lang.Throwable.fillInStackTrace() @bci=16, line=783 (Compiled frame) > - java.lang.Throwable.<init>(java.lang.String, java.lang.Throwable) @bci=24, > line=287 (Compiled frame) > - java.lang.Exception.<init>(java.lang.String, java.lang.Throwable) @bci=3, > line=84 (Compiled frame) > - org.json4s.package$MappingException.<init>(java.lang.String, > java.lang.Exception) @bci=13, line=56 (Compiled frame) > - org.json4s.reflect.package$.fail(java.lang.String, java.lang.Exception) > @bci=6, line=96 (Compiled frame) > - org.json4s.Extraction$.convert(org.json4s.JsonAST$JValue, > org.json4s.reflect.ScalaType, org.json4s.Formats, scala.Option) @bci=2447, > line=554 (Compiled frame) > - org.json4s.Extraction$.extract(org.json4s.JsonAST$JValue, > org.json4s.reflect.ScalaType, org.json4s.Formats) @bci=796, line=331 > (Compiled frame) > - org.json4s.Extraction$.extract(org.json4s.JsonAST$JValue, > org.json4s.Formats, scala.reflect.Manifest) @bci=10, line=42 (Compiled frame) > - org.json4s.Extraction$.extractOpt(org.json4s.JsonAST$JValue, > org.json4s.Formats, scala.reflect.Manifest) @bci=7, line=54 (Compiled frame) > - org.json4s.ExtractableJsonAstNode.extractOpt(org.json4s.Formats, > scala.reflect.Manifest) @bci=9, line=40 (Compiled frame) > - > org.apache.spark.util.JsonProtocol$.shuffleWriteMetricsFromJson(org.json4s.JsonAST$JValue) > @bci=116, line=702 (Compiled frame) > - > org.apache.spark.util.JsonProtocol$$anonfun$taskMetricsFromJson$2.apply(org.json4s.JsonAST$JValue) > @bci=4, line=670 (Compiled frame) > - > org.apache.spark.util.JsonProtocol$$anonfun$taskMetricsFromJson$2.apply(java.lang.Object) > @bci=5, line=670 (Compiled frame) > - scala.Option.map(scala.Function1) @bci=22, line=145 (Compiled frame) > - > org.apache.spark.util.JsonProtocol$.taskMetricsFromJson(org.json4s.JsonAST$JValue) > @bci=414, line=670 (Compiled frame) > - > org.apache.spark.util.JsonProtocol$.taskEndFromJson(org.json4s.JsonAST$JValue) > @bci=174, line=508 (Compiled frame) > - > org.apache.spark.util.JsonProtocol$.sparkEventFromJson(org.json4s.JsonAST$JValue) > @bci=389, line=464 (Compiled frame) > - > org.apache.spark.scheduler.ReplayListenerBus$$anonfun$replay$1.apply(java.lang.String) > @bci=34, line=51 (Compiled frame) > - > org.apache.spark.scheduler.ReplayListenerBus$$anonfun$replay$1.apply(java.lang.Object) > @bci=5, line=49 (Compiled frame) > - scala.collection.Iterator$class.foreach(scala.collection.Iterator, > scala.Function1) @bci=16, line=743 (Compiled frame) > - scala.collection.AbstractIterator.foreach(scala.Function1) @bci=2, > line=1177 (Interpreted frame) > - org.apache.spark.scheduler.ReplayListenerBus.replay(java.io.InputStream, > java.lang.String) @bci=42, line=49 (Interpreted frame) > - > org.apache.spark.deploy.history.FsHistoryProvider.org$apache$spark$deploy$history$FsHistoryProvider$$replay(org.apache.hadoop.fs.FileStatus, > org.apache.spark.scheduler.ReplayListenerBus) @bci=69, line=260 (Interpreted > frame) > - > org.apache.spark.deploy.history.FsHistoryProvider$$anonfun$6.apply(org.apache.hadoop.fs.FileStatus) > @bci=19, line=190 (Interpreted frame) > - > org.apache.spark.deploy.history.FsHistoryProvider$$anonfun$6.apply(java.lang.Object) > @bci=5, line=188 (Interpreted frame) > - > scala.collection.TraversableLike$$anonfun$flatMap$1.apply(java.lang.Object) > @bci=9, line=252 (Interpreted frame) > - > scala.collection.TraversableLike$$anonfun$flatMap$1.apply(java.lang.Object) > @bci=2, line=252 (Interpreted frame) > - > scala.collection.IndexedSeqOptimized$class.foreach(scala.collection.IndexedSeqOptimized, > scala.Function1) @bci=22, line=33 (Interpreted frame) > - scala.collection.mutable.WrappedArray.foreach(scala.Function1) @bci=2, > line=35 (Interpreted frame) > - > scala.collection.TraversableLike$class.flatMap(scala.collection.TraversableLike, > scala.Function1, scala.collection.generic.CanBuildFrom) @bci=17, line=252 > (Interpreted frame) > - scala.collection.AbstractTraversable.flatMap(scala.Function1, > scala.collection.generic.CanBuildFrom) @bci=3, line=104 (Interpreted frame) > - org.apache.spark.deploy.history.FsHistoryProvider.checkForLogs() @bci=110, > line=188 (Interpreted frame) > - org.apache.spark.deploy.history.FsHistoryProvider.initialize() @bci=38, > line=116 (Interpreted frame) > - > org.apache.spark.deploy.history.FsHistoryProvider.<init>(org.apache.spark.SparkConf) > @bci=214, line=99 (Interpreted frame) > - > sun.reflect.NativeConstructorAccessorImpl.newInstance0(java.lang.reflect.Constructor, > java.lang.Object[]) @bci=0 (Interpreted frame) > - sun.reflect.NativeConstructorAccessorImpl.newInstance(java.lang.Object[]) > @bci=72, line=57 (Interpreted frame) > - > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(java.lang.Object[]) > @bci=5, line=45 (Interpreted frame) > - java.lang.reflect.Constructor.newInstance(java.lang.Object[]) @bci=79, > line=526 (Interpreted frame) > - org.apache.spark.deploy.history.HistoryServer$.main(java.lang.String[]) > @bci=89, line=185 (Interpreted frame) > - org.apache.spark.deploy.history.HistoryServer.main(java.lang.String[]) > @bci=4 (Interpreted frame) > > > > On Thu, May 7, 2015 at 2:17 PM, Koert Kuipers <ko...@tresata.com> wrote: > good idea i will take a look. it does seem to be spinning one cpu at 100%... > > On Thu, May 7, 2015 at 2:03 PM, Marcelo Vanzin <van...@cloudera.com> wrote: > Can you get a jstack for the process? Maybe it's stuck somewhere. > > On Thu, May 7, 2015 at 11:00 AM, Koert Kuipers <ko...@tresata.com> wrote: > i am trying to launch the spark 1.3.1 history server on a secure cluster. > > i can see in the logs that it successfully logs into kerberos, and it is > replaying all the logs, but i never see the log message that indicate the web > server is started (i should see something like "Successfully started service > on port 18080." or "Started HistoryServer at http://somehost:18080"). yet the > daemon stays alive... > > any idea why the history server would never start the web service? > > thanks! > > > > > -- > Marcelo > > > > > > -- > Marcelo >
signature.asc
Description: Message signed with OpenPGP using GPGMail