* [email protected] ([email protected]) wrote: > Anyone can answer our questions? Mathieu?
sorry for the slow reply, I've been swamped in filtering implementation lately, > > From: [email protected] [mailto:[email protected]] > Sent: Wednesday, July 18, 2012 5:54 PM > To: [email protected] > Subject: [lttng-dev] some questions on lttng > > Hello the dev list, > We encounter some basic questions when try to adapt the LTTNG in our poject. > > 1. When the trace is enabled and all are well configurated, we get > trace messages collected under the session folder. The question is > whether it is possible that some traces will lost when the trace > messages are huge. How will LTTNG do if the consumer deamon cannot > fast enough to copy the trace message from trace buffer? There are currently two ways to configure the channels: discard and overwrite mode. In discard mode, upon buffer full condition, events are discarded, and we keep track of the number of events discarded in the packet headers, so the trace viewer can print warnings about discarded events within a specific time-frame. In overwrite mode, upon buffer full condition, the oldest subbuffer (packet) is overwritten. We will soon add a sequence counter to the packet header, so the trace viewer can show when a packet is missing in the stream (either due to being overwritten by the tracer or due to UDP packet loss in network streaming). If the message (event) is too large to fit within a packet, it is discarded, incrementing the event discarded counter accordingly (so the viewer can show this information from the packet header). It would be interesting to implement a "blocking" mode that makes the application block if buffer is full. This makes the tracer much more intrusive, and if something goes wrong in the session daemon or consumer daemon, the app hangs, but it might be interesting for logging purposes, if you care about _never_ losing an event. I would recommend to use this kind of feature in debugging setups, not in production, at the beginning, since it would make the sessiond/consumerd critical (if they die, the application hangs. I don't want to see this happen in production). > > 2. For user space trace, currectly seems we cannot set the flush > interval. How can we control the flush internal for UST? If no, is > there hardcoded or random flash interval? Or there is no time based > flush mechanism at all for UST? The flush interval will only be useful for live streaming, which is not supported yet. Currently, UST does a subbuffer switch (flush) each time the subbuffer cannot hold the next event to write. I plan to implement this periodic flush at the consumer daemon level, so it will be less intrusive within the application. Adding timers to a traced application without its knowledge would likely be a nightware waiting to happen. > > 3. For a severe kernel panic can we extrace buffered trace messages > from LTTNG internal buffer in full memory core dump file? Is there > tools on this we can leverage? Not yet, but it's a feature we look forward to see coming. At some point in the past, I remember there was a patch to lcrash that was supporting extraction of lttng 0.x buffers from a crashed kernel image. However, LTTng 2.0 data structure layout is quite different, so this work would have to be redone. I'm aware that Thomas Gleixner is working on "Shrinking core dump on the fly" (see the Tracing Summit schedule at http://tracingsummit.org/wiki/TracingSummit2012). Maybe he has a few words to say on this, and some recommendations. One related thing that would be interesting to implement is a libringbuffer backend that uses video card memory buffers that survive reboots. Given that lttng libringbuffer is very modular, it should be straightforward to implement. Thanks! Mathieu > > Thank you, > - Bingfeng > > _______________________________________________ > lttng-dev mailing list > [email protected] > http://lists.lttng.org/cgi-bin/mailman/listinfo/lttng-dev -- Mathieu Desnoyers Operating System Efficiency R&D Consultant EfficiOS Inc. http://www.efficios.com _______________________________________________ lttng-dev mailing list [email protected] http://lists.lttng.org/cgi-bin/mailman/listinfo/lttng-dev
