@Glenn
Your question is legit. When I was working with Philippe to QA the
BackendListener feature I was really excited thinking about all the
possible ways we can solve the exact requirement you mentioned in your
question. This is a standard requirement which I am sure the performance
test community has wanted.

To accomplish your requirements all we need in Jmeter is a plugin to store
the test event details such as test start time, test end time, test name,
test comment and few other meta data attributes of the test. It would be
awesome if someone develops a listener which would store the test events
also in a database which then can be used to retrieve the entire dashboard
metrics pertaining to a test. Grafana already supports a feature where you
can through a REST API retrieve retrospective dashboard based on start/stop
time of an event.

Now the question is what solution do we use to display all the test events
in another performance test management tool kind of an application where
you can search, sort, compare test based on test event etc. and I think
there is already a open source tool called "CentralPerf" which has the
potential to do it.

Check out :
1.http://www.centralperf.org/
2.https://github.com/centralperf/centralperf/wiki

Charles Le Gallic who is one of the developers of this tool posted a
message about CentralPerf a couple of weeks back to the Jmeter mailing list
talking about the usefulness of the tool.
The tool's persistence layer is currently built using PostgreSQL and I
think the framework can be extended to store test data events which in turn
can be used to retrieve data from either influxdb directly
 or view the data through Grafana using REST API.


Thanks
Chaitanya M Bhatt
http://www.performancecompetence.com








On Fri, Apr 10, 2015 at 11:54 AM, Philippe Mouawad <
[email protected]> wrote:

> Hi,
> What about playing on rootMetricsPrefix to do that ?
>
> Regarding SQL, do you know that you can now easily build a jdbc backend to
> store results in a database, you could contribute this to core.
>
>
> Regards
>
> On Friday, April 10, 2015, Glenn Caccia <[email protected]>
> wrote:
>
> >  I've successfully installed InfluxDB and Grafana and did some basic
> > testing where I can now see results in Grafana.  I'm beginning to wonder
> > about the benefits of this system.  A while ago I had toyed around with
> the
> > idea of using Elasticsearch as a backend for JMeter test results and
> using
> > Kibana to view results.  I ultimately dropped the idea because of the
> > limitations of how data is structured.  I see the exact same issue with
> > InfluxDB and Grafana (either that, or I don't fully understand what can
> be
> > done in these tools).
> > What I want when viewing results is the ability to work with results in
> > terms of projects, test plans, and results from a particular test run.
> For
> > example, I want to see results for project A, test plan B and compare
> > results from the prior run with the current run.  With InfluxDB/Grafana
> > solution, there is no concept of a run.  If I run a test one day and then
> > run the same test the subsequent day, I can't compare the results using
> the
> > same view.  I can certainly change my time filter to see both inline
> (with
> > a big gap inbetween) or view one and then view the other, but I can't
> stack
> > them in separate graphs and see them at the same time or display them in
> > the same graph.  Likewise, if I want to see what performance was like the
> > last time a test was run and I don't know when the last test was run, I
> > have to do a bit of searching by playing with the time filter.
> > A while ago I worked for a company that used SQL Server for a lot of
> their
> > data storage needs.  This gave me access to the SQL Server Report Builder
> > tool.  I was able to create a solution where JMeter results were loaded
> > into SQL Server and we had a report interface where you could choose your
> > project, choose your test plan and then see the dates/times for all prior
> > runs.  From this, you could choose which run(s) to view.  I don't have
> > access to tools like that with my current company, but I miss that kind
> of
> > ability to structure and access test results.  A similar approach
> > to storing and presenting results can be seen with loadosphia.
> > In short, it seems like this new solution is primarily useful for
> > analyzing results from a current test run (which can already be done with
> > existing listeners) but is not as useful a tool for comparing results or
> > checking on results from prior runs.  Am I missing something or is that a
> > fair conclusion?
> >
>
>
> --
> Cordialement.
> Philippe Mouawad.
>

Reply via email to