comp.lang.ada
 help / color / mirror / Atom feed
* performance benchmarking
@ 1989-08-09  0:27 roseman
  1989-08-10  8:22 ` Paul Warren
  0 siblings, 1 reply; 4+ messages in thread
From: roseman @ 1989-08-09  0:27 UTC (permalink / raw)


I'm involved in doing a bit of performance benchmarking work on a couple
of Ada compilers.  I'm wondering if there is anyone else out there doing
similar kind of work.  I've also got a few questions.

Right now we're using the PIWG (Performance Issues Working Group) test
suite to do the tests.  This seems to be "the" standard Ada test suite
out there.  I'm wondering first off if people are using other tests,
and if so, what?  (Furthermore, where did they come from, why are you
using them, etc?)

Second, the machine we're running PIWG on is a Unix based system.  With
the PIWG, we're running into some problems.  The tests themselves are
very very short - total times including iteration is well under a second
for most of them!

The problem with that is its almost impossible to get any accurate
measurements that way - you've got all the little Unix daemons popping in
and out and using up some time.  We have tests which vary from 0 usecs
to almost 4 (per iteration that is), which is most unacceptable!

What can you do to correct things?  Run tests 25 (e.g.) times and take
the best?  The average?  Increase the iteration count to some ridiculous
amount to try to compensate?

I guess this is getting into general benchmarking procedures (any digests
or lists devoted to this out there?).. but how are tests like this supposed
to be used?  Surely, this must be an old problem.  You have various companies
out there who are publishing PIWG numbers for their compilers, but what
are they measuring?  Is it reasonable to measure on a souped-up system
(e.g. high priority, kill the daeomons), or do people want to see results
on a real Unix system?

If anyone has any answers, comments, pointers to any papers covering these
issues, etc. I would very much like to hear from you.  I ask only that if
you post to the list that you also send a copy to my userid directly, as
my time is so tight these days I can't keep up with the digest.  Thanks.

Mark Roseman, University of Manitoba
<ROSEMAN@ccu.UManitoba.CA> or <ROSEMAN@UOFMCC.BITNET>

^ permalink raw reply	[flat|nested] 4+ messages in thread
* Performance Benchmarking
@ 1989-08-30  7:01 "Lennart M}nsson"
  0 siblings, 0 replies; 4+ messages in thread
From: "Lennart M}nsson" @ 1989-08-30  7:01 UTC (permalink / raw)


Mark Roseman, University of Manitoba, <ROSEMAN@ccu.UManitoba.CA>
 or <ROSEMAN@UOFMCC.BITNET>, asks for methodological advice when 
doing performance benchmarking. 

I did some heavy testing a couple of years ago. It was on a VAX and 
not on UNIX but the problem with disturbing processes seems to be 
analog. 

My conclusion was that the only comparable measure is achieved when 
trying to eliminate the measurment noise as far as possible. Thus I 
always runned the tests on night time when no other people was supposed
to use the system. I ALSO CONTROLLED THAT IT WAS IN FACT SO. Secondly
I turned of all possibly interfering process, which seems to correspond 
you turning of all deamons running around. (If they are needed for your 
Ada program to run, well then they should be part of the mesaure)

Thirdly I ran each test three times each night checking that I got 
stable measurment. ( I made up a script calculating means and issuing 
warnings for to big deviations, it also made printouts of the reports, 
be sure not to let them interfere with the next test, printing is often 
done in parallell with other program execution) Above this I repetead 
the tests another night to check long time stability.  

Now, if you only see to that resorces are as memory usage etc is returned to 
system for each test, you can fairly well rely that you mesure the capacity
of your Ada code, under no disturbancies. What you really is interested in
a how it perform on a normally loaded system. But then you can't compare
figures between different implementations if you can't set up a standard
load situation or do so many iterations that you can count the mean 
disturbancy as a standard situation.

Testing is not easily done !

Lennart Mansson
Telelogic AB, Box 4148, S-203 12 MALM\, SWEDEN 
Phone:  +46-40-25 46 36
Fax:    +46-40-25 46 25
E-mail: lm@logm.se

------

^ permalink raw reply	[flat|nested] 4+ messages in thread

end of thread, other threads:[~1989-08-30  7:01 UTC | newest]

Thread overview: 4+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
1989-08-09  0:27 performance benchmarking roseman
1989-08-10  8:22 ` Paul Warren
1989-08-12 14:25   ` roseman
  -- strict thread matches above, loose matches on Subject: below --
1989-08-30  7:01 Performance Benchmarking "Lennart M}nsson"

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox