%z Article
%K Banga97
%A Guarav Banga
%A Peter Druschel
%T Measuring the capacity of a web server
%B Proceedings USENIX Symposium on Internet Technologies and Systems
%C Monterey, CA
%D December 1997

%z Article
%K Banga98
%A Guarav Banga
%A Jeffrey C. Mogul
%T Scalable kernel performance for internet servers under realistic loads
%B Proceedings of the 1998 USENIX Annual Technical Conference
%C New Orleans, LA
%D June 1998

%K Bray90
%A Tim Bray
%T Bonnie benchmark
%D 1990
%O http://www.textuality.com/bonnie/

%z Article
%K Brown97
%A Aaron Brown
%A Margo Seltzer
%T Operating system benchmarking in the wake of lmbench: A case study of the performance of NetBSD on the Intel x86 architecture
%B Proceedings of the 1997 ACM SIGMETRICS Conference on Measurement and Modeling of Computer Systems
%C Seattle, WA
%D June 1997
%P 214-224
%O http://www.eecs.harvard.edu/~vino/perf/hbench/sigmetrics/hbench.html

%z Article
%K Chen93d
%A Peter M. Chen
%A David Patterson
%T Storage performance \- metrics and benchmarks
%J Proceedings of the IEEE
%V 81
%N 8
%D August 1993
%P 1151-1165
%x Discusses metrics and benchmarks used in storage performance evaluation.
%x Describes, reviews, and runs popular I/O benchmarks on three systems.  Also
%x describes two new approaches to storage benchmarks: LADDIS and a Self-Scaling
%x benchmark with predicted performance.
%k I/O, storage, benchmark, workload, self-scaling benchmark, 
%k predicted performance, disk, performance evaluation
%s staelin%cello@hpl.hp.com (Wed Sep 27 16:21:11 PDT 1995)

%z Article
%K Chen94a
%A P. M. Chen
%A D. A. Patterson
%T A new approach to I/O performance evaluation \- self-scaling I/O benchmarks, predicted I/O performance
%D November 1994
%J Transactions on Computer Systems
%V 12
%N 4
%P 308-339
%x Current I/O benchmarks suffer from several chronic problems: they
%x quickly become obsolete; they do not stress the I/O system; and they
%x do not help much in undelsi;anding I/O system performance. We
%x propose a new approach to I/O performance analysis. First, we
%x propose a self-scaling benchmark that dynamically adjusts aspects of
%x its workload according to the performance characteristic of the
%x system being measured. By doing so, the benchmark automatically
%x scales across current and future systems. The evaluation aids in
%x understanding system performance by reporting how performance varies
%x according to each of five workload parameters. Second, we propose
%x predicted performance, a technique for using the results from the
%x self-scaling evaluation to estimate quickly the performance for
%x workloads that have not been measured. We show that this technique
%x yields reasonably accurate performance estimates and argue that this
%x method gives a far more accurate comparative performance evaluation
%x than traditional single-point benchmarks. We apply our new
%x evaluation technique by measuring a SPARCstation 1+ with one SCSI
%x disk, an HP 730 with one SCSI-II disk, a DECstation 5000/200 running
%x the Sprite LFS operating system with a three-disk disk array, a
%x Convex C240 minisupercomputer with a four-disk disk array, and a
%x Solbourne 5E/905 fileserver with a two-disk disk array.
%s toc@hpl.hp.com (Mon Mar 13 10:57:38 1995)
%s wilkes%hplajw@hpl.hp.com (Sun Mar 19 12:38:01 PST 1995)
%s wilkes%cello@hpl.hp.com (Sun Mar 19 12:38:53 PST 1995)

%z Article
%K Fenwick95
%A David M. Fenwick
%A Denis J. Foley
%A William B. Gist
%A Stephen R. VanDoren
%A Danial Wissell
%T The AlphaServer 8000 series: high-end server platform development
%J Digital Technical Journal
%V 7
%N 1
%D August 1995
%P 43-65
%x The AlphaServer 8400 and the AlphaServer 8200 are Digital's newest high-end
%x server products.  Both servers are based on the 300Mhz Alpha 21164 
%x microprocessor and on the AlphaServer 8000-series platform architecture.
%x The AlphaServer 8000 platform development team set aggressive system data
%x bandwidth and memory read latency targets in order to achieve high-performance
%x goals.  The low-latency criterion was factored into design decisions made at
%x each of the seven layers of platform development.  The combination of 
%x industry-leading microprocessor technology and a system platform focused
%x on low latency has resulted in a 12-processor server implementation ---
%x the AlphaServer 8400 --- capable of supercomputer levels of performance.
%k DEC Alpha server, performance, memory latency
%s staelin%cello@hpl.hp.com (Wed Sep 27 17:27:23 PDT 1995)

%z Book
%K Hennessy96
%A John L. Hennessy
%A David A. Patterson
%T Computer Architecture A Quantitative Approach, 2nd Edition
%I Morgan Kaufman
%D 1996

%z Article
%K Howard88
%A J. Howard
%A M. Kazar
%A S. Menees
%A S. Nichols
%A M. Satyanrayanan
%A R. Sidebotham
%A M. West
%T Scale and performance in a distributed system
%J ACM Transactions on Computer Systems
%V 6
%N 1
%D February 1988
%P 51-81
%k Andrew benchmark

%z Book
%K Jain91
%A Raj Jain
%T The Art of Computer Systems Performance Analysis: Techniques for Experimental Design, Measurement, Simulation, and Modeling
%I Wiley-Interscience
%C New York, NY
%D April 1991

%z Article
%K McCalpin95
%A John D. McCalpin
%T Memory bandwidth and machine balance in current high performance computers
%J IEEE Technical Committee on Computer Architecture newsletter
%D December 1995

%z InProceedings
%K McVoy91
%A L. W. McVoy
%A S. R. Kleiman
%T Extent-like Performance from a Unix File System
%B Proceedings USENIX Winter Conference
%C Dallas, TX
%D January 1991
%P 33-43

%z Article
%K McVoy96
%A Larry McVoy
%A Carl Staelin
%T lmbench: Portable tools for performance analysis
%B Proceedings USENIX Winter Conference
%C San Diego, CA
%D January 1996
%P 279-284

%z InProceedings
%K Ousterhout90
%s wilkes%cello@hplabs.hp.com (Fri Jun 29 20:46:08 PDT 1990)
%A John K. Ousterhout
%T Why aren't operating systems getting faster as fast as hardware?
%B Proceedings USENIX Summer Conference
%C Anaheim, CA
%D June 1990
%P 247-256
%x This paper evaluates several hardware pplatforms and operating systems using
%x a set of benchmarks that stress kernel entry/exit, file systems, and
%x other things related to operating systems. The overall conclusion is that
%x operating system performance is not improving at the same rate as the base speed of the
%x underlying hardware. The most obvious ways to remedy this situation
%x are to improve memory bandwidth and reduce operating systems'
%x tendency to wait for disk operations to complete.
%o Typical performance of 10-20 MIPS cpus is only 0.4 times what
%o their raw hardware performance would suggest. HP-UX is
%o particularly bad on the HP 9000/835, at about 0.2x. (Although
%o this measurement discounted a highly-tuned getpid call.)
%k OS performance, RISC machines, HP9000 Series 835 system calls

%z Article
%K Park90a
%A Arvin Park
%A J. C. Becker
%T IOStone: a synthetic file system benchmark
%J Computer Architecture News
%V 18
%N 2
%D June 1990
%P 45-52
%o this benchmark is useless for all modern systems; it fits
%o completely inside the file system buffer cache.  Soon it may even
%o fit inside the processor cache!
%k IOStone, I/O, benchmarks
%s staelin%cello@hpl.hp.com (Wed Sep 27 16:37:26 PDT 1995)

%z Thesis
%K Prestor01
%A Uros Prestor
%T Evaluating the memory performance of a ccNUMA system
%I Department of Computer Science, University of Utah
%D May 2001

%z Thesis
%K Saavedra92
%A Rafael H. Saavedra-Barrera
%T CPU Performance evaluation and execution time prediction using narrow spectrum benchmarking
%I Department of Computer Science, University of California at Berkeley
%D 1992

%z Article
%K Saavedra95
%A R.H. Saavedra
%A A.J. Smith
%T Measuring cache and TLB performance and their effect on benchmark runtimes
%J IEEE Transactions on Computers
%V 44
%N 10
%D October 1995
%P 1223-1235

%z Article
%k Seltzer99
%A Margo Seltzer
%A David Krinsky
%A Keith Smith
%A Xiolan Zhang
%T The case for application-specific benchmarking
%B Proceedings of the 1999 Workshop on Hot Topics in Operating Systems
%C Rico, AZ
%D 1999

%z Article
%K Shein89
%A Barry Shein
%A Mike Callahan
%A Paul Woodbury
%T NFSSTONE: A network file server performance benchmark
%B Proceedings USENIX Summer Conference
%C Baltimore, MD
%D June 1989
%P 269-275

%z Article
%K Staelin98
%A Carl Staelin
%A Larry McVoy
%T mhz: Anatomy of a microbenchmark
%B Proceedings USENIX Annual Technical Conference
%C New Orleans, LA
%D June 1998
%P 155-166

%z Article
%K FSF89
%A Richard Stallman
%Q Free Software Foundation
%T General Public License
%D 1989
%O Included with \*[lmbench]

%z Book
%K Toshiba94
%A Toshiba
%T DRAM Components and Modules
%I Toshiba America Electronic Components, Inc.
%P A59-A77,C37-C42
%D 1994

%z Article
%K Tullsen96
%A Dean Tullsen
%A Susan Eggers
%A Joel Emer
%A Henry Levy
%A Jack Lo
%A Rebecca Stamm
%T Exploiting choice: Instruction fetch and issue on an implementable simultaneous multithreading processor
%C Proceedings of the 23rd Annual International Symposium on Computer Architecture
%D May 1996
%P 191-202
%O http://www.cs.washington.edu/research/smt/papers/ISCA96.ps

%z Article
%K Tullsen99
%A Dean Tullsen
%A Jack Lo
%A Susan Eggers
%A Henry Levy
%T Supporting fine-grain synchronization on a simultaneous multithreaded processor
%B Proceedings of the 5th International Symposium on High Performance Computer Architecture
%D January 1999
%P 54-58
%O http://www.cs.washington.edu/research/smt/papers/hpca.ps

%z Article
%K Weicker84
%A R.P. Weicker
%T Dhrystone: A synthetic systems programming benchmark
%J CACM
%V 27
%N 10
%P 1013-1030
%D 1984

%z Article
%K Wolman89
%A Barry L. Wolman
%A Thomas M. Olson
%T IOBENCH: a system independent IO benchmark
%J Computer Architecture News
%V 17
%N 5
%D September 1989
%P 55-70
%x IOBENCH is an operating system and processor independent synthetic
%x input/output (IO) benchmark designed to put a configurable IO and
%x processor (CP) load on the system under test.  This paper discusses
%x the UNIX versions.
%k IOBENCH, synthetic I/O benchmark, UNIX workload
%s vinton%cello@hplabs.hp.com (Fri Sep 20 12:55:58 PDT 1991)

