usenix.ol 3.2 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102
  1. Introduction
  2. What is it?
  3. A bunch of speed of light benchmarks,
  4. not MP, not throughput, not saturation, not stress tests.
  5. A microbenchmark suite
  6. Measures system performance
  7. Latency and bandwidth measurements
  8. Measurements focus on OS and hardware
  9. What is delivered to the application
  10. Not marketing numbers
  11. Benchmark performance predicts application performance
  12. Results for which systems?
  13. Sun, SGI, DEC, IBM, HP, PCs
  14. Useful information to whom?
  15. Performance engineers, system programmers, system architects.
  16. Motivation
  17. What are we measuring?
  18. Control / latecy operatins
  19. Bandwidth operations
  20. What aren't we measuring?
  21. Basic MIPS & MFLOPS. XXX - not unless I do it right.
  22. What can I learn?
  23. Cost of operations
  24. ****Operations per time unit****
  25. Compare speed of alternative paths (e.g. mmap vs. read)
  26. Performance problems = f(bw issues + latency issues)
  27. Give at least two examples
  28. NFS control & data: UDP lat, proc lat, & various BW metrics
  29. Oracle lock manager: TCP lat
  30. Verilog: mem lat
  31. AIM: fs ops XXX -ask Scott about pipes.
  32. Knowing the speeds of primitives can provide speeds of apps.
  33. An example here would be nice.
  34. Outline
  35. Describe benchmark
  36. Give results from current machines
  37. Discuss results
  38. Future changes, enhancements, etc.
  39. Tutorial on benchmarks
  40. For each metric
  41. what is it?
  42. why is it being measured?
  43. How is it measured?
  44. Measuring subtlities
  45. Interpreting the results
  46. Latency
  47. Process stuff
  48. networking stuff
  49. file system stuff
  50. memory stuff
  51. whatever
  52. Bandwidth
  53. networking
  54. file system
  55. memory
  56. Results
  57. Tabular results - XXX update that table to reflect the newer metrics
  58. Graphs of memory latency & context switches
  59. Discussion
  60. Memory stuff
  61. Maybe contrast AIX with the $100K IBM
  62. uniprocessor w/ killer memory perf and point out
  63. that it is the memory that is making AIX go
  64. fast, it certainly isn't AIX. A more politic
  65. observation would be that systems with good
  66. memory performace tend to have good system
  67. performance; the point being to shift people's
  68. attention to system performance, especially
  69. memory subsystem, as opposed to processor mips.
  70. Comparisons
  71. Maybe look at the table and draw attention to
  72. really good and really bad numbers for various
  73. platforms (like Linux' context switch time,
  74. Linux fs ops, solaris syscall, process stuff,
  75. 990 memory BW).
  76. Graphs
  77. A graph showing a range of really fast to really slow ops, all on the
  78. same graph. Do bandwidth stuff normalized on MB/sec.
  79. Carl sez: show both ops/sec and cost/op on two graphs.
  80. A graph showing processor slow down due to memory misses, assuming
  81. each instruction misses. Maybe a graph that shows # of clocks
  82. (or better yet, # of instructions - think super scalar) that you would
  83. have to have between each memory miss in order to run at the clock
  84. speed.
  85. War stories
  86. Sun page coloring bug
  87. SGI page coloring bug
  88. SGI hippi bug - XXX ask Thomas
  89. Sun bcopy bug
  90. Lmbench [optional?]
  91. how to get lmbench
  92. how to compile
  93. how to run
  94. how to show results
  95. Future work
  96. More hardware stuff - better latency measurements (write lat,
  97. cache to cache latency).
  98. add throughput & saturation measurements
  99. TODO
  100. get some similar papers for comparison
  101. Someday I need reasonable I/O benchmarks to show off good
  102. big SMP machines like Challenge.