Skip to content

07. Measuring Performance

$\gdef \N{\mathbb{N}} \gdef \Z{\mathbb{Z}} \gdef \Q{\mathbb{Q}} \gdef \R{\mathbb{R}} \gdef \C{\mathbb{C}} \gdef \setcomp#1{\overline{#1}} \gdef \sseq{\subseteq} \gdef \pset#1{\mathcal{P}(#1)} \gdef \covariant{\operatorname{Cov}} \gdef \of{\circ} \gdef \p{^{\prime}} \gdef \pp{^{\prime\prime}} \gdef \ppp{^{\prime\prime\prime}} \gdef \pn#1{^{\prime\times{#1}}} $

Important Terms

  • CPU: Central Processing Unit — Contains the ALU
    • I really hope this isn’t news to you
  • Clock Rate (CR)— The rate at which the CPU internal clock runs. It is measured in cycles per second (Hz).
  • Clock Cycle (CC) — The length of a single cycle. This is the inverse of the clock rate.
    • \(CC=\frac1{\text{Clock Rate}}\)
    • CC is usually measured in microseconds (\(1\times10^{-6}\) seconds), but sometimes milliseconds (\(\times10^{-3}\)) or nanoseconds (\(\times10^{-9}\))
  • Instruction Count — The number of instructions in a given program
  • Cycles Per Instruction — The number of clock cycles required to perform the entire program divided by the instruction count.
    • Basically, average cycle per instruction
  • CPIi — The number of cycles required to perform an I-Type instruction.
    • Cycles Per Instructionof type I
  • CPU Time (Execution Time) — The number of seconds it takes for the processor to execute a given program
    • CPU Time = Clock Cycle \(\times\) Cycles per Instruction \(\times\) Instruction Count
  • Ii — The number of I-Type instructions in the program
  • CPI — Not sure about this one, I’ll get back to it. it’s probably super important though.
  • MIPs — Million Instructions per Second
    • Measures roughly how many instructions can be executed in one second
    • \(\frac{\text{Instruction Count}}{\text{CPU TIME in seconds}\times10^6}\)
  • MFLOPs — Million Floating Point OPerations
    • \(\frac{\text{Number of FP Ops in program}}{\text{CPU TIME in seconds}\times10^6}\)

Parallelization

In order to calculate the speedup — how much faster the program can be — when parallelized, we employ this formula:

\[\text{Speedup}=\frac{1}{B+\frac{1-B}{N}}\]

Where \(B\) is the percentage of the program that can be parallelized, and \(N\) is the number of concurrent parallelizations.

General Speedup (Amdahl’s Law)

To calculate the new execution time of a program after modifying the speeds of some of the commands, we apply this scary lookin’ formula:

\[\text{Time}_\text{new}=\text{Time}_\text{old}\times\left[(1-F)+\frac {F}{\text{Speedup}}\right]\]

Where \(F\) is the Fraction of the program that can be made Faster, and Speedup is exactly how much faster that part is made to be.

Overall speedup factor is \(\frac{1}{\text{Time}_\text{new}}\) (Normalized old execution time to 1)