Question
> When making changes to optimize part of a processor, it is often the case that speeding up one type of instruction comes at the
> When making changes to optimize part of a processor, it is often the case that speeding up one type of instruction comes at the cost of slowing down something else. For example, if we put in a complicated fast floating-point unit, that takes space, and something might have to be moved farther away from the middle to accommodate it, adding an extra cycle in delay to reach that unit. The basic Amdahls law equation does not take into account this trade-off.
a. If the new fast floating-point unit speeds up floating-point operations by, on average, 3, and floating-point operations take 20% of the original programs execution time, what is the overall speedup (ignoring the penalty to any other instructions)?
b. Now assume that speeding up the floating-point unit slowed down data cache accesses, resulting in a 1.5 slowdown (or 2/3 speedup). Data cache accesses consume 20% of the execution time. What is the overall speedup now?
c. After implementing the new floating-point operations, what percentage of execution time is spent on floating-point operations? What percentage is spent on data cache accesses?
Step by Step Solution
There are 3 Steps involved in it
Step: 1
Get Instant Access to Expert-Tailored Solutions
See step-by-step solutions with expert insights and AI powered tools for academic success
Step: 2
Step: 3
Ace Your Homework with AI
Get the answers you need in no time with our AI-driven, step-by-step assistance
Get Started