Answer:
Stands for "Million Instructions Per Second." It is a method of measuring the raw speed of a computer's processor. ... The MIPS measurement has been used by computer manufacturers like IBM to measure the "cost of computing." The value of computers is determined in MIPS per dollar.