Final answer:
To determine the time it will take Larry's wife to reach home, we must calculate the distance Larry's wife needs to cover divided by her speed. It results in 2.5 hours.
Step-by-step explanation:
The student wants to find out how long it will take for Larry's wife to reach their home if Larry, who jogs home at 10 miles per hour, is half an hour late and his wife passes him 20 miles before their home.
Larry's wife drives at the same speed as Larry jogs since they usually arrive home at the same time. If Larry jogs at 10 miles per hour, and he is half an hour late, this means he would've covered 5 miles less than usual (because 10 miles/hour * 0.5 hour = 5 miles). Larry's wife passed him 20 miles before their home, which is, in essence, 25 miles she needs to cover because Larry is also 5 miles behind his usual spot. We simply divide the distance she needs to cover by her driving speed to get the time it will take her to reach home.
We have:
Distance = 25 miles,
Speed = 10 miles/hour,
Time = Distance / Speed = 25 miles / 10 miles/hour = 2.5 hours.
The correct answer is (c) 2.5 hours.