Best Answer

2hours 57 minutes = 177 minutes

177 minutes/26 miles= 6.80 minutes /miles

0.80 minutes x 60 seconds= 48 seconds

Answer= 6:48 /mile

Q: If you run 26 miles in 2 hours and 57 minutes about many minutes average each mile?

Write your answer...

Submit

Still have questions?

Continue Learning about Math & Arithmetic

26 miles in three hours is an average of about 6 minutes 55 seconds per mile.

If you have to drive 100 miles, and you have 2 hours to do it in, I think 50 miles in each hour ought to get you there.

Simply add your times and divide by the distance (or visa-versa).21 minutes + 63 minutes = 84 minutes84 minutes/10 miles = 8.4 minute miles.Note, this is a convenient form to calculate speed for runners. However, the "standard" would be in MPH (or the metric equivalent).(10 miles / 84 minutes)*(60 minutes/hour) = 7.14 MPHA common fallacy might be to calculate the speed in each leg of the race and average.3 miles in 21 minutes = 7 minute miles.7 miles in 63 minutes = 9 minute miles.If you average these two you get 8 minute miles WHICH IS INCORRECT.The problem with this is that the runner ran each pace for different distances. If the distances at each pace were the same, then this would work.

120 times 365.25 (days in an average year, approx) times 24 hours in each day times 60 minutes in each hour.

If I can travel 450 miles in 2 hours, I can travel 450 ÷ 2 = 225 miles in 1 hour. As I can travel 225 miles in each hour, I can travel 1200 miles in: 1200 ÷ 225 = 51/3 hours = 5 hours 20 minutes.

Related questions

26 miles in three hours is an average of about 6 minutes 55 seconds per mile.

7.5 456 miles divided by 75 miles is 6.08 hours. Then, 21 minutes times 4 rest stops equals 84 minutes. 84 min. divided by 60 min in an hour is 1.4 hours. So, 6.08 plus 1.4 is 7.48 hours, which rounds up to 7.5 hours.

7 hours, 28.8 minutes

It depends on how fast you drive. However, as an example, if you maintained an average speed of 65 mph, the time to cover 50 miles would be about 46 minutes. You would have to add time for stops, detours and other delays.

75 minutes is 1.25 hours. going 14.6 miles each hour means you multiply 1.25 by 14.5 and you get 18.125

25mph 100 miles / 4 hours = the amount of miles per each hour

If you have to drive 100 miles, and you have 2 hours to do it in, I think 50 miles in each hour ought to get you there.

60 miles per hour? Well since there are 60 minutes in each hour, then you go 1 mile for each minute. If you need to go 220 miles of distance, you will need 220 minutes of time. How many hours are there in 220 minutes? Divide 220 by the number of minutes in each hour. Then you will have the number of whole hours, but instead of having "minutes" left over, you'll have a decimal that is a percentage of an hour. You'll have to figure that part out yourself.

Simply add your times and divide by the distance (or visa-versa).21 minutes + 63 minutes = 84 minutes84 minutes/10 miles = 8.4 minute miles.Note, this is a convenient form to calculate speed for runners. However, the "standard" would be in MPH (or the metric equivalent).(10 miles / 84 minutes)*(60 minutes/hour) = 7.14 MPHA common fallacy might be to calculate the speed in each leg of the race and average.3 miles in 21 minutes = 7 minute miles.7 miles in 63 minutes = 9 minute miles.If you average these two you get 8 minute miles WHICH IS INCORRECT.The problem with this is that the runner ran each pace for different distances. If the distances at each pace were the same, then this would work.

55 miles each hour for 6 hours. 55 x 6 = 330 miles.

120 times 365.25 (days in an average year, approx) times 24 hours in each day times 60 minutes in each hour.

The average speed over the first 30 miles is equal to the average speed over the next 15 miles. You cannot guarantee that the speed was constant inbetween, though. The average speeds are 45 MPH for each leg.