Respuesta :
Well, in order to figure out the answer is to divide until you figure out how many miles they went per second. If it takes 5 seconds to reach 50 miles per hour it took 10 seconds per every 10 miles meaning each mile took 1 second. (Not actually possible but the answer) So, If it finished a 100 mile trip in 2 hours it took an hour for 50 miles. If it took 1 hour for 50 miles divide 60/50 which gets you 1.2 so it took 1.2 miles per minute meaning the car went 120 miles per hour I believe. I hope this helps :)
Answer:
Average speed is 50 mph
Average velocity is -50 mph . It is negative since it is towards south.
Explanation:
Distance is a scalar quantity and velocity is a vector. So for a vector, both magnitude and direction should be specified.
Average speed is the total distance over total time taken . So it is 100 miles over 2 hours which gives 50 mph.
Average velocity has the same magnitude and is also 100miles over 2 hours but it is negative since velocity is a vector and it is directed towards south.