If an arrow is fired from a bow with a perfectly horizontal velocity of 60 m/s and the arrow

was 2 m above the ground when it was released, how far will the arrow fly horizontally before it strikes the ground?

Respuesta :

Explanation:

This problem is solved by determining how long it takes the arrow to fall to the ground in exactly the same manner as if the arrow was dropped with no horizontal velocity. The time required for the arrow to fall to the ground will be the same time that the arrow flies horizontally at 60.0 m/s, so

d=1/2at^2

solved for t=√2d/a

t=√(2)(2.00m)/9.80m/s^2 =0.639s

The time required for the fall is multiplied by the horizontal velocity to get the horizontal distance.

d(horizontal)=v(horizontal)×time=60.0m×0.639s =38.3m

ACCESS MORE
EDU ACCESS