The general motion equation is given by:
d(t)=1/2 gt^2+v_ot +d_o where:
d(t) is the distance traveled, in metres
g is the acceleration due to gravity in m/s2
t is the time, in seconds (s)
vo is the initial velocity of the moving object
do is the initial height in metres.
An object is dropped from a height of 1,000 m, on Mars, where the acceleration due to gravity is 3.92 m/s2. Find, to two decimal places, the time it will take for the object to hit the surface of Mars.
i get 1.96t^2, is it right?