Originally Posted by

**william** The general motion equation is given by:

d(t)=1/2 gt^2+v_ot +d_o where:

d(t) is the distance traveled, in metres

g is the acceleration due to gravity in m/s2

t is the time, in seconds (s)

vo is the initial velocity of the moving object

do is the initial height in metres.

An object is dropped from a height of 1,000 m, on Mars, where the acceleration due to gravity is 3.92 m/s2. Find, to two decimal places, the time it will take for the object to hit the surface of Mars.

i get 1.96t^2, is it right?