Originally Posted by

**learningguy** OK, here's the problem: A ball (first ball) is dropped from rest from a height 20.0m above the ground. Another ball is thrown (second ball) vertically upward from the ground at the instant the first ball is released. Determine the intial speed of the second ball if the two balls are to meet at a height 10.0m above the ground.

The equations I am using are time = square root(2y/-g), and initial velocity = x/t.

For time I get square root((2*(-10m))/9.8m/s^2) = 1.43 seconds.

For initial velocity of the second ball I get 10m/1.43s = 6.99m/s.

However, at the end of the problem it says that the answer should be 14m/s but I can't figure out how to get that so am I doing something wrong, or is the answer I am given wrong?