113k views
4 votes
HELPPPP.

Suppose a bird is 500 feet above the ground. It descends at a steady rate. After 10 seconds, it is 250 feet above the ground.
a) Write an equation that gives the height of the bird as a function of time. Be sure to define your variables!
b) After how many seconds will the bird land on the ground?​

User Adsurbum
by
5.3k points

2 Answers

2 votes

Answer:

a)
h(t)=500-25t

b)
t=20

Explanation:

a)

Let
h = height above ground

Let
t = time

Because the bird descended down to 250 feet in 10 seconds with a steady rate, that means it descends 25 feet per second.


h(t) = 500-25t

b)

To find the time it takes until the bird reaches the ground,
h must equal
0:


0=500-25t


25t=500


t=20

∴ It takes 20 seconds for the bird to reach the ground.

User Sdedelbrock
by
5.9k points
1 vote

Answer:

500-10x=250

x=how much the bird descended in a second

the bird will land on the ground in another 10 seconds.

Explanation:

the bird will land on the ground in 10 seconds since it descended 250 ft in 10 seconds so that is 25 ft per second. then do 250/25 and you get 10. that is how you know the bird will land on the ground in 10 seconds.

User Jason Renaldo
by
5.4k points