Before reading the below content, it is important to read the final note at the bottom first.
Take this excerpt from Wikipedia regarding the Doppler effect: if the siren approached the observer directly, the pitch would remain constant until the vehicle hit him, and then immediately jump to a new lower pitch. Is that really true? I believe, from the sound of a car honking its horn at me, the pitch gradually increases as the horn approaches, the frequency at rest of the horn will match my perception of it once it reaches me, and then gradually decrease in pitch as it moves away. This is owing to the idea that as the transmitter approaches, the distance between it and me decreases, and thus the sound waves, or whatever, has less distance to travel and thus the sound of the horn will gradually increase, not suddenly.
Suppose the horn is moving towards the observer. As the distance between the source and the observer changes lineally with uniform velocity, the change in frequency with respect to distance travelled should also be lineal as follows:
Let $D$ be some arbitrary distance from the source (horn) to the observer (me)
Let ${f_\Gamma }$ be the apparent frequency of the horn at distance $D$ from the observer, while it is moving with uniform velocity.
Let ${f_o}$ be the frequency of the horn in its own inertial frame (i.e. rest frequency).
Let $f$ be the apparent frequency at some particular distance travelled (i.e. $d$).
Using the y-intercept form of a line,
$y = mx + b$
then
$$f = md + {f_\Gamma }$$
$$m = {{({f_o} - {f_\Gamma })} \over D}$$
$$f = \left( {{{({f_o} - {f_\Gamma })} \over D}} \right)d + {f_\Gamma }$$
$$f = \left( {{{({f_o} - {f_\Gamma })} \over D}} \right)vt + {f_\Gamma }$$
If the horn is moving away, then
$$f = \left( {{{({f_\Gamma } - {f_o})} \over D}} \right)vt + {f_o}$$
If I've made a mistake, please let me know. I can't afford a lab to verify this myself.
Allow me please some time to modify my thinking,
I have done some research looking up the answer. Please see the following diagram,
As such,
Thus in general, the time it takes for pulse $n$ to get to the observer is given by,
$${t_n} = {{d - (n - 1)x} \over {{V_m}}}$$ $${t_n} = {{d - (nx - x)} \over {{V_m}}}$$ $${t_n} = {{d - nx + x} \over {{V_m}}}$$ $${t_n} = {{d + x - nx} \over {{V_m}}}$$ $${t_n} = {{d + x(1 - n)} \over {{V_m}}}$$ and since $$x = {t_s}{v_s}$$ then, $${t_n} = {{d + {t_s}{v_s}(1 - n)} \over {{V_m}}}$$
Here it shows that for every pulse, the time it takes to reach the observer is decreasing, thus the wave period recorded by the observer is decreasing and thus the frequency, the observer records, is increasing.
The above is just nonsense, For the correct derivation, please see the help I received: Doppler Effect YouTube