Sync the gameplay with audio and music¶
In any application or game, sound and music playback will have a slight delay. For games, this delay is often so small that is negligible. Sound effects will come out a few milliseconds after any play() function is called. For music this does not matter as in most games it does not interact with the gameplay.
Still, for some games (mainly, rhythm games), it may be required to synchronize player actions with something happening in a song (usually in sync with the BPM). For this, having more precise timing information for an exact playback position is very welcome.
Achieving very low playback timing precision is very difficult. This because many factors are at play during audio playback:
- Audio is mixed in chunks (not continuously), depending on the size of audio buffers used (check latency in project settings).
- Audio has latency after mixed, it doesn’t immediately play after mixed.
- Graphics APIs display two or three frames late.
- When playing on TVs, some delay may be added due to image processing.
The most common way to reducing latency is to shrink the audio buffers (again, by editing the latency setting in the project settings). The problem is that when latency is too small, it will require considerably higher amount of CPU for sound mixing. this increases the risk of skpping (a crack in sound because a mix callback was lost).
This is common a tradeoff, so Godot ships with sensible defaults that should not need to be commonly altered.
The problem, in the end, is dealing with synchronization for games tha require it, and not so much this slight delay. Beginning Godot 3.2, some helpers were added to obtain more precise playback timing.
Using the system clock to sync¶
As mentioned before, If you call AudioStreamPlayer.play(), sound will not begin immediately, but when the audio thread processes the next block.
This delay can’t be avoided but it can be estimated by calling AudioServer.get_time_to_next_mix().
The output latency (what happens after the mix) can also be estimated by calling AudioServer.get_output_latency().
Add these two and it’s possible to guess almost exactly when sound or music will begin playing in the speakers:
This way, obtaining the actual playback position during _process() is possible:
In the long run, though, as the sound hardware clock is never exactly in sync with the system clock, they timing information will slowly drift away.
For a rhythm game where a song begins and ends after a few minutes, this approach is fine (and it’s the recommended approach). For a game where playback can last a much longer time, the game will eventually go out of sync and a different approach is needed.
Using the sound hardware clock to sync¶
Using AudioStreamPlayer.get_playback_position() to obtain the current position for the song sounds ideal, but it’s not that useful as-is. This value will increment in chunks (every time the audio callback mixed a block of sound), so many calls can return the same value. Added to this, the value will be out of sync with the speakers too because of the previously mentioned reasons.
To compensate for the “chunked” output, there is a function that can help: AudioServer.get_time_since_last_mix().
Adding the return value from this function to get_playback_position() increases precision:
To increase precision, substract the latency information (how much it takes for the audio to be heard after it was mixed):
The result may be a bit jittery due how multiple threads work. Just check that the value is not less than in the previous frame (discard it if so). This is also a less precise approach than the one before, but it will work for songs of any length, or synchronizing anything (sound effects, as an example) to music.
Here is the same code as before using this approach: