No announcement yet.

Music Timing System

  • Filter
  • Time
  • Show
Clear All
new posts

    Music Timing System

    Hi, I've been trying to get a stable timing system to use to trigger musical transitions between one piece of music and another. Obviously these transitions need to be accurately timed - but I am finding this impossible at the moment. Basically what I need is a reliable and consistent 'metronome' so that I can trigger different musical events on defined beat points.

    I've tried setting up a looping retriggerable delay node which then triggers a music cue, but the timing of the delay loops are off so there are inconsistent gaps between the musical events. My next step was to build a blueprint system using the event tick node and the deltatime value to calculate the timing intervals, but I'm getting widely inconsistent timings between ticks (this is based on a custom Kismet object that I created in UDK that works fairly reliably). Just in case I’m doing something wrong, here’s a description of my tick-based system along with screenshots… When the system is activated it stores the current time using the Get Real Time Seconds node and uses a pre-defined bpm value to calculate the various different beat division intervals (ie. 120bpm = 2bps = 500ms for a 1/4 beat, 250ms for a 1/8 beat, etc…), and then calculates what the time point for each of these intervals. So, if the current time is 5.385secs and the beat interval time for a 1/4 beat is 500ms, the next time point for a 1/4 beat is 5.885secs (5.385 + 0.5). Then every tick the system checks the Get Real Time Seconds node and compares this to the ‘ideal’ beat time. If the actual time falls within a window (half the deltatime) of the ‘ideal’ time then an event is fired off to trigger the musical event and the next ‘ideal’ beat time is calculated. So, if the deltatime is 0.03secs and the ‘ideal’ time for a 1/4 beat is 5.885secs then an ‘actual’ time of 5.885+/- 0.015 would trigger an event.
    Click image for larger version

Name:	musicalTiming01.png
Views:	1
Size:	108.1 KB
ID:	1136407
    Click image for larger version

Name:	musicalTiming02.png
Views:	1
Size:	80.7 KB
ID:	1136406

    Even if the ticks were consistent, the likelihood is that the ‘actual’ (accurate) time point of a musical cue might be inbetween ticks and so using a frame-based timing system is going to be problematic as we are very sensitive to timing issues within music.

    Does anyone have any suggestions on how to get either a reliable and consistent timing system within UE4, or more ideally how to use the sample accurate timings from the audio soundcard through the creation of a custom C++ based object/actor.

    You could see the ammount of 'off time' and include it for whatever you are doing.

    Could you explain what you are doing that need this accuracy?


      I'm attempting to create a metronome system so that I can trigger musical events at specific (and accurate) musical timings. With a view to creating a generative music generator.
      Because we are extremely sensitive to timings within music this needs to be fairly accurate, but more importantly it needs to be consistent and that is the main problem I'm having at the moment...
      So, lets say we are using a musical tempo of 120bpm, then every beat (pulse) would occur at 500ms intervals, and then common sub-divisions would be at 250ms and 125ms.
      If the tick events were consistent then half a frame's worth of 'error' in the timings may not be that noticeable at the larger interval times, but could become problematic as the intervals get smaller.
      In a ideal world I would be able to fire off an event at a specific time interval, but using frame-based systems for timings means that you can't do this as the 'ideal' time point may well fall between frames.

      Does that make sense?


        I don't think you'll be able to get such accurate timing from the tick system, it's not designed for that. I would do it on a separate thread if I were in your position. I don't know much about UE4's sound system and if sounds can be started from different threads. If not, at worst case you'd need to write your own sound mixer and output the result into a buffer that's streamed by UE4 (that should be possible, since vorbis audio files are decoded in separate threads).


          Pedro is probably right. Have delta time print to screen every tick to see what I mean. It's not all that stable.

          I'm in the planning phase of adding music to my game also. The best way I can see to sync tracks is to have them all start at the same time and fade them in and out. This might be a bit of a weight on the processing and memory, but I'm hoping it will be OK. Portal 2 has several tracks all playing at the same time, that just fade in and out.
          Please let us know if you do manage to work something else out! Audio in UE4 still seems pretty primitive (as it was in UE3).


            maybe this could help you


            good luck


              Alas, there is not currently a rock-solid timing mechanism for music and audio playback to a tempo. The suggestion of timers is a good one; we used a system like this on Bulletstorm to allow branching on a measure-by-measure basis. This was done at the C++ level though. Audio will at some point get its own thread, at which point this functionality will be possible natively. Sorry I couldn't be of more help.
              Zak Belica
              Audio Director, Epic Games


                Hi guys,
                Thanks for the replies!
                The frustrating thing is that I built a custom Kismet object for UDK using UnRealScript which works in the same way as my Blueprint attempt (ie. based on Tick and deltatimes) and while it was not 100% accurate it was good enough for most purposes... Guess I may have to look into C++ scripting...

                @SuperNovaBen - Yes, your idea would work fine for parallel music stems where you just want to fade tracks in and out. But I want to be able to transition between different pieces of music, and have the occasional stinger and so need a system that keeps the musical pulse accurately.

                @ZakBelica - Are you able to offer any advice about how you went about this at the C++ level?

                If I get anywhere I'll be sure to share...!



                  Hello again Teed,

                  I'm really interested in this (I used to do a lot of sound design and studio engineering). Having a good tempo system would of course be ideal, but there are other work arounds.

                  Let's say your music is comprised of various phrases and passages. You could cut them up and loop them a certain number of times then move on to the next if required. If you needed to change the music to fit the context (e.g some action started or the mood changed), you could wait until the end of the loop currently playing and change the music then. Of course, depending on your loop lengths, this will have some delay between the situation changing and the music.

                  My first idea about cross fading (which I think you understood) was that if all the pieces are at the same tempo, they will be synced automatically as long as you start them all at the same time. You can cross fade at a speed of 1ms to make the music jump to the next piece depending on the situation. This would work if you only needed say 4 pieces within a level/environment.

                  If you are trying to make a music based game, these aren't really good options. Anyway, just thought I'd share some ideas.
                  By the way, how accurate is 'Real world time'?


                    To be honest, I think that having a reliable tempo system is a bit more than 'ideal' I think it's pretty much 'essential' if you want to be doing anything interesting with music..! But given that that's my thing - I suppose I probably would do...

                    Your ideas are good ones and would work for some situations, but I want to be able to create a system whereby I can specify transition times between different musical events (be these loops, stingers or transitions). So, yes you could just wait until the end of the loop, but if your loops are a reasonable length (in order to avoid a sense of repetition) then the delay between the context changing and the music changing to reflect this could be relatively long. If we had a reliable tempo system you could choose to apply the transition on either the bar line or even smaller beat sub-divisions depending on fast you wanted to respond and/or the musical material.

                    Cross-fading works very well for some applications although you do need to bear in mind the cost of having multiple long pieces of music all playing simultaneously. have you tried this yet? How does a fade time of 1ms work?
                    My initial concern with a fade time of this speed would be that given that things can only be processed/applied in terms of frame (deltatime) time and given the fact that 1ms is considerably shorter than frame time, this would produce the effect of the audio instantaneously jumping between volume levels, thus producing audible artefacts within the audio...

                    As far as I can tell the [Get Real World Time] and [Get Real Time Seconds] nodes are fairly accurate (although, I have found some very small inconsistencies between the two) I haven't done too much in the way of rigorous testing... The main problem with checking all of this is that you can only check their outputs at frame time which is inconsistent!


                      Hi teed,

                      As has been stated, the Delta time is not ideal as it affected by framerate, which can certainly throw off a metronome if you have a faster/slower pc than others. As a musician myself I understand the need for consistency and accuracy with a metronome system, but unfortunately I have yet to find a solution that is consistent and exact. Please let me know what you come up with as I'd be very interested in seeing what you are able to do!
                      Adam Davis | Marketplace Support | Epic Games
                      How to report a bug? | Installation & Setup issues? | Answerhub Bug Reports | Twitter


                        Hi teed,

                        I'm trying to develop something similar and have run into the same problems. I did find some interesting BPs in this PlugIn from Rama:


                        but they don't actually solve the triggering conundrum. Am interested to see how this progresses with future builds!


                          i did that simply blueprint for someone who needed a radio. : Click image for larger version

Name:	RadioArraySet.png
Views:	4
Size:	148.8 KB
ID:	1053195 the thing is that by setting sound it stopped the old one. In hope it helps. Maybe with differents ambient sound system at same place you can mix different timelines delay and maybe var them separately and select the sound to set in each timeline so you know where it is in time...
                          Last edited by Fen; 08-04-2014, 05:43 AM.


                            I'm not able to test this yet, as I'm not at my workstation, so at the risk of looking silly... has anyone tried using Timeline objects?

                            Set up a timeline object to run for 2 seconds, and loop, and add an event track that fires off an execution pulse once every 0.5 seconds. That should be giving you quarter notes at 120BPM, no..?

                            That at least decouples you from the tick pulse, so it's no longer framerate-dependent, and using a timeline node over a simple delay should give you more stability in terms of slowdowns accumulating.

                            I'd think the biggest problem with audio sync wouldn't even be the timing itself, it would be queueing; in audio software programming, when handling real-time audio, there's always some sort of buffer (on a good card usually like 2-4ms, on others it's substantially higher) because audio buffers fill up inconsistently owing to the other processes running at the same time. Even if you could "fire off" execution pulses at the right time, Unreal still has to queue all those audio playback commands and execute them, which won't always have correct timing. An audio sequencing system is usually looking ahead to the events that come in the future and prepping them, loading audio into memory, decoding, etc., and then playing the audio at the predetermined appropriate time. I would suspect that most dynamic music systems in games would tend to start all audio playing back in simultaneous loops and blending between them, as mentioned ITT, rather than trying to build it in real-time; if you need the game to do something in response to player input, you can't look ahead to it. Even with a good audio card and using a dedicated spec like ASIO, it's a challenge getting a machine to reliably fire off musical events in response to real-time user inputs. Trying to do so while ALSO doing everything else that the game needs to do in real-time seems like it would be nightmare-tier, especially considering most gaming rigs are not built with audio performance in mind. If all the audio is already synchronized, then all that needs doing is some very quick multiplication for volume control, rather than all the other tasks associated with playing back an audio file.


                              Just tried the timeline method and seems to be working well so far! Haven't had much chance to test it out properly but will let you know how I get on. Good shout RhythmScript!