Search Issue Tracker
By Design
Votes
0
Found in
3.5.2f2
Issue ID
478463
Regression
No
AudioSource.Play(offsetInSamples) with best latency audio settings on iOS is not always accurate
By design:
This is expected. The playback time is always relative to the next frame start as this is where all sources played within a frame are scheduled to start (also those not started using PlayScheduled). The reason you're seeing 500 samples is probably because this is the default processing block size used on iOS (actually 512 samples), so that defines the maximum delay. Obviously the lower the sample rate gets, the more likely this jitter is to happen, which is probably why you are getting the best results when running in the editor
To reproduce:
1. Create new project
2. Attach AudioTest.js to camera
3. Change audio settings>DSPbuffer size to best latency
4. Build and run on iOS
All about bugs
View bugs we have successfully reproduced, and vote for the bugs you want to see fixed most urgently.
Latest issues
- Texture2D hash changes inside of an AssetBundle when rebuilding a SpriteAtlas bundle with an empty AssetPostprocessor Script enabled
- Aniso Level still applies when Generate MipMap is disabled in Texture Import Settings
- Mipmap Limit Groups long names are not truncated when creating a new Mipmap Limit Group with a long name
- “ArgumentException: Invalid double parameter.” error is thrown when Infinity is typed into the Fixed Timestep field
- GameObject becomes gray when using HDRP and STP together on macOS
Add comment