Search Issue Tracker
By Design
Votes
0
Found in
3.5.2f2
Issue ID
478463
Regression
No
AudioSource.Play(offsetInSamples) with best latency audio settings on iOS is not always accurate
By design:
This is expected. The playback time is always relative to the next frame start as this is where all sources played within a frame are scheduled to start (also those not started using PlayScheduled). The reason you're seeing 500 samples is probably because this is the default processing block size used on iOS (actually 512 samples), so that defines the maximum delay. Obviously the lower the sample rate gets, the more likely this jitter is to happen, which is probably why you are getting the best results when running in the editor
To reproduce:
1. Create new project
2. Attach AudioTest.js to camera
3. Change audio settings>DSPbuffer size to best latency
4. Build and run on iOS
All about bugs
View bugs we have successfully reproduced, and vote for the bugs you want to see fixed most urgently.
Latest issues
- RenderPipelineManager.activeRenderPipelineTypeChanged does not detect changes when changing the Render Pipeline Asset outside of Play Mode
- The Scene View becomes gray and "NullReferenceException" errors are thrown when attaching the Unity Editor process to RenderDoc
- Memory usage rapidly rises when changing colors in the color wheel
- Physically Based Sky's Horizon and Zenith Tint Color Selection Bars don't have immediate access to color pickers
- EmptyAdditionalLightShadowmapTexture leaks from memory when URP quality level is changed using QualitySettings.SetQualityLevel
Add comment