diff --git a/test/Datadog.Trace.Tests/ApiTests.cs b/test/Datadog.Trace.Tests/ApiTests.cs index 7078ea0af49a..7f981ef3a3da 100644 --- a/test/Datadog.Trace.Tests/ApiTests.cs +++ b/test/Datadog.Trace.Tests/ApiTests.cs @@ -61,7 +61,7 @@ public async Task SendTracesAsync_500_ErrorIsCaught() sw.Stop(); Assert.Equal(5, handler.RequestsCount); - Assert.InRange(sw.ElapsedMilliseconds, 1500, 10000); // should be ~ 1600ms + Assert.InRange(sw.ElapsedMilliseconds, 1000, 12000); // should be ~ 1600ms // TODO:bertrand check that it's properly logged } diff --git a/test/Datadog.Trace.Tests/Sampling/RateLimiterTests.cs b/test/Datadog.Trace.Tests/Sampling/RateLimiterTests.cs index b5662b295aa0..311fc1c95349 100644 --- a/test/Datadog.Trace.Tests/Sampling/RateLimiterTests.cs +++ b/test/Datadog.Trace.Tests/Sampling/RateLimiterTests.cs @@ -74,9 +74,9 @@ private static void Run_Limit_Test(int? intervalLimit, int numberPerBurst, int n var totalMilliseconds = result.TimeElapsed.TotalMilliseconds; var expectedLimit = totalMilliseconds * actualIntervalLimit / 1_000; - - var upperLimit = expectedLimit + (actualIntervalLimit * 0.8); - var lowerLimit = expectedLimit - (actualIntervalLimit * 0.8); + var acceptableDifference = (actualIntervalLimit * 0.80); + var upperLimit = expectedLimit + acceptableDifference; + var lowerLimit = expectedLimit - acceptableDifference; Assert.True( result.TotalAllowed >= lowerLimit && result.TotalAllowed <= upperLimit, @@ -88,7 +88,7 @@ private static void Run_Limit_Test(int? intervalLimit, int numberPerBurst, int n var totalExpectedAllowed = 2 * actualIntervalLimit; var expectedRate = totalExpectedAllowed / (float)totalExpectedSent; - var maxPercentVariance = 0.20f; + var maxPercentVariance = 0.35f; var lowestRate = expectedRate - maxPercentVariance; var highestRate = expectedRate + maxPercentVariance;