Blog

More Than You Ever Wanted to Know About Calibrations, Part 8 – Low Level Replicates

17 Nov 2025

After Part 6 of this blog, someone reached out to ask, “In our lab, it’s been typical practice for decades to perform extra replicates near the LOQ to offset the effects of variation in the RIAs (notably non-linear calibration curves). I wonder now how beneficial it would be to the calibrations you’ve presented here. Does changing from duplicate to triplicate (or from triplicate to quadruplicate) determination in the highest variation region improve the calibration performance?” My assumption was that this is basically adding more calibration points to the low end, so adding replicates would have a similar effect to simply adding more discrete points. I did a quick check that seemed to confirm it, but let’s dig into this more deeply here.

The general idea behind this seems to be that for points that may be inaccurate, running them in replicate would average them out closer to the true value. On paper this makes sense, but the potential to add more error by measuring more high-uncertainty points is also there. To check this, I did my tried-and-true method of modeling calibrations by using Excels random number generator and playing with the variance, similar to what I’ve done in previous posts. I tried both an equal spaced calibration and a calibration following the EPA method 1633 levels, with both higher variance at the low end and with equal variance at all points. A summary of the levels and variance is shown in Table I below. 

Table I: Modified Calibration Levels and Variance

Method SpacingRelative ErrorEqual SpacingRelative Error
0.250%0.250%
0.540%1020%
1.2530%2020%
2.530%3020%
520%4010%
12.520%5010%
62.510%62.510%
0.210%0.210%
0.510%1010%
1.2510%2010%
2.510%3010%
510%4010%
12.510%5010%
62.510%62.510%

To evaluate how the calibrations performed, I modeled 5 calibrations for each parameter and looked at how the % relative standard error (%RSE) compared between a calibration with single points, and one with 3 replicates at the lowest calibration point. Table II shows the results using the method spacing and the higher variance at the low end.

Table II: %RSE results for high variance calibrations with single points and replicate low points using 1633 calibration spacing.

Method Spacing Relative Error Equal Spacing Relative Error
0.2 50% 0.2 50%
0.5 40% 10 20%
1.25 30% 20 20%
2.5 30% 30 20%
5 20% 40 10%
12.5 20% 50 10%
62.5 10% 62.5 10%
0.2 10% 0.2 10%
0.5 10% 10 10%
1.25 10% 20 10%
2.5 10% 30 10%
5 10% 40 10%
12.5 10% 50 10%
62.5 10% 62.5 10%

 

To evaluate how the calibrations performed, I modeled 5 calibrations for each parameter and looked at how the % relative standard error (%RSE) compared between a calibration with single points, and one with 3 replicates at the lowest calibration point. Table II shows the results using the method spacing and the higher variance at the low end.

Table II: %RSE results for high variance calibrations with single points and replicate low points using 1633 calibration spacing.

  Force 0 Ignore 0 Force 0 Ignore 0
Calibration Linear Equal Linear 1/x Linear 1/x2 Linear Equal Linear 1/x Linear 1/x2 Quad Equal Quad 1/x Quad 1/x2 Quad Equal Quad 1/x Quad 1/x2
Single 1 131.70% 19.08% 11.09% 25.05% 24.71% 25.25% 10.77% 10.24% 10.21% 30.54% 29.56% 28.30%
Single 2 105.52% 12.44% 11.95% 23.84% 24.04% 24.33% 42.58% 14.05% 12.77% 32.30% 29.59% 26.89%
Single 3 44.07% 13.87% 13.90% 14.46% 14.55% 14.56% 37.98% 17.31% 15.59% 19.12% 17.42% 16.19%
Single 4 43.95% 16.00% 15.96% 22.97% 21.92% 18.80% 31.80% 18.17% 17.83% 20.59% 21.05% 20.03%
Single 5 24.95% 20.64% 19.88% 19.83% 19.78% 20.01% 33.05% 24.35% 22.33% 22.21% 22.06% 22.44%
Average 70.04% 16.80% 14.56% 21.23% 21.00% 20.59% 31.23% 16.82% 15.75% 24.95% 23.94% 22.77%
Replicate 1 60.16% 12.94% 11.90% 25.38% 24.16% 20.00% 34.81% 12.55% 12.64% 23.60% 21.24% 19.10%
Replicate 2 27.17% 27.02% 27.58% 27.33% 27.35% 28.11% 43.19% 29.21% 30.17% 29.47% 29.63% 30.53%
Replicate 3 38.74% 23.77% 23.32% 25.36% 25.37% 26.02% 33.30% 27.43% 24.98% 28.18% 27.92% 28.23%
Replicate 4 148.14% 19.74% 14.46% 22.28% 23.38% 22.33% 38.86% 13.29% 13.39% 45.10% 37.48% 23.95%
Replicate 5 34.70% 26.12% 25.57% 28.69% 28.69% 29.65% 72.67% 29.55% 27.36% 32.48% 31.29% 32.20%
Average 61.78 21.92% 20.57% 25.81% 25.79% 25.22% 44.57% 22.41% 21.71% 31.77% 29.51% 26.80%
Difference 8.26% -5.11% -6.01% -4.58% -4.79% -4.63% -13.33% -5.58% -5.96% -6.82% -5.57% -4.03%

The linear equal weighted curves that ignores 0 shows a slight improvement on average, with the %RSE being 8.26% lower on average when using replicates. However, the %RSE is still generally very high, with some calibrations having over 100% RSE on both types, so adding replicates won’t turn a poor calibration into a good one. The %RSE on all other types of curves is slightly better when not using replicates, but the difference is still very small. From the results above, weighting a curve with no replicates is a much better choice than adding replicates at the low end.

Table III shows the same results, but with an equal spaced curve.

Table III: %RSE results for high variance calibrations with single points and replicate low points using equal spacing.

  Force 0 Ignore 0 Force 0 Ignore 0
Calibration Linear Equal Linear 1/x Linear 1/x2 Linear Equal Linear 1/x Linear 1/x2 Quad Equal Quad 1/x Quad 1/x2 Quad Equal Quad 1/x Quad 1/x2
Single 1 271.30% 13.45% 10.12% 11.19% 10.96% 10.97% 165.11% 9.51% 7.77% 12.29% 13.22% 12.08%
Single 2 28.01% 8.98% 7.73% 12.85% 12.87% 12.45% 230.72% 9.64% 7.38% 13.25% 14.61% 13.14%
Single 3 80.56% 13.63% 12.18% 17.62% 17.48% 16.98% 287.43% 14.02% 11.44% 21.26% 18.07% 12.44%
Single 4 337.96% 12.02% 10.98% 11.73% 11.72% 11.75% 513.83% 11.25% 10.84% 22.92% 16.72% 13.07%
Single 5 102.71% 5.03% 5.02% 10.29% 10.40% 10.12% 73.35% 6.22% 5.62% 13.72% 12.89% 9.55%
Average 164.11% 10.62% 9.21% 12.74% 12.69% 12.45% 254.09% 10.13% 8.61% 16.69% 15.10% 12.06%
Replicate 1 159.41% 22.62% 22.81% 22.89% 22.84% 23.32% 127.18% 25.22% 26.74% 24.63% 25.15% 25.54%
Replicate 2 68.40% 26.86% 26.66% 27.49% 27.64% 26.29% 233.79% 30.02% 28.11% 36.78% 31.19% 27.14%
Replicate 3 64.17% 7.68% 7.38% 15.63% 15.73% 14.97% 110.14% 7.85% 7.75% 16.50% 17.99% 13.34%
Replicate 4 232.75% 11.65% 11.03% 18.72% 19.47% 17.06% 60.45% 11.35% 11.56% 28.53% 29.69% 16.75%
Replicate 5 209.19% 21.77% 21.34% 25.20% 24.38% 21.74% 28.95% 21.49% 21.86% 21.76% 21.94% 21.05%
Average 146.78% 18.12% 17.85% 21.98% 22.01% 20.67% 112.10% 19.19% 19.20% 25.64% 25.19% 20.76%
Difference 17.33% -7.49% -8.64% -9.25% -9.33% -8.22% 141.99% -9.06% -10.59% -8.95% -10.09% -8.71%


The results are very similar, with a slight improvement on average for the linear equal weighted ignore 0 fit, but overall high %RSE that would make the fits unacceptable. The quadratic equal weighted ignore 0 fit shows significant improvement in average %RSE, but only one of the five calibrations has a %RSE result under 60%, so it’s still better to move to weighted fits instead of adding replicates.

In the interest of space, I won’t show the tables for the calibrations with 10% variance at all points, but the results match what Tables II and III show, with some improvement on %RSE on average for unweighted curves, but the best results are still weighted curves without replicates. 

While %RSE is a good overall metric, it’s not quite an apples-to-apples comparison in what I’m showing here, especially for the models with higher variance at the low end. By adding replicates to the low end I was adding more inaccurate measurements to the calculation, which could drive the %RSE higher even if the calibration fit was similar. To check this, I took 20 modeled points at the low calibration level with high variance and plugged them into the 5 replicate calibrations for each type. From there I calculated the residual error for each measurement and averaged the results across the 5 replicate calibrations to see how accurate on average each calibration type is. I also calculated the standard deviation to see if any of the calibration types were more precise. Again, in the interests of space I only show results for the high variance linear equal weighted and linear 1/x2 weighted ignore 0 fits, but the quadratic and force 0 fits followed similar trends.

Table IV: Average Accuracy and Precision at Low Calibration Level

 Linear Equal WeightLinear 1/x2
 Single PointsReplicateSingle Point Replicate
Average71.01%77.73%19.97%19.60%
Stdev.43.96%36.90%12.84%12.51%

Table IV shows that on average there is no significant difference in accuracy or precision when adding replicate points. The best improvement in accuracy and precision comes from weighting the calibration curve appropriately. The results show that my general recommendation of weighting the curve when possible is the best way to properly define the calibration curve at the low end, and that adding replicates does not add any value.

View all of the posts in the “More Than You Ever Wanted to Know About Calibrations” series.

Author

  • Jason Hoisington

    Jason Hoisington received his bachelor’s degree in general science with a focus on chemistry from the University of Alaska, Fairbanks. He worked for SGS Environmental for seven years in environmental soil and water testing, developing methods for the analysis of volatiles and semivolatile organics to include pesticides and polychlorinated biphenyls (PCBs). In 2012, Jason moved on to lab and application support for Dow Chemical Company, providing advanced analytical troubleshooting and method development. In 2019, Jason joined Restek and has focused on air applications.

    View all posts
GNBL4924