Page 1 of 1
Smoothing settings
Posted: Mon Oct 26, 2015 9:26 pm
by John
I have low s/n HPLC peaks that I would like to smooth (radiochromatography). My options in Chromeleon are moving average and Savitzky Golay (and some others). The smoothing has a number setting, which I presume is the number of points over which the smoothing is performed on a rolling basis. Can anyone suggest limitations on the parameters? Obviously I want sufficient smoothing to remove noise but not so much to remove small impurities and affect peak shape. This can be done by eye, but I wanted something a little more numerically based. Thanks.
Re: Smoothing settings
Posted: Tue Oct 27, 2015 10:15 am
by carlo.annaratone
I would suggest setting up a design-of-experiments test:
- inject a set of calibration samples from your top concentration down to below the approximate detection limit
- define a reasonable range of smoothing parameters by eye, then define combinations of parameters to cover the parameter space
- each time smooth with a combination of parameters, integrate, build calibration curve, calculate linearity, R2, limit of detection, any other parameter you are interested in.
- Repeat for each combination
- Find the one giving the best results and use it
Re: Smoothing settings
Posted: Tue Oct 27, 2015 5:46 pm
by lmh
I strongly recommend Savitsky-Golay rather than moving average, because it won't affect peak-shape unless you seriously over-do it. Basically it fits a polynomial curve through a succession of points, and chooses the point in the middle of the curve as the best-fit best-point. This means it can curve itself round the curvy starts and ends of peaks, where moving-average will smudge the peak and make it wider.
Apart from this, it's actually really difficult to recommend ideal parameters, because the smoothing actually interacts with the integrator. The main benefit of the smoothing is that it helps the integrator to find the start and end of the peak more reliably, and not get misled by a dip in the middle that might be mistaken for a valley between two peaks. Basically I smooth enough that a reasonable selection of low-abundance peaks are integrated "correctly" (i.e. not split), and I don't worry about being too quantitative about assessing how well it's done, though I do then insist that I use the same integrator and smoothing settings for all samples (otherwise, in effect, you're doing manual integration, but at arm's length by pretending that manual changes to integration on individual files aren't actually manual...). It is always difficult to assess the accuracy of inherently inaccurate data (weak peaks are always less accurate!) without a definite correct answer to which to compare the measurements.