Académique Documents
Professionnel Documents
Culture Documents
1.0 Introduction Regulated laboratories must perform method validation in order to be in compliance with 1) US Food and Drug Administration (FDA) regulations 2) US Pharmacopeia 3) International Conference on Harmonization Unfortunately, there is no single source or final guideline on analytical method validation Validation is customized by choosing necessary tests and acceptance criteria for a given method. The comprehensiveness of validation is based upon the type of method and its requirements. Due to legal status of USP and NF standards, it is essential, therefore, that proposals for adoption of new or revised compendial analytical procedures be supported by sufficient laboratory data to document their validity. 2.0 Pre-validation requirements The first step in the development of a method validation protocol is to determine the objective of the method. What is the method intended to demonstrate? 2.1 Analytical equipment qualification Vendors site Structural and software validation ISO / 21 CFR 11 Vendors Responsibility Users site Functional validation Users site Calibration Maintenance System suitability PQ IQ OQ PQ
IQ
OQ
Users Responsibility
OQ - specific modules of the system are operating according to the defined specifications
PQ - PQ testing is conducted under actual running conditions across the anticipated working. For HPLC, the PQ test should use a method with a well-characterized analyte mixture. After the instrument is placed on-line in the laboratory, and after a set period of use, regulations require maintenance followed by calibration and standardization. A system suitability test provides assurance that a systems performance still is appropriate for use.
The parameters to be measured and their recommended limits obtained from the analysis of the system suitability are shown as follows
Resolution (Rs )
Rs of >2 between the peak of interest and the closest Eluting potential interference ( impurity, excipient, degradation product, internal standard, etc.
T of 2 In general should be > 2000
Validation of an analytical procedure is the process by which it is established, by laboratory studies, that the performance characteristics of the procedure meet the requirements for the intended analytical applications.
3.1 Analytical Performance Characteristics 3.1.1 Accuracy Definition The accuracy of an analytical procedure is the closeness of test results obtained by that procedure to the true value.
The accuracy of an analytical procedure should be established across specified range of the analytical procedure
Accuracy is usually determined in one of four ways. 1) In the case of assay of drug substance accuracy can be assessed by analyzing a sample of known purity (reference materials) and comparing the measured value to the true value. 2) Comparison of the results of the proposed analytical procedure with those of a second well-characterized procedure, the accuracy of which is stated and/or defined
3) The third approach, based on the recovery of known amounts of analyte, is performed by spiking analyte in blank matrices. The analyte levels in the spiked samples should be determined using the same quantitation procedure as will be used in the final method procedure. The percent recovery is then calculated. For assay methods Spiked samples are prepared in triplicate at three levels over a range of 50150% of the target concentration. For impurity methodsSpiked samples are prepared in triplicate at three levels over a range that covers the expected impurity content of the sample, such as 0.12.5% (v/w). 4) The fourth approach is the technique of standard additions, which can also be used to determine recovery of spiked analyte. This approach is used if it is not possible to prepare a blank sample matrix without the presence of the analyte. For assay methods The mean recovery should be 100 2% at each concentration over the range of 80120% of the target concentration. For impurity methodsThe mean recovery should be within 0.1% absolute of the theoretical concentration or 10% relative, whichever is greater, for impurities in the range of 0.12.5% (v/w) To document accuracy the ICH guideline recommends collecting data from a minimum of nine determinations over a minimum of three concentration levels covering the specified range (for example, three concentrations, three replicates each). 3.1.2 Precision The precision of an analytical procedure expresses the closeness of agreement (degree of scatter) between a series of measurements obtained from multiple sampling of the same homogeneous sample under the prescribed conditions. It is normally expressed as the percent relative standard deviation (RSD) for a statistically significant number of samples. According to the ICH precision should be performed at three different levels: repeatability, intermediate precision, and reproducibility.
3.1.2.1 Repeatability Repeatability expresses the precision under the same operating conditions over a short interval of time. Repeatability is also termed intra-assay precision. It should be determined from a) A minimum of 9 determinations covering the specified range for the procedure (e.g., 3 concentrations/3 replicates each) b) A minimum of 6 determinations at 100% of the test concentration. 3.1.2.2 Intermediate precision Intermediate precision expresses within-laboratories variations : different days, different analysts, different equipment, etc. In determining intermediate precision, experimental design should be employed so that the effects (if any) of the individual variables can be monitored. For assay method The instrument precision (RSD) will be 1% and the intra-assay precision will be 2%.
will be within 10% (relative) of the value obtained by the primary testing lab, % (w/w) within 25% for impurities from 0.1 to 1.0 % (w/w).
product with appropriate levels of impurities and demonstrating the separation of these impurities individually and/or from other components in the sample matrix.
3.1.4.3 Based on the Standard Deviation of the Response and the Slope
The slope S may be estimated from the calibration curve of the analyte. The estimate of may be carried out in a variety of ways, for example: a) Based on the Standard Deviation of the Blank Measurement of the magnitude of analytical background response is performed by analyzing an appropriate number of blank samples and calculating the standard deviation of these responses. b) Based on the Calibration Curve A specific calibration curve should be studied using samples containing an analyte in the range of LOD. The residual standard deviation of a regression line or the standard deviation of y-intercepts of regression lines may be used as the standard deviation. Recommended Data The detection limit and the method used for determining the detection limit should be presented. If DL is determined based on visual evaluation or based on signal to noise ratio, the presentation of the relevant chromatograms is considered acceptable for justification. In cases where an estimated value for the detection limit is obtained by calculation or extrapolation, this estimate may subsequently be validated by the independent analysis of a suitable number of samples known to be near or prepared at the detection limit. 3.1.5 Limit of quantification The quantitation limit of an individual analytical procedure is the lowest amount of analyte in a sample which can be quantitatively determined with suitable precision and accuracy. The quantitation limit is a parameter of quantitative assays for low levels of compounds in sample matrices, and is used particularly for the determination of impurities and / or degradation products. 3.1.5.1 Based on Visual Evaluation Visual evaluation may be used for non-instrumental methods but may also be used with instrumental methods.
The quantitation limit is generally determined by the analysis of samples with known concentrations of analyte and by establishing the minimum level at which the analyte can be quantified with acceptable accuracy and precision. 3.1.5.2 Based on Signal-to-Noise Approach
This approach can only be applied to analytical procedures that exhibit baseline noise. Determination of the signal-to-noise ratio is performed by comparing measured signals from samples with known low concentrations of analyte with those of blank samples and by establishing the minimum concentration at which the analyte can be reliably quantified. A typical signal-to-noise ratio is 10:1.
3.1.5.3 Based on the Standard Deviation of the Response and the Slope The quantitation limit (LOQ may be expressed as: 10.0 / S where = the standard deviation of the response S = the slope of the calibration curve
The slope S may be estimated from the calibration curve of the analyte. The estimate of may be carried out in a variety of ways, for example:
a) Based on the Standard Deviation of the Blank Measurement of the magnitude of analytical background response is performed by analyzing an appropriate number of blank samples and calculating the standard deviation of these responses. b) Based on the Calibration Curve A specific calibration curve should be studied using samples containing an analyte in the range of LOD. The residual standard deviation of a regression line or the standard deviation of y-intercepts of regression lines may be used as the standard deviation. Recommended Data The detection limit and the method used for determining the detection limit should be presented. If DL is determined based on visual evaluation or based on signal to noise ratio, the presentation of the relevant chromatograms is considered acceptable for justification. In cases where an estimated value for the detection limit is obtained by calculation or extrapolation, this estimate may subsequently be validated by the independent analysis of a suitable number of samples known to be near or prepared at the detection limit.