Open access peer-reviewed chapter

Method Validation Approaches for Pharmaceutical Assessments – Highlights with High Performance Thin Layer Chromatographic (HPTLC) Techniques

By David Jenkins, Cherif Diallo, Ed Bethea, Eliangiringa Kaale and Thomas Layloff

Submitted: June 29th 2017Reviewed: October 18th 2017Published: December 20th 2017

DOI: 10.5772/intechopen.71765

Downloaded: 363

Abstract

Method validation is an important activity for pharmaceutical evaluations to ensure that analytical methods are suitable for their intended use. With particular focus on active ingredient and impurities, the implementation of different categories of method validation are explained for qualitative and quantitative methods. Detailed explanations with example approaches are provided for the key aspects of method validation, namely specificity, accuracy, linearity, limits of detection/quantitation, precision, robustness, and method range. While all of the sections outlined for method validation are generally applicable for a variety of techniques commonly used in pharmaceutical analysis (i.e., UV and HPLC instrumentation), focused attention is provided for examples that have been implemented using high performance thin layer chromatographic techniques.

Keywords

  • method validation
  • pharmaceuticals
  • HPTLC
  • assay
  • active pharmaceutical ingredient
  • impurities

1. Introduction

Method Validation (MV) is a development process undertaken to establish, within acceptable statistical bounds, that an assessment procedure or method consistently yields a “true” result both in “within laboratory” and “among laboratories” testing. Pharmaceutical product quality assessments are focused on methods for the active pharmaceutical ingredient (API) and related impurities. Being able to perform methods of analysis to assess product quality is critical in law enforcement and regulating commerce. In addition, for new drug products, these quality determinations are surrogate performance indicators for assuring the safety and efficacy of a pharmaceutical product. The safety and efficacy of a pharmaceutical product are established with a “pivotal lot” production of the product and the characterization of this lot with well validated methods with acceptable performance characteristics is critical to assure that future production lots have the same quality characteristics as the “pivotal lot”, thereby assuring they have equivalent safety and efficacy.

In the United States of America (USA), there are both private and public standards; the private standards are created through a USA Food and Drug Administration (FDA) approval process of industry method submissions that can be used for law enforcement, and public standards, which are promulgated in the monographs of the United States Pharmacopeia (USP) [1], that may be used in law enforcement or to support commercial agreements. The private standards, which are not publicly available, are private agreements between the approving government body and the submitting industry on the methods and standards to be used in law enforcement. The method validation protocols for the establishment of private standards are provided in the guidance of the “International Conference On Harmonisation Of Technical Requirements For Registration Of Pharmaceuticals For Human Use” (ICH) which have been incorporated into the laws and regulations in the European Union, Japan and the USA; these procedures are required for the assessments of new drug entities [2]. The method validation protocols for the establishment of monographs to support public standards are provided in USP <1225> [3] and ICH Q2 [4]. Both protocols cite the same analytical performance characteristics and test procedures except that the public standard must be able to be applied to all legally marketed products containing the specific API whereas the private standard applies only to the approved API in the specific product.

The analytical performance characteristics which must be assessed in both the ICH and USP are Accuracy, Precision (both Repeatability and Intermediate in ICH), and Specificity. Detection Limit, Quantitation Limit, Linearity and Range depending on which attributes are to be assessed. The USP presents the characteristics as noted below [3]:

  1. “Category I — Analytical procedures for quantitation of major components of bulk drug substances or active ingredients (including preservatives) in finished pharmaceutical products.

  2. Category II — Analytical procedures for determination of impurities in bulk drug substances or degradation compounds in finished pharmaceutical products. These procedures include quantitative assays and limit tests.

  3. Category III — Analytical procedures for determination of performance characteristics (e.g., dissolution, drug release, etc.).

  4. Category IV — Identification tests.”

These can be categorized further into Assay procedures for Category I, Category II, impurity determinations, and Category III (dissolution and drug release are different procedures for preparing a solution of the API), all performance characteristics except Detection Limit must be validated. For the Category II limit tests only the Specificity and Detection Limits must be validated and for Category IV, Identification Tests, only the Specificity needs to be validated. The following sections will provide approaches toward the various aspects of method validation. Although the approaches are generally applicable to common techniques used in pharmaceutical analysis (such UV-VIS and HPLC quantifications), particular emphasis will be placed on high performance thin layer chromatography (HPTLC) techniques.

2. Specificity

Specificity is the ability of a method to distinguish an analyte from all substances that are present or likely to be present in test samples [3, 4]. When possible, these substances should include future degradation products and other ingredients (i.e., excipients). An analytical procedure is specific when placebo and impurity spots do not overlap partially with and are not buried under the analyte spot. In addition, the calculated amount of analyte does not depend on the quantity of other substances.

Various approaches are possible when evaluating method specificity [5, 6, 7, 8]. Ideal demonstration of specificity for an HPTLC analytical procedure requires chromatographing simultaneously three types of samples: sample type 1 is the pure analyte or its reference standard, sample type 2 is the analyte mixed with a representative blank and all likely impurities, and sample type 3 is the representative blank mixed with all likely impurities. Likely impurities include degradation products, reagents, intermediates, excipients, side products, and analyte isomers. The mixtures can be created by spiking test samples (API substances or finished products) or placebos with likely impurities.

In practice, the unavailability of one or more of these types of samples can pose a significant challenge. In some cases, it is often difficult to know all likely impurities. There can be several sample deficiency scenarios. If the pure analyte or its reference standard is not available, demonstration of specificity can be quite challenging if not impossible. If a representative blank is available, but some or all likely impurities are missing, the typical test sample is subjected to stress testing environments. It should be noted however that stress testing is unlikely to produce some analyte isomers, reagents, intermediates and side products. If a representative blank is unavailable, but some likely impurities are available, spiking the typical test sample with impurities can show that increasing impurities will not change analyte signal. In addition, efforts should be made (perhaps by contacting the manufacturer), whenever possible, to create a representative blank even if it’s not exactly in the same dosage form as the test sample. If neither a representative blank nor impurity standards are available, the typical test sample is subjected to stress testing to alleviate some of the deficiencies. Once again, the limitations of stress testing should be acknowledged because it may not produce all likely impurities, it may not account for impurities that are completely buried under the analyte signal, and it may not indicate whether some excipients or impurities can react with the analyte.

In general, stress testing, impurity spiking, and peak-purity analysis are the common tools used to address certain sample deficiencies. To demonstrate method specificity, validation reports typically discuss several measures of performance. One measure of specificity is resolution of the analyte spot relative to the closest non-analyte spot. For HPTLC, the resolution should be a least 1 [5].

Analyte peak purity is another measure of specificity that is typically reported. Often, analyte peak purity in the analyte reference standard is compared to analyte peak purity in the other test samples mentioned above. The analysis is performed by comparison of peak spectra at the start, apex, and end of the analyte peak. Some authors use correlation coefficients [9] as a measure of peak purity, and others rely on software algorithms that may involve Matrix Algebra. It should be noted that while peak purity can detect the presence of some impurities in the analyte peak, it does have some limitations. For example, peak-purity analysis does not account for missing impurities that could overlap with the analyte peak, and it does not account for impurities having a spectrum that is similar to that of the analyte. In addition, peak-purity analysis is not applicable for detectors that do not register the entire analyte spectrum for each time point.

A third measure of peak purity is an overlay of chromatograms. This measure is especially useful for showing the analyte peak stability during impurity spiking or stress testing. For example, the chromatograms of a finished pharmaceutical product, before and after accelerated aging, can be overlaid to support method specificity.

3. Accuracy

A succinct definition of accuracy is “nearness to truth”. The ICH guidelines [4] provide the following definition:

“The accuracy of an analytical procedure expresses the closeness of agreement between the value which is accepted either as a conventional true value or an accepted reference value and the value found.”

In other words, accuracy of a method represents the agreement between an expected value and the value generated by the candidate method (the method value). Therefore, accuracy determination involves determining the expected value, finding the method value and calculating the agreement between the two values [3, 4].

In pharmaceutical testing, accuracy is mainly relevant to quantitative methods, such as assay, content uniformity, dissolution, and impurity quantitation. To determine the accuracy of a quantitative HPTLC method, there are typically four major options, which differ mainly on how the expected value is determined. Unfortunately, the most preferable options are not always feasible due to the non-availability of appropriate reference standards or placebo samples. For each option, we will explain how to determine the expected value and the method value. Agreement between the two values will be addressed later.

3.1. Options for determining the expected value and the method value

3.1.1. Option 1 (using a certified reference standard)

The first option involves using a representative, certified reference standard. We say representative because the certified reference standard needs to have a chemical matrix that is the same as the matrix of a typical unknown sample. So, if the method is intended for API quantitation in an API substance, the reference standard could contain only the API. However, if the method is intended for finished product testing or for impurity quantitation in API substance, the reference standard should contain the appropriate amount of all the substances typically found in the finished product or API substance. The reference standard should be prepared by an ISO certified reference material manufacturer.

3.1.1.1. Determining the accepted value

If a representative certified reference standard is available, the accepted value is the certified amount of analyte (e.g. API) per given sample of the reference standard.

3.1.1.2. Determining the method value

To obtain the method value, at least nine reference standard solutions are prepared and tested as if they were unknown samples, using the candidate HPTLC method. The average and standard deviations of the results will represent the method value. The standard solutions should contain the following analyte concentrations:

  • The first three standard solutions should contain analyte concentrations between 50 and 90% of the analyte’s label claim (or quantitation limit for an impurity).

  • The next three standard solutions should contain analyte concentrations between 90 and 110% of the analyte’s label claim.

  • The last three standard solutions should contain analyte concentrations between 110 and 150% of the label.

  • Due to the unavailability of representative certified reference standards for most pharmaceutical products, option 1 is rarely used in method validation.

3.1.2. Option 2 (Using a representative blank)

The second option involves using a representative blank, which means a substance or mixture that contains all the chemical components of a typical unknown sample except the analyte. It is important to note that the chemical composition of a representative blank depends on both the analyte and the composition of a typical unknown. For assay, or content uniformity, the representative blank is a placebo. For API quantitation in the API substance, the representative blank is typically the solvent used to dissolve the standard. For impurity quantitation in an API substance, the representative blank is typically the API substance. For impurity quantitation in the finished product, the representative blank is a mixture of the placebo plus all the APIs plus all the typical impurities. During testing, the representative blank should be treated the same way as an unknown sample would be. Care must be taken so that only the absence of analyte distinguishes the representative blank from a typical unknown sample.

3.1.2.1. Determining the accepted value

If a representative blank is available, at least nine samples are prepared by spiking the blank with various amounts of analyte. The accepted value can be represented as the average and standard deviation of all the amounts of analyte spiked to the representative blank samples. At least nine difference samples should be tested covering a minimum of three different concentrations across the expected range of analyte concentration (50–150% of label claim or impurity limit).

3.1.2.2. Determining the method value

Once the spiked blank samples are prepared, they can be analyzed in parallel using the candidate HPTLC method. The average and standard deviation of the results (expressed in the same unit as the accepted value) can represent the method value.

3.1.3. Option 3 (Using a reference method)

If options 1 and 2 are not feasible, a reference method can be used to determine the accuracy of a candidate method. The reference method must be independent of the candidate method, have been well validated with a stated accuracy, and have the same intended use as the candidate method.

3.1.3.1. Determining the accepted value

To obtain the accepted value for option 3, the reference method can be used to test 6 or more unknown samples. The average and standard deviation of the results will represent the accepted value.

3.1.3.2. Determining the method value

To obtain the method value, each of the samples used to determine the accepted value is tested using the candidate method. The average and standard deviation of the results will represent the method value.

3.1.4. Option 4 (Using standard addition to unknown)

In lieu of option 3, method accuracy can be estimated using the standard addition method [10]. In this case the test sample is an unknown finished product or an API substance, whose analyte amount has been predetermined using the candidate method.

3.1.4.1. Determining the accepted value

To obtain the accepted value for option 4, at least 6 or more stock solutions of unknown samples should be prepared and tested per the candidate method. The average and standard deviation of the results will represent the accepted value.

3.1.4.2. Determining the method value

To obtain the method value, each of the stock solutions used to determine the accepted value is tested once again using the standard addition method [10]. So, each stock solution should have its own standard addition curve with 5 or more data points. The average and standard deviations of the absolute values of the x-intercepts will represent the method value.

3.2. Agreement between expected and method values

Several calculation methods are used to determine the agreement between the expected and method values. The percent recovery method is the simplest. It involves dividing the average method value by the average expected value and multiplying the result by 100. Although this method is considered acceptable in the ICH guidelines [3, 4], and it is found in many publications, it does not take the standard deviations into account.

One method that takes variation into account is the expanded uncertainty interval method [11]. It involves combining the expected and method uncertainties to obtain the expanded uncertainty, which is then compared to the difference between the average expected value and the average method value. If the expanded uncertainty is greater or equal to the difference, the candidate method is considered accurate.

A more statistically rigorous method to calculate accuracy is the t-test for two equal means [12]. It can be performed using MS Excel or other statistical software, but it requires an understanding of hypothesis testing. Interested readers can consult any general Statistics book for more details on t-test and hypothesis testing.

The concept of accuracy profile, which is different from the concept of accuracy described herein, is described by Shewiyo et al. It aims to describe method performance using a single statistic [13].

4. Linearity

Linearity evaluations demonstrate measurements from a test method are proportional to the amount of analyte within a particular concentration range [3, 4]. Responses from samples containing different amounts of analyte are obtained from the test method. Generally, a minimum of five different concentrations should be used where multiple (i.e., ≥3) responses are obtained at each analyte level. The method response (y-axis) is plotted as a function of the analyte concentration (x-axis) for subsequent analysis with linear regression techniques, where slope, intercept, and correlation coefficient are reported. The concentration range should cover the upper and lower levels anticipated during an analysis.

In the following example, a graphical representation of a linear calibration model is demonstrated, where the raw data is provided in Table 1 and Figure 1 shows the corresponding linear regression curve.

[Concentration], w/v[Response], Instrument reading (triplicate results)
20.06, 0.06, 0.06
40.12, 0.12, 0.11
60.17, 0.17, 0.16
80.22, 0.23, 0.22
100.28, 0.28, 0.28

Table 1.

Data to demonstrate a linear calibration model.

Figure 1.

Graphical representation of linear calibration model data.

However, a linear model may not be the best calibration fit for the data as is the case of the data listed in Table 2, and plotted in Figure 2. When the linear model is applied to the data, the resulting correlation coefficient (R2 = 0.98) is less than ideal.

VariableData
[Concentration], w/v12345678
[Response], Instrument reading1020304146556065

Table 2.

Data to demonstrate a non-linear calibration model.

Figure 2.

Graphical representation of a less than ideal linear calibration model.

Further examination of the data indicates that a polynomial fit can provide a better calibration model from the data (Figure 3). It should be noted that most pharmaceutical analysis methods commonly use a one-point standard during routine use of the method (after validation has been established).

Figure 3.

Graphical representation of polynomial approach to calibration model.

If the analysis range for the method only requires concentrations from 1 to 4 (w/v), a linear model for just that concentration range provides a r2 of 0.9994 (Figure 4) and would be easier to implement in future analysis (note-an additional standard should be added within that range during final validation).

Figure 4.

Graphical representation of the more linear range of the preceding polynomial calibration model.

5. Limits of detection/quantitation

Various options are possible for determining limits of detection (LOD) and limits of quantitation (LOQ) [3, 4, 14, 15]. The section below will provide some key example approaches for tests that generate instrument based responses.

The signal to noise ratio can be used to determine both the LOD and LOQ, where responses are obtained from blank and from an array of samples at lower concentrations. A ratio of signal (from analyte samples) to noise (from blank) of 3 is an accepted concentration level for the LOD. Likewise, a concentration level that provides a signal to noise of 10 can be used as the LOQ.

Another approach first involves the determination of the standard deviation of the response and the slope of calibration (linearity) curve. Although other options are possible [3, 4, 14, 15], the standard deviation of the response can be estimated from replicate injections from blank samples or from the standard deviation of y-intercepts from multiple regression lines. Multiplying the ratio of the standard deviation of the responses to the slope of the curve by 3.3 or 10 provides the LOD or LOQ, respectively.

6. Precision

For an analytical method, precision is an assessment of the consistency of results obtained with multiple measurements from the same sample [3, 4, 16]. There are three categories of precision for an analytical method, namely repeatability, intermediate precision, and reproducibility, which can be assessed through variations with different equipment, testing times (conducted on different days), analysts, and/or laboratories.

Repeatability is often evaluated with replicate measurements of a sample on the same day in the same laboratory, where the analyst and equipment are not changed. Intermediate precision can be evaluated from replicate measurements of a sample within the same laboratory, but with systematic variations with different analysts, times of analysis, and equipment (such as different instruments). Reproducibility is commonly determined from replicate measurements of the same sample but within different labs, which will inherently incorporate different analysts, equipment, and time of analysis.

ICH Q2 provides several recommendations for number of replicates and concentration levels for each of the three types of precision [4]. Recommended approaches for repeatability are at least nine measurements that span the method’s range (such as three replicates for each of three analyte levels) or at least six measurements at the target analyte level. ICH Q2 does not specify a minimum number of samples for intermediate precision and reproducibility but encourages that the effects of variables (i.e., analysts, days, instruments) be systematically evaluated. The following section will provide possible approaches for evaluating repeatability and intermediate precision, followed by references for examples for reproducibility will be provided.

To perform the appropriate precision assessments, the following equations are indicated [16] and will be used for further development of subsequent examples. The average (x¯) of n replicates is provided in Eq. (1),

x¯=i=1nxinE1

where xirepresents the individual replicates measurements. The standard deviation (s) of a data set can be determined through Eq. (2),

s=i=1nxix¯2n1E2

and the % relative standard deviation (%RSD) is provided in Eq. (3).

%RSD=100×sx¯E3

The %RSD is often used in method validation assessments because it normalizes the standard deviation to the average.

From Eqs. (1)–(3), an evaluation of repeatability can be determined. In the following example, assume that an analyst has performed six replicate analysis (within the same laboratory) from a method capable of quantifying the amount of active ingredient in a pharmaceutical product in units of % label claim (assay) and obtained the following results (102.1%, 100.5%, 98.2%, 99.1%, 101.8%, 99.8%). Using Eqs. (1)–(3), the average (x¯), standard deviation (s), and %RSD would be 100.25%, 1.52%, and 1.52%, respectively, where s (or more commonly %RSD), is a measure of method repeatability.

Intermediate precision involves an evaluation of variations “within runs” and “between runs” [17]. Consider the data in Table 3 containing replicate runs (n = 3, indexed with j) obtained on each of multiple days (p = 5, indexed with i), where each day used a different analyst with separate solution preparations but using the same method as above for the repeatability analysis.

n = 3(j)/p = 5(i)Day 1Day 2Day 3Day 4Day 5
Replicate 1102.298.799.3101.9102.1
Replicate 2100.3101.898.1100.1101.4
Replicate 399.9102.398.799.1101.2
x¯i100.8100.998.7100.4101.6
x¯¯100.5

Table 3.

Example data for intermediate precision determination.

With each entry in the data representing a separate xij, the repeatability or within-run standard deviation (days) can be determined using Eq. (4), where x¯iis provided in Eq. (5).

sr=i=1pj=1nxijx¯i2pn1E4
x¯i=j=1nxijnE5

The between-run standard deviation (days) can be calculated with Eq. (6), where x¯¯is provided in Eq. (7).

sB=i=1px¯ix¯¯2p1sr2nE6
x¯¯=i=1pj=1nxijpnE7

Subsequently, the intermediate precision standard deviation can be calculated with Eq. (8).

sIP=sr2sB2E8

From the data presented in Table 3 and using Eqs. (4)–(8) [17], the standard deviations for repeatability (within-run), between-run, and intermediate precision are calculated as 1.26, 0.80, and 1.49, respectively.

Evaluations for reproducibility utilize interlaboratory trials, and are commonly employed when a procedure requires further standardization for use among a more extended array of laboratories. ISO 5725 [18] provides the necessary approach and management structure needed to properly plan, conduct, and interpret the results of an interlaboratory trial that will involve multiple laboratories conducting replicate analysis of a sample(s) at a particular analyte level(s). Approaches are provided to graphically (Mandel’s statistics) and quantitatively (Cochran/Grubb) identify outliers so that the most accurate assessments of repeatability and reproducibility variance (standard deviations) are possible. The calculations involved in these types of trials are fairly extensive. Several examples are provided within ISO 5725, and Vander Heyden et al. provides a detailed example for an interlaboratory trial for an HPLC procedure [19].

Overall, desired levels for precision for pharmaceutical analysis are commonly on the order of ~2% RSD. However, different ranges can be necessary depending on the concentration level of the analyte (i.e., higher levels of %RSD can be allowed as the analyte concentration decreases) [17].

7. Robustness

Robustness is a measure of how much a method is impacted by deliberate (small) changes in method conditions [3, 4]. The following are a listing of the types of parameters that can be evaluated to assess method robustness; solution stability (to heat and or time), extraction conditions during sample preparation (time, temperature, mechanical shaking time, sonication time), type of filters used during final standard/sample preparation, minor adjustments in mobile phase composition, and adjustments in other chromatographic conditions (flow rate, different suppliers of columns, temperature). Commonly, robustness is evaluated during the development stages of the method.

An approach for evaluating robustness could be to compare an analysis using the primary method compared to an analysis where a certain parameter is adjusted. Depending on the method and sample type, adjustments in parameters that generate less than ~2% difference relative to the primary method can provide a reasonable measure of how sensitive the method is to various types of adjustments. During the development of the method for example, a study could be conducted to evaluate the sensitivity of the method on the type of filter by comparing the results from a sample solution that was centrifuged (without filtration) to those filtered with different filter types (PTFE, PVDF, nylon) from different manufacturers. Solution stability could be evaluated by comparing results from freshly prepared solutions compared to the same solutions stored at room temperature over several days.

There are a variety of approaches that can be incorporated to evaluate method robustness. Dejaegher and Vander Heyden provide an extensive review for a variety of approaches to systematically evaluate method robustness (ruggedness) [20].

8. Range

The range of the method corresponds to the lower and upper analyte concentration where satisfactory levels of linearity, precision, and accuracy have been achieved during the method validation process. The range is indicated in the same units as that of the results obtained from the method.

For analysis of pharmaceutical products [3, 4], the following ranges (in percentage relative to the target level) are often required for the respective types of tests; assay (80–120%), content uniformity (approximately 70–130%), impurities (approximately 50–120% of the acceptance limit), dissolution (± 20% of the required range).

9. Application with HPTLC techniques

High-performance thin layer chromatography (HPTLC), an extension of TLC, is a robust, simple, rapid, and efficient tool in quantitative and qualitative analysis of compounds [21], and a variety of applications can be found in the literature [22, 23, 24, 25, 26, 27, 28, 29]. In this section an overview of applications of HPTLC in typical pharmaceutical testing protocol is highlighted with examples. HPTLC is one of the sophisticated instrumental techniques based on the full capabilities of thin layer chromatography. The advantages of automation, scanning, full optimization, selective detection principle, minimum sample preparation, hyphenation, and so on enable it to be a powerful analytical tool for chromatographic information of complex mixtures of pharmaceuticals, natural products, clinical samples, and food stuffs [21]. HPTLC is one of the ideal TLC technique for the analytical purposes because of its increased accuracy, reproducibility, and ability to document the results, compared with standard TLC. Because of this, HPTLC technologies are also the most appropriate TLC technique for conformity with GMPs [30].

9.1. Identification test

In a pharmaceutical testing protocol, identification tests are intended to ensure the identity of an analyte in an API or finished pharmaceutical product sample. This is normally achieved by comparison of a chromatographic behavior of unknown sample to that of a reference standard. The identity of the test substance is confirmed if the migration distance of the test substance matches that of the reference substance. Thin layer chromatography experiments are among the key identity tests in most pharmacopeia monographs. Pharmacopeia standards are typically used by industry as a basis for meeting QC requirements and current good manufacturing practices (cGMPs). Many identification tests in the major pharmacopeia (e.g., USP, Ph. Int., and Ph. Eur. [1, 31, 32]) use planar chromatography (TLC), however HPTLC is a superior technology. Figure 5 below represent a typical densitogram obtained in the identification of sulfamethoxazole (SMX) and trimethoprim (TPM). In this example, the migration distances are 0.35 and 0.90 for TMP and SMX respectively.

Figure 5.

An example of overlaid densitogram for identification of sample 1 and a reference 2 containing sulfamethoxazole (SMX) and trimethoprim (TPM). Conditions Mobile Phase: (Methanol: Ethyl Acetate: Toluene 6: 9:15 v/v) Detection Wavelength: 275 nm and Application Volume: 5 μl and aluminum plates precoated with silica gel 60 F254 as the stationary phase.

9.2. Assay content determination

A second most important critical quality attribute for pharmaceutical products testing is assay or determination of content. The procedure intended to measure the analyte present in a sample. In this context, the assay represents a quantitative measurement of the major component(s) in the drug substance. This is done by comparing the area under the peak of test substance to that of reference standard material. For a drug product, containing paracetamol an overlaid densitogram is presented in Figure 6. Similar validation characteristics also apply when assaying for the active or other selected component(s). The same validation characteristics may also apply to assays associated with other analytical procedures (e.g., dissolution) [33].

Figure 6.

An example of overlaid densitograms for assay of paracetamol in sample tracks 2, 3, 5, 8 and 9, and a reference in tracks 1, 4 and 7. Conditions: Mobile Phase: Acetone; Methanol; Toluene: 6:6:16 v/v/v acidified with three drops of Glacial Acetic Acid, Detection Wavelength: 274 nm, Application Volume: 5 μl and aluminum plates precoated with silica gel 60 F254 as the stationary phase.

9.3. Impurities and related substances

The principal requirement is that an analytical method for assessing impurities should be a stability indicating and meeting specificity criteria described in Section 2 above. Stability indicating method (SIM) is defined as a validated analytical procedure that accurately and precisely measures the active ingredients (drug substance or drug product) free from process impurities, excipients and degradation products. This can be demonstrated by forced degradation study of the drug substance and subjecting the resultant solution to the chromatographic conditions [34].

Testing for impurities can be either a quantitative test or a limit test for the impurity in a sample. Either test is intended to accurately reflect the purity characteristics of the sample. Quantitative tests for impurities are meant to quantify the exact amount of impurity. This is done by comparing the response from a single or multi-level calibration curve. Whereas the limit test is an estimative test where the impurity is controlled not to exceed certain limit. In this case an impurity standard is prepared at the control level and compared to the response from the sample (which should not exceed this level).

With the improved resolution powers of HPTLC (enhanced by reduced particle sizes), it is possible to perform both tests by using HPTLC. In the literature, there are many stability indicating method for various drug substances for example pseudoephedrine and cetirizine in pharmaceutical formulations [35], clopidogrel bisulphate [36] timolol maleate [37] simultaneous determination of ezetimibe and simvastatin [38], piroxicam [39], and estradiol [40].

9.4. Dissolution testing

Dissolution testing is a performance characterizing test and a requirement for all solid oral dosage forms and is used in all phases of development for product release and stability testing [41, 42, 43]. It is a key analytical test used for detecting physical changes in an active pharmaceutical ingredient (API) and in the formulated product. It is a multi-unit test and multi-point sampling, making it very tedious. HPTLC offers a multi-channel capability where a total of 18–25 samples can be applied on one plate in form of bands and analyzed simultaneously. One lot of a product will require 6 units (tested in duplicate), plus calibrators (in triplicates at single or multiple levels). HPTLC methods have been successfully deployed for monitoring dissolution profile of diclofenac and acetaminophen [44], and the stability of rifampicin in dissolution medium in presence of isoniazid [45].

9.5. Content uniformity

The test for Content Uniformity (CU) is the assay of the individual content of drug substance(s) in a number of individual dosage units to determine whether the individual contents are within the set limits [46]. Multiple capsules or tablets are selected at random and each are analyzed to determine the active ingredient in each capsule or tablet. The performance efficiency of this method can benefit from the HPTLC multi-channel capabilities. HPTLC has been successful applied in content uniformity of atorvastatin calcium tablets [47], diazepam tablets [48], diosgenin and levodopa [49], nicorandil tablets [50] and rosiglitazone in tablets [51]. All of these HPTLC method examples provide a faster, more cost efficient approach to quantitative testing for routine analysis.

10. Conclusions

Classic method validations for pharmaceuticals involve techniques such as UV-VIS, TLC, and HPLC. This chapter highlights different examples with High Performance Thin Layer Chromatography (HPTLC). General approaches are provided for method validation, as applicable to pharmaceutical assessments, outlined for each of the key aspects (i.e., specificity, accuracy, linearity, limits of detection/quantitation, precision, robustness, and range). Although classical application of pharmaceutical method validation uses techniques such as UV-VIS or HPLC, important examples are provided using HPTLC techniques that provide high accuracy/precision with minimal use of reagents and other resources.

Acknowledgments

This work was funded through support from USAID Contract No. AID-OAA-C-15-00001, Global Health Supply Chain Quality Assurance Program (GHSC-QA).

© 2017 The Author(s). Licensee IntechOpen. This chapter is distributed under the terms of the Creative Commons Attribution 3.0 License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

How to cite and reference

Link to this chapter Copy to clipboard

Cite this chapter Copy to clipboard

David Jenkins, Cherif Diallo, Ed Bethea, Eliangiringa Kaale and Thomas Layloff (December 20th 2017). Method Validation Approaches for Pharmaceutical Assessments – Highlights with High Performance Thin Layer Chromatographic (HPTLC) Techniques, Calibration and Validation of Analytical Methods - A Sampling of Current Approaches, Mark T. Stauffer, IntechOpen, DOI: 10.5772/intechopen.71765. Available from:

chapter statistics

363total chapter downloads

More statistics for editors and authors

Login to your personal dashboard for more detailed statistics on your publications.

Access personal reporting

Related Content

This Book

Next chapter

Introductory Chapter: The Many Faces of Calibration and Validation in Analytical Methodology in the Present Day

By Mark T. Stauffer

Related Book

Applications of Molecular Spectroscopy to Current Research in the Chemical and Biological Sciences

Edited by Mark Stauffer

First chapter

Fourier Transform Infrared and Raman Characterization of Silica-Based Materials

By Larissa Brentano Capeletti and João Henrique Zimnoch

We are IntechOpen, the world's leading publisher of Open Access books. Built by scientists, for scientists. Our readership spans scientists, professors, researchers, librarians, and students, as well as business professionals. We share our knowledge and peer-reveiwed research papers with libraries, scientific and engineering societies, and also work with corporate R&D departments and government entities.

More about us