Quantitative quality control in microarray experiments and the application in data filtering, normalization and false positive rate prediction
Open Access
- 22 July 2003
- journal article
- research article
- Published by Oxford University Press (OUP) in Bioinformatics
- Vol. 19 (11) , 1341-1347
- https://doi.org/10.1093/bioinformatics/btg154
Abstract
Data preprocessing including proper normalization and adequate quality control before complex data mining is crucial for studies using the cDNA microarray technology. We have developed a simple procedure that integrates data filtering and normalization with quantitative quality control of microarray experiments. Previously we have shown that data variability in a microarray experiment can be very well captured by a quality score qcom that is defined for every spot, and the ratio distribution depends on qcom. Utilizing this knowledge, our data-filtering scheme allows the investigator to decide on the filtering stringency according to desired data variability, and our normalization procedure corrects the qcom-dependent dye biases in terms of both the location and the spread of the ratio distribution. In addition, we propose a statistical model for false positive rate determination based on the design and the quality of a microarray experiment. The model predicts that a lower limit of 0.5 for the replicate concordance rate is needed in order to be certain of true positives. Our work demonstrates the importance and advantages of having a quantitative quality control scheme for microarrays. Contact: xujing@mcw.eduKeywords
This publication has 0 references indexed in Scilit: