Proportions in the relevant data set, such as the recounted number of indicator’s data by the reported number at the next tier in the reporting system. A ratio less than 100% indicates “over-reporting”; a ratio over 100% suggests “under-reporting”
Characteristics of the methods for assessment of data use reported in the 10 publications included in the review.
Authors Year | Attributes Major measures | Study design | Data collection methods | Data analysis methods | Contribution | Limitations |
---|---|---|---|---|---|---|
Freestone 2012 [ ] | Trends in use Actioned requests from researchers in a set period of time | Analysis of actioned requests from researchers in a period of time | Abstracted data from the database for the study period | Trend analysis of proportion of requests | Quantifiable measures | Limit attributes |
Hahn 2013 [ ] | Use of data The usage of aggregated data for monitoring, information processing, finance and accounting, and long-term business decisions | Qualitative methods: structured interviews with purposive sample of 44 staff and in-depth interviews with 15 key informants | Structured survey and key informant interview to assess five structured statements. Five-point scales were used for each statement | Responses were processed manually, classified and grouped by facility and staff class | Identified indicators of use of data | Lack of quantifiable results for assessment of data use |
Iguiñiz-Romero and Palomino 2012 [ ] | Data use Data dissemination: identify whether data used for decision making, the availability of feedback mechanisms | Qualitative exploratory study including interview and review of documentations | Open-ended, semi-structured questionnaire interviews with 15 key decision-makers. Review national documents and academic publications | Interview data recorded, transcribed, organized thematically and chronologically. The respondents were identified by positions but not named | Most respondents held key positions and a long period of the reviewed publications | Purposive sample lack of representativeness |
Matheson 2012 [ ] | Clinical use of data: the number of summaries produced. Use of data for local activities to improve care. Data entry: the number of active sites. Report use: the percentage of active sites using prebuilt queries to produce data for each type of report in a given month over time | Qualitative and quantitative methods: key informant interview, documentation review, database query. | Personal interviews by phone and through internet telephony; follow up in person or by email; running SQL queries against the central database. External events were identified by reviewing news reports and through personal knowledge of the authors | Descriptive statistics using charts on number of clinics using the system in a given month, percentage of active clinics | Multiple methods | Lack of verification of data source |
ME PRISM 2010 [ ] | Checklist of use of information Report production, display of information, discussion and decisions about use of information, promotion and use of information at each level | Quantitative method to complete a predesigned checklist diagnostic tool | Checklist and non-anonymous interviewing staff, asking, manual counting, observation and recording results or circling “yes or no” | Two Likert score and descriptive statistics | Quantitative terms help set control limits and targets and monitor over time | |
Petter and Fruhling 2011 [ ] | System use, intention to use, user satisfaction | Quantitative methods to use DeLone & McLean IS success model. Survey respondents with a response rate of 42.7% and with representative demographics | Use an online survey in structured questionnaire with 7 Likert scale for all quantitative questions, in addition to facsimile and mail | Summative score for each construct, and each hypothesis was tested using simple regression, in addition to mean, standard deviation, the Spearman’s correlation coefficients | Use is dictated by factors outside of the control of the user, and it is not a reasonable measure of IS success. The quality does not affect the depth of use | Lack of objective assessments |
Qazi and Al 2011 [ ] | Use of data Non-use, misuse, disuse of data | Descriptive qualitative interviews | In-depth, face to face and semi structured interviews with an interview guide, 26 managers (all men, ages ranging from 26 to 49 years; selected from federal level (2), provincial (4) and seven selected districts (20) from all four provinces) | Data transcription, analysis based on categorization of verbatim notes into themes and a general description of the experience that emerged out of statements | A qualitative study allows getting close to the people and situations being studied, identified a number of hurdles to use of data | Convenience sample only one type of stakeholders has been covered. |
Saeed 2013 | Usefulness of the system Data linked to action, feedback at lower level, data used for planning, detect outbreaks, data used for the development and conduct of studies | Quantitative and qualitative methods, including interview, consultation, and documentation review | 10 key informants interview; consultation with stakeholders, document review of each system | Predefined scoring criteria for attributes: poor, average, or good | Mixed methods | Purposive sampling |
WHO HMN 2008 [ ] | Information dissemination and use, demand and analysis, policy and advocacy, planning and priority-setting, resource allocation, implementation and action | Mixed methods: quantitative and qualitative. Use 10 out of 197 questions among stakeholders at national and subnational levels | Use group discussions (100 major stakeholders), self-assessment approach, individual (less than 14) or group scoring to yield a percentage rating for each category | An overall score for each question, quartiles for the overall report | Expert panel discussion, operational indicators with quality assessment criteria | Lack of field verification of data use |
Wilkinson and McCarthy | Extent of data recognition and use, strategies and routines, specific uses, dissemination | Quantitative and qualitative methods to use standardized semi-structured questionnaire telephone interviews of key informants from the management teams of the system | Telephone structured questionnaire interviews of 68 key informants from the 29 out of 34 management teams of the networks. Response options for most of the questionnaire items were yes/no or five or seven point Likert and semantic differential response scales | Quantitative and qualitative analysis of survey results. Qualitative data transcribed, ordered by question number, and common themes, then content analyzed to indicate frequencies and percentages. Correlational analyses used Pearson’s r for parametric data and Spearman’s Rho for non-parametric data | Quantification of qualitative data | Statistical analysis is limited by the size of the sample as there were only 29 networks and 68 individual participants, statistical power to detect an effect is weak, and general trends are mainly reported. |
Characteristics of the methods for assessment of data collection process reported in the 16 publications included in the review.
Authors Year | Attributes Major measures | Study design | Data collection methods | Data analysis methods | Contribution | Limitations |
---|---|---|---|---|---|---|
Ancker 2011 [ ] | Group discussion about root causes of poor data quality and strategies for solving the problems | Qualitative method by focus group discussion | Held a series of weekly team meetings over about 4 months with key informants involved in the data collection | Theme grouping to each data quality issue | Initiated by and related to identified poor data quality issues | Implicitly focused. Only analyzed causes not assessed the magnitude |
Bosch-Capblanch 2009 [ ] | Quality scores Recording and reporting of data, keeping of vaccine ledgers and information system design | Quantitative method by user’s survey based on WHO DQA. A multistage weighted representative sampling procedure | Questionnaire based on a series of 19 questions and observations undertaken at each level (national, district and health units) | Each question 1 point. Average score, summary score, medians, inter-quartile ranges, confidence intervals, P value, bubble scatter chart, Rho value | Combined with data quality | Implicitly focused, the number of questions surveyed was less than that of the WHO DQA |
CIHI 2009 [ ] | Metadata documentation Data holding description, methodology, data collection and capture, data processing, data analysis and dissemination, data storage, and documentation. | Quantitative method by surveying users | Questionnaire | Undefined | 7 categories, with subcategories and definition and/or example | Implicitly focused |
Corriols 2008 [ ] | Identification of underreporting reasons by reviewing information flow chart and non-reporting in physicians | Qualitative method to review documentations | Review the national reports on the system related to deficiency in the information flow chart and non-reporting in physicians | Undefined | Initiated by identified data quality issues | Implicitly focused |
Dai 2011 [ ] | Data collection, data quality management, statistical analysis and data dissemination | Qualitative method, review documentations | Document review | Theme grouping | Desk review | Implicitly focused |
Forster 2008 | Routine data collection, training and data quality control | Quantitative method by online survey | Questionnaire | Descriptive statistics. | Examine associations between site characteristics and data quality | Implicitly focused. Convenience sample |
Freestone 2012 [ ] | Data collection and recording processes | Qualitative method to review current processes about identification, code, geocode of address or location data. Staff consulted to establish and observe coder activities and entry processes | Review the processes; consultation with staff; observation of coder activities and entry processes to identify any potential cause of errors which then grouped thematically | Thematically grouping data | Identify each of the key elements of the geocoding process are factors that impact on geocoding quality | Differences in software and system settings need to be aware of. |
Hahn 2013 [ ] | Data flow The generation and transmission of health information | Qualitative method to use workplace walkthroughs on 5 subsequent working days at each site | Informal observations of the generation and transmission of health information of all kinds for the selection of data flows | Undefined | Observation of walkthroughs | Undefined indicators |
Iguiñiz-Romero and Palomino 2012 [ ] | Data flow or data collection process: data collectors, frequencies, data flow, data processing and sharing, | Qualitative exploratory study including interview and review documentations | Open-ended, semi-structured questionnaire interviews with 15 key decision-makers. Review national documents and academic publications | Data recorded, transcribed, organized thematically and chronologically | Most respondents held key positions and a long period of reviewed publications | Purposive sample |
Lin 2012 [ ] | Data collection and reporting | Qualitative methods based on CDC’s Guidelines, | Review guidelines and protocols using a detailed checklist; direct observation; focus group discussions and semi-structured interviews | Theme grouping | Field visits or observations of data collection to identify impact on the data quality | Undefined indicators |
ME DQA 2008 [ ] | Five functional areas: M&E structures, functions and capabilities, indicator definitions and reporting guidelines, data collection and reporting forms and tools, data management processes, and links with national reporting system | Quantitative and qualitative methods by 13 system assessment summary questions based on 39 questions from five functional areas. Score the system combined with a comprehensive audit of data quality | Off-site desk review of documentation provided by the program/project; on-site follow-up assessments at each level of the IS, including observation, interviews, and consultations with key informants | Using summary statistics based on judgment of the audit team. Three-point Likert scale to each response. Average scores for per site between 0 and 3 continuous scale | DQA protocol and system assessment protocol | Implicitly focused. The scores should be interpreted within the context of the interviews, documentation reviews, data verifications and observations made during the assessment. |
ME PRISM 2010 [ ] | ProcessesData collection, transmission, processing, analysis, display, quality checking, feedback | Quantitative method by questionnaire survey including data transmission, quality check, processing and analysis and assessing the respondent’s perceptions about the use of registers, data collection forms and information technology | Non-anonymous interviewing staff with identified name and title, including asking, observation and circling “yes or no” | Using a data entry and analysis tool (DEAT), described in quantitative terms rather than qualitative. Yes or No tick checklist | A diagnostic tool. Quantitative terms help set control limits and targets and monitor over time | Indicators are not all inclusive; tool should be adapted and pre-test and make adjustments |
Ronveaux 2005 [ ] | Quality index (QI) Recording practices, storing/reporting practices, monitoring and evaluation, denominators used at district and national levels, and system design at national level | Quantitative and qualitative methods by external on-site evaluation after a multi-stage sampling based on WHO DQA. | Questionnaires and observations. Survey at national level (53 questions), district level (38 questions) and health-unit level (31 questions). Observations to workers at the health-unit level. They were asked to complete 20 hypothetical practices. | Descriptive statistics (aggregated scores, mean scores): 1 point each question or task observed. Correlational analyses by zero-order Pearson correlation coefficients | Implicitly focused. The chosen sample size and the precision of the results were dictated by logistical and financial considerations | |
Venkatarao 2012 [ ] | Accuracy of case detection, data recording, data compilation, data transmission | Quantitative method by using a 4-stage sampling method to conduct field survey (questionnaire) during May-June 2005 among 178 subjects | Questionnaires of 2 study instruments: the first focused on the components of disease surveillance; the second assessed the ability of the study subject in identifying cases through a syndromic approach | Descriptive statistics analysis | Assessment from user’s viewpoint. | Implicitly focused. Lack of field verification of data collection process |
WHO DQA 2003 [ ] | Quality questions checklist, quality index Five components: recording practices, storing/reporting practices, monitoring and evaluation, denominators, system design (the receipt, processing, storage and tabulation of the reported data) | Quantitative and qualitative method using questionnaire checklists for each level (three levels: national, district, health unit level) of the system including 45, 38, 31 questions respectively | Questionnaires and discussions. Observations by walking around the health unit for field observation to validate the reported values | Percentage of the items answered yes. The target is 100% for each component | Describe the quality of data collection and transmission | Implicitly focused. The chosen sample size was dictated by logistical and financial considerations |
WHO HMN 2008 [ ] | Data management or metadata A written set of procedures for data management including data collection, storage, cleaning, quality control, analysis and presentation for users, an integrated data warehouse, a metadata dictionary, unique identifier codes available | Mixed methods: quantitative and qualitative. Use 5 out of 197 questions, at various national and subnational levels | Use group discussions around 100 major stakeholders, self-assessment approach, individual (less than 14) or group scoring to yield a percentage rating for each category | An overall score for each question, quartiles for the overall report | Expert panel discussion, operational indicators with quality assessment criteria | Lack of field verification of data collection process |
PY conceptualized the study. HC developed the conceptual framework with the guidance of PY, and carried out the design of the study with all co-authors. HC collected the data, performed the data analysis and appraised all included papers as part of her PhD studies. PY reviewed the papers included and the data extracted. PY, DH and NW discussed the study; all participated in the synthesis processes. HC drafted the first manuscript. All authors made intellectual input through critical revision to the manuscript. All authors read and approved the final manuscript.
The authors declare no conflict of interest.
DAS Slides: Data Quality Best Practices from DATAVERSITY To view just the On Demand recording of this presentation, click HERE>> About the Webinar Tackling data quality problems requires more than a series of tactical, one off improvement projects. By their nature, many data quality problems extend across and often beyond an organization. Addressing these issues […]
Tackling data quality problems requires more than a series of tactical, one off improvement projects. By their nature, many data quality problems extend across and often beyond an organization. Addressing these issues requires a holistic architectural approach combining people, process and technology. Join Nigel Turner and Donna Burbank as they provide practical ways to control data quality issues in your organization.
Donna Burbank
Managing Director, Global Data Strategy, Ltd
Donna Burbank is a recognized industry expert in information management with over 20 years of experience helping organizations enrich their business opportunities through data and information. She currently is the Managing Director of Global Data Strategy Ltd , where she assists organizations around the globe in driving value from their data. She has worked with dozens of Fortune 500 companies worldwide in the Americas, Europe, Asia, and Africa and speaks regularly at industry conferences. She has co-authored several books on data management and is a regular contributor to industry publications. She can be reached at [email protected] and you can follow her on Twitter @donnaburbank.
Nigel Turner
Principal Information Management Consultant/EMEA, Global Data Strategy, Ltd
Nigel Turner has over 20 years of experience in Information Management (IM) with specialization in Information Strategy, Data Quality, Data Governance, and Master Data Management. He has created and led large IM & CRM consultancy & delivery practices in multiple consulting organizations including British Telecommunications Group (BT), IPL, and FHO. Nigel also has experience in the data quality tools space as Vice President of Information Management Strategy at Harte Hanks Trillium Software, a leading global provider of Data Quality & Data Governance tools and consultancy where he engaged with over 150 customer organizations from all parts of the globe. Nigel is a well-known thought leader in Information Management and has presented at many international conferences in addition to writing numerous white papers and blogs on Information Management topics. Nigel provides education across the IM community, having lectured at Cardiff University on Data Governance and as an active member of DAMA International’s mentoring program, which he was instrumental in founding. He can be reached at [email protected] .
You must be logged in to post a comment.
JavaScript seems to be disabled in your browser. For the best experience on our site, be sure to turn on Javascript in your browser.
You must be logged in to download this file*
item details (16 Editable Slides)
(16 Editable Slides)
If you wish to deliver a compelling presentation on how Data Quality impacts organizational efficiency in today's data-driven landscape, our presentation template is the perfect solution. Lay your hands on this deck now and present important aspects related to this topic in a meaningful and impactful manner. Available for MS PowerPoint and Google Slides!
Quality experts and data scientists will find this deck helpful in showcasing the characteristics and elements of data quality. You can depict the key areas of data quality management. You can also demonstrate the stepwise process of improving the quality of data. Visualize the key factors for data quality and how they contribute to effective data management. Besides this, you can illustrate the process of data quality assessment.
Size | XS | S | S | M | M | L |
---|---|---|---|---|---|---|
EU | 32 | 34 | 36 | 38 | 40 | 42 |
UK | 4 | 6 | 8 | 10 | 12 | 14 |
US | 0 | 2 | 4 | 6 | 8 | 10 |
Bust | 79.5cm / 31" | 82cm / 32" | 84.5cm / 33" | 89.5cm / 35" | 94.5cm / 37" | 99.5cm / 39" |
Waist | 61.5cm / 24" | 64cm / 25" | 66.5cm / 26" | 71.5cm / 28" | 76.5cm / 30" | 81.5cm / 32" |
Hip | 86.5cm / 34" | 89cm / 35" | 91.5cm / 36" | 96.5cm / 38" | 101.5cm / 40" | 106.5cm / 42" |
Size | XS | S | M | L | XL | XXL |
---|---|---|---|---|---|---|
UK/US | 34 | 36 | 38 | 40 | 42 | 44 |
Neck | 37cm / 14.5" | 38cm /15" | 39.5cm / 15.5" | 41cm / 16" | 42cm / 16.5" | 43cm / 17" |
Chest | 86.5cm / 34" | 91.5cm / 36" | 96.5cm / 38" | 101.5cm / 40" | 106.5cm / 42" | 111.5cm / 44" |
Waist | 71.5cm / 28" | 76.5cm / 30" | 81.5cm / 32" | 86.5cm / 34" | 91.5cm / 36" | 96.5cm / 38" |
Seat | 90cm / 35.4" | 95cm / 37.4" | 100cm / 39.4" | 105cm / 41.3" | 110cm / 43.3" | 115cm / 45.3" |
Jul 27, 2014
520 likes | 866 Views
Assessment of data quality. Mirza Muhammad Waqar Contact: [email protected] +92-21-34650765-79 EXT:2257. RG610. Course: Introduction to RS & DIP. Contents. Hard vs Soft Classification Supervised Classification Training Stage Field Truthing
Assessment of data quality Mirza Muhammad Waqar Contact: [email protected] +92-21-34650765-79 EXT:2257 RG610 Course: Introduction to RS & DIP
Contents • Hard vs Soft Classification • Supervised Classification • Training Stage • Field Truthing • Inter class vs Intra Class Variability • Classification Stage • Minimum Distance to Mean Classifier • Parallelepiped Classifier • Maximum Likelihood Classifier • Output Stage • Supervised vs Unsupervised Classification
Positional and Attribute Accuracies • Positional and attributeaccuracies are the most critical factors in determining the quality of geographic data. • Can be quantified by sample data (a portion of whole data set) against reference data. • The concepts and methods of spatial data quality are applicable to both raster and vector data.
Evaluation of Positional Accuracy • Made up of two elements: • Planimetric accuracy, and • This is done by comparing the coordinates (x and y) of sample points on maps to the coordinates (x and y) of corresponding reference points. • Height accuracy • Involves comparison of elevation values of sample and reference data points.
Reference Data • To be used as a sample point, the point must be well defined, which means that it can be unambiguously identified both on the map and on the ground. • Survey monuments • Bench marks • Road intersections • Corner of building • Lampposts • Fire hydrants etc.
Reference Data • It is important for both the sample and reference data to be in the same map projection and based on the same datum. • The Accuracy Standards for Large-scale Maps however, specifies that: • A minimum of 20 check points must be established throughout the area covered by the map. • These sample points should be spatially distributed in such a way that at least 20% of the points be located in each quadrant of the map. • with individual points spaced at intervals equal to at least 10% of the diagonal of the map sheet.
Standard to take sample Points
Root Mean Square Error The discrepancies between the coordinate values of the sample points and their corresponding reference coordinate values are used to compute the overall accuracy of the map as represented by the root mean-square error (RMSE) The RMSE is defined as the square root of the average of the squared discrepancies. The RMSE for discrepancies in the X coordinate direction (rmsx) Y coordinate direction (rmsy) and elevation (rmst ) are computed from:
RMS for discrepancies • Where • dx = discrepancies in X coordinate direction = Xreference – Xsample
RMS for discrepancies • dy = discrepancies in Y coordinate direction = Yreference – Ysample • e = discrepancies in elevation = E reference – E sample • n = total number of points checked (sampled)
RMS for discrepancies • From rmsx and rmsy, a single RMSE of planimetry (rmsp) can be computed as follows.
RMS as Overall Accuracy The RMSEs of planimetry and elevation have now been generally accepted as the overall accuracy of the map. RMSE is used as the index to check against specific standards to determine the fitness for use of the map. The major drawback of the RMSE is that it provides information of only the overall accuracy. It does not give any indication of the spatial variation of the errors.
RMS as Overall Accuracy • For users who require such information, a map showing the positional discrepancies at the sample points can be generated. • Separate maps can be generated for discrepancies in easting and northing. • Alternatively a map showing the vectors of discrepancies at each point can be plotted
Evaluation of Attribute Accuracy • Attribute accuracy is obtained by comparing values of sample spatial data units with reference data obtained either by field checks or from sources of data with a higher degree of accuracy. • These sample spatial units can be raster cells; raster image pixels; or sample points, lines, and polygons.
Error Matrix • An error matrix is constructed to show the frequency of discrepancies between encoded values (i.e., data values on a map or in a database) and their corresponding actual or reference values for a sample of locations. • The error matrix has been widely used as a method for assessing classification accuracy of remotely sensed images
Error/Confusion Matrix • An error matrix, also known as classification error matrix or confusion matrix, is a square array of values, which cross-tabulates the number of sample spatial data units assigned to a particular category relative to the actual category as verified by the reference data.
Error Matrix • Conventionally, the rows of the error matrix represent the categories of the classification of the database, while the columns indicate the classification of the reference data. • In the error matrix, the element ij represents the frequency of spatial data units assigned to category i that actually belong to category j.
An Error Matrix A = Exposed soil B = Cropland C = Range D = Sparse woodland E = Forest F = water body
Error Matrix • The numbers along the diagonal of the error matrix (i.e. when i = j) indicate the frequencies of correctly classified spatial data units in each category; and the off-diagonal numbers (when I j) represent the frequencies of misclassification in the various categories.
Error Matrix • The error matrix is an effective way to describe attribute accuracy of geographic data. • If in a particular error matrix, all the nonzero entries lie on the diagonal. it indicates that no misclassification at the sample locations has occurred and an overall accuracy of 100% is obtained.
Commission or Omission • When misclassification occurs, it can be identified either as an error of commission or an error of omission. • Any misclassification is simultaneously an error of commission and an error of omission.
Error of Commission and Omission • Errors of commission, also known as errors of inclusion, are defined as wrongful inclusion of a sample location in a particular category due to misclassification. • When this happens, it means that the same sample location is omitted from another category in the reference data, which is an error of omission.
Commission vs Omission • Errors of commission are identified by off-diagonal values across the rows. • Errors of omission. also known as errors of exclusion, are identified by those off-diagonal values down the columns.
An Error Matrix Error of Commission A = Exposed soil B = Cropland C = Range D = Sparse woodland E = Forest F = water body
An Error Matrix Error of Omission A = Exposed soil B = Cropland C = Range D = Sparse woodland E = Forest F = water body
Indices to check Accuracy • In addition to the interpretation of errors of commission and omission, the error matrix may also be used to compute a series of descriptive indices to quantify the attribute accuracy of the data. • These include: • Overall Accuracy • Producer's Accuracy • User's Accuracy
Overall Accuracy • The PCC (Percent Correctly Classified) index represents the overall accuracy of the data. • In the case of simple random sampling, the PCC is defined as the trace of the error matrix (i.e., the sum of the diagonal values) divided by n, the total number of sample locations.
Overall Accuracy • PCC = (Sd / n) * 100% • Where • Sd = sum of values along diagonal • n = total number of sample locations
PCC – Overall Accuracy A = Exposed soil B = Cropland C = Range D = Sparse woodland E = Forest F = water body PCC = (1+5+5+4+4+1) x 100/35 PCC = 20 x 100 / 35 = 57.1%
Overall Accuracy • The maximum value of the PCC index is 100 when there is perfect agreement between the database and the reference data. The minimum value is 0, which indicates no agreement.
Deficiencies in PCC index • In the first place, since the sample points are randomly selected, the index is sensitive to the structure of the error matrix. This means that if one category of data dominates the sample (this occurs when the category covers a much larger area than others), the PCC index can be quite high even if the other classes are poorly classified.
Deficiencies in PCC index • Second, the computation of the PCC index does not take into account the chance agreements that might occur between sample and reference data. The index therefore always tends to overestimate the accuracy of the data. • Third, the PCC index does not differentiate between errors of omission and commission. Indices of these two types of errors are provided by the producer's accuracy and the user's accuracy.
Producer’s Accuracy • This is the probability of a sample spatial data unit being correctly classified and is a measure of the error of omission for the particular category to which the sample data belong. • The producer's accuracy is so-called because it indicates how accurate the classification is at the time when the data are produced.
Producer’s Accuracy • Producer’s accuracy is computed by: • Producer’s accuracy = (Ci/Ct) * 100 • Where • Ci = Correctly classified sample locations in column • Ct = Total number of sample locations in column • Error of omission = 100 – producer’s accuracy
User’s Accuracy • This is the probability that a spatial data unit classified on the map or image actually represents that particular category on the ground. • This index of attribute accuracy, which is actually a measure of the error of commission, is of more interest to the user than the producer of the data.
User’s Accuracy • User’s accuracy is computed by: • User’s accuracy = (Ri/Rt) * 100 • where • Rj = correctly classified sample locations in row • Rt = total number of sample locations in row • error of commission = 100 – user's accuracy
An Error Matrix PCC = (1+5+5+4+4+1) x 100/35 = 57.1% Producer’s accuracy: A = 1/1 = 100% D = 4/7 = 57.1% B = 5/10 = 50% E = 4/7 = 57.1% C = 5/9 = 55.6% F = 1/1 = 100% User’s Accuracy: A = 1/3 = 33.3% D = 4/8 = 50% B = 5/10 = 50% E = 4/4 = 100% C = 5/9 = 55.6% F = 1/1 = 100% A = Exposed soil B = Cropland C = Range D = Sparse woodland E = Forest F = water body
Kappa Coefficient (k) • Another useful analytical technique is the computation of the kappa coefficient or Kappa Index of Agreement (KIA) • It is capable of controlling the tendency of the PCC index to overestimate by incorporating all the off-diagonal values in its computation • The use of the off-diagonal values in the computation of the kappa coefficients also makes them useful for testing the statistical significance of the differences in different error matrices
Kappa Coefficient (k) • The coefficient (K), first developed by Cohen (1960) for nominal scale data • K = Po – Pc / 1 – Pc • Po is the proportion of agreement between the reference and sample data (PCC) • Kappa coefficient varies from a minimum of 1 to a maximum of 0.
Tau Coefficient • Kappa coefficient tends to overestimate the agreement between data sets. • Foody (1992) described a modified kappa coefficient based on equal probability of group membership that resembles and is derived more properly from the tau coefficient.
Tau Coefficient • = Po – Pr / 1 – Pr • It was demonstrated that the tau coefficient, which is based on the a priori probabilities of group membership, provides an intuitive and relatively more precise quantitative measure of classification accuracy than the kappa coefficient, which is based on the a posteriori probabilities
Questions & Discussion
Quality of life Assessment introduction. Jan J. v. Busschbach, Ph.D. www.xs4all.nl/~jannetvb/busschbach Department of Medical Psychology and Psychotherapy, Erasmus MC [email protected] +31 10 4087807 Psychotherapeutic Centrum ‘De Viersprong’, Halsteren
626 views • 30 slides
Background. In late 2008 the NQC commissioned Learning Australia to undertake a scoping study to: review what constitutes and contributes to quality assessment and how perceptions or understandings of this may have changed in the last five years from the point of view of key stakeholders, including
393 views • 19 slides
Assessment of Interchangeable Multisource Medicines Quality of BE Data . Dr. Henrike Potthast ([email protected]). Training workshop: Assessment of Interchangeable Multisource Medicines, Kenya, August 2009. Quality of Bioequivalence Studies. Citation from Directive 2001/20/EC:
490 views • 32 slides
The Data Quality Assessment Framework. OECD Meeting of National Accounts Experts October 2001. Purpose of this Presentation. To describe: The IMF’s Data Quality Assessment Framework (DQAF), and
685 views • 41 slides
Data Quality Assessment and Measurement. Laura Sebastian-Coleman, Ph.D., IQCP Optum Data Management EDW April 2014 – AM5 April 28, 8:30 – 11:45. Agenda. Welcome and Thank You for attending! Agenda Introductory materials Abstract Information about Optum and about me
1.12k views • 84 slides
Assessment of data quality. Dr Venanzio Vella , Epidemiologist Key Expert CEEN. Overview of the analysis. Individual data were available from all the regions of the RS , BD, 6 cantons in the Federation (4 provided weekly data) for 2008-11;
543 views • 38 slides
Ontologias e Web Semântica Profs.: Fred Freitas e Bernadette Farias Lóscio. Quality Assessment Methodologies for Linked Open Data. Aluno: Walter Travassos Sarinho [email protected]. Summary. Introduction What is Data Quality? The Data Quality Problem Quality Dimensions
733 views • 59 slides
Quality of Assessment Practices . Industry perceptions of VET assessment ‘. Source: NQC: Investigation into industry expectations of vocational education and training assessment, 2009 High level of satisfaction with assessment in organisation with a VET qualification – relied on:
326 views • 14 slides
Bioinformatics Unit. OVERVIEW
120 views • 1 slides
Quality Assessment. Clinical Research Compliance Office (CRCO). QA activities include:. Auditing Not-for-cause For-cause Federal Incident Reporting Assistance to Investigators who hold an Investigational New Drug (IND) or Investigational Device Exemption (IDE). What is auditing? .
488 views • 17 slides
Quality Assessment. What is Quality Control?. Quality Control in the clinical laboratory is a system designed to increase the probability that each result reported by the laboratory is valid and can be used with confidence by the physician making a diagnostic or therapeutic decision.
537 views • 18 slides
SQU College of Education Instructional & Learning Technologies Department TECH4211. Models of Quality Assessment. Asma Al Yahyaei Ateka Al Saqri Sara Al Hattali Maytha Al Omari. Outline:. Baldrige criteria ISO 900-2000. There are five popular models of quality assurance :.
411 views • 23 slides
Models of Quality Assessment. Done By: Asila AL- harthi Fatma AL- shehhi Fakhriya AL- Omieri Safaa AL- Mahroqi. 3. Capability Maturity Model. I s based on the concept of “ key Process Areas” that achieve a set of goals important for enhancing process capability.
293 views • 18 slides
Tsunami Recovery Program Data Quality Assessment. American Red Cross Tsunami Recovery Program Regional Technical Team, Bangkok 15 September 2009. Overview.
312 views • 20 slides
Quality Assessment. Lecture 2. Laboratory Analysis. The goal of laboratory analysis is to provide the reliable laboratory data to the health-care provider in order to assist in clinical decision-making.
805 views • 42 slides
The Data Quality Assessment Manager is a Data Quality product specifically designed to manage data quality assessments, manage data quality scores, review and correct quality issues and manage the workflow across all stakeholders involved in a data quality assessment. DQAM is the industry’s first platform designed to put data quality in the hands of data stewards and business owners who know and understand the data the best.
147 views • 2 slides
Assessment of quality of standards. Jolanta Szutkowska Methodology, Standards and Registers Division GUS. Content. The purpose of quality assessment of standards Intrinsic nature of standard and its links with quality References with quality components ESS and other aspects of quality
379 views • 37 slides
721 views • 41 slides
Newly Launched - AI Presentation Maker
AI PPT Maker
Powerpoint Templates
Icon Bundle
Kpi Dashboard
Professional
Business Plans
Swot Analysis
Gantt Chart
Business Proposal
Marketing Plan
Project Management
Business Case
Business Model
Cyber Security
Business PPT
Digital Marketing
Digital Transformation
Human Resources
Product Management
Artificial Intelligence
Company Profile
Acknowledgement PPT
PPT Presentation
Reports Brochures
One Page Pitch
Interview PPT
All Categories
Category // require(['jquery'], function ($) { $(document).ready(function () { //removes paginator if items are less than selected items per page var paginator = $("#limiter :selected").text(); var itemsperpage = parseint(paginator); var itemscount = $(".products.list.items.product-items.sli_container").children().length; if (itemscount.
Introduction HIV is a major global public health issue. The risk of sexual transmission of HIV in serodiscordant couples when the partner living with HIV maintains a suppressed viral load of <200 copies of HIV copies/mL has been found in systematic reviews to be negligible. A recent systematic review reported a similar risk of transmission for viral load<1000 copies/mL, but quantitative transmission risk estimates were not provided. Precise estimates of the risk of sexual transmission at sustained viral load levels between 200 copies/mL and 1000 copies/mL remain a significant gap in the literature.
Methods and analysis A systematic search of various electronic databases for the articles written in English or French will be conducted from January 2000 to October 2023, including MEDLINE, Embase, the Cochrane Central Register of Controlled Trials via Ovid and Scopus. The first step of a two-step meta-analysis will consist of a systematic review along with a meta-analysis, and the second step will use individual participant data for meta-analysis. Our primary outcome is the risk of sexual HIV transmission in serodiscordant couples where the partner living with HIV is on antiretroviral therapy. Our secondary outcome is the dose-response association between different levels of viral load and the risk of sexual HIV transmission. We will ascertain the risk of bias using the Risk Of Bias in Non-randomised Studies of Interventions (ROBINS-I) and Quality in Prognostic Studies (QUIPS), the risk of publication bias using forest plots and Egger’s test and heterogeneity using I 2 . A random effects model will estimate the pooled incidence of sexual HIV transmission, and multivariate logistic regression will be used to assess the viral load dose-response relationships. The Grading of Recommendations, Assessment, Development and Evaluation system will determine the certainty of evidence.
Ethics and dissemination The meta-analysis will be conducted using deidentified data. No human subjects will be involved in the research. Findings will be disseminated through peer-reviewed publications, presentations and conferences.
PROSPERO registration number CRD42023476946.
This is an open access article distributed in accordance with the Creative Commons Attribution Non Commercial (CC BY-NC 4.0) license, which permits others to distribute, remix, adapt, build upon this work non-commercially, and license their derivative works on different terms, provided the original work is properly cited, appropriate credit is given, any changes made indicated, and the use is non-commercial. See: http://creativecommons.org/licenses/by-nc/4.0/ .
https://doi.org/10.1136/bmjopen-2023-082254
Request permissions.
If you wish to reuse any or all of this article please use the link below which will take you to the Copyright Clearance Center’s RightsLink service. You will be able to get a quick price and instant permission to reuse the content in many different ways.
The proposed individual participant data (IPD) meta-analysis will be conducted using raw data from individual participants with advantages, including greater quantity of data, more flexibility in analytic approaches, the ability to conduct subgroup analyses and improved ability to detect and address biases.
This innovative two-step meta-analysis will still collect and synthesise evidence and answer the research questions even if the IPD part is not feasible.
Studies collected in the review may have differences in the timing and frequency of viral load testing, adherence to antiretroviral therapy and patient follow-up, causing imprecision within the data.
There may be insufficient data across the full range of viral load levels to fully assess the association and/or potential lack of agreement by authors to share data for the IPD meta-analysis.
Globally, an estimated 39 million people are currently living with HIV (PLHIV), of whom 29.8 million (76%) are on treatment and 21.2 million (54% of all PLHIV and 71% of PLHIV on treatment) are living with suppressed HIV. 1 Antiretroviral therapy (ART) can improve the lives of PLHIV and help protect their sexual partners from sexual HIV transmission. People who are on HIV treatment can achieve an undetectable viral load with effectively no risk of transmitting HIV to their sexual partners. 2 This concept is referred to as Undetectable equals Untransmittable, or U=U, 3 and it was initiated in 2016 by the Prevention Access Campaign, a health equity initiative with the goal of ending the HIV/AIDS pandemic and associated HIV-related stigma. 4 The U=U concept is based on a substantial body of scientific evidence demonstrating that for PLHIV who have achieved a sustained suppressed and undetectable viral load, there is effectively no risk of sexual HIV transmission. 5 Furthermore, treatment as prevention is one of the effective strategies to prevent HIV transmission, with high uptake of ART suggested as an effective approach to reduce HIV incidence. 6 7
A systematic review and meta-analysis published by the Public Health Agency of Canada (PHAC) in 2018, concluded, using criteria defined by the Canadian AIDS Society framework to characterise HIV transmission risk, 8 that the risk of sexual transmission of HIV is negligible when the PLHIV is on ART with a suppressed viral load of <200 copies of HIV RNA/ml with consecutive testing every 4–6 months. 2 A rapid review published by the Canadian Agency for Drugs and Technologies in Health (CADTH) in 2023 as well as a 2023 PHAC rapid communication confirmed these findings, with the PHAC report providing an estimated risk of HIV sexual transmission of 0.00 transmissions per 100 person-years (95% CI 0.00 to 0.10) in this specific situation. 9 10 In 2023, a systematic review by Broyles et al 11 concluded that the risk of sexual transmission of HIV is almost zero when the PLHIV is under ART and has a suppressed viral load of <1000 copies of HIV RNA, 11 but no quantitative risk estimate was calculated. Furthermore, the WHO concluded in its 2023 policy brief that PLHIV who have a suppressed but detectable viral load have almost zero or negligible risk of sexual transmission of HIV to their partner as long as they continue to take their ART as prescribed. 12 The WHO also revised the operational definition for undetected viral load from ‘≤ 50 copies/ml’ to ‘not detected by the test or sample type used’ and suppressed viral load from ‘≤200 copies/ml’ to ‘≤1000 copies/ml’ and recommended a viral suppression threshold of 1000 copies/mL because persistent viral load levels above 1000 copies/mL are associated with treatment failure. 12
Most of the literature demonstrating that a suppressed, undetectable viral load is associated with effectively no risk of sexual HIV transmission uses a viral load threshold of 200 copies/mL. 3 5 Precise estimates of the risk of sexual transmission at sustained viral load levels between 200 and 1000 copies/mL remain a significant gap in the literature. Addressing this gap by quantifying these risks is needed to evaluate the strength of the association between different viral load levels and the risk of HIV transmission, and to better understand considerations of viral load levels with respect to HIV treatment and prevention programmes.
The primary objective of this review is to quantify the risk estimate of HIV transmission and determine the association between different levels of viral load (primarily in the range 200–1000 copies/mL) and the risk of sexual HIV transmission among serodiscordant couples where the PLHIV is on ART.
The specific hypotheses include: (a) there will be a significant difference in the risk of sexual HIV transmission between viral load levels and (b) there will be a dose-response relationship between different viral load levels (200 copies/mL, 400 copies/mL, 1000 copies/mL or >1000 copies/mL) and risk of sexual HIV transmission.
Q1 : What is the risk of sexual transmission of HIV with suppressed viral load<1000 copies/ml and at different levels of viral load>1000 copies/ml?
Q1.1: What is the risk of sexual HIV transmission in serodiscordant couples when the PLHIV is on ART with different levels of suppressed viral load between 200 to 1000 copies/ml (new potential evidence on risk of HIV transmission with viral load<200 copies/ml will also be assessed and reported if available)?
Q1.2: What is the risk of sexual HIV transmission in serodiscordant couples when the PLHIV is on ART with different levels of viral load>1000?
Q2 : Is there a dose-response association between different levels of viral load and the risk of sexual HIV transmission?
Patient and public involvement.
In designing this meta-analysis protocol, neither patients nor public were involved.
This systematic review will follow a two-step meta-analysis approach. First, a systematic review and meta-analysis will be conducted. Second, an individual participant data (IPD) meta-analysis will be performed if feasible. IPD is considered as the gold standard of reviews and has several advantages compared with aggregate data systematic reviews and meta-analyses. These advantages include a greater quantity of data, the ability to standardise outcomes across trials, more flexibility in analytic approaches, the ability to conduct subgroup/moderator analyses and an enhanced ability to detect and address biases. 13 This protocol is based on the Preferred Reporting Items for Systematic Review and Meta-Analysis Protocol (PRISMA-P) statement 14 ( online supplemental appendix table 1 ). This systematic review and meta-analysis will follow the methodology outlined in the Cochrane Handbook for Systematic Reviews of Intervention. 15 16 The reporting of results will follow the PRISMA 2020 and PRISMA-IPD meta-analysis guidelines ( online supplemental appendix tables 1,2 ). 17 18 IPD meta-analysis will be done using data from studies already published or ongoing on this topic. Although such an approach would produce the best theoretical result, there are some limitations with this method, 19 namely the potential for insufficient data across the full range of viral load levels and/or potential lack of agreement by authors to share such data. Based on the results of the included studies from full-text screening of studies, we will be selecting collaborators for IPD requests. If the proposed IPD meta-analysis is not possible, the systematic review will assess the extent to which these research questions can be answered from existing published literature alone.
Protocol registration.
This study has been registered with the International Registration of Systematic Reviews (PROSPERO) on 11 November 2023 with the registration number CRD42023476946 . Any future changes or modifications to the review procedures will be documented and updated to the PROSPERO registration.
Original studies (randomised controlled trials and non-randomised studies), case reports and conference abstracts will be included if they report on longitudinal studies of couples with one partner living with HIV and document the number of HIV infections in previously seronegative sexual partners and provide information about viral load levels in the HIV-seropositive partner and/or use of ART. For studies that report any HIV infections in the seronegative partner, they will need to be linked to the partner living with HIV through phylogenetic analysis to rule out infection from outside the couple. Considering the difficulty of doing individualised randomisation in public health interventions, cluster Randomized Controlled Trials (RCTs) and quasi-experimental studies with self-control will also be considered for inclusion. Studies reporting a sex partner living with HIV who takes ART and has a viral load measurement provided will be included. Articles written in English and French will be retrieved from electronic English and French databases with full-text access, and published within the timeframe of 1 January, 2000 to Oct 2023 will be included. 20 Studies involving condom use or pre-exposure prophylaxis will be excluded. Studies where HIV is not primarily transmitted through sex will also be excluded. Reviews, editorials, letters and conference proceedings without detailed results will be excluded. Search types and patterns are featured in online supplemental appendix 2 .
A comprehensive and systematic search of the following databases will be conducted: MEDLINE, Embase, the Cochrane Central Register of Controlled Trials via Ovid and Scopus. The search strategy, developed by a health information professional in collaboration with the other authors, uses text words and relevant indexing to identify studies on viral load, ART and transmission of HIV between serodiscordant couples. The MEDLINE search strategy (see Appendix) will be applied to all databases with appropriate modifications. The search will be limited to publications from January 2000 to 2023.
In addition, a thorough examination will be performed of the reference lists of identified relevant studies, experts in the field of HIV sexual transmission will be contacted to identify any additional studies or results, and ClinicalTrials.gov and International Clinical Trials Registry Platform will be examined to identify planned, ongoing or unpublished trials. To retrieve any grey literature, Google Scholar and Baidu Scholar will also be searched. Clinical trial registries will also be searched, including the US National Institutes of Health’s clinicaltrials.gov and Health Canada’s Clinical Trials Database. Search types and patterns are featured in online supplemental appendix table 3 .
Articles will be imported and deduplicated using EndNote20 (Clarivate, Philadelphia, Pennsylvania, USA) and then imported into Covidence systematic review software (Veritas Health Innovation, Melbourne, Australia) for screening. Reviewers (GB, JD and PVN) will do pilot screening with a sample of 100 abstracts to ensure consistency of use and clarity of the inclusion and exclusion criteria. To measure the inter-rater reliability, a Cohen’s kappa statistic will be used. Screening will begin when >70% agreement is achieved. 21 In duplicate, the authors (GB, JD and PVN) will conduct all screening, data extraction and quality assessment procedures. Disagreements will be resolved by consensus. Situations where consensus cannot be reached will be resolved by a third author who will arbitrate (PD and HB). Eligible articles identified by title and abstract screening based on inclusion criteria will be selected for full-text screening. Two independent reviewers will review the full texts. References of the included studies will be hand searched to identify additional relevant studies for inclusion. Conflicts between reviewers will be resolved through discussion, and if no resolution can be achieved, a third reviewer (PD and HB) will be consulted. In case of missing data or information, authors will be contacted.
A third reviewer (PD and HB) will confirm the excluded publications and their respective reasons for elimination. A PRISMA flow chart adapted from the PRISMA 2020 and the PRISMA IPD flow diagram ( figure 1 ) 17 18 will be used to show the process of study selection.
PRISMA flow diagram. IPD, individual participant data; PRISMA, Preferred Reporting Items for Systematic Review and Meta-Analysis.
After the full-text screening and study selection process, the selected studies will undergo data extraction, wherein information from the studies will be extracted after a thorough reading of the full text. The list of variables to be extracted is presented in table 1 . The data extraction form will be created using Microsoft Excel 2016. Data extraction will be conducted by two independent reviewers using the designed data extraction form. Following this process, the records extracted by the reviewers will be cross-checked, and any disputed points will be resolved through a third reviewer (PD and HB).
List of variables for data extraction.
For non-RCTs, the ROBINS-I (the Risk Of Bias in Non-randomised Studies of Interventions) tool will be used by the reviewers to determine the quality of the study. The ROBINS-I tool is concerned with evaluating the risk of bias in estimates of the effectiveness or safety (benefit or harm) of an intervention from studies that did not use randomisation to allocate interventions. 22 This will influence how the data are interpreted. For prognostic studies, QUIPS tool will be used. 23 Biases will be measured as ‘critical risk’, ‘serious risk’, ‘moderate risk’, ‘low risk’ and ‘no information’.
The risk of publication bias will be assessed by visual inspection of funnel plots and using the Egger’s test (with 10 or more included articles). 24
Descriptive statistics from included studies will be extracted and summarised in tables. When there is a difference in data units across studies, we will perform data conversion for the meta-analysis. The main statistical analysis of the study will involve two steps:
Incidence data will be summarised for meta-analysis. A pooled estimate of the incidence of sexual HIV transmission will be generated and reported with 95% CI. Heterogeneity will be examined using 25 26 the I 2 and the H² statistics since they both relate to the percentage of variability that is due to true differences between studies (heterogeneity). I² will be quantified as low (≤25%), moderate (25%–50%) or high (>50%). Fixed-effect model will be used for heterogeneity <50%. Where heterogeneity is >50%, we will use the random-effect model to examine the association between varying viral loads and risk of HIV transmission among serodiscordant couples and create summary forest plots.
The variation for moderate or higher heterogeneity will be explored by conducting meta-regression and sensitivity analyses, including sample size, study year and demographic characteristics, or excluding studies to examine heterogeneity. Furthermore, we will also attempt to explain the heterogeneity by conducting subgroup analyses to compare the risk of HIV transmission between groups, including gay, bisexual and other men-who-have-sex-with-men (gbMSM), women who have sex with women and heterosexuals.
The presence of publication bias will be assessed using a funnel plot and Egger’s test, provided we have at least 10 studies included in the meta-analysis. 24
A data sharing agreement will be established outlining the nature of the project, collaboration and responsibilities of each party. Deidentified and anonymised participant data will be confidentially collected from collaborators. Descriptive analyses will be performed to examine the participants’ demographic characteristics.
We will analyse all the studies separately to compare our results with the original study. Any discrepancies will be resolved. Analysis will include all study participants following the intention-to-treat approach. Summary statistics will be presented as mean (SD) or median (IQR) for continuous variables and per cent for categorical variables. Effect size will be computed for different thresholds of viral load. χ 2 test will be used to evaluate the association of viral load to the risk of sexual HIV transmission by comparing the various viral load levels. We will also compute ORs and corresponding 95% CI to assess the strength of the association of viral load to the risk of sexual HIV transmission. The level of statistical significance α will be 0.05 for all tests. An individual random-effect meta-analysis will be conducted to determine the overall effect of viral load on sexual HIV transmission. Furthermore, a multivariable logistic regression for binary outcomes will be done to predict the risk of HIV transmission among serodiscordant couples at different levels of viral load at the baseline level from each study. Additional adjustments with sociodemographic characteristics, including age, sex, education and location, will also be included. Effect sizes and standard errors can be obtained from this analysis including covariate adjustment which could potentially address bias concerns.
Additionally, viral load levels will be categorised into a contingency table to investigate whether different viral load categories are associated with different levels of HIV transmission risk among sexual partners. A dose-response relationship will also be examined between different viral load levels in PLHIV and the incidence of HIV among their partners using multivariate logistic regression and incidence frequencies of sexual HIV transmission.
All analyses will be done in R V.4.2.3, REVMAN and SPSS V.28 as needed.
Missing data will be addressed depending on the specific characteristics of the missing data. An effort will be made to discuss with collaborating teams the possibility of collecting missing data from their studies. If the data are missing completely at random for the entire study, a list-wise or pair-wise deletion to obtain valid and complete cases will be performed. However, this step may reduce the sample and power of the study. For the remaining non-random missing data, multiple imputations by chained equations will be used. 27 In this method, missing data is computed on a case-by-case basis. A regression model will also be conditionally applied to the other variables in the dataset.
Summary of findings will be presented via tables, including tables for each of the prespecified outcomes (eg, number of cases of HIV transmission). The Grading of Recommendations, Assessment, Development and Evaluation will be used to assess the certainty of evidence considering the bias risk of the trials, consistency of effect, imprecision, indirectness, publication bias, dose response and residual confounding. 28
The meta-analysis will be conducted using deidentified and anonymised data. No human subjects will be directly involved in this research. Dissemination of results of this review will be done through peer-reviewed publications and presentations, as well as international conferences.
We understand that effort, resources and international cooperation are required to perform meta-analysis based on IPD. We will produce a meta-analysis based on the number of collaborators interested in this review and the quality of data collected. We will attempt to establish quantitative risk estimates of sexual HIV transmission at viral load levels between 200 copies/mL and 1000 copies/mL and potentially also at levels >1000 copies/mL. This two-step systematic review (SR) and individual participant data (IPD) meta-analysis will also evaluate the strength of the association of viral load to the risk of sexual HIV transmission. The findings of this SR and IPD meta-analysis will help patients, researchers and policymakers to better understand the risk of sexual HIV transmission in the context of ART and the associated considerations for HIV treatment and prevention programmes.
Patient consent for publication.
Not applicable.
Contributors PD, HB, CA and AF participated in the conception and design of the study. PD, HB, CA, and TE developed the search strategy and assessed the feasibility of the study. PD, HB, GB, JD and PVN wrote the manuscript. CA improved the manuscript. PD, HB, CA and AF are the guarantors. All the authors critically reviewed this manuscript and approved the final version.
Funding This research was funded, conducted and approved by the Public Health Agency of Canada.
Competing interests None declared.
Patient and public involvement Patients and/or the public were not involved in the design, conduct, reporting or dissemination plans of this research.
Provenance and peer review Not commissioned; externally peer reviewed.
Supplemental material This content has been supplied by the author(s). It has not been vetted by BMJ Publishing Group Limited (BMJ) and may not have been peer-reviewed. Any opinions or recommendations discussed are solely those of the author(s) and are not endorsed by BMJ. BMJ disclaims all liability and responsibility arising from any reliance placed on the content. Where the content includes any translated material, BMJ does not warrant the accuracy and reliability of the translations (including but not limited to local regulations, clinical guidelines, terminology, drug names and drug dosages), and is not responsible for any error and/or omissions arising from translation and adaptation or otherwise.
IMAGES
VIDEO
COMMENTS
Steps for Conducting a DQA. Step 1: Conduct PMP Desk Review using PMP Tool to establish status & quality of PMP. Step 2: Meet with Partners. Use DQA Tool No.1 to assess Outcome Indicators; and Tool No.2 for Output Indicators. Step 3: Generate DQA Assessment Summaries for Outcome & Output indicators. Note: For multiple partners, use the Summary ...
data quality assessment is a precondition for informing the users about the possible uses of the data, or which results could be published with or without a warning. ... After a short presentation of the basic quality components for products, processes and user perception, chapters 2 and 3 give concise descriptions of each of the methods. The ...
5. Share the draft Report with the DO team, COR/AOR, and with the IP for comments and formal acknowledgement of the report's findings. HOW TO CONDUCT A DATA QUALITY ASSESSMENT 6. 6. File the DQA report in the official award files and send to the Mission's M&E Specialist for the Program Office files.
Introduction. This resource describes how to conduct a Data Quality Assessment (DQA). How-To Notes provide guidelines and practical advice to USAID staff and partners related to the Program Cycle. This note was produced by the Bureau for Policy, Planning and Learning (PPL). This How-To Note supplements ADS 201.3.5.7.
Monitoring experts responsible for data management and reporting are increasingly questioned about data validity, reliability, integrity, and timeliness. In short, people want to know whether they can trust the data. In order to answer these questions, a Data Quality Assessment (DQA) is carried out.
The data quality assessment1 literature is extensive,2 but it lacks practical applicability in 1. Throughout this paper, we use the word "assessment" to include what is o˙en referred to as a "validation" exercise ... but rather a theoretical presentation of assessment concepts. The classic work of Carmines and Zeller (1979) and Zeller ...
Download ppt "Introduction to Data Quality". Learning Objectives and Topics Covered Understand the data quality conceptual framework Become familiar with the dimensions and metrics of data quality Understand what different RHIS management levels can do to ensure data quality Identify the main types of data quality problems Topics Covered Define ...
It also discusses the relationship between data assessment and requirements definition, and the application of criticality and risk assessment to make better decisions about what data to measure and how to measure it. ... In this presentation, data quality and Lean Six Sigma author, Rajesh Jugulum, provides practical ideas for aligning your ...
Presentation on theme: "Data Quality Assessment"— Presentation transcript: 1 Data Quality Assessment Good morning and welcome to this Data Quality Session. It is great to be with you here at USAID/Jordan. As you are aware, the Mission is in the final phases of updating its PMP. You have been working on this for sometime now and you were busy ...
The course is intended for managers or analysts that either use (analyze) data themselves or review the use of data by others. "Introduction to Data Quality Assessment" Course. Introduction to Data Quality Assessment Presentation (PPT) (ppt) (828.5 KB) Data Quality Assessment Handout 1 - Project Table (doc) (642 KB)
The DQA Tool focuses exclusively on (1) verifying the quality of reported data, and (2) assessing the underlying data management and reporting systems for standard program-level output indicators. The DQA Tool is not intended to assess the entire M&E system of a country's response to HIV/AIDS, Tuberculosis or Malaria.
Data quality assessment methods generally base on the measurement theory [35,36,37,38]. Each dimension of data quality consists of a set of attributes. ... quality control, analysis and presentation for users, an integrated data warehouse, a metadata dictionary, unique identifier codes available: Mixed methods: quantitative and qualitative. Use ...
Presentation Transcript. The Data Quality Assessment Framework OECD Meeting of National Accounts Experts October 2001. Purpose of this Presentation To describe: • The IMF's Data Quality Assessment Framework (DQAF), and • Experience to date with the DQAF for Reports on Observance of Standard and Codes (ROSCs) and beyond.
Summary. Data quality issues can significantly impact the business, but they may not be hard to fix if you address them early. This presentation provides 12 pragmatic actions to help data and analytics leaders holistically improve their data quality to attain long-term, sustainable business value.
Tackling data quality problems requires more than a series of tactical, one off improvement projects. By their nature, many data quality problems extend across and often beyond an organization. Addressing these issues requires a holistic architectural approach combining people, process and technology. Join Nigel Turner and Donna Burbank as they ...
PowerPoint presentation slides: Presenting this set of slides with name - Data Quality In 6 Step Process Showing Assessment And Control. This is a six stage process. The stages in this process are Data Quality, Data Governance, Data Management.
ÐECÍãP"J`oí"ìê‹Z'oÂȾƒÉ‹ ">m UýÀ(Ž¡"þ.›ô^*™ô] UŸZYè» × ™X µ1&iàG8úFdr \æЀ™Gø.@ !l ¦ ¯€V¡ƒG}Ïõ÷ècÏQüÀ À BWÇúÅØÇŽ†žGˆcSä½EE¾ ä` 2ŒuK) ÿRicš±Ì†NA¹Ê¢™"YQ„CËM¼ ÛIìg882̽ ¸rŒîÊQUÇs(* ¶ ÁN;¾öÙöÁ°¾y ˆ=`©ºFòL,n„yø 'CÞªg®`ùÏ ...
Use our PPT slide to represent data profiling, data reporting, data repair, and other pillars of information quality management. Consolidate data quality assessment criteria like uniqueness, completeness, timeliness, accuracy, validity, and consistency through this PowerPoint template. Highlight the data cleansing system with the help of this ...
Quality experts and data scientists will find this deck helpful in showcasing the characteristics and elements of data quality. You can depict the key areas of data quality management. You can also demonstrate the stepwise process of improving the quality of data. Visualize the key factors for data quality and how they contribute to effective ...
Presentation Transcript. Assessment of data quality Mirza Muhammad Waqar Contact: [email protected] +92-21-34650765-79 EXT:2257 RG610 Course: Introduction to RS & DIP. Contents • Hard vs Soft Classification • Supervised Classification • Training Stage • Field Truthing • Inter class vs Intra Class Variability • Classification ...
Data Assessment found in: Data Assessment Powerpoint Ppt Template Bundles, Six dimensions for data quality assessment, Data Quality Assessment Metrics For Operational Records, Enterprise Data And Analytics Maturity Assessment..
Introduction HIV is a major global public health issue. The risk of sexual transmission of HIV in serodiscordant couples when the partner living with HIV maintains a suppressed viral load of <200 copies of HIV copies/mL has been found in systematic reviews to be negligible. A recent systematic review reported a similar risk of transmission for viral load<1000 copies/mL, but quantitative ...