MacKay, Robert
(2018).
Calibration with confidence: A principled method for panel assessment.
[Data Collection]. Colchester, Essex:
UK Data Archive.
10.5255/UKDA-SN-852889
This network project brings together economists, psychologists, computer and complexity scientists from three leading centres for behavioural social science at Nottingham, Warwick and UEA. This group will lead a research programme with two broad objectives: to develop and test cross-disciplinary models of human behaviour and behaviour change; to draw out their implications for the formulation and evaluation of public policy.
Foundational research will focus on three inter-related themes:
understanding individual behaviour and behaviour change; understanding social and interactive behaviour; rethinking the foundations of policy analysis.
The project will explore implications of the basic science for policy via a series of applied projects connecting naturally with the three themes. These will include: the determinants of consumer credit behaviour; the formation of social values; strategies for evaluation of policies affecting health and safety.
The research will integrate theoretical perspectives from multiple disciplines and utilise a wide range of complementary methodologies including: theoretical modeling of individuals, groups and complex systems; conceptual analysis; lab and field experiments; analysis of large data sets.
The Network will promote high quality cross-disciplinary research and serve as a policy forum for understanding behaviour and behaviour change.
Data description (abstract)
Frequently, a set of objects has to be evaluated by a panel of assessors, but not every object is assessed by every assessor. A problem facing such panels is how to take into account different standards among panel members and varying levels of confidence in their scores. Here, a mathematically based algorithm is developed to calibrate the scores of such assessors, addressing both of these issues. The algorithm is based on the connectivity of the graph of assessors and objects evaluated, incorporating declared confidences as weights on its edges. If the graph is sufficiently well connected, relative standards can be inferred by comparing how assessors rate objects they assess in common, weighted by the levels of confidence of each assessment. By removing these biases, ‘true’ values are inferred for all the objects. Reliability estimates for the resulting values are obtained. The algorithm is tested in two case studies: one by computer simulation and another based on realistic evaluation data. The process is compared to the simple averaging procedure in widespread use, and to Fisher's additive incomplete block analysis. It is anticipated that the algorithm will prove useful in a wide variety of situations such as evaluation of the quality of research submitted to national assessment exercises; appraisal of grant proposals submitted to funding panels; ranking of job applicants; and judgement of performances on degree courses wherein candidates can choose from lists of options.
Data creators: |
Creator Name |
Affiliation |
ORCID (as URL) |
MacKay Robert |
University of Warwick |
|
|
Contributors: |
|
Sponsors: |
Economic and Social Research Council
|
Grant reference: |
ES/K002201/1
|
Topic classification: |
Economics
|
Keywords: |
calibration, evaluation, assessment, confidence, uncertainty, model comparison
|
Project title: |
Network for Integrated Behavioural Science
|
Grant holders: |
Chris Starmer, Nick Chater, Daniel John Zizzo, Gordon Brown, Anders Poulsen, Martin Sefton, Neil Stewart, Uwe Aickelin, John Gathergood, Robert Sugden, Abigail Barr, Graham Loomes, Daniel Read, Robin Cubitt, Robert MacKay, Enrique Fatas, Shaun Hargreaves-Heap, Simon Gaechter, Theodore Turocy
|
Project dates: |
From | To |
---|
31 December 2012 | 30 September 2017 |
|
Date published: |
04 Dec 2017 16:46
|
Last modified: |
07 Feb 2018 15:17
|
Collection period: |
Date from: | Date to: |
---|
31 December 2012 | 30 September 2017 |
|
Country: |
United Kingdom |
Data collection method: |
Experimental data. We have tested the approach in three contexts. We report in detail on two case studies here. In the first case study, we use a computer-generated set of data containing true values of assessed items, assessor biases and confidences for the assessments and resulting scores. This has the advantage of allowing us to compare the values obtained by the new approach with the true underlying value of each item. The second case study is an evaluation of grant proposals using realistic data based on a university's internal competition. In this test, of course, there is no possibility to access ‘true’ values, so instead we compare the evidence for the models using a Bayesian approach (appendix E), and we compare their posterior uncertainties (appendix D). The third context in which we tested our method was assessment of students; we report briefly on this at the end of the section. |
Observation unit: |
Individual |
Kind of data: |
Numeric |
Type of data: |
Experimental data
|
Resource language: |
English |
|
Rights owners: |
Name |
Affiliation |
ORCID (as URL) |
MacKay Robert |
University of Warwick |
|
|
Contact: |
Name | Email | Affiliation | ORCID (as URL) |
---|
MacKay, Robert | R.S.MacKay@warwick.ac.uk | University of Warwick | Unspecified |
|
Notes on access: |
The Data Collection is available from an external repository. Access is available via Related Resources.
|
Publisher: |
UK Data Archive
|
Last modified: |
07 Feb 2018 15:17
|
|
Data collections
Publications
Software
Website
Edit item (login required)
 |
Edit Item |