Why rate when you could compare? Using the “EloChoice” package to assess pairwise comparisons of perceived physical strength (original) (raw)

Open Access

Peer-reviewed

Research Article

Why rate when you could compare? Using the “EloChoice” package to assess pairwise comparisons of perceived physical strength

PLOS

x

Figures

Abstract

We introduce “EloChoice”, a package for R which uses Elo rating to assess pairwise comparisons between stimuli in order to measure perceived stimulus characteristics. To demonstrate the package and compare results from forced choice pairwise comparisons to those from more standard single stimulus rating tasks using Likert (or Likert-type) items, we investigated perceptions of physical strength from images of male bodies. The stimulus set comprised images of 82 men standing on a raised platform with minimal clothing. Strength-related anthropometrics and grip strength measurements were available for each man in the set. UK laboratory participants (Study 1) and US online participants (Study 2) viewed all images in both a Likert rating task, to collect mean Likert scores, and a pairwise comparison task, to calculate Elo, mean Elo (mElo), and Bradley-Terry scores. Within both studies, Likert, Elo and Bradley-Terry scores were closely correlated to mElo scores (all _r_s > 0.95), and all measures were correlated with stimulus grip strength (all _r_s > 0.38) and body size (all _r_s > 0.59). However, mElo scores were less variable than Elo scores and were hundreds of times quicker to compute than Bradley-Terry scores. Responses in pairwise comparison trials were 2/3 quicker than in Likert tasks, indicating that participants found pairwise comparisons to be easier. In addition, mElo scores generated from a data set with half the participants randomly excluded produced very comparable results to those produced with Likert scores from the full participant set, indicating that researchers require fewer participants when using pairwise comparisons.

Citation: Clark AP, Howard KL, Woods AT, Penton-Voak IS, Neumann C (2018) Why rate when you could compare? Using the “EloChoice” package to assess pairwise comparisons of perceived physical strength. PLoS ONE 13(1): e0190393. https://doi.org/10.1371/journal.pone.0190393

Editor: Iratxe Puebla, Public Library of Science, UNITED KINGDOM

Received: June 15, 2016; Accepted: December 14, 2017; Published: January 2, 2018

Copyright: © 2018 Clark et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.

Data Availability: All relevant data from this study is shared in a public repository (figshare: https://figshare.com/articles/Why_rate_when_you_could_compare_Elo_strength_data/4902977).

Funding: Leverhulme Trust (www.leverhulme.ac.uk) provided a Research Grant [F/00 182/CJ (2011-2013)] to ISP-V and APC. The funder had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript, and did not provide support in the form of salaries. The specific roles of these authors are articulated in the ‘author contributions’ section. Xperiment, a commercial company founded by ATW, had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript. Neither Xperiment nor ATW were paid to perform the study - he provided his services and the platform that he created for free. The specific role of this author is articulated in the ‘author contributions’ section.

Competing interests: One of our authors, ATW, is the founder of Xperiment, but neither he nor the company were paid in relation to this study. This affiliation did not interfere with the full and objective presentation of our results, nor alter our adherence to PLOS ONE policies on sharing data and materials. We declare no further competing interests.

Introduction

Likert-type rating vs pairwise comparisons

When a researcher wants to quantify some perceived characteristic for stimuli within a set, one way to achieve this is to ask participants to rate each stimulus along an integer scale, anchored on either end with bivalent labels. For example, a rater could be asked to evaluate the attractiveness (“characteristic”) of an image depicting a face (i.e. “stimulus”) along a gradient between not attractive at all and highly attractive that could be scored along integers between 1 and 7. This is often referred to as a Likert scale [13] but this is technically inaccurate as Likert himself used “scale” to refer to a group of multiple such items [4]. Discrete Visual Analogue Scale is suggested as a general alternative [5], but we will simply refer to the practice as “Likert rating”. When inter-rater agreement is high, and average responses are calculated from a reasonable sample of raters, Likert rating tends to produce repeatable results that convey useful information about the stimuli [6]. This represents a mathematically straight-forward process that is easy for researchers with even minimal technical ability to complete using very basic and readily available hardware and software tools.

From an individual participants’ perspective, however, the rating experience could be better. The scale itself can be a challenging concept for users with a weak grasp of number lines; for instance, small children [7], or members of some traditional, pre-literate cultures [8]. More generally, using a rating scale can increase cognitive demand as it requires conscientious raters to monitor their responses to each stimuli and retrospectively compare them to responses from previous perceptive states, whereas paired comparisons imposes no such constraint [9].

One alternative to Likert rating is to show stimuli in pairs and ask participants to choose which one better expresses some characteristic, for example, which of two faces is more attractive. The results of these pairwise comparisons can be used to quantify perceptions of a stimulus according to a mathematical model that calculates its probability of being judged to be more or less expressive than a given set of alternatives. This method was first developed in the context of psychological research by Thurstone [10], but developed further by Bradley & Terry [11]. A class of mathematically-related models are now often known as Bradley-Terry models [1213], and include a variety of sophisticated extensions and approaches to model-fitting [9, 1416]. Also related are models that were independently derived to rank chess players—the earliest by Zermelo [17] and the most famous by Elo [18], which in turn has led to more complex derivatives such as Glicko [19] and TrueSkill™ [20].

There are a number of advantages associated with pairwise comparisons. They are less complicated for participants: there are only two choices for each decision, the choices do not require an understanding of number lines, and there is no need to track responses from previous decisions. Besides this, each decision made provides information about two stimuli, therefore representing a more efficient use of participant time. In previous contrasts, results from Likert ratings correlate strongly with pairwise comparison methods [2122]. Nevertheless, pairwise comparisons are much less used, perhaps because many Bradley-Terry models, although simple by the standards of mathematicians, are comparatively complex and computationally intensive. However, this is no longer a serious constraint as the hardware and software capabilities to overcome it are readily available. Elo rating, in particular, is a relatively simple approach that offers conceptual accessibility, widespread familiarity and use [2327], and recent programming implementations [2829]. The purpose of this paper is to demonstrate the application of Elo rating to pairwise comparisons of stimuli, and introduce “EloChoice”, an R package optimized for this use and designed to be accessible even to those with limited experience of the free, and increasingly popular, R programming environment [30]. The package including its source code is available from https://cran.r-project.org/web/packages/EloChoice/index.html and includes a detailed manual (https://cran.r-project.org/web/packages/EloChoice/vignettes/tutorial.pdf).

Elo rating

The following paragraphs outline the Elo rating process in brief, but further details and more in-depth discussion are readily available [18, 23, 31, 29].

Elo rating was developed to quantify and track over time skill distribution amongst chess players using information available from their previous matches. After a match between two players, points are exchanged depending on both players’ prior probability of winning the match. Each player enters with a score based on their match records, and the disparity between scores is used to predict the probability of outcomes, according to the following formulas, which assume a logistical distribution [18]:(1) (2)where _E_1 and _E_2 are the estimated probabilities of a win for players 1 and 2 respectively, and _S_1 and _S_2 are their current scores. Players with equal scores are expected to be equally likely to win. After each new match, more information is available and is used to update each player’s scores, according to the following formulas:(3) (4)where _O_1 and _O_2 are the actual outcomes for players 1 and 2 respectively (win = 1, loss = 0), _S_’1 and _S_’2 are their new scores, and k is a constant representing maximum point exchange. The exact value of k affects how quickly scores change and their eventual range, but, given a sufficient number of matches, will have a limited effect on player rankings [29, 3132].

A key feature of the system is that the less expected the win, the more scores change (both for winner and loser). An unsurprising victory over a far weaker opponent will not result in a large score change, but an upset win will. Therefore, the highest scores require wins over most other opponents, including those with above-average ability.

Elo rating can be readily adapted to quantify the distribution of perceived attributes for a set of stimuli on an interval scale, allowing simple side-by-side comparisons while taking into account disparity between stimuli. Each pairwise comparison can be seen as a contest between two stimuli, with the one chosen as a better exemplar of some attribute being the winner. All stimuli begin with the same starting score, but diverge as successive comparisons take place. Because the number of points exchanged in any comparison is symmetrical (the winner’s gains equal the loser’s losses), the mean score will always equal the starting score, so long as the number of stimuli in the set remains stable. When stimuli are paired randomly, the scores should quickly converge to a stable ranking, assuming sufficient variation among stimuli and a reasonable degree of shared perceptions among rating participants. As with chess players, there is no requirement that all combinations of stimuli are observed, and, indeed, this is one of the system’s chief advantages [33].

There are at least two problems associated with the Elo rating system in the context of this application, but both are readily surmountable. The first problem arises from the fact that final scores can be influenced by the sequence order of contests, because the same contest outcomes will produce different results depending on the order in which they are observed. This is appropriate when the targets being scored are entities that may become weaker or stronger with time (such as players gaining or losing skill), but psychological stimuli are typically static in this regard and hence the order of events is not capturing any useful information. As there is nothing special about the sequence of contests presented to participants, any variation in final scores introduced by sequence variation is undesirable noise. To counter this problem, we propose a simple fix; randomly shuffle the original sequence presented to participants multiple times to create multiple virtual sequences, and average Elo scores for each stimuli from the different sequences to generate mean Elo scores that are free from sequence noise.

A second problem is that there is currently no commonly accepted way to measure consistency across paired comparison trials. It is often desirable to measure the consistency of ratings for stimuli across multiple raters (to gauge the extent to which raters share perceptions about targets), and in the case of Likert rating this is most appropriately achieved by calculating the intraclass correlation coefficient (ICC) [3435]. To provide a conceptually (but not directly) comparable value, our solution is to propose a novel measure, the Elo consistency index, which tracks how often outcomes deviate from predictions based on previous judgments (that is, those trials in which the stimulus with the lower Elo score was chosen–in other words, an upset or reversal). Because the preceding Elo score is generated by the judgements made in previous comparisons, violations of expected outcomes can be seen as inconsistency between judgements. This inconsistency can occur both between and within raters. The index will be biased towards tracking one form of inconsistency or the other depending on the ratio of raters to the times each rater makes a judgement on each stimulus. In the case that each rater makes only one judgement on each stimulus, the index will be entirely tracking between-rater consistency (because each preceding Elo score was generated by judgements made by previous raters only). The index can be calculated for any sequence of trials (across any number of raters) and is defined by the following formula:(5)where u is a vector of 1’s and 0’s, in which a 1 indicates that the expected outcome was violated and a 0 indicates that it was not, and N is the total number of trials for which an expectation existed (that is, for all trials in which the preceding Elo score difference was not 0). The resulting index value will vary between 0 and 1. Index values approach 1 as fewer expectation violations (or upsets/reversals) are observed, and approach 0 as more are observed. An index value of 0.5 indicates random choice, but values of less than 0.5 are technically possible. Index values for each sequence generated can be averaged to produce a mean consistency index score to correspond with mean Elo scores.

A weighted version of the Elo consistency index can be calculated by taking into account the score disparity observed in expectation violations, according to the following formula:(6)where u is the same vector described above and w is the absolute difference in preceding Elo scores between members of a trial. This can be understood as the proportion of all points exchanged in a sequence that were correctly predicted by preceding Elo scores. The weighted consistency index will again vary between 0 and 1, and larger expectation violations (that is, where the difference in preceding Elo scores is greater) will negatively impact values of R more. Generally the weighted consistency index value is expected to be greater than the unweighted value, as the magnitude of difference between Elo scores should negatively predict probability of an expectation violation.

To demonstrate Elo rating in action and to compare it both to Likert rating and to a basic Bradley-Terry model, we use the example of perceived physical strength from stimuli composed of images of human male bodies. Previous research using Likert-type ratings have shown that participants judging strength from photographs of men’s bodies display high inter-rater agreement, and mean ratings correlate robustly with men’s actual measured strength and strength-related body measurements [36]. We use both laboratory (Study 1) and online participants (Study 2) to demonstrate that the Elo-rating method is appropriate in both contexts.

Study 1—Laboratory rating

Methods

Target stimuli.

Stimuli consisted of digital images (399 x 710 pixels) of 82 men (mean age = 21.4, sd = 2.5, min = 18, max = 30) depicted standing on a raised platform and facing the camera. The men are wearing only a pair of black boxer briefs. Their heads are digitally blurred, making positive identification from face alone impossible. Images were all captured from the same camera, mounted on a tripod at the same height, location and orientation. Anthropometric measures were collected from each man on the day he was image was captured. The men who posed for the images gave written informed consent beforehand, and were reimbursed for their time (£40 GBP). Approval for their participation was given by the Faculty of Science Human Research Ethics Committee, at the University of Bristol.

Stimuli measures.

We used hand grip strength as an assay of overall physical strength, as it easy to collect and previous research has shown it to be highly correlated with other measures, such as chest and shoulder strength [36]. Each man who posed for stimulus images was instructed to squeeze a hand dynamometer as hard as he could using one hand. Trials were completed for both the right hand and the left hand separately. For each man, the average of these two squeezes (measured in kg) was taken and used as the grip strength score associated with the stimulus image (mean = 39.5, sd = 7.1).

Body size has also been found to correlate with upper-body strength in males [36]. We used six anthropometric measures to construct a proxy of body size: height, weight, shoulder circumference, chest circumference, bicep circumference and forearm circumference. We performed a principal components analysis on these measures and used scores for the first principal component (explaining 72.4% of total variance) as our body size score (PC1 of body size; mean = 0, sd = 2.1). Grip strength and PC1 of body size were moderately correlated (r = 0.46).

Rating participants.

56 participants (28 male, 28 female; mean age = 19.9, sd = 1.98) took part in the rating study. These participants were students recruited from the University of Bristol either by opportunity sampling (word of mouth) or for partial course credit in the undergraduate Psychology programme. Participants gave written informed consent before completing any task. Approval for their participation was given by the Faculty of Science Human Research Ethics Committee, at the University of Bristol.

Rating procedure.

Participants viewed each image three times, once within a Likert rating task, and twice within a paired comparison task. The order of these tasks was counterbalanced. The tasks were presented on a laboratory computer using e-Prime software. Each participant completed the tasks in private, and no time requirement was specified.

For the Likert rating task, images were presented sequentially in random order. Each image was presented for 2 seconds before text appeared, reading, “How strong is this man?” accompanied by a 7-point scale anchored by “very weak” at point 1 and “very strong” at point 7. Participants used the keyboard to indicate a response, and time to make a response was recorded. 82 responses were required to complete the task.

The paired comparison task was presented in two blocks. In each block, each image was randomly paired with a different image and each pair was presented in random order. The paired images were presented side by side. Each pair was presented for 2 seconds before text appeared, reading, “Which man is stronger?”. Participants used the keyboard to indicate a response by pressing ‘z’ to choose the image on the left, and ‘m’ to choose the image on the right. Again, response time was recorded. 41 responses were required to complete each block, and 82 responses were required to complete the task.

Rating measures.

For each stimulus image, Likert responses from all rating participants were averaged to generate a mean Likert rating score.

To generate Elo scores for each stimulus image, all paired comparison trials were arranged in their original sequence, ordered first by rating participant number and then by trial number. Beginning Elo scores were set to 0, and k was arbitrarily set to 100, following precedent [29]. Using Formulas 1–4, updates to Elo scores were calculated for each trial sequentially—each trial involved updating scores for two images. Final Elo scores were the scores for each image after the calculation performed for the last trial within the sequence.

Elo scores were calculated for each image using the original sequence, as described above, and then for 99 randomized trial sequences. Final Elo scores from each of the 100 sequences were averaged together to generate mean Elo (mElo) scores for each image. For comparative purposes, an additional set of mElo scores was generated by averaging Elo scores across 1000 sequences (the original sequence plus 999 randomized sequences). These are referred to as mElo(M) scores in the subsequent analyses.

Scores were collected by extracting worth parameters from a Bradley-Terry model fit using the paired comparison data. As these scores are highly skewed, they are log-transformed for the purposes of correlations.

Statistical software.

All analyses were conducted in R version 3.2.1 [30].

Principal components analysis was conducted using the “FactoMineR” package [37].

Elo and mElo scores and the Elo consistency indices were calculated using the “EloChoice” package [38].

Bradley-Terry models were fit and worth parameters extracted using the “psychotools” package [39].

Comparisons between overlapping and non-overlapping correlations from dependent groups were conducted using the “cocor” package [40] and Williams’s t [41].

Intraclass correlation coefficients were calculated using “irr” package [42].

Comparisons between means from dependent groups were conducted using the “perfect-t-test” script [43].

Results

All analyses shown incorporate responses from male and female rating participants together, as we were not interested in sex differences for this project.

Descriptive statistics of rating measures.

Table 1 displays descriptive statistics for Likert ratings of perceived strength, and Elo scores (from the original sequence), mElo scores, mElo(M) scores, and Bradley-Terry scores from perceived strength comparisons for the sample of 82 stimulus images.

thumbnail

Table 1. Descriptive statistics for mean Likert ratings, Elo scores, mElo scores (using 100 iterations), mElo(M) scores (using 1000 iterations) and Bradley-Terry scores for the 82 stimulus images.

https://doi.org/10.1371/journal.pone.0190393.t001

Correlations between rating measures and stimuli measures.

All rating measures were strongly correlated with each other (see Table 2), particularly in the cases of mElo, mElo(M) and Bradley-Terry scores (log-transformed), which were almost perfect correlated (all _r_s > 0.999).

thumbnail

Table 2. Correlations between mean Likert ratings, Elo scores, mElo scores (100 iterations), mElo(M) scores (1000 iterations), and log-transformed Bradley-Terry scores for the 82 stimulus images.

https://doi.org/10.1371/journal.pone.0190393.t002

All ratings measures were correlated with stimuli measures (grip strength and PC1 of body size) to a similar extent (see Table 3), although these relationships were weakest for mean Likert ratings.

thumbnail

Table 3. Correlations between stimuli measures (grip strength and PC1 of body size) and rating measures (mean Likert ratings, Elo scores, mElo scores (100 iterations), mElo(M) scores (1000 iterations), and log-transformed Bradley-Terry scores) for the 82 stimulus images.

https://doi.org/10.1371/journal.pone.0190393.t003

The very similar results obtained by mElo and mElo(M) suggest that there is little to be gained by generating mElo scores from greater than 100 sequences.

Comparing Elo and mElo scores.

The 100 sequences used to calculate the mElo scores each produced their own set of Elo scores. Considered separately, each set significantly correlated with grip strength, but the strength of these correlations varied (0.35 < all _r_s < 0.49). This was also the case for correlations with PC1 of body size (0.56 < all _r_s < 0.68). These 100 sets of Elo scores were all strongly correlated with each other, but again with considerable variation (0.86 < all _r_s < 0.95).

Fig 1 shows Elo scores from the original sequence and mElo scores for individual stimulus images on the same plot for comparison.

thumbnail

Fig 1. Elo scores from the original sequence (grey circles) and mElo scores (black circles) are depicted for each of the stimulus images, arranged from left to right in order of decreasing mElo score.

The whiskers depict the range of Elo scores from the 100 sequences generated to calculate the mElo scores.

https://doi.org/10.1371/journal.pone.0190393.g001

Comparing Likert ratings and mElo scores from reduced data sets.

To show how the relationships between measures of perceived strength and stimulus variables is affected by the number of raters, Likert ratings of perceived strength and mElo scores from perceived strength comparisons were recalculated using just the first half of rating participants (n = 28; 14 males and 14 females), and then again using just the first quarter of rating participants (n = 14; 7 males and 7 females). Correlation coefficients (_r_s) for these relationships using the full, half, and quarter participant sets are show in Table 4.

For the Likert ratings, correlations from the halved set were significantly smaller than those from the full set (grip strength: t = -2.100, df = 79, p = 0.039*; PC1 of body size: t = -2.412, df = 79, p = 0.018*), and correlations from the quartered set were significantly smaller than those from the halved set (grip strength: t = -3.616, df = 79, p < 0.001***; PC1 of body size: t = -3.203, df = 79, p = 0.004**).

For the mElo scores, correlations from the halved set were significantly smaller than those from the full set (grip strength: t = -2.144, df = 79, p = 0.035*; PC1 of body size: t = -4.524, df = 79, p < 0.001***), but correlations from the quartered set were not significantly smaller than those from the halved set.

Consistency of Likert ratings and mElo scores.

To measure inter-rater consistency of Likert ratings, an average score intraclass correlation coefficient (model: two-way; type: consistency) was calculated [34]. The result (ICC = 0.984, 95% CI: 0.979 < ICC < 0.989) indicates very high agreement between raters about the perceived strength of targets.

To measure inter-trial consistency of perceived strength comparisons used to calculate mElo scores, we used the novel consistency indices described (formulas 5 and 6). The mean unweighted consistency index was 0.766, while the mean weighted consistency index was 0.864. It is important to note that these values are not directly comparable with ICC values. The unweighted value indicates that 76.6% of trial outcomes were concordant with the predictions of preceding Elo scores, and the weighted value indicates that the direction of 86.4% of all points exchanged was predicted by preceding Elo scores. Both values indicate high consistency of judgements between trials.

How many raters are needed to establish stable consistency indices?

Values for the consistency indices are expected to become more accurate as the number of raters increases. This is because the number of trials with an expected result will increase, effectively increasing the sample size from which to estimate consistency. To demonstrate how many raters are needed before these estimates become stable, the weighted consistency index was re-calculated multiple times starting with data from just the first rater (according to the original sequence), then adding the second rater, then the third rater, and so on until all 56 raters were included. Fig 2 depicts these results, showing that index values reached a stable plateau after roughly 30 raters. This could also be used as an indication of the minimal number of raters required to produce relatively stable rankings of mElo scores.

thumbnail

Fig 2. Weighted consistency indices for the original sequence (small grey circles) and mean weighted consistency indices (large black circles) calculated for an increasing number of Study 1 raters (1–56, in increments of 1).

The whiskers represent inter-quartile ranges of indices from the 100 sequences generated to calculate the mean index.

https://doi.org/10.1371/journal.pone.0190393.g002

Comparing decision times for Likert ratings and paired comparison tasks.

Mean decision times (in milliseconds) for Likert ratings (mean = 995.6, sd = 392.6) were compared to mean decision times (in milliseconds) for paired comparisons (mean = 635.8, sd = 233.7) using a paired t-test. The result indicates a significant difference between the two (t 55 = 7.65, p < 0.001***), and a large effect size (Hedges’ g = 1.1, 95% CI: 0.76 < g < 1.46).

Fig 3 shows mean decision times for Likert ratings and paired comparisons.

thumbnail

Fig 3. Violin plots depict the distribution of response times for paired comparison trials (left), and Likert trials (right).

Individual values are depicted by small grey circles. Error bars depict within-subject 95% confidence intervals. Means are depicted by large black circles.

https://doi.org/10.1371/journal.pone.0190393.g003

Comparing computation times for mElo and Bradley-Terry scores.

Both mElo and Bradley-Terry models involve more computation than either Likert or Elo. To compare the time it takes to compute scores via these methods, we timed each procedure on the same laptop computer (2.2 GHz processor speed, 8 GB RAM), using R’s system.time function. The elapsed time for computing mElo scores for the entire data set was 3 seconds. The elapsed time for computing Bradley-Terry scores for the entire data set was 1017 seconds.

Study 2—Online rating

The results of Study 1 suggest that mElo scores correspond well with Likert ratings for tasks completed within the lab, but with the advantages offered by web-based data collection (speed, cost, diversity of participants) it is important to check whether comparable results are obtained for tasks completed online. Previous research indicates that online performance is generally good compared to performance in labs [44] and that attention to instructions is actually better [45].

Methods

Target stimuli, stimuli measures, rating measures and statistical software were identical to Study 1, but mElo(M) scores (based on 1000 sequences) were not calculated.

Rating participants.

96 participants (59 male, 37 female; mean age = 32.6, sd = 9.31) took part in and completed the rating study. These participants were recruited from mTurk (US residents only) and paid $3 USD for participation. Before completing any task, participants gave informed consent by ticking a check box stating that they had read and understood the procedure. Approval for their participation was given by the Faculty of Science Human Research Ethics Committee, at the University of Bristol.

Rating procedure.

Participants viewed each image three times, once within a Likert rating task, and twice within a paired comparison task. The order of these tasks was counterbalanced. The tasks were presented using Xperiment online presentation software (www.experiment.mobi). Unlike Study 1, no response times were recorded.

For the Likert rating task, images were presented sequentially in random order. Each image was presented for 2 seconds before text appeared, reading, “How strong is this man?” accompanied by a 7-point scale anchored by “very weak” at point 1 and “very strong” at point 7. Participants clicked a point on the scale to indicate a response. 82 responses were required to complete the task.

The paired comparison task was presented in two blocks. In each block, each image was randomly paired with a different image and each pair was presented in random order. The paired images were presented side by side for 2 seconds before text appeared, reading, “Which man is stronger?”. Participants indicated a response by clicking on the button corresponding to either the left or right image. 41 responses were required to complete each block, and 82 responses were required to complete the task.

Results

All analyses shown incorporate responses from male and female rating participants together, as we were not interested in sex differences for this project.

Descriptive statistics.

Table 5 displays descriptive statistics for Likert ratings of perceived strength, and Elo scores (from the original sequence), mElo scores, and Bradley-Terry scores for the sample of 82 stimulus images.

Correlations between rating measures and stimuli measures.

All rating measures were strongly correlated with each other (see Table 6), particularly mElo and Bradley-Terry scores (log-transformed), which were almost perfect correlated (r = 0.9992).

All ratings measures were correlated with stimuli measures (grip strength and PC1 of body size) to a similar extent (see Table 7).

thumbnail

Table 7. Correlations between stimuli measures (grip strength and PC1 of body size) and rating measures (mean Likert ratings, Elo scores, mElo scores and log-transformed Bradley-Terry scores) for the 82 stimulus images.

https://doi.org/10.1371/journal.pone.0190393.t007

Correlations between rating measures from Study 1 and Study 2.

To examine whether lab participants and online participants produce similar results using these rating measures, we correlated ratings and scores from Studies 1 and 2 (see Table 8). The measures were strongly correlated with one another, indicating that mode of delivery (lab or online) did not have a great impact.

Comparing Elo and mElo scores.

Fig 4 shows Elo scores from the original sequence and mElo scores for individual stimulus images on the same plot for comparison.

thumbnail

Fig 4. Elo scores from the original sequence (grey circles) and mElo scores (black circles) are depicted for each of the stimulus images, arranged from left to right in order of decreasing mElo.

The whiskers depict the range of Elo scores from the 100 sequences generated to calculate the mElo scores.

https://doi.org/10.1371/journal.pone.0190393.g004

Consistency of Likert ratings and mElo scores.

To measure inter-rater consistency of Likert ratings, an average score intraclass correlation coefficient (model: two-way; type: consistency) was calculated. The result (ICC = 0.99, 95% CI: 0.985 < ICC < 0.994) indicates very high agreement between raters about the perceived strength of targets.

To measure inter-trial consistency of perceived strength comparisons used to calculate mElo scores, we used the novel consistency indices described (formulas 5 and 6). The mean unweighted consistency index was 0.740, while the mean weighted consistency index was 0.837. Again, note that these values are not directly comparable with ICC values. The unweighted value indicates that 74.0% of trial outcomes were concordant with the predictions of preceding Elo scores, and the weighted value indicates that the direction of 83.7% of all points exchanged was predicted by preceding Elo scores. Both values indicate high consistency of judgements between trials.

How many raters are needed to establish stable consistency indices?

As in Study 1, to demonstrate how many raters are needed before estimates of consistency become stable, the weighted consistency index was re-calculated multiple times, starting with data from just the first rater (according to the original sequence), then adding the second rater, then the third rater, and so on until all 96 raters were included. Fig 5 depicts these results, showing that index values changed little after about 40 raters. Again, this could also be used as an indication of the minimal number of raters required to produce relatively stable rankings of mElo scores.

thumbnail

Fig 5. Weighted consistency indices for the original sequence (small grey circles) and mean weighted consistency indices (large black circles) calculated for an increasing number of Study 1 raters (1–96, in increments of 1).

The whiskers represent inter-quartile ranges of indices from the 100 sequences generated to calculate the mean index.

https://doi.org/10.1371/journal.pone.0190393.g005

Comparing computation times for Bradley-Terry and mElo scores.

To compare the time it takes to compute scores via mElo and Bradley-Terry methods, we timed each procedure on the same laptop computer (2.2 GHz processor speed, 8 GB RAM), using R’s system.time function. The elapsed time for computing mElo scores for the entire data set was 5 seconds. The elapsed time for computing Bradley-Terry scores for the entire data set was 904 seconds.

Discussion

Likert ratings, Elo scores and Bradley-Terry scores were closely correlated with mElo scores, indicating that they were tracking perceptions similarly. Each measure of perceived strength was also similarly correlated with actual strength and strength-related measurements of the men depicted in the stimuli, indicating that these perceptions reliably track actual physical differences between men, as reported by Sell et al. [36]. In addition, these relationships were observed in both the laboratory, using a primarily UK undergraduate participant base, and online, using a more diversely aged US participant base. On the basis of these facts alone, there is little to separate one measure from another, and little reason to favour one above the other options. However, we argue that mElo scores represent a good alternative option to Elo scores, Bradley-Terry scores and Likert ratings, for distinct reasons.

Figs 1 and 4 demonstrates the range of Elo scores that can result from different sequences of the same trial comparisons with the same outcomes, and the range of correlations with grip strength and PC1 of body size (Study 1) demonstrates why this is problematic. As argued earlier, this variation in final scores represents an undesirable feature in the context of our rating temporally stable features, and we proposed mElo scores to address this issue. When multiple mElo scores are generated, each using the same trial data but using different randomly shuffled sequence orders, the results are far less variable. Each mElo score was almost perfectly correlated with all others, and there was no detectable variation between them when predicting actual strength and strength-related measurements. This was even the case for the one mElo score that was generated using 1000 different sequence orders, which indicates that, in this case at least, using 100 sequence orders is sufficient and saves computing time without cost.

We argue that the mElo approach is conceptually simpler than standard Bradley-Terry models, and thus more likely to appeal to researchers who are new to pairwise comparison. Our results suggest that mElo is computationally simpler as well, calculating score in 5 seconds or less, whereas Bradley-Terry models took over 15 minutes. This could be related to the number of stimuli used in our example, so studies using an even greater number of stimuli may suffer even greater time disparities.

Although results from Likert ratings and mElo scores were largely similar for this task, the procedure for obtaining them was quite different, leading to a very divergent participant experience. Participants completed pairwise comparison trials significantly quicker than rating trials, and the effect size for this difference was large (see Fig 3). Taking response time as a proxy of cognitive load [46], the speed difference between the trial types indicates that participants found the pairwise comparisons to be easier. This is likely to be a particularly important consideration for more difficult tasks (e.g. rating characteristics which are not readily apparent, such as personality traits from faces [47], or kinship in monkeys [48]), or for participant groups that have either compromised capacity or limited experience with stimulus tasks, particularly those involving novel concepts.

In addition, because participants complete pairwise comparison tasks more quickly, they might be willing to be paid less to provide similarly useful information. This is not the only way in which pairwise comparisons may prove to be more efficient. Note that correlations with actual strength and strength-related measures deteriorated as more raters were excluded for both Likert ratings and mElo scores, despite strong correlations with ratings and scores from the full participant set. This indicates that collecting data from more participants is likely to beneficial for both methods. However, also note that the mElo scores obtained from the halved set had strikingly similar associations with stimulus variables as Likert ratings from the full set did, and similarly mElo scores from the quartered set produced similar associations as Likert ratings from the halved set (see Table 2). If this pattern is typical of other data sets, this suggests that only half as many pairwise comparison participants are required to produce results equivalent to those from ratings tasks. This could represent a considerable savings for researchers both in time and participant payments.

In conclusion, using pairwise comparisons in conjunction with the “EloChoice” package represents a simple alternative option for researchers interested in quantifying perceived characteristics of stimuli, offering participants a less demanding experience while requiring fewer of them, and representing a minimal learning cost even for researchers inexperienced with R.

Acknowledgments

This research was supported by a research project grant awarded to A.P.C. and I.S.P-V from the Leverhulme Trust. Thanks to Tim Fawcett for contributing idea to plot consistency by number of raters.

References

  1. 1.Brown TA, Cash TF, Noles SW. Perceptions of physical attractiveness among college students: selected determinants and methodological matters. The Journal of Social Psychology. 1986; 126(3): 305–316.
  2. 2.Perrett DI, May KA, Yoshikawa S. Facial shape and judgements of female attractiveness. Nature. 1994; 368(6468): 239–242. pmid:8145822
  3. 3.Brooks RC, Shelly JP, Jordan LA, Dixson BJW. The multivariate evolution of female body shape in an artificial digital ecosystem. Evolution and Human Behavior. 2015; 36(5): 351–358.
  4. 4.Likert R. A technique for the measurement of attitudes. Archives of Psychology. 1932; 22(140): 5–55.
  5. 5.Uebersax JS. Likert scales: dispelling the confusion. Statistical Methods for Rater Agreement; 2015. http://john-uebersax.com/stat/likert.htm.
  6. 6.Cunningham MR, Roberts AR, Barbee AP, Druen PB, Wu C-H. “Their ideas of beauty are, on the whole, the same as ours”: Consistency and variability in the cross-cultural perception of female physical attractiveness. Journal of Personality and Social Psychology. 1995; 68(2): 261–279.
  7. 7.Mellor D, Moore KA. The use of Likert scales with children. Journal of pediatric psychology. 2014; 39(3): 369–379. pmid:24163438
  8. 8.Dressler WW, Oths KS. Social survey methods. In: Bernard HR, Gravlee CC, editors. Handbook of methods in cultural anthropology. 2nd ed. Lanham, MD USA: Altamira Press; 2014. p. 497–515.
  9. 9.Böckenholt U. Thresholds and intransitivities in pairwise judgments: A multilevel analysis. Journal of Educational and Behavioral Statistics. 2001; 26(3): 269–82.
  10. 10.Thurstone LL. A law of comparative judgment. Psychological Review. 1927; 34(4): 273–286.
  1. 11.Bradley RA, Terry ME. Rank analysis of incomplete block designs: I. The method of paired comparisons. Biometrika. 1952; 39(3/4): 324–345.
  1. 12.Glickman ME. Parameter estimation in large dynamic paired comparison experiments. Journal of the Royal Statistical Society: Series C (Applied Statistics). 1999; 48(3): 377–394.
  1. 13.Shev A, Hsieh F, Beisner B, McCowan B. Using Markov chain Monte Carlo (MCMC) to visualize and test the linearity assumption of the Bradley–Terry class of models. Animal Behaviour. 2012; 84(6): 1523–1531. pmid:24052665
  1. 14.Maydeu-Olivares A, Böckenholt U. Structural equation modeling of paired-comparison and ranking data. Psychological methods. 2005; 10(3): 285. pmid:16221029
  1. 15.Strobl C, Wickelmaier F, Zeileis A. Accounting for individual differences in Bradley-Terry models by means of recursive partitioning. Journal of Educational and Behavioral Statistics. 2011; 36(2): 135–53.
  1. 16.Tsukida K, Gupta MR. How to analyze paired comparison data. UWEETR-2011-0004. University of Washington Seattle Department of Electrical Engineering; 2011.
  2. 17.Zermelo E. Die Berechnung der Turnier-Ergebnisse als ein Maximumproblem der Wahrscheinlichkeitsrechnung. Mathematische Zeitschrift. 1929; 29(1): 436–460.
  1. 18.Elo AE. The rating of chess players past and present. New York: Arco; 1978.
  2. 19.Glickman ME. Parameter estimation in large dynamic paired comparison experiments. Applied Statistics. 1999; 48(3): 377–94.
  1. 20.Herbrich R, Minka T, Graepel T. Trueskill™: A Bayesian skill rating system. Advances in neural information processing systems. 2006; 19: 569–576.
  1. 21.Elliott LL. Reliability of judgments of figural complexity. Journal of Experimental Psychology. 1958; 56(4): 335. pmid:13587863
  1. 22.Mueser KT, Grau BW, Sussman S, Rosen AJ. You’re only as pretty as you feel: facial expression as a determinant of physical attractiveness. Journal of Personality and Social Psychology 1984; 46(2): 469–478.
  1. 23.Glickman ME. A comprehensive guide to chess ratings. American Chess Journal. 1995; 3: 59–102.
  1. 24.Hvattum LM, Arntzen H. Using ELO ratings for match result prediction in association football. International Journal of forecasting. 2010; 26(3): 460–70.
  1. 25.Doebler P, Alavash M, Giessing C. Adaptive experiments with a multivariate Elo-type algorithm. Behavior research methods. 2015; 47(2): 384–94. pmid:24878597
  1. 26.Ziemba B. The 2014–2015 NFL Season, Playoffs, and the Super Bowl. Wilmott. 2015; 2015(77): 24–43.
  1. 27.Goodspeed R. Research note: An evaluation of the Elo algorithm for pairwise visual assessment surveys. Landscape and Urban Planning. 2017; 157: 131–7.
  1. 28.Luon Y, Aperjis C, Huberman BA. Rankr: A mobile system for crowdsourcing opinions. InInternational Conference on Mobile Computing, Applications, and Services 2011 Oct 24 (pp. 20–31). Springer Berlin Heidelberg.
  2. 29.Neumann C, Duboscq J, Dubuc C, Ginting A, Irwan AM, Agil M, Widdig A, Engelhardt A. Assessing dominance hierarchies: validation and advantages of progressive evaluation with Elo-rating. Animal Behaviour. 2011; 82(4): 911–921.
  1. 30.R Core Team. R: A language and environment for statistical computing (version 3.2.1). Vienna, Austria: R Foundation for Statistical Computing; 2015. https://www.R-project.org.
  2. 31.Albers P, de Vries H. Elo-rating as a tool in the sequential estimation of dominance strengths. Animal Behaviour. 2001; 61: 489–495.
  1. 32.Franz M, McLean E, Tung J, Altmann J, Alberts SC. Self-organizing dominance hierarchies in a wild primate population. Proceedings of the Royal Society B: Biological Sciences. 2015; 282(1814): 20151512. pmid:26336168
  1. 33.Reid DA, Nixon MS, Stevenage SV. Soft biometrics; human identification using comparative descriptions. Pattern Analysis and Machine Intelligence, IEEE Transactions on. 2014; 36(6): 1216–28.
  1. 34.Hallgren KA. Computing inter-rater reliability for observational data: an overview and tutorial. Tutorials in quantitative methods for psychology. 2012; 8(1): 23. pmid:22833776
  1. 35.Gwet KL. Handbook of inter-rater reliability: The definitive guide to measuring the extent of agreement among raters. Advanced Analytics, LLC; 2014.
  2. 36.Sell A, Cosmides L, Tooby J, Sznycer D, von Rueden C, Gurven M. Human adaptations for the visual assessment of strength and fighting ability from the body and face. Proceedings of the Royal Society B. 2009; 276: 575–584. pmid:18945661
  1. 37.Lê S, Josse J, Husson F. FactoMineR: an R package for multivariate analysis. Journal of Statistical Software. 2008; 25(1): 1–18.
  1. 38.Neumann C. EloChoice: Preference rating for visual stimuli based on elo ratings (version 0.29). http://cran.r-project.org/web/packages/EloChoice.
  2. 39.Zeileis A, Strobl C, Wickelmaier F, Komboz, B, Kopf J. psychotools: Infrastructure for Psychometric Modeling (version 0.4–2). http://cran.r-project.org/web/packages/psychotools.
  3. 40.Diedenhofen B, Musch J. cocor: A comprehensive solution for the statistical comparison of correlations. PLoS One. 2015; 10(4): e0121945. pmid:25835001
  1. 41.Williams EJ. The comparison of regression variables. Journal of the Royal Statistical Society, Series B. 1959; 21(2): 396–399.
  1. 42.Gamer M, Lemon J, Fellows I, Singh P. irr: Various coefficients of interrater reliability and agreement (version 0.84). http://cran.r-project.org/package=irr.
  2. 43.Lakens D. The perfect t-test (version 0.1.0). 2015. http://github.com/Lakens/perfect-t-test.
  3. 44.Crump M, McDonnell J, Gureckis T. Evaluating Amazon’s Mechanical Turk as a tool for experimental behavioral research. PLoS One. 2013; 8(3): e57410. pmid:23516406
  1. 45.Ramsey SR, Thompson KL, McKenzie M, Rosenbaum A. Psychological research in the internet age: The quality of web-based data. Computers in Human Behavior. 2016; 58: 354–360.
  1. 46.Luce RD. Response times: Their role in inferring elementary mental organization. New York: Oxford University Press; 1986.
  2. 47.Penton-Voak IS, Pound N, Little AC, Perrett DI. Personality judgments from natural and composite facial images: More evidence for a “kernel of truth” in social perception. Social Cognition. 2006; 24(5): 607–40.
  1. 48.Kazem AJ, Widdig A. Visual phenotype matching: cues to paternity are present in rhesus macaque faces. PLoS One. 2013; 8(2): e55846. pmid:23451032