Mapping the Bid Behavior of Conference Referees

05/24/2006 ∙ by Marko A. Rodriguez, et al. ∙ Los Alamos National Laboratory 0

The peer-review process, in its present form, has been repeatedly criticized. Of the many critiques ranging from publication delays to referee bias, this paper will focus specifically on the issue of how submitted manuscripts are distributed to qualified referees. Unqualified referees, without the proper knowledge of a manuscript's domain, may reject a perfectly valid study or potentially more damaging, unknowingly accept a faulty or fraudulent result. In this paper, referee competence is analyzed with respect to referee bid data collected from the 2005 Joint Conference on Digital Libraries (JCDL). The analysis of the referee bid behavior provides a validation of the intuition that referees are bidding on conference submissions with regards to the subject domain of the submission. Unfortunately, this relationship is not strong and therefore suggests that there exists other factors beyond subject domain that may be influencing referees to bid for particular submissions.



There are no comments yet.


page 7

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

1 Introduction

The peer-review process is the most widely accepted method for validating research results within the scientific community. However, its credibility as a valid certification mechanism has come under scrutiny. There exists a rich body of literature that points to many of the inadequacies of the current system [EvansEvans1995, El-MunchidEl-Munchid2001, Bence OppenheimBence Oppenheim2004], but of particular interest to this paper is the issue concerned with ensuring that referees are in fact reviewing manuscripts within their domain of expertise [Kassirer CampionKassirer Campion1994, EisenhartEisenhart2002]. There exists a series of stages within the peer-review process that ultimately lead up to a referee review. One of the first and potentially most important stage is the one that attempts to distributed submitted manuscripts to competent referees. Unfortunately, it is difficult to study many of the stages of the peer-review process due to its confidential nature. Therefore, much of the peer-review process, including referee assignment, remains sheltered from the rigors of the scientific method. Fortunately, the program chairs and steering committee of the 2005 Joint Conference on Digital Libraries111JCDL 2005 is located at: (JCDL) has provided the Los Alamos National Laboratory (LANL) Digital Library Research and Prototyping team the referee bid data used for their 2005 conference peer-review process so that referee assignment could be analyzed for this study.

In conference situations, where there exist a large number of submissions at one particular point in time (near the submission deadline date), conference organizers tend to rely on a pool of pre-selected referees to review the submission archive. The conference organizers require each referee to briefly look over each submission (e.g. read each submission abstract or ACM classification codes) and place submission bids. A referee bid states the referee’s subjective opinion of their level of expertise with regards to a submission. Furthermore, conflict of interest situations are usually identified at this point. Once all the referee bids have been collected, the conference organizers can use any number of the many documented manuscript-to-referee matching algorithms to distribute each submission to a set of competent referees [Wei, Hartvigsen CzuchlewskiWei 1999]. These stages are represented in Figure 1. The data set provided by the 2005 JCDL program chair does not state which referees reviewed which submission, only the subjective opinion of the referee’s level of expertise with respect to each submission.

Figure 1: Typical conference review stages

Since conference organizers ask their referees to bid on submissions with regard to their domain of expertise, it is hypothesized that referee bidding is based on two factors: 1) the subject domain of the submission and 2) the expertise of the referee. The validity of this hypothesis is investigated using various statistical techniques that rely on a keyword analysis of submission abstracts and the location of each referee within the greater scientific community’s co-authorship network. In short, the analysis demonstrates that the referees of the 2005 JCDL program committee are, in fact, bidding for submissions with respect to the subject domain of the submissions. Unfortunately, the strength of this relationship is not strong enough to conclude that submission subject domain is the only, or even the most significant, factor influencing referee bidding behavior.

2 The 2005 JCDL Bid Data Set

The JCDL is an international forum that focuses on the technical, practical, and social issues concerning digital libraries. Each year the JCDL hosts a conference to present technical papers, posters, demonstrations, tutorials, etc. that present recent developments in the digital library community. From June 7 to June 10 of 2005, the JCDL was held in Denver, Colorado in the United States [SumnerSumner2005]. The bid data provided by the 2005 JCDL program chair is considered extremely sensitive, therefore careful handling and analysis of this data was the first priority of this research endeavor. All information that is not publicly available from the JCDL website is, to the best of our knowledge, indeterminable from the presented results. Information such as which submissions were rejected is not provided. The names of the referees have been anonymized by assigning each referee a unique random identifier. This section will discuss the bid data provided by the 2005 JCDL program chair as well as the various manipulations necessary to appropriately represent this information for analysis.

There were submissions to the 2005 JCDL. Of those submissions, were full technical articles, were short technical articles, were posters, were demonstrations, were panel talks, were tutorials, were workshop talks, and were doctoral presentations. The JCDL program committee provided the authors a table containing each submission’s unique identification number, title, authors, type, and acceptance/rejection status. An example subset of this data is provided in Table 1. The submission titles and authors of those submission that were rejected by the committee have been replaced with the ### notation in order to protect the privacy of the submitters. Since accepted submissions are freely accessible, information pertaining to accepted publications is provided222JCDL 2005 proceedings located at: ley/db/conf/jcdl/jcdl2005.html. Furthermore, note that the title and authors have been truncated to ensure that the table fits within the margins of this paper.

sub id submission title submission authors submission type submission status
13 ### ### Full Technical Article REJECTED
14 ### ### Full Technical Article REJECTED
15 Creating an Infrastructure for Collaboration… R. David Lankes, … Short Technical Article ACCEPTED
16 Graph-based Text Representation Model… Hidekazu Nakawatase, … Full Techinical Article ACCEPTED
17 An Evaluation of Automatic Ontologies… Aaron Krowne, … Full Techinical Article ACCEPTED
Table 1: Sample of the 2005 JCDL submission data

Each referee on the 2005 JCDL program committee was asked to bid on which submissions they wished to review in terms of their expertise in the subject domain of the submission. Therefore, accompanying the submission data table there also exists an associated bid matrix, , where is the set of submissions, is the set of referees, and . It is important to note that . The rows of the bid matrix refer to the unique id of each of the submissions. The columns of the bid matrix refer to the referees of the program committee. The matrix entries are the bid values provided by each referee for each submission. Therefore, refers to referee ’s bid for submission , where . Table 2 is an artificial example of the supplied bid information. Note that the bid values for Table 2 were randomly generated and the referee names are not provided. The actual program committee for the JCDL is public information333JCDL 2005 program committee available at:

, but their respective bid vectors are not.

sub/ref 1 2 3 4 5
13 1 2 2 3 3
14 2 3 2 3 3
15 4 2 3 1 1
16 3 3 1 2 0
17 1 3 2 3 3
Table 2: Example bid matrix for each submission for each program committee referee,

The values of the bid matrix, , are not on an interval scale, but instead are nominal (i.e. each value symbolizes a particular bid type). Table 3 provides the meaning for each of the bid values.

bid meaning of the bid value
0 did not provide a bid
1 expert in the domain of the submission and wants to review
2 expert in the domain of the submission
3 not an expert in the domain of the submission
4 conflict of interest between referee and submission
Table 3: The meaning of the bid values within the bid matrix,

The bid matrix provided by the 2005 JCDL, , contains extraneous information such as ’wants to review’ () and ’conflict of interest’ (). Since this study focuses specifically on referee expertise, this information will be discarded. Therefore, bid categories and will be considered the same and bid categories and will be considered wildcards. The modified bid matrix used throughout the remainder of this study has the properties of where . Table 4 has the bid meanings of the modified bid matrix.

bid meaning of the bid value
0 unknown expertise (wildcard)
1 expert in the domain of the submission
2 not an expert in the domain of the submission
Table 4: The meaning of the bid values within the modified bid matrix,

The original artificial bid matrix provided in Table 2 is thus transformed into the one shown in Table 5.

sub/ref 1 2 3 4 5
13 1 1 1 2 2
14 1 2 1 2 2
15 0 1 2 1 1
16 2 2 1 1 0
17 1 2 1 2 2
Table 5: Example modified bid matrix for each submission for each program committee referee,

Of the submissions, only of the submissions have actual bid data. This means that submissions had bids of all . Therefore, only those submissions with a complete set of bid data will be analyzed for the remainder of this study. In addition, of the program committee members of the JCDL, members gave no bid information. No bid information is defined as an individual whose bid vector is all ’s. These referees were removed from the analysis. Finally, since a portion of this analysis is based on co-authorship behavior, those referee committee members not located within the DBLP444Digital Bibliography and Library Project available at: ley/db/ were not included in this study. Of the remaining referees, were not in the DBLP. Therefore, the bid matrix as defined for the remainder of this study has rows (submissions), and columns (referees), .

3 The Methodology

Intuitively, when ignoring conflict of interest situations, referee bidding should be based on two factors: 1) the domain of the submission and 2) the domain of expertise of the referee. Therefore, the referee bid matrix should be the result of each referees analysis of the submission abstracts and the referee’s area of expertise (their location in the scientific community’s co-authorship network). This idea, which is the hypothesis of this study, is represented by the arced dotted lines at the top of Figure 2. To verify or falsify this hypothesis, a collection of statistical techniques are used to determine the relationship between referee bidding and submission subject domain. The two factors of the hypothesis are explored according to Track and Track of Figure 2.

Track provides a correlation between two submission similarity matrices. The first similarity matrix is constructed using referee bid data, , and the second is constructed according to an submission abstract term analysis, (Section 4). If referees are in fact bidding according to the subject domain of the submissions, then the correlation between and should be high. If the correlation is negative, or extremely low, then other factors that may not include submission subject domain are influencing referee bidding. Furthermore, it is possible to cluster submissions according to referee bid behavior. A intra-term analysis of these clusters provide an entropy value for each of the clusters. If the clusters created by referee bidding maintain a low entropy for their highest weighted terms and a low correlation between their term vectors, then it can be argued that referee bidding is driven by submission subject domain.

Track provides the correlation between a referee similarity matrix created according to referee bidding behavior, , and a referee similarity matrix created using a relative rank algorithm within a co-authorship network, (Section 5). A high correlation means that referees who are similar in expertise, as determined by their place in the co-authorship network, are also bidding similarly. A high correlation would be expected if referee bidding is based solely on submission subject domain. If this correlation is low, then other factors besides submission subject domain are influencing referee bidding. This paper will first explore Track and then Track .

Figure 2: Experiment outline

4 The Bid Matrix and Submission Similarity

This section will present the Track analysis represented in Figure 2. In order to determine the relationship between referee bidding and submission subject domain, the submissions are related according to the bid behavior of the program committee referees, , and are related according to their abstract term-frequency inverse document-frequency (TFIDF) term weight distributions, [SaltonSalton1998]. In short, a TFIDF calculation determines the most descriptive words within a document (or document cluster) with respect to the entire document corpus. This section will first discuss the construction of and then .

Since the values of the bid matrix, , refer to semantic categories and not a gradient scale, a Hamming distance function is used to determine the similarity of any two submissions [HammingHamming1950]. Hamming distance is defined as the amount of characters that differ between two strings of equal length. For example, if there exists the strings ”2212” and ”1212”, the Hamming distance is since only their first characters differ. Given the Hamming distance between two bid vectors, , and the length of a vector, , the similarity between any two submissions is calculated according to Eq. 1. To account for wildcard bids (), if any one of the two bid vectors being compared has an entry that contains a , that particular entry on both vectors is ignored and both their vector lengths, , are reduced by . For example, when comparing the two bid vectors ”0121” and ”2120”, their length, , is and their Hamming distance, , is because both their first and last entries are ignored and their second and third entries are equal. Therefore, their similarity is .


Eq. 1 ensures a symmetrical submission similarity matrix, , whose diagonal values are . According to the sample bid matrix presented in Table 5, the submission similarity matrix shown in Table 6 is constructed using Eq. 1.

id 13 14 15 16 17
13 1.0 0.8 0.25 0.25 0.8
14 0.8 1.0 0.0 0.5 1.0
15 0.25 0.0 1.0 0.0
16 0.25 0.5 1.0 0.5
17 0.8 1.0 0.0 0.5 1.0
Table 6: Submission similarity determined according to their Hamming distance

4.1 Submission Similarity and the Dendrogram

Once a submission similarity matrix, , has been constructed it is possible to hierarchically structure the submissions into a dendrogram in order to visualize the relationship between the various submissions. The submission dendrogram constructed from is presented in Figure 3. Note that the titles of the rejected submissions have been left out. Accepted submission titles have been truncated to ensure readability. Furthermore, larger cluster patterns are represented as the boxed sections and are denoted C1 through C8. These clusters were extracted from the dendrogram by setting a threshold on the dendrogram tree height. The threshold, which is , was arbitrarily selected to expose enough clusters to make the following analysis interesting.

Figure 3:

Submission similarity represented according to a hierarchical cluster

A manual review of the clusters with respect to the submissions they contain demonstrates a congruency between submission topic and referee bidding. To validate this qualitative claim, three statistical techniques are used. The first involves analyzing the abstracts of the submissions of each of the clusters in order to determine cluster subject domain. The second involves determining the entropy value of each cluster. Clusters that are more strict with respect to a particular subject domain will tend to have a lower entropy. The third technique provides a correlation between a similarity matrix constructed from the cosine similarity of the TFIDF term weight vectors of each submission,

, and the matrix constructed from the referee bid behavior, . This final correlation provides a single quantitative value expressing the relationship between submission subject domain and referee bidding.

4.2 Entropy in the Submission Clusters

The major clusters of the submission dendrogram derived according to referee bid data can be validated as meaningful categorizations by analyzing the terms of the submission abstracts. This requires that all submission abstracts be parsed to determine the full collection of keywords across all submission abstracts. Each abstract is processed by removing stop words and then applying the Porter stemming algorithm [PorterPorter1980]. These two processes remove overly frequent words (i.e. the, and, it) and perform suffix stripping (i.e. computer and computation stem to comput), respectively. Each time a particular term in the full collection of keywords is used in an abstract of one of the cluster submissions, the term frequency for that term in that cluster is incremented by . An example feature vector is provided in Table 7. For example, for all the submissions in cluster , the term built was used times.

cluster/term browser built bureau bush
3 3 7 3 1
4 4 3 2 0
5 1 0 1 0
Table 7: Cluster feature vectors of the keywords in the submission abstracts

For each cluster it is possible to determine how specific a particular term is to that cluster according to Eq. 2 where is the frequency of term in cluster , is the total number of terms in cluster , is the number of clusters (which is always for this experiment), and is the number of clusters for which term appears [SaltonSalton1998].


The higher the TFIDF weight for term in cluster , the more specific term is to the cluster and therefore the more suited it is as a description of the cluster’s subject domain. The following table presents the TFIDF calculations for the sample feature vector presented in Table 7.

cluster/term browser built bureau bush
3 0.00 0.08 0.00 0.03
4 0.00 0.05 0.00 0.00
5 0.00 0.00 0.00 0.00
Table 8: Cluster TFIDF term weight vectors of the keywords in the submission abstracts

In order to determine the subject domain of each of the clusters, the top 10 TFIDF weighted terms were extracted. Table 9 provides these terms ordered by their TFIDF weight where term has a higher weight than term .

C1 C2 C3 C4 C5 C6 C7 C8
1 webcast behaviour extract patent name hidden ecl photo
2 cyberinfrastructur drew powerpoint tobacco surrog crawler morf preserv
3 ncknow note handel invent music subschema relev video
4 interview overload mainli determin disambigu flora item-level european
5 teacher engag train hidden segment ontos network alert
6 descriptor factor graph control candid queri citat dark
7 faculti gather step american network expans circleview region
8 lesson school weight chemic tempor homepag dlii busi
9 survei teamsearch algebra compani citat plant extract addit
10 transcript visualis basi searchabl genet reusabl meta-inform mobil
Table 9: Top 10 terms for the 8 clusters defined in Figure 3

The term weight distributions derived from the TFIDF calculation of the cluster abstracts can now be represented according to their internal cluster information content. Internal cluster information content can be calculated using the standard entropy equation as defined according to its information theoretic sense [ShannonShannon1948]. The lower the entropy, the more specialized, or focused, the cluster. The higher the entropy, the less specialized. Since clusters vary in size, the entropy for a cluster is calculated only for the top term weights presented in Table 9

. Furthermore, since an entropy calculation is defined for a probability distribution, Eq.

3 normalizes the top term weights.


The entropy of a cluster is then calculated over the probability distribution as described by Eq. 4, where is the entropy for cluster .


The entropy values for the clusters of the dendrogram presented in Figure 3 are presented in Table 10.

cluster entropy
1 3.2668
2 3.2840
3 3.2148
4 3.2213
5 3.2025
6 3.2281
7 3.2610
8 3.2442
Table 10: Entropy values for the 8 clusters defined in Figure 3

It is interesting to note that C5, the lowest entropy cluster, is composed mainly of submissions associated with name disambiguation and music in digital-library research. On the other hand, the highest entropy cluster C2 has a mix of more unrelated submissions ranging from digital libraries in educational settings to infectious diseases and terrorism. Figure 4 and 5 present the distribution of the term weights for the top terms of the clusters. The steeper the distribution tail, the lower the cluster entropy and therefore the more focused the cluster is towards its higher weighted terms. The analysis of the terms for each cluster points to a qualitative relationship between referee bidding and submission subject domain.

Figure 4: Normalized TFIDF weights for the 10 terms of Table 9 for the clusters 1 through 4 defined in Figure 3
Figure 5: Normalized TFIDF weights for the 10 terms of Table 9 for the clusters 5 through 8 defined in Figure 3

A more quantitative validation can be determined when the TFIDF term weight vectors of the clusters are compared using a Pearson correlation. The correlations are performed on the cluster’s TFIDF term weight vectors which contain the entire abstract dictionary, , where . Table 11 provides the Pearson correlations for each cluster comparison. What is noticeable from Table 11 is that all the correlations are less than . The fact that the clusters, which are organized by referee bidding, yield very low correlations between their TFIDF term weight vectors means that the clusters are well separated according to their term distributions. If these correlations were high, then it would be difficult to claim that the clusters are organized according to subject domain and thus referee bidding would not be related to submission subject domain. Since the correlations are all less that , this confirms the hypothesis that there does exist a relationship between the bidding behavior of the conference referees and the subject domain of the submission abstracts. The next section will further explore the strength of this relationship.

C1 C2 C3 C4 C5 C6 C7 C8
C1 1.0 -0.0318 -0.0396 -0.0631 -0.0616 -0.0477 -0.0720 -0.0527
C2 -0.0318 1.0 -0.0214 -0.0639 -0.0392 -0.0502 -0.0323 -0.0804
C3 -0.0396 -0.0214 1.0 -0.0343 0.01064 0.02349 0.04540 -0.0321
C4 -0.0631 -0.0639 -0.0343 1.0 -0.0437 -0.0531 -0.0540 -0.0850
C5 -0.0616 -0.0392 0.01064 -0.0437 1.0 -0.0410 0.03928 -0.0684
C6 -0.0477 -0.0502 0.02349 -0.0531 -0.0410 1.0 -0.0156 -0.0780
C7 -0.0720 -0.0323 0.04540 -0.0540 0.03928 -0.0156 1.0 -0.0687
C8 -0.0527 -0.0804 -0.0321 -0.0850 -0.0684 -0.0780 -0.0687 1.0
Table 11: Pearson correlations for the TFIDF term weights of the 8 clusters defined in Figure 3

4.3 Cosine Similarity Correlation

To further quantify the relationship between referee bidding and a submission’s subject domain, it is possible to correlate the relationship between submissions based on referee bidding, on the one hand, and the relationship between submissions based on their TFIDF term weight vectors, on the other. This requires the construction of the similarity matrix , which denotes the cosine similarity between every submission with respect to their complete TFIDF term weight vector. This means that each term in the abstracts of each submission is analyzed according to the TFIDF equation presented in Eq. 2. This results in a matrix where is the size of the submission archive and is the size of the full collection of terms of all abstracts in the submission archive. For this particular experiment is therefore defined as . The TFIDF term weight vector of each submission can be compared against every other submission’s TFIDF term weight vector using the standard cosine similarity function presented in Eq. 5, where is the TFIDF term weight vector for submission . This equation guarantees a symmetrical matrix with a diagonal of .


The correlation between and

can now be calculated. With 13,922 degrees of freedom and a

-value , the Pearson correlation was determined to be . This means that submissions categorized according to a TFIDF analysis of their abstracts and submissions categorized according to the referee bid behavior are in fact positively correlated, though not strongly. Therefore, it can be concluded that there are other factors besides submission subject domain that influence referee bid behavior.

@R=0.1pt *+[F-]df = 13922, p ¡ 2.2^-16, r = 0.357

5 The Bid Matrix and Referee Similarity

This section will overview the experiment as described by Track of Figure 2. If referees are deemed similar in expertise, as determined by their relative location to one another within the scientific community’s co-authorship network, then similar referees should be bidding similarly. To test this hypothesis, two referee similarity matrices are created. The first referee similarity matrix, , is constructed from the transpose of the modified bid matrix, . Each referee is compared to each other referee with respect to their bidding behavior. Based on the transpose of the artificial data from Table 2, the same similarity equation used to construct the submission similarity matrix, Eq. 1, can be used to construct the referee similarity matrix presented in Table 12. The next section will present a dendrorgam of before discussing the second referee similarity matrix, .

ref 1 2 3 4 5
1 1.0 0.5 0.75 0.0 0.0
2 0.5 1.0 0.2 0.4 0.75
3 0.75 0.2 1.0 0.2 0.0
4 0.0 0.4 0.2 1.0 1.0
5 0.0 0.75 0.0 1.0 1.0
Table 12: Referee similarity determined according to their Hamming distance

5.1 Referee Similarity and the Dendrogram

Given the referee similarity matrix, , the dendrogram in Figure 6 can be constructed. Unfortunately, due to privacy issues, the referee names are not provided. What is noticeable from the dendrogram is the collection of nearly identical referees on the upper branch. When reviewing the modified bid matrix, , it becomes apparent that of the referees stated themselves to be expert in the domain of every submission (excluding their wildcard bids).

Figure 6: Referee similarity represented according to a hierarchical cluster

5.2 Relative-Rank Correlation

In order to provide a quantitative evaluation of the similarity of referees with respect to their bidding behavior and their domain of expertise, a relative-rank algorithm within a co-authorship network is computed to determine referee similarity. It has been widely accepted that co-authorship networks represent the relationship of individuals with respect to their domain of expertise [NewmanNewman2004]. The relative-rank algorithm will determine the similarity of each referee with respect to each other referee as defined by their relative location to one another within the greater scientific community’s co-authorship network. The similarity of the referees as determined by their relative-rank, , and their similarity as determined by their bid behavior, , can then be correlated. A high correlation means that referees of similar expertise are bidding in a similar manner. A low correlation means that referees of similar expertise are not bidding in a similar manner. The co-authorship network, , used for this experiment was constructed from the DBLP database as of October 2005. The DBLP co-authorship network has 284,082 nodes (authors) and 2,167,018 edges (co-authorship relationships). This section will first formalize the co-authorship network data structure and relative-rank algorithm before discussing the results.

A co-authorship network is defined by a graph composed of nodes that represent authors and edges that represent a joint publication. Therefore, a co-authorship network is represented by the tuple , where is the set of authors in the network, is the set of edges relating the various authors, and is the set of weights associated with the strength of tie between any two collaborating authors. Any edge, , connects two authors, and , with a respective weight of . Furthermore, and . The edge weight between any two authors is determined by Eq. 6, where the summation is over the set of all manuscripts registered with the DBLP, , expressing a collaboration between authors and , and the function returns the total number of authors for manuscript , where and [Liu, Bollen, Nelson SompelLiu 2005, NewmanNewman2001].


To provide the reader with an understanding of the relationship between the 2005 JCDL program committee members, a subset of the DBLP co-authorship network which contains the program committee’s co-authorship relationships is presented in Figure 7. Note that this network was not constructed using referee bid data, but from information that is publicly available through the DBLP database. Furthermore, the co-authorship edge weights have been left out to improve readability.

Figure 7: Subset of the DBLP co-authorship network containing only connected JCDL referees

The final analysis to be performed is to rank each of the referees relative to one another so as to construct , Eq. 7. For each referee in the JCDL program committee that provided valid bid data and is located in the DBLP, a similarity value to every other member in the committee was computed using a relative-rank algorithm (sometimes called a ’personalized’ rank) [Rodriguez BollenRodriguez Bollen2005, White SmythWhite Smyth2003] within the DBLP co-authorship network. Since is a weighted graph, the ranking algorithm actually used in this experiment is the weighted relative-rank implementation described in [Rodriguez BollenRodriguez Bollen2005].


An example of relative-ranking is as follows. Given a network such as the one displayed in Figure 7, the relative-rank algorithm would rank FOX more strongly to NELSON than to RAY since there exists a clique relationship between FOX, NELSON, and their co-authors. This network structure does not exist between FOX and RAY. Since co-authorship networks relate individuals with respect to similar domains of expertise, the conclusion to be drawn is that the stronger ranking of FOX to NELSON implies that FOX is more related by expertise to NELSON than he is to RAY. A simplified version of the pseudo-code for constructing , Eq. 7, is presented in Algorithm 1. For a more indepth, and formal, review of relative-rank algorithms for network analysis, refer to [Rodriguez BollenRodriguez Bollen2005, White SmythWhite Smyth2003].

foreach do  foreach do  ; AlgoLine0.1
end end
Algorithm 1 Constructing the referee similarity matrix

Given and , with 3,598 degrees of freedom and a -value , the Pearson correlation was calculated to be . The positive correlation indicates that referees are bidding with respect to their domain of expertise, but the low correlation again hints that there may be other factors contributing to referee bidding.

@R=0.1pt *+[F-]df = 3598, p ¡ 2.2^-16, r = 0.220

6 Conclusion

This paper provided an exploration of the bidding behavior of the 2005 JCDL program committee. The various analysis techniques used demonstrate that the 2005 JCDL program committee did, in fact, bid for conference submissions with respect to the subject domain of the submission. On the other hand, the strength of this relationship is low and therefore demonstrates that other factors may be involved in referee bidding. One such factor seems to be referee fatigue. With submissions having no bid data and with referees stating themselves to be an expert in the domain of all submissions, human-driven referee bidding in conference settings may not be the most optimal technique for performing conference peer-review. Since bidding is the preliminary component of the manuscript-to-referee matching algorithm, sloppy bidding can have dramatic effects on which referees actually review which submissions, Figure 8. In general, the stages that follow from the inclusion of noisy data in the peer-review chain can severely effect the quality of the peer-review process. It is speculated that referee fatigue not only influences the bidding and manuscript dissemination stages of the review cycle, but potentially more damaging, fatigued referees could be rejecting acceptable manuscripts or accepting fraudulent or faulty manuscripts in the review stage.

Figure 8: Human-factor noise in review stages

When the referees that provided no variation in their bid vector (referees that stated themselves to be experts in the domain of all submissions) are removed from the analysis, the correlation between and was determined to be (originally ). Likewise, for and the correlation was determined to be (originally ). Though these are both higher correlations, they still are not strong correlations. It can only be concluded that, contrary to the hypothesis, submission subject domain and referee expertise are not the only factors involved in the referee bidding process. Future work in this area will focus on expanding this study’s hypothesis in order to develop a mathematical model of the factors influencing referee bidding. Furthermore, an application of this methodology to bid data from other conferences can help to provide a broader perspective (and confirmation) of the factors influencing submission bidding in the peer-review process.

7 Acknowledgments

This research could only have been conducted with the help of the 2005 JCDL program chairs (Mary Marlino, Tamara Sumner, Frank Shipman) and steering committee (Erich Neuhold). Furthermore, the DBLP has once again provided the authors of this paper a thorough dataset for exploring the structure of science. The authors would like to thank the producers of R Statistics, GraphViz, Dia, and LaTeX for their freely available software. Finally, Xiaoming Liu and Karin Verspoor contributed by reviewing drafts of this manuscript. This research was financially supported by the Los Alamos National Laboratory.


  • [Bence OppenheimBence Oppenheim2004] Bence, V. Oppenheim, C. 2004. The influence of peer review on the research assessment exercise. Journal of Information Science, 30(4), 347–368.
  • [EisenhartEisenhart2002] Eisenhart, M. 2002. The paradox of peer review: Admitting too much or allowing too little? Research in Science Education, 32(2), 241-255.
  • [El-MunchidEl-Munchid2001] El-Munchid, H. 2001. Evaluation of peer review in biomedical publications. Annals of Saudi Medicine, 21, 5–6.
  • [EvansEvans1995] Evans, P. 1995. The peer-review process. Literati Newsline: Special Issue for Authors and Editors, 2.
  • [HammingHamming1950] Hamming, R. W. 1950. Error-detecting and error-correcting codes. Bell System Technical Journal, 29(2), 147–160.
  • [Kassirer CampionKassirer Campion1994] Kassirer, J. P. Campion, E. W. 1994. Peer review: Crude and understudied, but indispensable. The Journal of the American Medical Association, 272, 96-97.
  • [Liu, Bollen, Nelson SompelLiu 2005] Liu, X., Bollen, J., Nelson, M. L. Sompel, H. V. de. 2005. Co-authorship networks in the digital library research community. Information Processing and Management, In press.
  • [NewmanNewman2001] Newman, M. E. J. 2001. Scientific collaboration networks: I. network construction and fundamental results. Physical Review, 64.
  • [NewmanNewman2004] Newman, M. E. J. 2004. Coauthorship networks and patterns of scientific collaboration. Proceedings of the National Academy of Science, 5200–5205.
  • [PorterPorter1980] Porter, M. F. 1980. An algorithm for suffix stripping. Program: Automated Library and Information Systems, 14(3), 130–137.
  • [Rodriguez BollenRodriguez Bollen2005] Rodriguez, M. A. Bollen, J. 2005. Simulating network influence algorithms using particle-swarms: Pagerank and pagerank-priors. [submitted].
  • [SaltonSalton1998] Salton, G. 1998. Term-weighting approaches in automatic text retrieval. Information Processing and Management, 24(5), 513–523.
  • [ShannonShannon1948] Shannon, C. E. 1948. A matematical theory of communication. Bell System Technical Journal, 27, 379-423.
  • [SumnerSumner2005] Sumner, T. 2005. Report on the fifth ACM/IEEE joint conference on digital libraries - cyberinfrastructure for research and education. D-Lib Magazine, 11(7/8).
  • [Wei, Hartvigsen CzuchlewskiWei 1999] Wei, J. C., Hartvigsen, D. Czuchlewski, R. 1999. The conference paper-reviewer assignment problem. Decision Science, 30(3).
  • [White SmythWhite Smyth2003] White, S. Smyth, P. 2003.

    Algorithms for estimating relative importance in networks.

    In Kdd ’03: Proceedings of the ninth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining ( 266–275). New York, NY, USA: ACM Press.