Analysis of scientific society-awarded honors reveals disparities

A DOI-citable version of this manuscript is available at https://doi.org/10.1016/j.cels.2021.07.007.

This manuscript (permalink) was automatically generated from greenelab/iscb-diversity-manuscript@23502a0 on September 7, 2021.

Authors

Highlights

Summary

Delivering a keynote talk at a conference organized by a scientific society, or being named as a fellow by such a society, indicates that a scientist is held in high regard by their colleagues. To explore if the distribution of such indicators of esteem in the field of bioinformatics reflects the composition of this field, we compared the gender, name-origin and country of affiliation of 412 honorees from the International Society for Computational Biology (75 fellows and 337 keynote speakers) with over 170,000 last authorships on computational biology papers between 1993 and 2019. The proportion of honors bestowed on women was similar to that of the field’s overall last authorship rate. However, names of East Asian origin have been persistently underrepresented among honorees. Moreover, there were roughly twice as many honors bestowed on scientists with an affiliation in the United States as expected based on literature authorship.

A record of this paper’s Transparent Peer Review process is included in the Supplemental Information.

Introduction

Scientists’ roles in society include identifying important topics of study, undertaking an investigation of those topics, and disseminating their findings broadly. The scientific enterprise is largely self-governing: scientists act as peer reviewers on papers and grants, comprise hiring committees in academia, make tenure decisions, and select which applicants will be admitted to doctoral programs. A lack of diversity in science could lead to pernicious biases that hamper the extent to which scientific findings are relevant to minoritized communities. Furthermore, even though minoritized groups innovate at higher rates, their novel contributions are discounted [1]. One first step to address this systemic issue is to directly examine peer recognition in different scientific fields.

Gender bias among conference speakers has been recognized as an area that can be improved with targeted interventions [2,3,4,5]. Having more female organizers on conference committees is associated with having more female speakers [6]. At medical conferences in the US and Canada, the proportion of female speakers is increasing at a modest rate [7]. Gender bias appears to also influence funding decisions: an examination of scoring of proposals in Canada found that reviewers asked to assess the science produced a smaller gender gap in scoring than reviewers asked to assess the applicant [8].

Challenges extend beyond gender: an analysis of awards at the NIH found that proposals by Asian, black or African-American applicants were less likely to be funded than those by White applicants [9]. There are also potential interaction effects between gender and race or ethnicity that may particularly affect women of color’s efforts to gain NIH funding [10]. Another recent analysis found that minority scientists tend to apply for awards on topics with lower success rates [11]. This finding might be the result of minority scientists selecting topics in more poorly funded areas. Alternatively, reviewing scientists may not recognize the scientific importance of these topics, which may be of particular interest to minority scientists.

We sought to understand the extent to which honors and high-profile speaking invitations were distributed equitably among gender and name origin groups by an international society and its associated meetings. As computational biologists, we focused on the International Society for Computational Biology (ISCB), its honorary Fellows as well as its affiliated international meetings that aim to have a global reach: Intelligent Systems for Molecular Biology (ISMB) and Research in Computational Molecular Biology (RECOMB).

Existing methods were relatively US-centric because most of the data was derived in whole or in part from the US Census. We scraped more than 700,000 entries from English-language Wikipedia that contained nationality information and built machine learning classifiers to predict name origin. We also examined last authorships for more than 170,000 computational biology publications to establish a field-specific baseline using the same metrics. We used methods to predict the gender and name origins of honor recipients and also examined the affiliations of authorships and honor recipients. Analysis of affiliations by country revealed disparities between authorships and honoree affiliations. We also observed fewer honors to scientists with East Asian name origin than expected from authorship set, an effect that persisted even after we controlled for affiliation by restricting analysis to only US-affiliated scientists.

Results

We curated a dataset of ISCB honors that included 412 keynote speakers at international ISCB-associated conferences (ISMB and RECOMB) as well as ISCB Fellows. The ISCB Fellows set contained the complete set of Fellows named (2009–2019). Keynote speakers were available for ISMB for all years from 1993–2019. Keynote speakers for RECOMB were available for all years from 1997–2019. We included individuals who were honored multiple times as separate entries.

We sought to compare this dataset with a background distribution of potential speakers, which we considered to be last authors of bioinformatics and computational biology manuscripts (see Methods). We used authorships instead of authors as our metric for the field’s composition under the expectation that honorees would be drawn in a manner weighted by the number of last author contributions. Using authorships also does not require accurate name disambiguation, which is an open challenge. We scraped PubMed for manuscripts written in English from 1993–2019 with the MeSH term “computational biology”. We downloaded the metadata of manuscripts published in these journals from PubMed, which provided 176,110 articles for evaluation. For each article, we extracted its last author’s fore name and last name for analysis.

Defining metrics and a field-specific background

We were faced with a number of choices as we set out to examine representation within the field. Here, we outline the choices that we made and the rationale for these choices.

For one step, we needed to determine whether the most appropriate unit for analysis was researchers or honor and authorship events. We elected to perform the analysis at the level of honors and authorships, not individuals. Our rationale was that, if a scientist was honored three times, these three honoree slots represented distinct selection processes and should be considered separately. We also elected to use authorships instead of authors because name disambiguation can be error prone, and because the overall design of the work was to examine authorship distributions, the precise linking of researchers was not necessary to address the research question. As such, we estimated the gender and origin of each last name of each given paper and honoree slot, but not how many papers an individual scientist has authored or how many times they were invited as speakers or named fellows. We also needed to determine whether or not authorships should be weighted by some property of the resulting manuscript, for example the number of citations. We elected not to weight based on the concern that certain honors - in particular keynotes - could increase the exposure of the work and consequently increase citations which would add circularity to the analysis.

At another point, we needed to determine what the appropriate set of publications were to define a field-specific background distribution. Defining this distribution is key to determining the extent to which representation diverges from that background. Practical options included selecting authors in key society-endorsed journals or selecting those based on some article metadata. We elected to perform the analysis to all 176,110 PubMed articles with a computational biology MeSH term to generate the largest possible set of relevant literature.

We assumed that, among the authors of a specific paper, the senior author (often research advisors) would be most likely to be invited for a keynote or honored as a fellow. Based on field-specific conventions, we could have selected the last or corresponding author. We found that corresponding authors were often first and/or last suggesting that they were highly involved with more corresponding authors being last authors than first authors. Given the convention and the observation that corresponding authors were more often last, we selected the last author as the most appropriate in this setting.

We present the results with the selected strategy (Fig. 1); however, we performed this work through multiple iterations varying these parameters to examine the extent to which they influenced the results. Our findings across different combinations of the above choices were consistent with respect to the broad conclusions, though the numerical results differed (see Methods).

Figure 1: Study framework. We extracted full names and affiliations of the last authors of 176,110 computational biology PubMed articles and those of 412 honorees. We estimated the gender, name-origin group and country of affiliation of each scientist and compared the probability values between these two groups (see Methods).

Similar gender proportion between ISCB’s honorees and the field

We predicted the gender of honorees and authors using the https://genderize.io API, which was trained on over 100 million name-gender pairings collected from the web (see the STAR Methods for more details) and is one of the three widely-used gender inference services [12]. The predictions represent the estimated probability of an honoree or author being male or female based on their first name; we did not convert probabilities to a hard group assignment. For example, a query to https://genderize.io on January 26, 2020 for “Casey” returns a probability of male of 0.74 and a probability of female of 0.26, which we would add for an author with this first name. Because of technical limitations, our analysis only considered two binary gender categories, and we used “male” and “female” to refer to the gender of the scientists. However, as described in the Discussion, we recognize the limitation of not accounting for non-binary gender categories and only considered predictions in aggregate and not as individual values for specific scientists.

We observed a gradual increase of the proportion of predicted female authorships, arriving at an average of approximately 28% in 2017-2019 (Fig. 2, left). In recent years, ISCB Fellows and keynote speakers appear to have similar gender proportions compared to the population of authors published in computational biology and bioinformatics journals (averaged around 30% in the last three years, Fig. 2, right). Examining each honor category, we observed in 10.visualize-gender an increasing trend of honorees who were women, especially in the group of ISCB Fellows, which markedly increased after 2015. Through 2019, there were a number of years when meetings or ISCB Fellow classes have a high proportion of honors for male honorees and none that appeared to have exclusively female honorees. We sought to examine whether or not there was a difference in the proportion of female names between authorships and honors. A multiple logistic regression of this proportion for the groups and year did not reveal a significant difference (p = 0.19). Interaction terms did not predict the group of scientists over and above the main effect of gender probability and year.

Figure 2: Estimated composition of gender prediction over the years of all PubMed computational biology and bioinformatics journal authorships (left), and all ISCB honors (right). Male proportion (yellow) was computed as the average of the probability of being male of last authors (weight accordingly) or ISCB honorees each year. Female proportion (blue) was the complement of the male proportion. ISCB honors appear to have similar gender proportions compared to that of PubMed authorships.

Honorees with Celtic/English names are overrepresented while honorees with East Asian names are underrepresented

We inferred the geographical region of origin of authors’ names using a Long Short-Term Memory (LSTM) neural network trained on a dataset of 708,493 names called Wiki2019 (see the STAR Methods for details); the resulting model was called Wiki2019-LSTM. We found that the proportion of authorships with Celtic/English names had decreased (Fig. 3A, left). Among keynote speakers and fellows, we found that the majority were predicted to have Celtic/English or European names (Fig. 3A, right). When we directly compared honor composition with PubMed, we observed discrepancies between the two groups (Fig. 3B). Compared to other names, there was an overabundance of Celtic/English name among the honors (ORCeltic/English = 2.39, \(\beta\)Celtic/English = 0.86, p < 10-5). Meanwhile, an East Asian name has significantly lower odds of being selected for an honor compared to other names (OREast Asian = 0.17, \(\beta\)East Asian = -1.75, p < 10-5). The two groups of scientists did not have a significant association with names predicted to be European and in Other categories (p = 0.11 and p = 0.29, respectively). Interaction terms did not predict the group of scientists over and above the main effect of name origin probability and year.

Figure 3: (A) Estimated composition of name origin prediction over the years of all PubMed computational biology and bioinformatics journal authorships (left), and all ISCB honors (right). (B) For each region, the mean predicted probability of PubMed authorships is shown as the teal generalized additive model curve. The mean probability and 95% confidence interval of the ISCB honor predictions are shown as dark circles and vertical lines. Compared to PubMed authorships, honors were more frequently associated with Celtic/English names and less frequently with East Asian names. We did not observe a statistically significant difference in other categories (see STAR Methods Table 1).

Consistent disparities were also apparent with an alternative approach that considered each pair of forename and last name was the unit of measurement. In addition, a time-lagged model, which might be appropriate if we assume that honors accrue ten years after an author’s most prolifically cited year, results in a similar underrepresentation of East Asian scientists’ names in the group of honorees, though the effect size is smaller. For example, the proportion of honor associated with East Asian name origins in 2019 is still substantially less than the proportion of senior authorships associated with East Asian names in 2009. The 11.visualize-name-origins analysis notebook for this portion provides the results in these scenarios.

We sought to disentangle geography from other factors by examining results for the country with the most affiliated scientists receiving honors, the US. When applying the Wiki2019-LSTM model to the name origins of only US-affiliated scientists, we found a similar underrepresentation of honors to scientists with East Asian names (OREast Asian = 0.15, \(\beta\)East Asian = -1.89, p = 3.4×10-5). We observed no statistically significant difference between the proportion of honors given to authors with Celtic/English names, European names, or names in Other categories (p = 0.15, p = 0.02, and p = 0.65, respectively). Please see the 14.us-name-origin notebook for more details of the US-specific analysis.

Overrepresentation of US-affiliated honorees

We analyzed the countries of affiliation between last authorships and ISCB honors. For each country, we report a value of log2 enrichment (LOE) and its 95% confidence intervals (see Methods). The full table with all countries and their corresponding enrichment can be browsed interactively in the corresponding 12.analyze-affiliation analysis notebook. A positive value of LOE indicates a higher proportion of honorees affiliated with that country compared to authors. A LOE value of 1 indicates that observed number of honors is twice as much as expected. In the 20 countries with the most publications, we found an overrepresentation of honorees affiliated with institutions and companies in the US (119 speakers more than expected, LOE = 1.0, 95% CI (0.8, 1.2)) and Israel (15 speakers more than expected, LOE = 2.7, 95% CI (1.9, 3.4)), and an underrepresentation of honorees affiliated with those in China, France, Italy, India, South Korea, and Brazil (Fig. 4).

Figure 4: Each country’s log2 enrichment (LOE) and its 95% confidence interval (left), and the absolute difference between observed (triangle) and expected (circle) number of honors (right). Positive value of LOE indicates a higher proportion of honorees affiliated with that country compared to authors. Countries are ordered based on the proportion of authorships in the field. The overrepresentation of honorees affiliated with institutions and companies in the US and Israel contrasts the underrepresentation of honorees affiliated with those in China, France, Italy, India, South Korea, and Brazil.

Improvements to Honoree Diversity Subsequent to Our Primary Analysis

While our study was primarily designed to assess the diversity of honor recipients, the findings raise important questions about what can be done to address the disparities. We examined changes subsequent to our initial report for suggestions that increased awareness may drive improvements in the practice of honoree suggestion. We released version 1.0 of our manuscript on 2020-01-30. Early indications suggested an increased the diversity of honorees. In 2020, among 12 ISCB Fellows and 5 ISMB keynote speakers, the mean predicted probability of each honoree having an East Asian name was 33%, higher than any estimate in previous years (see the 15.analyze-2020 notebook). The set of honorees also included the first ISCB Fellow from China. Compared to past years, the 2020 honorees appeared to better reflect the diversity of scientists in the computational biology field. These new results suggested: 1) deserving honorees who were members of under-recognized groups existed but had not been recognized, and 2) examining honoree distribution’s alignment with the field may trigger changes that begin to address issues of unequal representation. However, we note that this analysis dealt only with more senior scientists (the last authors on scientific manuscripts) in the context of honors and that many years of changed honoree distributions will be required for the set of honored scientists to better reflect the field’s senior author contributions.

Discussion

There are significant technical and ethical challenges that one faces in carrying out retrospective work to examine the fairness of scientific practices. A major technical challenge was to narrow down geographic origins for some groups of names. Specific name origin groups, such as Hispanic names, are geographically disparate. We were unable to construct a classifier that could distinguish names from Iberian countries (Spain and Portugal) from those in Latin America in the group of Hispanic names. Discrepancies in representation between these groups are thus undetectable by our classifier. Honoree counts of those with Hispanic names are influenced from Spain as well as Latin America. In such cases, our analyses may understate the extent to which minoritized scientists are underrepresented among honorees and authors. Another technical challenge is that supervised machine learning approaches are neither error free nor bias free. By integrating different lines of evidence and preserving uncertainty by analyzing prediction probabilities rather than applying a hard assignment for each prediction, we aimed to alleviate method-specific biases and discover insightful findings that correctly reflect the current representation diversity at conferences.

A key ethical challenge with retrospective work to examine disparities is that algorithmic approaches to infer characteristics are often required. This leads to significant limitations such as considering gender as a binary variable. While this situation limits retroactive examination, there is a substantial opportunity for scientific societies, grant-making organizations, publishers, and others to proactively collect self-identified demographic information. It is equally crucial that this information is properly used to continuously evaluate inclusion practices.

A difficulty that straddles the ethical and technical divide for studies examining the representation of honorees is that the background that can be best assessed, which we assess here, is the current field composition. Scientific societies exist to promote the discipline, and many, including ISCB, include diversity as a value. In these cases, the ideal background distribution would be the population of senior computational biologists in the absence of systemic barriers to participation. The implication would be that scientific societies exist to reflect what the field could be, not just what it is. However, we are limited to measuring the field as it is. Furthermore, authorships, which we use to assess the field’s composition, are also affected by systemic barriers to participation. We estimated the composition of the field using last author status, but in neuroscience [13] and other disciplines [14], women are underrepresented in this position. Such an effect would cause us to underestimate the number of women in the field. Similarly, other studies have showed that underrepresented groups are less likely to be last authors [15], and Hispanic and Black scientists were underrepresented in academic publishing in general [16]. Thus, systemic barriers that reduce representation within our estimation of the field would reduce apparent disparities in honor distributions as long as those systemic barriers did not also have a particular influence on the honoree selection process as well.

An important ethical question to ask when measuring representation is what the right level of representation is. Societies should examine their processes to determine whether the process of selecting honorees should be equal or equitable. For example, we found similar representation of women between authors and honorees, which suggests honoree diversity is similar to that of authors and that there may be equality during the honoree selection process. However, if fewer women are in the field because of systemic factors that inhibit their participation, reaching equality is not equivalent to reaching equity. In addition to holding fewer corresponding authorship positions, on average, female scientists of different disciplines are cited less often [17,18,19,20], invited by journals to submit papers less often [14], suggested as reviewers less often [21], and receive significantly worse review scores [18]. Meanwhile, a review of women’s underrepresentation in math-intensive fields argued that today’s underrepresentation is not explained by historic forms of discrimination but factors surrounding fertility decisions and lifestyle choices, whether freely made or constrained by biology and society [22]. A recent analysis of gender inequality across different disciplines showed that, although both gender groups have equivalent annual productivity, women scientists have higher dropout rates throughout their scientific careers [23]. Therefore, although we found that ISCB’s honorees and keynote speakers appear to have similar gender proportion to the field as a whole, the gender proportions have not reached parity. To the extent that this gap is due to systemic barriers, the process may have reached equality but not equity.

It is also possible to have processes that reach neither equality nor equity. We find that honorees include significantly fewer people of color than the field as a whole, and Asian scientists are dramatically underrepresented among honorees. Because invitation and honor patterns could be driven by biases associated with name groups, geography, or other factors, we cross-referenced name group predictions with author affiliations to disentangle the relationship between geographic regions, name groups and invitation probabilities. We found that disparities persisted even within the group of honorees with a US affiliation. Societies’ honoree selection process failing to reflect the diversity of the field can play a part in why minoritized scientists’ innovations are discounted [1]. Although we estimate the fraction of non-White and non-Asian authors to be relatively similar to the estimated honoree rate, we note that both are represented at levels substantially lower than in the US population.

Societies, both through their honorees and the individuals who deliver keynotes at their meetings, can play a positive role in improving the presence of female STEM role models, which can boost young students’ interests in STEM [22] and, for example, lead to higher persistence for undergraduate women in geoscience [24]. Efforts are underway to create Wikipedia entries for more female [25] and black, Asian, and minority scientists [26], which can help early-career scientists identify role models. Societies can contribute toward equity if they design policies to honor scientists in ways that counter these biases such as ensuring diversity in the selection committees.

The central role that scientists play in evaluating each other and each other’s findings makes equity critical. Even many nominally objective methods of assessing excellence (e.g., h-index, grant funding obtained, number of high-impact peer-reviewed publications, and total number of peer-reviewed publications) are subject to the bias of peers during review. These could be affected by explicit biases, implicit biases, or pernicious biases in which a reviewer might consider a path of inquiry, as opposed to an individual, to be more or less meritorious based on the reviewer’s own background [11,27]. Our efforts to measure the diversity of honorees in an international society suggests that, while a focus on gender parity may be improving some aspects of diversity among honorees, contributions from scientists of color are underrecognized.

Acknowledgments

We would like to thank the Gordon and Betty Moore Foundation whose support makes the study possible (GBMF4552 to D.S.H. and GBMF4552 to C.S.G.). We thank the reviewers and editors for their insightful input to improve the quality of the study.

Author Contributions

Conceptualization, C.S.G.; Methodology, T.T.L., D.S.H., A.A.H., M.R.G. and C.S.G.; Software, T.T.L., D.S.H., A.A.H. and M.R.G.; Investigation T.T.L., D.S.H., A.A.H., M.R.G. and C.S.G.; Data Curation, T.T.L., D.S.H., A.A.H., M.R.G. and C.S.G.; Writing – Original Draft, T.T.L, D.S.H and C.S.G; Writing – Review & Editing, T.T.L and C.S.G; Visualization, T.T.L and D.S.H; Supervision, C.S.G; Project Administration, C.S.G; Funding Acquisition, D.S.H and C.S.G

Declaration of Interests

The authors declare no competing interests.

STAR Methods

Key resources table

REAGENT or RESOURCE SOURCE IDENTIFIER
Software and algorithms
Data and code Zenodo deposit This paper https://doi.org/10.5281/zenodo.5014756
Source code on GitHub This paper https://github.com/greenelab/iscb-diversity
Browsable Python and R notebooks This paper https://greenelab.github.io/iscb-diversity/
PubmedPy This paper https://github.com/dhimmel/pubmedpy
Manubot Himmelstein et al., 2019 https://doi.org/10.1371/journal.pcbi.1007128
Name origin prediction Wiki2019-LSTM This paper https://github.com/greenelab/wiki-nationality-estimate
Geotext Yaser Martinez Palenzuela https://github.com/elyase/geotext
Geopy Nominatim https://github.com/geopy/geopy https://geopy.readthedocs.io/en/stable/#nominatim
Genderize.io https://genderize.io/ https://genderize.io/
Other
Manuscript repository This paper https://github.com/greenelab/iscb-diversity-manuscript

Resource Availability

Lead Contact

Further information and requests for resources and reagents should be directed to and will be fulfilled by the Lead Contact, Casey Greene (casey.s.greene@cuanschutz.edu).

Materials Availability

This study did not generate new materials.

Data and Code Availability

All data have been deposited at https://github.com/greenelab/iscb-diversity and are publicly available as of the date of publication. DOIs are listed in the key resources table. Our Wikipedia name dataset is dedicated to the public domain under CC0 License at https://github.com/greenelab/wiki-nationality-estimate, with source code to construct the dataset available under a BSD 3-Clause License.

All original code has been deposited at Zenodo and is publicly available as of the date of publication. DOIs are listed in the key resources table. Specifically, our analysis of authors and ISCB-associated honorees is available under CC BY 4.0 at https://github.com/greenelab/iscb-diversity, with source code also distributed under a BSD 3-Clause License [28]. Rendered Python and R notebooks from this repository are browsable at greenelab.github.io/iscb-diversity. Our analysis of PubMed, PubMed Central, and author names relies on the Python pubmedpy package, developed as part of this project and available under a Blue Oak Model License 1.0 at https://github.com/dhimmel/pubmedpy and on PyPI. No additional information is required to reanalyze the data reported in this paper.

This manuscript was written openly on GitHub at github.com/greenelab/iscb-diversity-manuscript using Manubot [29]. The Manubot HTML version is available under a Creative Commons Attribution (CC BY 4.0) License at greenelab.github.io/iscb-diversity-manuscript.

Method details

Honoree Curation

From ISCB’s webpage listing ISCB Distinguished Fellows, we found recipients listed by their full names for the years 2009–2019. We gleaned the full name of the Fellow as well as the year in which they received the honor. We identified major ISCB-associated conferences as those designated flagship (ISMB) or those that had been held on many continents (RECOMB). To identify ISMB Keynote Speakers, we examined the webpage for each ISMB meeting. The invited speakers at ISMB before 2001 were listed in the Preface pages of each year’s proceedings, which were archived in the ISMB collection of the AAAI digital library. We found full names of all keynote speakers for the years 1993–2019.

For the RECOMB meeting, we found conference webpages with keynote speakers for 1999, 2000, 2001, 2004, 2007, 2008, and 2010–2019. We were able to fill in the missing years using information from the RECOMB 2016 proceedings, which summarizes the first 20 years of the RECOMB conference [30]. This volume has two tables of keynote speakers from 1997–2006 (Table 14, page XXVII) and 2007–2016 (Table 4, page 8). Using these tables to verify the conference speaker lists, we arrived at two special instances of inclusion/exclusion. Although Jun Wang was not included in these tables, we were able to confirm that he was a keynote speaker in 2011 with the RECOMB 2011 proceedings [31], and thus we included this speaker in the dataset. Marian Walhout was invited as a keynote speaker but had to cancel the talk due to other obligations. Because her name was neither mentioned in the 2015 proceedings [32] nor in the above-mentioned tables, we excluded this speaker from our dataset.

Name processing

When extracting honoree names, we began with the full name as provided on the site. Because our prediction methods required separated first and last names, we chose the first non-initial name as the first name and the final name as the last name. We did not consider a hyphen to be a name separator: for hyphenated names, all components were included. For metadata from PubMed and PMC where first (fore) and last names are coded separately, we applied the same cleaning steps. We created functions to simplify names in the pubmedpy Python package to support standardized fore and last name processing.

Last author extraction

We assumed that, in the list of authors for a specific paper, last authors (often research advisors) would be most likely to be invited for keynotes or to be honored as fellows. Therefore, we utilized PubMed to retrieve last author names to assess the composition of the field. PubMed is a search engine resource provided by the US National Library of Medicine and index scholarly articles. PubMed contains a record for every article published in journals it indexes (30 million records total circa 2020), within which we were able to extract author first and last names and their order using the E-Utilities APIs. To automate and generalize these tasks, we created the pubmedpy Python package.

From PubMed, we compiled a catalog of 176,773 journal articles that were published from 1993 through 2019 that were written in English and tagged with the MeSH term “computational biology”, which is equivalent to “bioinformatics” and includes categories such as genomics and systems biology (via PubMed’s term explosion to include subterms). Excluding 663 articles with no author information and years with less than 200 articles/year, we analyzed 176,110 articles from 1998–2019. We extracted the number of times an article has been cited by PubMed Central articles from the PmcRefCount of the PubMed DocSum XML records.

Countries of Affiliations

Publications often provide affiliation lists for authors, which generally associate authors with research organizations and their corresponding physical addresses. We implemented affiliation extraction in the pubmedpy Python package for both PubMed and PMC XML records. These methods extract a sequence of textual affiliations for each author.

We relied on two Python utilities to extract countries from text: geotext and geopy.geocoders.Nominatim. The first, geotext, used regular expressions to find mentions of places from the GeoNames database. To avoid mislabeling, we only mapped the affiliation to a country if geotext identified 2 or more mentions of that country. For example, in the affiliation string “Laboratory of Computational and Quantitative Biology, Paris, France”, geotext detected 2 mentions of places in France: Paris, France. In this case, we assign France to this affiliation.

This country extraction method accommodates multiple countries. Although ideally each affiliation record would refer to one and only one research organization, sometimes journals deposit multiple affiliations in a single structured affiliation. In these cases, we assigned multiple countries to the article. For more details on this approach, please consult the accompanying 07.affiliations-to-countries notebook and label dataset.

When geotext did not return results, we use the geopy approach, which returns a single country for an affiliation when successful. Its geocoders.Nominatim function converts names / addresses to geographic coordinates using the OpenStreetMap’s Nomatim service. With this method, we split a textual affiliation by punctuation into a list of strings and iterate backward through this list until we found a Nomatim search result. For the above affiliation, the search order would be “France”, “Paris”, and “Laboratory of Computational and Quantitative Biology”. Since Nomatim would return a match for the first term “France” (matched to France), the search would stop before getting to “Paris”, and “France” would be assigned to this affiliation.

Our ability to assign countries to authors was largely driven by the availability of affiliations. The country-assignment-rate for last authors from PubMed records was approximately 47%. This reflects the varying availability of affiliation metadata by journal.

For ISCB honorees, during the curation process, if an honoree was listed with their affiliation at the time, we recorded this affiliation for analysis. For ISCB Fellows, we used the affiliation listed on the ISCB page. Because we could not find affiliations for the 1997 and 1998 RECOMB keynote speakers’ listed for these years, they were left blank. If an author or speaker had more than one affiliation, each was inversely weighted by the number of affiliations that individual had.

Estimation of Gender

We predicted the gender of honorees and authors using the https://genderize.io API, which was trained on over 100 million name-gender pairings collected from the web and is one of the three widely-used gender inference services [12]. We used author and honoree first names to retrieve predictions from genderize.io. The predictions represent the probability of an honoree or author being male or female. We used the estimated probabilities and did not convert to a hard group assignment. For example, a query to https://genderize.io on January 26, 2020 for “Casey” returns a probability of male of 0.74 and a probability of female of 0.26, which we would add for an author with this first name. Because of technical limitations, our analysis only considered two binary gender categories, and we used “male” and “female” to refer to the gender of the scientists. However, we recognized the limitation of not accounting for non-binary gender categories and only considered predictions in aggregate and not as individual values for specific scientists.

Of 412 ISCB honorees, genderize.io fails to provide gender predictions for one name. Of 176,110 last authors, 1,014 were missing a fore name in the raw paper metadata and 11,498 had fore names consisting of only initials. Specifically, the metadata for most papers before 2002 (2,566 out of 2,601 papers) only have initials for first and/or middle author names. Without gender predictions for these names, we consider only articles from 2002 on when comparing gender compositions between two groups. Of the remaining authors, genderize.io failed to predict gender for 10,003 of these fore names. We note that approximately 42% of these NA predictions are hyphenated names, which is likely because they are more unique and thus are more difficult to find predictions for. This bias of NA predictions toward non-English names has been previously observed [33] and may have a minor influence on the final estimate of gender compositions.

Estimation of Name Origin Groups

We developed a model to predict geographical origins of names. The existing Python package ethnicolr [34] produces reasonable predictions, but its international representation in the data curated from Wikipedia in 2009 [35] is still limited. For instance, 76% of the names in ethnicolr’s Wikipedia dataset are European in origin.

To address these limitations in ethnicolr, we built a similar classifier, a Long Short-term Memory (LSTM) neural network, to infer the region of origin from patterns in the sequences of letters in full names. We applied this model on an updated, approximately 4.5 times larger training dataset called Wiki2019 (described below). We tested multiple character sequence lengths and, based on this comparison, selected tri-characters for the primary results described in this work. We trained our prediction model on 80% of the Wiki2019 dataset and evaluated its performance using the remaining 20%. This model, which we term Wiki2019-LSTM, is available in the online file LSTM.h5.

To generate a training dataset for name origin prediction that reflects a modern naming landscape, we scraped the English Wikipedia’s category of Living People. This category, which contained approximately 930,000 pages at the time of processing in November 2019, is regularly curated and allowed us to avoid pages related to non-persons. For each Wikipedia page, we used two strategies to find a full birth name and location context for that person. First, we looked for nationality mention in the first sentence in the body of the text. In most English-language biographical Wikipedia pages, the first sentence usually begins with, for example, “John Edward Smith (born 1 January 1970) is an American novelist known for …” This structure comes from editor guidance on biography articles and is designed to capture:

… the country of which the person is a citizen, national or permanent resident, or if the person is notable mainly for past events, the country where the person was a citizen, national or permanent resident when the person became notable.

Second, if this information is not available in the first sentence of the main text, we used information from the personal details sidebar; the information in this sidebar varied widely but often contained a full name and a place of birth.

We used regular expressions to parse out the person’s name from this structure and checked that the expression after “is a” matched a list of nationalities. We were able to define a name and nationality for 708,493 people by using the union of these strategies. This process produced country labels that were more fine-grained than the broader patterns that we sought to examine among honorees and authors. We initially grouped names by continent, but later decided to model our categorization after the hierarchical taxonomy used by NamePrism [36]. The NamePrism taxonomy was derived from name-country pairs by producing an embedding of names by Twitter contact patterns and then grouping countries using the similarity of names from those countries. The countries associated with each grouping are shown in supplementary figure S1. NamePrism excluded the US, Canada and Australia because these countries have been populated by a mix of immigrant groups [36].

In an earlier version of this manuscript, we also used category names derived from NamePrism, but a reader pointed out the titles of the groupings were problematic; therefore, in this version, we renamed these groupings to reflect that the NamePrism approach primarily identifies groups based on linguistic patterns from name etymology rather than religious or racial similarities. We note that our mapping from nationality to name origins was not without error. For example, a scientist of Israeli nationality may not bear a Hebrew name. These mismatches were assessed via the heatmap of the model performance (supplementary figure S2) and complemented by the affiliation analysis below. An alternative approach is to assign arbitrary names to these groups such as via letter coding (e.g., A, B, C, etc.), but we did not choose this strategy because ten arbitrary letters for ten groups can greatly reduce the paper’s readibility.

Predicting Name Origin Groups with LSTM Neural Networks and Wikipedia

Table 1 shows the size of the training set for each of the name origin groups as well as a few examples of PubMed author names that had at least 90% prediction probability in that group. We refer to this dataset as Wiki2019 (available online in annotated_names.tsv).

Table 1: Predicting name-origin groups of names trained on Wikipedia’s living people. The table lists the 10 groups and the number of living people for each region that the LSTM was trained on. Example names shows actual author names that received a high prediction for each region. Full information about which countries comprised each region can be found in the online dataset country_to_region.tsv.
Group Training Size Example Names
Celtic/English names 154,890 Adam O Hebb, Oliver G Pybus, David W Ritchie, James WJ Anderson, James W MacDonald, Robert Clarke
European names 78,157 Tracey M Filzen, Jos H Beijnen, Caroline Louis-Jeune, Christian Lorenzi, Boris Vassilev, Verena Heinrich
Hispanic names 66,931 Ramón Latorre, Antonio J Jimeno-Yepes, Felipe A Simão, Paulo S L de Oliveira, Juan Carlos Rodríguez-Manzaneque, Natalia Acevedo-Luna
East Asian names 54,588 Heejoon Chae, Wenchao Jiang, Haizhou Liu, Miho Uchida, Wenxuan Zhang, Jiali Feng
Arabic/Turkic/Persian names 31,418 Hamidreza Chitsaz, Farzad Sangi, Habib Motieghader, Berke Ç Toptas, Ali Aliyari, Bülent Arman Aksoy
Nordic names 28,978 Cecilia M Lindgren, Ellen Larsen, Jesper R Gådin, Janne H Korhonen, Johan Åqvist, Jens Nilsson
South Asian names 20,025 Amitabh Chak, Matthew G Seetin, Matrika Gupta, Sumudu P Leelananda, VS Kumar Kolli, Swanand Gore
African names 17,826 Timothy Kinyanjui, Jammbe Z Musoro, Nyaradzo M Mgodi, Magambo Phillip Kimuda, Probhonjon Baruah, Adaoha E C Ihekwaba
Hebrew names 4,549 Alexander J Sadovsky, Boris Shraiman, Gil Goldshlager, Eytan Adar, Aviva Peleg, Nir Esterman
Greek names 4,138 Gianni Panagiotou, Themis Lazaridis, Eleni Mijalis, Nikolaos Tsiantis, Konstantinos A Kyritsis, Dimitris E Messinis

We next aimed to predict the name origin groups of honorees and authors. We constructed a training dataset with more than 700,000 name-nationality pairs by parsing the English-language Wikipedia. We trained a LSTM neural network on n-grams to predict name groups. We found similar performance across 1, 2, and 3-grams; however, because the classifier required fewer epochs to train with 3-grams, we used this length in the model that we term Wiki2019-LSTM. Our Wiki2019-LSTM returns, for each given name, a probability of that name originating from each of the specified 10 groups. We observed a multiclass area under the receiver operating characteristic curve (AUC) score of 95.9% for the classifier, indicating that the classifier can recapitulate name origins with high sensitivity and specificity. For each individual group, the high AUC (above 92%) suggests that our classifier was sufficient for use in a broad-scale examination of disparities. We also observed that the model was well calibrated. We also examined potential systematic errors between pairs of name origin groupings with a confusion heatmap and did not find off-diagonal enrichment for any pairing (supplementary figure S2).

Applying Wiki2019-LSTM on the author and honoree datasets, we obtained name origin estimates for all honorees’ and authors’ name, except the 12,512 that did not have fore names (see breakdown in the Estimation of Gender section above). Once again, because the large majority of author fore names prior to 2002 were recorded with initials only, predictions were not possible, and we excluded 1998–2001 when comparing name origin compositions between two groups.

Affiliation Analysis

For each country, we computed the expected number of honorees by multiplying the proportion of authors whose affiliations were in that country with the total number of honorees. We then performed an enrichment analysis to examine the difference in country affiliation proportions between ISCB honorees and field-specific last authors. We calculated each country’s enrichment by dividing the observed proportion of honorees by the expected proportion of honorees. The 95% confidence interval of the log2 enrichment was estimated using the Poisson model method [37].

Statistical Analysis

We estimated the levels of representation by performing the following logistic regression of the group of scientists on each name’s prediction probability while controlling for year: \[g = \beta_0 + \beta_1 prob + \beta_2 y + \epsilon.\] The variable \(prob\) is the prediction probablity of a demographic variable (gender and name origin) for names of scientists in group \(g\) (honoree or author) during year \(y\). \(\epsilon \sim N(0, \sigma^2)\) accounts for random variation. An effect was deemed statistically significant if its corresponding p < \(\alpha\) = 0.01.

We emphasize that the units in this analysis are honors and authorships. Therefore, each row of the input data frame represents either an honor or authorship with the scientist’s name’s probability value from the gender or name origin classifier (\(prob\)). We also performed an alternative approach in which each pair of forename and last name as the unit of measurement and the citations were totaled across different papers whose last author had that name. Controlling for each name’s citations, this method tested the demographic effects but did not account for names honored more than once and may be affected by name collisions.

To reiterate, we only consider the prediction probabilities in aggregate and not as individual values for specific scientists. Moreover, although the average of the probabilities is not exactly “proportion”, we use the phrase “estimated proportion” for readability. For example, the average of the probabilities of authors having an East Asian name origin is the estimate for the proportion of authors with East Asian names.

Iterative Research Process

Parallel analyses for the other versions are available in supplementary figure S3. In our first version of the analysis pipeline, we sought to characterize the distribution of authorships in the field using field-specific journals. This resulted in an analysis set of 29,755 authorships. We also examined corresponding authors, as we considered that senior authors may occasionally occupy a different position [38], and only fell back to last authors in cases where corresponding author annotations were unavailable.

In the next version of the analysis, we extended the analysis to all 176,110 computational biology PubMed articles, substantially increasing the sample size. We also extracted names of last authors instead of other potential selections to better capture the honoree population. Our assumption was that, among the authors of a specific paper, the last author (often research advisors) would be most likely to be invited for a keynote or honored as a fellow. Also, the availability of information on corresponding authors was limited, and extracting last author became a more consistent approach.

In version 3, instead of weighting all articles equally as in the earlier versions, we used citation count to weight articles to control for the differential impact of research contributions. Using citation counts has key limitations: female scientists of different disciplines are cited less often than their male counterparts [17,18,19]. Furthermore, the act of being honored, particularly with a keynote at an international meeting, could lead to work being more recognized and cited, which would reverse the arrow of causality. In version 4, we returned to the equal weight for all articles.

Finally, in all versions of the analysis, rather than applying a hard assignment for each prediction, we analyzed the raw prediction probability values to capture the uncertainty of the prediction model. Although we expect our estimates of disparities for citation-weighted analyses to be conservative, through each analysis, the overall findings remained consistent. Examining the literature with and without citation weighting, we learned that disparities exist and these disparities are large enough to overcome existing disparities in citation patterns.

References

1.
The Diversity–Innovation Paradox in Science
Bas Hofstra, Vivek V Kulkarni, Sebastian Munoz-Najar Galvez, Bryan He, Dan Jurafsky, Daniel A McFarland
Proceedings of the National Academy of Sciences (2020-04-28) https://doi.org/ggskr7
2.
Ten Simple Rules to Achieve Conference Speaker Gender Balance
Jennifer L Martin
PLoS Computational Biology (2014-11-20) https://doi.org/gf853n
3.
How scientists are fighting against gender bias in conference speaker lineups
Katie Langin
Science (2019-02-11) https://doi.org/gghp8v
4.
Speaking out about gender imbalance in invited speakers improves diversity
Robyn S Klein, Rhonda Voskuhl, Benjamin M Segal, Bonnie N Dittel, Thomas E Lane, John R Bethea, Monica J Carson, Carol Colton, Susanna Rosi, Aileen Anderson, … Anne H Cross
Nature Immunology (2017-05-01) https://doi.org/gghp8s
DOI: 10.1038/ni.3707 · PMID: 28418385 · PMCID: PMC5775963
5.
Addressing the underrepresentation of women in mathematics conferences
Greg Martin
arXiv (2015-02-24) https://arxiv.org/abs/1502.06326
6.
The Presence of Female Conveners Correlates with a Higher Proportion of Female Speakers at Scientific Symposia
Arturo Casadevall, Jo Handelsman
mBio (2014-02-28) https://doi.org/qsh
DOI: 10.1128/mbio.00846-13 · PMID: 24399856 · PMCID: PMC3884059
7.
Trends in the Proportion of Female Speakers at Medical Conferences in the United States and in Canada, 2007 to 2017
Shannon M Ruzycki, Sarah Fletcher, Madalene Earp, Aleem Bharwani, Kirstie C Lithgow
JAMA Network Open (2019-04-12) https://doi.org/gghp8r
8.
Are gender gaps due to evaluations of the applicant or the science? A natural experiment at a national funding agency
Holly O Witteman, Michael Hendricks, Sharon Straus, Cara Tannenbaum
The Lancet (2019-02) https://doi.org/djc5
9.
Race, Ethnicity, and NIH Research Awards
DK Ginther, WT Schaffer, J Schnell, B Masimore, F Liu, LL Haak, R Kington
Science (2011-08-18) https://doi.org/csf8j8
10.
Gender, Race/Ethnicity, and National Institutes of Health R01 Research Awards
Donna K Ginther, Shulamit Kahn, Walter T Schaffer
Academic Medicine (2016-08) https://doi.org/f8zzw4
11.
Topic choice contributes to the lower rate of NIH awards to African-American/black scientists
Travis A Hoppe, Aviva Litovitz, Kristine A Willis, Rebecca A Meseroll, Matthew J Perkins, BIan Hutchins, Alison F Davis, Michael S Lauer, Hannah A Valantine, James M Anderson, George M Santangelo
Science Advances (2019-10-09) https://doi.org/gghp8t
12.
Comparison and benchmark of name-to-gender inference services
Lucía Santamaría, Helena Mihaljević
PeerJ Computer Science (2018-07-16) https://doi.org/ggpbn6
DOI: 10.7717/peerj-cs.156 · PMID: 33816809 · PMCID: PMC7924484
13.
Persistent Underrepresentation of Women’s Science in High Profile Journals
Yiqin Alicia Shen, Jason M Webster, Yuichi Shoda, Ione Fine
Cold Spring Harbor Laboratory (2018-03-08) https://doi.org/cmh5
14.
The gender gap in science: How long until women are equally represented?
Luke Holman, Devi Stuart-Fox, Cindy E Hauser
PLOS Biology (2018-04-19) https://doi.org/gdb9db
15.
Last Place? The Intersection of Ethnicity, Gender, and Race in Biomedical Authorship
Gerald Marschke, Allison Nunez, Bruce A Weinberg, Huifeng Yu
AEA Papers and Proceedings (2018-05-01) https://doi.org/gjg9k8
16.
Disparities in publication patterns by gender, race and ethnicity based on a survey of a random sample of authors
Allison L Hopkins, James W Jawitz, Christopher McCarty, Alex Goldman, Nandita B Basu
Scientometrics (2012-11-10) https://doi.org/gffmpv
17.
The extent and drivers of gender imbalance in neuroscience reference lists
Jordan D Dworkin, Kristin A Linn, Erin G Teich, Perry Zurn, Russell T Shinohara, Danielle S Bassett
arXiv (2020-08-28) https://arxiv.org/abs/2001.01002
18.
Gender differences in peer review outcomes and manuscript impact at six journals of ecology and evolution
Charles W Fox, CETimothy Paine
Ecology and Evolution (2019-03-04) https://doi.org/gfwjjb
DOI: 10.1002/ece3.4993 · PMID: 30962913 · PMCID: PMC6434606
19.
Quantitative evaluation of gender bias in astronomical publications from citation counts
Neven Caplar, Sandro Tacchella, Simon Birrer
Nature Astronomy (2017-05-26) https://doi.org/b8rn
20.
Bibliometrics: global gender disparities in science.
Vincent Larivière, Chaoqun Ni, Yves Gingras, Blaise Cronin, Cassidy R Sugimoto
Nature (2013-12-12) https://www.ncbi.nlm.nih.gov/pubmed/24350369
21.
Journals invite too few women to referee
Jory Lerback, Brooks Hanson
Nature (2017-01-26) https://doi.org/gf4jjz
22.
Understanding current causes of women's underrepresentation in science
SJ Ceci, WM Williams
Proceedings of the National Academy of Sciences (2011-02-07) https://doi.org/dsrczk
23.
Historical comparison of gender inequality in scientific careers across countries and disciplines
Junming Huang, Alexander J Gates, Roberta Sinatra, Albert-László Barabási
Proceedings of the National Academy of Sciences (2020-03-03) https://doi.org/ggk89f
24.
Role modeling is a viable retention strategy for undergraduate women in the geosciences
Paul R Hernandez, Brittany Bloodhart, Amanda S Adams, Rebecca T Barnes, Melissa Burt, Sandra M Clinton, Wenyi Du, Elaine Godfrey, Heather Henderson, Ilana B Pollack, Emily V Fischer
Geosphere (2018-10-31) https://doi.org/gghp9d
25.
Why we’re editing women scientists onto Wikipedia
Jess Wade, Maryam Zaringhalam
Nature (2018-08-14) https://doi.org/gdz52z
26.
Why we’re creating Wikipedia profiles for BAME scientists
Nicola O’Reilly
Nature (2019-03-07) https://doi.org/gfwhcr
27.
Author-Reviewer Homophily in Peer Review
Dakota Murray, Kyle Siler, Vincent Larivière, Wei Mun Chan, Andrew M Collings, Jennifer Raymond, Cassidy R Sugimoto
Cold Spring Harbor Laboratory (2019-08-04) https://doi.org/ddzr
28.
greenelab/iscb-diversity: ISCB analysis v4.0 release
Trang Le, Daniel Himmelstein, Casey Greene
Zenodo (2021-06-22) https://doi.org/gkspbs
29.
Open collaborative writing with Manubot
Daniel S Himmelstein, Vincent Rubinetti, David R Slochower, Dongbo Hu, Venkat S Malladi, Casey S Greene, Anthony Gitter
PLOS Computational Biology (2019-06-24) https://doi.org/c7np
30.
Research in Computational Molecular Biology
Lecture Notes in Computer Science
Springer Science and Business Media LLC (2016) https://doi.org/gghp87
31.
Research in Computational Molecular Biology
Lecture Notes in Computer Science
Springer Science and Business Media LLC (2011) https://doi.org/dvwjgg
32.
Research in Computational Molecular Biology
Lecture Notes in Computer Science
Springer Science and Business Media LLC (2015) https://doi.org/gghwjh
33.
Gender Prediction Methods Based on First Names with genderizeR
Kamil Wais
The R Journal (2016) https://doi.org/gf4zqx
34.
Predicting Race and Ethnicity From the Sequence of Characters in a Name
Gaurav Sood, Suriyan Laohaprapanon
arXiv (2018-07-31) https://arxiv.org/abs/1805.02109
35.
Name-ethnicity classification from open sources
Anurag Ambekar, Charles Ward, Jahangir Mohammed, Swapna Male, Steven Skiena
Association for Computing Machinery (ACM) (2009) https://doi.org/fs3pr8
36.
Nationality Classification Using Name Embeddings
Junting Ye, Shuchu Han, Yifan Hu, Baris Coskun, Meizhu Liu, Hong Qin, Steven Skiena
Association for Computing Machinery (ACM) (2017-11-06) https://doi.org/ggjc78
37.
Statistics in epidemiology: methods, techniques, and applications
Hardeo Sahai, Anwer Khurshid
CRC Press (1996)
ISBN: 9780849394447
38.
Contributorship and division of labor in knowledge production
Vincent Larivière, Nadine Desrochers, Benoît Macaluso, Philippe Mongeon, Adèle Paul-Hus, Cassidy R Sugimoto
Social Studies of Science (2016-06-21) https://doi.org/gc6xqv