Estimating social bias in data sharing behaviours: an open science experiment

Estimating social bias in data sharing behaviours: an open science experiment

Play all audios:

Loading...

ABSTRACT Open data sharing is critical for scientific progress. Yet, many authors refrain from sharing scientific data, even when they have promised to do so. Through a preregistered,


randomized audit experiment (N = 1,634), we tested possible ethnic, gender and status-related bias in scientists’ data-sharing willingness. 814 (54%) authors of papers where data were


indicated to be ‘available upon request’ responded to our data requests, and 226 (14%) either shared or indicated willingness to share all or some data. While our preregistered hypotheses


regarding bias in data-sharing willingness were not confirmed, we observed systematically lower response rates for data requests made by putatively Chinese treatments compared to putatively


Anglo-Saxon treatments. Further analysis indicated a theoretically plausible heterogeneity in the causal effect of ethnicity on data-sharing. In interaction analyses, we found indications of


lower responsiveness and data-sharing willingness towards male but not female data requestors with Chinese names. These disparities, which likely arise from stereotypic beliefs about male


Chinese requestors’ trustworthiness and deservingness, impede scientific progress by preventing the free circulation of knowledge. SIMILAR CONTENT BEING VIEWED BY OTHERS DATA SHARING


PRACTICES AND DATA AVAILABILITY UPON REQUEST DIFFER ACROSS SCIENTIFIC DISCIPLINES Article Open access 27 July 2021 PERCEIVED BENEFITS OF OPEN DATA ARE IMPROVING BUT SCIENTISTS STILL LACK


RESOURCES, SKILLS, AND REWARDS Article Open access 20 June 2023 BOUNDED RESEARCH ETHICALITY: RESEARCHERS RATE THEMSELVES AND THEIR FIELD AS BETTER THAN OTHERS AT FOLLOWING GOOD RESEARCH


PRACTICE Article Open access 06 February 2024 INTRODUCTION Scientific discoveries are made through cumulative and collective efforts, ideally based on full and open communication. For


science to work, published claims must be subject to organized skepticism1. Yet, science’s ethos of rigorous, structured scrutiny is contingent on data sharing. Lack of data prevents results


from being reexamined with new techniques, and samples from being pooled for meta-analysis2,3. This ultimately hinders the cumulative knowledge-building that drives scientific progress4.


Open data improves the credibility of scientific claims, and while journal editors increasingly acknowledge the importance of disclosing data5,6, many authors refrain from sharing their


data, even when they have promised to do so7,8,9,10,11,12,13. Previous research has focused on the supply-side determinants of data-sharing. Surveys find that scientists’ decisions to share


data depend on (i) contextual factors such as journal requirements, funding incentives and disciplinary competition, (ii) individual factors such as perceived risks of data misuse, lost


publication opportunities, and efforts associated with making data available, and (iii) demographic factors such as experience level, tenure-status and gender7,14,15,16,17,18,19. Much less


attention is given to the demand-side issues of data sharing. Ideally, everyone, irrespective of background, should be able to contribute to science1. As such, data access should not differ


depending on who is asking for the data. Yet, research indicates persistent gender, ethnic and status-related bias in science20,21,22,23,24,25,26,27 that likely also affects data-sharing


practices. Social bias in data-sharing may arise from scientists’ stereotypic beliefs about data requestors. According to status characteristics theory, nationality, ethnicity, gender and


institution prestige are diffuse cues that, when salient, may influence scientists’ impressions of requestors’ trustworthiness, competence or deservingness28,29. Such status cues are more


likely to guide people’s judgments in ambiguous situations, where information is scarce30,31. Further, status cues may be critical for data sharing, as knowledge transfer is generally more


likely in high-trust situations, here including the potential data sharer’s trust in the requestor’s competences and intentions32,33. We conducted a pre-registered (https://osf.io/378qd),


randomized audit experiment to examine possible ethnic, gender and status bias in data sharing. We requested data from authors of papers published in widely recognized journals that commit


to the FAIR data principles (_the Proceedings of the National Academy of Sciences_ [PNAS] and _Nature_-portfolio outlets), where data were indicated to be available upon request. We varied


the identity of the fictitious data requestor on four factors: (i) country of residence (China vs. United States [US]), (ii) institution prestige (high status vs. lower status university),


(iii) ethnicity (putatively Chinese vs. putatively Anglo-Saxon), and (iv) gender (masculine-coded vs. feminine-coded name). Motivated by evidence of gender, ethnic and status bias in trust


games34,35, in correspondence tests of employers, legislators, educators and citizens27,36,37,38,39, and in field- and survey experiments of peer evaluation22,40,41, we hypothesized that


scientists would be less willing to share data, when a requestor (i) was from China (compared to the US); (ii) was affiliated with a lower-status university (compared to a higher status


university); (iii) had a Chinese-sounding name (compared to a typical Anglo-Saxon name); and (iv) had a feminine-coded name (compared to a masculine-coded name). In addition to gender and


institution status, which have previously been covered in correspondence tests of scientists42,43, we were interested in the specific disadvantages facing researchers with Chinese names and


university affiliations. China is currently the world’s largest producer of scientific outputs44 and Chinese expatriates by far outnumber any other group of foreign graduate students at US


universities45,46. Considering these figures, a study of the possible bias facing Chinese nationals and descendants in a globalized science system seems timely. MATERIALS AND METHODS


SAMPLING Our data collection included four steps, summarized in Figure S1. The experimental population consisted of authors of scientific papers published in PNAS and Nature-portfolio


journals (between 2017 and 2021), where data were indicated to be ‘available upon request’. We queried journal websites to identify all peer-reviewed papers that included the text-string


“available upon request”. This resulted in an initial sample of 6,798 papers. C.A. and M.W.N. manually checked and coded the data availability section of each paper to identify cases where


data were unambiguously made available upon request (Table S1). If the primary author contact listed in a data statement occurred multiple times in our sample (due to >1 publications), we


only included the author’s most recent publication. We removed all retracted papers and sources with corrections. We matched the authors listed as primary data contacts in the papers with


bibliographic metadata from Clarivate’s Web of Science (WoS) to retrieve up-to-date information on emails and affiliations (Figure S2). Our final sample consists of 1,826 author-paper pairs.


Due to bounced emails and authors withdrawing from the study after receiving a debriefing statement (in total 78 authors decided to withdraw), our analysis sample is further reduced to


1,634 author-paper pairs. According to our registered power analysis, this sample size should be sufficient for detecting a small effect size of Cohen’s f2 = 0.02, with α = 0.01 and a power


of 0.95. PROCEDURES Efforts to measure ethnic, gender and status bias are complicated by observer effects and issues of social desirability47. Audit studies can mitigate such biases by


allowing the experimenter to estimate participants’ responses to the treatment in a realistic setting47,48,49. Our unpaired audit study randomized participants across twelve treatment


conditions (Figure S3). Given our four treatments, a 4 × 4 factorial design would be the typical set-up for our study. Yet, to keep the treatments realistic, we decided not to include a


putatively Anglo-Saxon (male or female) treatment associated with a (higher or lower status) Chinese University (during the COVID-19 pandemic, the number of international students in China


decreased enormously). Thus, we adopted 12, instead of 16, treatment conditions. Previous research indicates that data-sharing practices differ depending on the author’s gender and


disciplinary field11,13,50. Hence, we block-randomized the sample population according to scientific field and gender. Inspired by previous audits on discrimination in science42,43, our


data-sharing requests were emailed from fictitious “about-to-become” PhD students. We created Gmail accounts for all of the gender-ethnicity combinations (in total 4 email addresses). In the


emails, the fictitious data requestors asked the participants to share data related to a specific publication (Figure S4). If participants did not respond to our initial email, we sent out


a follow-up request after one week. Data collection was completed two weeks after the follow-up request. All data were collected in April and May 2022. Email correspondences were managed


through YAMM (https://yamm.com/), a private email service provider (google sheet add-on for Gmail and Google Workspace users). This tool allowed us to track the email delivery metrics and


provided information on whether an email had been received (or bounced) and opened (or left unopened). Our sampling and analysis plan was preregistered at the Open Science Framework. We have


followed all steps presented in the registered plan, with one minor deviation. In the preregistration we planned to run two linear probability models with cluster-robust standard errors at


the field level, while in the results section, we report the outcomes of these models without the cluster-robust standard errors. Cluster robust standard errors proved unsuitable given the


low number of clusters and because randomized treatments were assigned at the individual level as opposed to the group-level51. MANIPULATIONS Our treatment conditions varied on the following


four factors: gender, ethnicity, country of residence and institutional affiliation. We used first and last names to signal the fictitious requestors’ gender and ethnicity (Figure S4).


Emails from putatively Chinese treatments included a masculine- or feminine-coded Anglo-Saxon middle name in a parenthesis to signal the requestor’s gender (e.g., ‘Yadan (Cecilia) Xing’).


This is a widely used name-practice among transnational Chinese students52. We created four email addresses for our treatments. All accounts were opened and used for at least 90 days before


the experiment started. Warm-up activities of email domains (e.g., sending small volumes of emails and slowly increasing) helped build a positive sender reputation for the accounts. Such


standard email activity from the newly created account can reduce the risk of an account being labeled as a fake, hence reducing the email’s likelihood of being filtered as spam. To select


relevant names, we relied on a historical dataset of Olympic athletes53. We limited our focus to Chinese and American athletes participating in the Olympic Games between 1932 and 2016. For


each country treatment (US and China), we randomly selected separate first and last names until finding appropriate combinations. For the putatively Anglo-Saxon requestor conditions, we used


the following names: Jeffrey Killion and Hilary Witmer. For the putatively Chinese requestor conditions, we used the following names: 张嘉实 (Jia Shi Zhang) and 邢雅丹 (Yadan Xing). We used the R


package “rethnicity”54 to ensure that the selected Anglo-Saxon first names were typical Anglo-Saxon names and that the selected Chinese first names were well-known Asian names. In addition,


we manually verified the rethnicity estimates by looking up the first names on Linkedin. To ensure that participants perceived the Chinese names as Chinese, we wrote the signature names (at


the bottom of each email) in both English letters and Chinese characters (‘Yadan (Cecilia) Xing | 邢雅丹’ and ‘Jiashi (Wilson) Zhang | 张嘉实’) (see also Figure S4). According to the US Census of


Frequently Occurring Surnames55, 96% of individuals with the surname “Witmer” identified as white, 87% of individuals with the surname Killion identified as white, 97% of individuals with


the surname Xing identified as Asian, while 98% with the surname Zhang identified as Asian. Previous research suggests that some race and ethnicity-based manipulations also include class


primes56. In our design, we followed Block _et al_.36 and attempted to reduce the potential effect of socio-economic status by holding occupation constant. Specifically, all participants


received an email from an ‘about to become’ Ph.D. student at a Chinese or US research institution. Emails from putatively Chinese treatments (located in the US or China) that targeted


authors outside of China were written in English. Emails from Chinese nationals that targeted authors located in China were written in Mandarin. We used four international university


rankings to identify appropriate university affiliations that varied in institution status (high status [Carnegie Mellon and Zhejiang University] vs. lower status [Baylor University and


Chongqing University) (SI _Appendix_, Table S2). This manipulation also signaled the fictitious requestors’ country of residence (China or the US). Appropriate university affiliations were


identified using a combination of different university rankings (Times Higher Education, Shanghai national and international rankings, QS ranking, and the PP-top 10% indicator from the


Leiden Ranking). To select high status affiliations, we focused on universities that scored consistently high across the rankings. When selecting lower-status affiliations, we focused on


universities that scored consistently low across rankings. We decided to exclude top-ranked (top 10%) and bottom-ranked universities (bottom 10%) to lower the likelihood that participants


would discern the purpose of the experiment. In addition, we restricted our selection to universities with multiple faculties and active Ph.D. programs within each faculty. MEASURES We


measured the experimental treatments using four dichotomous variables: country (US = 0, China = 1), ethnicity (Anglo-Saxon name = 0, Chinese name = 1), institution status (high status


university = 0, lower status university = 1), and gender (masculine coded name = 0 and feminine coded name = 1). In addition, all statistical models included five dichotomous variables to


adjust for scientific field and publication outlet (Table S3). Our preregistered dependent variable is a binary measure of data-sharing willingness. This measure is based on a systematic


coding of participants’ email responses (respondents did not share data nor indicate willingness to share data = 0, respondents shared data or indicated willingness to share data = 1). Two


authors systematically coded all responses. The codebook (Table S4) was first tested by coding 10% of the sample. The pilot phase was repeated, and the codebook was further adjusted, until


coding reliability measures reached a satisfactory level (Kappa coefficient ≥ 0.8). In the manual coding, the coders were blinded to the treatments. As a second outcome (not registered), we


measured whether participants responded to our data requests or not (non-response = 0, response = 1). This outcome measure is widely used in previous email-based correspondence tests57. As


opposed to data-sharing willingness, our measure of email responses did not involve any coding of textual content and can thus be viewed as a more objective indicator of bias in data


sharing. In accordance with our registered analysis plan, we measured and reported the dependent variables in two ways. In one case, we excluded all unopened emails from the analysis. In the


results section, we refer to this sample as the sample of “opened emails”. In the other case, we included unopened emails and coded them as indicating unwillingness to share data or


non-responses. In the results section, we refer to this sample as the “full sample”. Given that all participants in the sample of opened emails have been directly exposed to a treatment, we


would expect the treatment effects to be larger for this sample. In contrast, given that some participants in the full sample were not directly exposed to a treatment, we would expect the


noise-to-signal ratio to be larger and the treatment effects to be smaller for this sample. Thus, while the sample of opened emails gets us closer to the direct effect of the treatments, the


full sample gives us a better sense of the real-world disadvantages associated with a given treatment. Due to a minor error in our email management, responses to one of the treatment emails


were associated with an alias during the first wave of data collection. Specifically, participant responses to emails sent from Yadan Xing were forwarded to a different alias address with a


similar manipulation condition. Only two participants noticed this issue in their responses, but nonetheless positively engaged with our request. The rest of the respondents exposed to this


error either responded to the alias email while addressing their message to Yadan (e.g., “Dear Yadan”) or responded directly to the correct email manipulation. Based on this evidence, we


find it reasonable to assume that the vast majority of respondents exposed to the error either did not notice the issue or perceived it to be irrelevant. This mistake was corrected in the


follow-up email addressed to all recipients who did not respond to the first email. Table S5 presents the response rates across treatments for the first and the second wave. The treatment


including the alias mistake (Yadan Xing) had the highest response rate of the four treatment emails after wave one. Note also that our main findings concerning a bias in responsiveness and


data-sharing willingness (Figs. 1,4) pertain to the male Chinese treatment (Jiashi Zhang) and not the female Chinese treatment (Yadan Xing). STATISTICAL ANALYSES Given that no participants


in our study were exposed to the putatively Anglo-Saxon treatments located in China, we estimated two groups of linear probability models. In one group, we estimated the direct effects of


gender, ethnicity and institution prestige on data sharing and email responses among participants exposed to treatments affiliated with US universities (Fig. 1 and S5). In another group, we


estimated the direct effects of gender, country-location, and institution prestige on data-sharing willingness among putatively Chinese treatments located in the US and China (Fig. 2 and


S6). Despite the unidirectional nature of our hypotheses, we report all outcomes with two-directional 95% and 99% confidence intervals. All analyses were conducted in R58, and we used the


estimator package59 to perform the linear probability models. ETHICS Since the social biases examined in our study may operate subconsciously, a pre-experimental consent process could damage


the validity of the experiment. For this reason, we decided to operate with ex-post consent and information disclosure. Our study was approved by the Ethics Review Board at the Department


of Sociology, University of Copenhagen (UCPH-SAMF-SOC-2022-03). At the end of the experiment, a debriefing email was sent to all participants (non-respondents as well as respondents). In


this debriefing, we explained the general purpose of the study, its experimental manipulations, the potential risks and benefits to participants, and the principles of data management and


anonymization. Moreover, we informed participants about their right to withdraw from the study without penalty. In total, 78 authors (5%) decided to withdraw from our study after receiving


the debriefing statement. RESULTS Eight-hundred-and-eighty-four scientists responded to our data requests, and 226 either shared or indicated willingness to share all or some of their data.


This corresponds to 54% (884 of 1634) and 14% (226 of 1634) of the full sample, and 75% (884 of 1179) and 20% (226 of 1179) of the sample of opened emails. In Fig. 1, we estimate how


institution prestige, gender and ethnicity influence participants’ responsiveness and data-sharing willingness in the sub-sample (that opened emails) exposed to treatments with US


affiliations. As shown in the figure (Panel A), neither university status nor gender affected participants’ likelihood of responding to data requests from treatments with US affiliations.


Yet, treatments with Chinese-sounding names were 7 percentage points less likely to receive a response than treatments with putatively Anglo-Saxon names (β = −0.07, 95% CI: −0.13:−0.01; 99%


CI: −0.15:0.01). This corresponds to an odds ratio of 0.66, or 34% lower odds of obtaining a response for the treatments with Chinese-sounding names compared to putatively Anglo-Saxon


treatments (Table S16). Results are similar, though associated with larger uncertainties, when estimates are based on the full sample as opposed to the sample of opened emails (Figure S5).


Our pre-registered analysis of ethnic, gender and status-related bias in scientists’ willingness to share data with treatments located in the US proved inconclusive in both the full sample


(Figure S5, Panel B) and in the sample of individuals that opened emails (Fig. 1, Panel B). In conflict with one of our hypotheses, participants seemed slightly more willing to share data,


when the request came from a lower status US institution (Baylor University) compared to a higher status US institution (Carnegie Mellon), but the confidence intervals for this effect all


include zero (β = 0.05, 95% CI: −0.01:0.11; 99% CI: −0.02:0.13). In Fig. 2, we estimate how university status, gender and country matter for participants’ responsiveness and data-sharing


willingness in the sub-sample exposed to treatments with Chinese-sounding names, located in China vs. the US. As shown in Fig. 2 (Panel A), we do not find any association between the


treatment conditions and participants’ responsiveness to data-requests in this subsample. All coefficients are close to zero, indicating no discernible effects. Similarly, for data-sharing


willingness (Fig. 2, Panel B), results are weak and inconclusive. The effects of university status and country on data-sharing willingness are close to zero, albeit women, not men


(contrasting our pre-registered hypothesis), are met with slightly higher data sharing willingness in this subsample. Yet, the confidence intervals for this gender effect include zero (β = 


0.05, 95% CI: −0.01:0.11; 99% CI: −0.02:0.13). Again, the results are similar, when models are based on the full sample as opposed to the sample of opened emails (Figure S6). In Fig. 3, we


explore if the ethnicity bias indicated in Fig. 1 (Panel A) is specific to the US treatments with Chinese-sounding names, or whether it affects putatively Chinese data requestors more


generally. In this analysis, which covers both the US-located and China-located treatments, we obtain results comparable to those reported in Fig. 1 (Panel A). The effects for university


status and gender remain inconclusive, but treatments with Chinese-sounding names have a 7 percentage points lower likelihood of receiving a response than treatments with typically


Anglo-Saxon names (β = -0.07, 95% CI: -0.12:-0.02; 99% CI: -0.14:-0.00). This corresponds to an odds ratio of 0.67, or 33% lower odds of response for putatively Chinese treatments compared


to putatively Anglo-Saxon treatments (Table S20). This estimated effect is smaller and associated with larger uncertainties when the analysis is based on the full sample compared to the


sample of opened emails (Figure S7). Given the indication in Fig. 2 (Panel B) that data-sharing willingness is lower for men with Chinese names compared to women with Chinese names, we also


estimated the conditional effects of ethnicity on data-sharing behaviors for male and female treatments. To secure the highest possible statistical power, we ran interaction-analyses on the


combined samples of US-located and China-located treatments. Figure 4 plots the conditional coefficients from four interactions between Ethnicity and Gender in the sample of authors that


opened emails. Male treatments with Chinese-sounding names (compared to male treatments with putatively Anglo-Saxon names) face consistent disadvantages both with respect to responsiveness


(β = −0.10, 95% CI: −0.17:-0.03; 99% CI: −0.19:-0.01) and willingness (β = −0.07 95% CI: −0.15:-0.00; 99% CI: -0.17:0.02) when requesting data, while this is not the case for female


treatments with Chinese-sounding names. We obtain comparable results when the conditional coefficients are estimated based on the full sample (Figure S8). DISCUSSION Previous research on


social bias in data sharing is scarce. Several studies document low response rates and low data-sharing willingness among scientists who agreed to make their data available upon


request7,9,11,12,15. Yet, demand-side determinants of data sharing remain largely unexamined. In a small-scale field experiment (N = 200), Krawczyk and Reuben60 tested whether economists’


willingness to share supplementary materials differed depending on the prestige of the requestor’s university and position level (Columbia University vs. University of Warsaw) and found


negligible effects. In our audit experiment, which draws on a larger multi-disciplinary sample of participants, none of the four preregistered hypotheses predicting national, ethnic, gender


and institutional prestige bias in data-sharing willingness were supported. Yet, tests for differences in scientists’ responsiveness to data requests (our most objective measure of


disparities in data-sharing) indicated lower response rates for Chinese treatments compared to Anglo-Saxon treatments. This may indicate that ethnic bias is more likely to occur at the


initial stage of a (potential) data exchange when scientists make rapid and unreflective judgments on whether to engage with a requestor or not. Indeed, previous research into ethnic bias in


pro-sociality also emphasizes the role of implicit attitudes (activated quickly and spontaneously) in discriminatory behaviors61,62. Contrary to our expectations, scientists exposed to


US-located treatments seemed slightly less willing to share data with requestors from prestigious universities compared to requestors from lower status universities (although the 95%


confidence bounds spanned zero for this estimate). One possible explanation for this may be that the perceived career risks associated with data sharing (in terms of lost publication


opportunities and lowered competitive advantages), on average, are higher, when requests are made from prestigious US universities compared to lower-status US universities or Chinese


universities. Indeed, previous research finds that scientists’ data-sharing willingness tends to be lower when perceived competition is high7,50,63,64. Such risks and concerns could


potentially be reduced through the use of data licensing on public repositories like OSF. Importantly, the negative prestige effect was only salient for scientists exposed to treatments from


US universities. This may be because the participants in our sample, which primarily reside in Europe or North America (Table S26), are more familiar with the prestige hierarchy of US


institutions and less knowledgeable about the relative standing of Chinese institutions. Also contrary to our predictions, Chinese treatments with feminine-coded names were met with slightly


higher data-sharing willingness than Chinese treatments with masculine-coded names (although the 95% confidence bounds spanned zero for this estimate). Importantly, this finding reflects an


underlying pattern of male-specific ethnic discrimination. Conditional effects derived from interaction analyses suggested a clear bias in responsiveness and data-sharing willingness


against male Chinese treatments, while results for female Chinese treatments were inconclusive. Given the “double-burden hypothesis” in intersectional theory65, which states that minority


women are most likely to face discrimination, these findings may seem counter-intuitive. Nevertheless, they have some bearing in previous studies on trust and discrimination. Indeed,


evidence suggests that women are stereotypically seen as more trustworthy than men66, and studies on helping behavior also indicate a greater taste for helping women than men in various


social situations67,68,69,70. Further, while field experiments have generally neglected intersectional perspectives on ethnic and gender discrimination71, studies that _do_ cover this aspect


typically find that minority males are subject to larger ethnic penalties than minority females in job-markets, housing markets, and sharing economies72,73,74,75,76,77,78,79. Building on


research on gender and nationality stereotypes80, Arai and colleagues theorize that when stereotypes against specific ethnic groups are negative, they are more likely to disadvantage men


than women, because it is men who are primarily presumed to embody these stereotypes72. Additional research is required to determine the root causes of the observed bias towards Chinese men.


However, we hypothesize that it likely arises from stereotypic beliefs about the group’s trustworthiness and deservingness in data exchange relationships. Such beliefs may have been


particularly salient during 2022, when we collected our data, due to recurring discussions about China’s alleged intellectual property theft in the US and Europe81, but also in the wake of


COVID-19, where prejudice and discriminatory intent against Asians aggravated82,83. While our field-experiment has clear advantages over survey-based approaches to measuring data-sharing


behaviors, it is not without limitations. Our data requests were sent from Gmail accounts, and this may have increased the likelihood that emails ended up in the recipients’ spam filters.


Further, some recipients may have found our data requests more suspicious than they would have, if the same emails were sent via institutional accounts. We have attempted to account for this


issue by tracking opened and unopened emails throughout the study-period and by reporting results for samples that include and exclude unopened emails (SI _Appendix_). Compared to previous


correspondence studies, where data-requests were made from institutional email accounts, the response rate in our study is quite high. Participants ignored 49% of the requests made by


Tedersoo and colleagues11 concerning data for recent papers in _Nature_ and _Science_; while 86% of the data requests made by Gabelica and colleagues to authors publishing in BioMed Central


journals were also ignored12. In comparison 54% of scientists in our sample responded to the data requests. This suggests that the drawbacks of using Gmail compared to an institutional


account have been small. Another limitation of our study design concerns the generic nature of our data request, which may have increased the level of suspicion among some recipients. After


the study was completed, a few authors approached us with concerns that the email request lacked detail and was not tailored to the specific practices of their discipline, which made them


hesitant to respond. While we acknowledge this critique, experiments like ours will always be subject to trade-offs between ecological validity and treatment bias. In this case, we decided


to keep the emails generic to hold constant all other components than our manipulations. A fourth limitation concerns our sampling strategy. Because we only targeted authors of papers in


Nature portfolio journals and PNAS, our results are limited to scientists publishing in these journals. In the future, researchers should examine whether data sharing behaviors differ for


authors publishing in journals that are less committed to open science and the FAIR data-sharing principles. Finally, given that none of our four preregistered hypotheses were directly


confirmed, our results concerning gender-specific ethnic discrimination in data sharing can only be seen as suggestive. Despite these limitations, our paper offers important new insights on


scientific data-sharing practices in science. Compared to unregistered experiments, our preregistered analysis has the advantage of providing a clear record of what ideas our study was


designed to evaluate, how we planned to examine them, and how our most notable finding of ethnic bias in data-sharing relate to these ideas84. Put differently, the preregistered analysis


plan has limited our degrees of freedom as researchers and thereby increased the validity and reliability of our study. Our paper has important implications for open science policies.


Despite clearly indicating intent to make their data available upon request, only around half of the targeted authors responded to our data requests, and only 14% indicated willingness to


share all, or some, of their data. While some participants may have had good reasons not to share, this behavior conflicts with the FAIR principles adopted by PNAS and Nature portfolio


journals, hence demonstrating the drawbacks of enabling researchers to make data available upon request. Our study further complicates this issue by exposing potential inequalities in who


can benefit from data-sharing, when disclosure decisions are left to the discretion of individual scientists7. In accordance with previous work, our study shows that data requests often


require more than trivial efforts from the side of the requestor. These efforts could be reduced if funders and publishers required authors to release all relevant data, whenever


possible11,12. Unfortunately, the reality is that most journals do not incentivize data sharing. In a review of editorial policies for 318 biomedical journals5, only 12% explicitly required


data sharing as a condition for publication, 9% required data sharing without stating it as a condition for publication, while around one third of the journal sample did not mention


data-sharing at all. Under such conditions, we expect that “data availability upon request” will remain a widespread practice in many disciplines. Importantly, disclosure is sometimes also


challenged by practical issues (e.g., data size and propriety rights) or ethical issues (e.g., sensitive information on human subjects), and publishers could do more to help mitigate these


challenges. From our experience, it seems that many authors that cannot share their data for practical or ethical reasons currently opt to indicate data availability upon request to


circumvent a journal’s data-sharing requirements. Assistance from the side of publishers in providing the necessary storage space or easy-to-use methods for making synthetic datasets for


sensitive populations, could help mitigate these problems85. In summary, our field experiment extends on research about scientists’ compliance with open data principles by indicating that


sharing behaviors may differ depending on who is asking for the data. These disparities, which likely arise from stereotypic beliefs about specific requestors’ trustworthiness and


deservingness, hamper the core principle of mutual accountability in science and impede scientific progress, by preventing the free circulation of knowledge. DATA AVAILABILITY Data available


at this link: osf.io/kzrc486 CODE AVAILABILITY Code can be accessed at this link: osf.io/kzrc4 REFERENCES * Merton, R. K. The normative structure of science. in _In_ _Norman W_. _Storer_


_(Ed.)_, _The sociology of science: Theoretical and empirical investigations_. 267–278 (The University of Chicago Press., 1942). * Nosek, B. A. & Bar-Anan, Y. Scientific Utopia: I.


Opening Scientific Communication. _Psychol. Inq._ 23, 217–243 (2012). Article  Google Scholar  * Murray-Rust, P. Open Data in Science. _Nat. Preced._ https://doi.org/10.1038/npre.2008.1526.1


(2008). Article  Google Scholar  * Bird, A. What Is Scientific Progress? _Noûs_ 41, 64–89 (2007). Article  Google Scholar  * Vasilevsky, N. A., Minnier, J., Haendel, M. A. & Champieux,


R. E. Reproducible and reusable research: are journal data sharing policies meeting the mark? _PeerJ_ 5, e3208 (2017). Article  PubMed  PubMed Central  Google Scholar  * Wilkinson, M. D. _et


al_. The FAIR Guiding Principles for scientific data management and stewardship. _Sci. Data_ 3, 160018 (2016). Article  PubMed  PubMed Central  Google Scholar  * Andreoli-Versbach, P. &


Mueller-Langer, F. Open access to data: An ideal professed but not practised. _Res. Policy_ 43 (2014). * Federer, L. M. _et al_. Data sharing in PLOS ONE: An analysis of Data Availability


Statements. _PLoS One_ 13, e0194768 (2018). Article  PubMed  PubMed Central  Google Scholar  * Savage, C. J. & Vickers, A. J. Empirical study of data sharing by authors publishing in


PLoS journals. _PLoS One_ 4 (2009). * Roche, D. G. _et al_. Slow improvement to the archiving quality of open datasets shared by researchers in ecology and evolution. _Proc. R. Soc. B Biol.


Sci._ 289, 20212780 (2022). Article  Google Scholar  * Tedersoo, L. _et al_. Data sharing practices and data availability upon request differ across scientific disciplines. _Sci. Data 2021


81_ 8, 1–11 (2021). Google Scholar  * Gabelica, M., Bojčić, R. & Puljak, L. Many researchers were not compliant with their published data sharing statement: a mixed-methods study. _J.


Clin. Epidemiol._ 150, 33–41 (2022). Article  PubMed  Google Scholar  * Tenopir, C. _et al_. Data sharing by scientists: Practices and perceptions. _PLoS One_ 6, (2011). * Feigenbaum, S.


& Levy, D. M. The market for (ir)reproducible econometrics. _Soc. Epistemol._ 7, 215–232 (1993). Article  Google Scholar  * Campbell, H. A., Micheli-Campbell, M. A. & Udyawer, V.


Early Career Researchers Embrace Data Sharing. _Trends Ecol. Evol._ 34, 95–98 (2019). Article  PubMed  Google Scholar  * Tenopir, C., Christian, L., Allard, S. & Borycz, J. Research Data


Sharing: Practices and Attitudes of Geophysicists. _Earth Sp. Sci._ 5, 891–902 (2018). Article  ADS  Google Scholar  * Stieglitz, S. _et al_. When are researchers willing to share their


data? – Impacts of values and uncertainty on open data in academia. _PLoS One_ 15, e0234172 (2020). Article  CAS  PubMed  PubMed Central  Google Scholar  * Houtkoop, B. L. _et al_. Data


Sharing in Psychology: A Survey on Barriers and Preconditions. _Adv. Methods Pract. Psychol. Sci._ 1, 70–85 (2018). Article  Google Scholar  * Linek, S. B., Fecher, B., Friesike, S. &


Hebing, M. Data sharing as social dilemma: Influence of the researcher’s personality. _PLoS One_ 12, e0183216 (2017). Article  PubMed  PubMed Central  Google Scholar  * Weisshaar, K. Publish


and Perish? An Assessment of Gender Gaps in Promotion to Tenure in Academia. _Soc. Forces_ 96, 529–560 (2017). Article  Google Scholar  * Ross, J. S. _et al_. Effect of blinded peer review


on abstract acceptance. _J. Am. Med. Assoc._ 295, 1675–1680 (2006). Article  CAS  Google Scholar  * Tomkins, A., Zhang, M. & Heavlin, W. D. Reviewer bias in single- versus double-blind


peer review. _Proc. Natl. Acad. Sci. USA_ 114, 12708–12713 (2017). Article  ADS  CAS  PubMed  PubMed Central  Google Scholar  * Krawczyk, M. & Smyk, M. Author’s gender affects rating of


academic articles: Evidence from an incentivized, deception-free laboratory experiment. _Eur. Econ. Rev._ 90, 326–335 (2016). Article  Google Scholar  * Card, D., DellaVigna, S., Funk, P.


& Iriberri, N. Are Referees and Editors in Economics Gender Neutral?*. _Q. J. Econ._ 135, 269–327 (2020). Article  Google Scholar  * Peng, H., Teplitskiy, M. & Jurgens, D. Author


Mentions in Science News Reveal Wide-Spread Ethnic Bias. _ArXiv Prepr_. ABS/2009.0, (2020). * Peng, H., Lakhani, K. & Teplitskiy, M. Acceptance in Top Journals Shows Large Disparities


across Name-inferred Ethnicities. _SocArXiv_ https://doi.org/10.31235/osf.io/mjbxg (2021). Article  Google Scholar  * Milkman, K. L., Akinola, M. & Chugh, D. What happens before? A field


experiment exploring how pay and representation differentially shape bias on the pathway into organizations. _J. Appl. Psychol._ 100, 1678–1712 (2015). Article  PubMed  Google Scholar  *


Ridgeway, C. L. Why Status Matters for Inequality. _Am. Sociol. Rev._ 79, 1–16 (2013). Article  Google Scholar  * Berger, J., Cohen, B. P. & Zelditch, M. Status Characteristics and


Social Interaction. _Am. Sociol. Rev._ 37, 241–255 (1972). Article  Google Scholar  * Correll, S. J., Weisshaar, K. R., Wynn, A. T. & Wehner, J. D. Inside the Black Box of Organizational


Life: The Gendered Language of Performance Assessment. _Am. Sociol. Rev._ 85, 1022–1050 (2020). Article  Google Scholar  * Melamed, D. & Savage, S. V. Status, Numbers and Influence.


_Soc. Forces_ 91, 1085–1104 (2013). Article  Google Scholar  * Hsu, M.-H. & Chang, C.-M. Examining interpersonal trust as a facilitator and uncertainty as an inhibitor of


intra-organisational knowledge sharing. _Inf. Syst. J._ 24, 119–142 (2014). Article  Google Scholar  * Rutten, W., Blaas-Franken, J. & Martin, H. The impact of (low) trust on knowledge


sharing. _J. Knowl. Manag._ 20, 199–214 (2016). Article  Google Scholar  * Fershtman, C. & Gneezy, U. Discrimination in a Segmented Society: An Experimental Approach*. _Q. J. Econ._ 116,


351–377 (2001). Article  MATH  Google Scholar  * Cettolin, E. & Suetens, S. Return on Trust is Lower for Immigrants. _Econ. J._ 129, 1992–2009 (2019). Article  Google Scholar  * Block,


R., Crabtree, C., Holbein, J. B. & Monson, J. Q. Are Americans less likely to reply to emails from Black people relative to White people? _Proc. Natl. Acad. Sci. USA_ 118, (2021). *


Booth, A. L., Leigh, A. & Varganova, E. Does Ethnic Discrimination Vary Across Minority Groups? Evidence from a Field Experiment*. _Oxf. Bull. Econ. Stat._ 74, 547–573 (2012). Article 


Google Scholar  * Baert, S. Hiring Discrimination: An Overview of (Almost) All Correspondence Experiments Since 2005 BT - Audit Studies: Behind the Scenes with Theory, Method, and Nuance. in


(ed. Gaddis, S. M.) 63–77. https://doi.org/10.1007/978-3-319-71153-9_3 (Springer International Publishing, 2018). * Gaddis, S. M. & Ghoshal, R. Searching for a Roommate: A


Correspondence Audit Examining Racial/Ethnic and Immigrant Discrimination among Millennials. _Socius_ 6, 2378023120972287 (2020). Article  Google Scholar  * Ross, J. S. _et al_. Effect of


Blinded Peer Review on Abstract Acceptance. _JAMA_ 295, 1675–1680 (2006). Article  CAS  PubMed  Google Scholar  * Harris, M. _et al_. Explicit bias toward high-income- country research: A


randomized, blinded, crossover experiment of English clinicians. _Health Aff._ 36, 1997–2004 (2017). Article  Google Scholar  * Milkman, K. L., Akinola, M. & Chugh, D. Temporal Distance


and Discrimination: An Audit Study in Academia. _Psychol. Sci._ 23, 710–717 (2012). Article  PubMed  Google Scholar  * Gerhards, J., Hans, S. & Drewski, D. Global inequality in the


academic system: effects of national and university symbolic capital on international academic mobility. _High. Educ._ 76, 669–685 (2018). Article  Google Scholar  * Tollefson, J. China


declared world’s largest producer of scientific articles. _Nature_ 553, 390–391 (2018). Article  ADS  CAS  PubMed  Google Scholar  * Brumfiel, G. Chinese students in the US: Taking a stand.


_Nature_ 438, 278–280 (2005). Article  ADS  CAS  PubMed  Google Scholar  * Bartlett, T., & Fischer, K. The China Conundrum. _The New York Times. (Retrieved October 2022)_ (2011). *


Pager, D. & Quillian, L. Walking the Talk? What Employers Say Versus What They Do: Am. _Sociol._ 70, 355–380 (2005). Google Scholar  * Riach, P. A. & Rich, J. Field Experiments of


Discrimination in the Market Place*. _Econ. J._ 112, F480–F518 (2002). Article  Google Scholar  * Gaddis, S. M. Understanding the “How” and “Why” Aspects of Racial-Ethnic Discrimination: A


Multimethod Approach to Audit Studies: Sociology of Race and Ethnicity 5, 443–455 (2019). * Thursby, J. G., Haeussler, C., Thursby, M. C. & Jiang, L. Prepublication disclosure of


scientific results: Norms, competition, and commercial orientation. _Sci. Adv._ 4, eaar2133 (2022). Article  ADS  Google Scholar  * Abadie, A., Athey, S., Imbens, G. W., & Wooldridge, J.


When should you adjust standard errors for clustering? (_No. w24003)_. _Natl. Bur. Econ. Res_. (2022). * Diao, W. Between Ethnic and English Names: Name Choice for Transnational Chinese


Students in a US Academic Community. _J. Int. Students_ 4, 205–222 (2014). Article  Google Scholar  * Griffin, R. 120 years of Olympic history: athletes and results. * Xie, F. rethnicity: An


R package for predicting ethnicity from names. _SoftwareX_ 17, 100965 (2022). Article  Google Scholar  * U.S. Census Bureau:


https://www.census.gov/topics/population/genealogy/data/2010_surnames.html (2022). * Gaddis, S. M. Signaling class: An experiment examining social class perceptions from names used in


correspondence audit studies. Preprint at: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3350739 (2019). * Crabtree, C. An Introduction to Conducting Email Audit Studies BT - Audit


Studies: Behind the Scenes with Theory, Method, and Nuance. in (ed. Gaddis, S. M.) Ch. 6 (Springer International Publishing, 2018). * R Core Team. _R: A language and environment for


statistical computing_. (2016). * Blair, G., Cooper, J., Coppock, A., Humphreys, M. & Sonnet, L. Estimatr: Fast estimators for design-based inference. _R Packag. version_ (2019). *


Krawczyk, M. & Reuben, E. (Un)Available upon Request: Field Experiment on Researchers’ Willingness to Share Supplementary Materials. _Account. Res_. 19, (2012). * Bhati, A. Does Implicit


Color Bias Reduce Giving? Learnings from Fundraising Survey Using Implicit Association Test (IAT). _Volunt. Int. J. Volunt. Nonprofit Organ._ 32, 340–350 (2021). Article  Google Scholar  *


Stepanikova, I., Triplett, J. & Simpson, B. Implicit racial bias and prosocial behavior. _Soc. Sci. Res._ 40, 1186–1195 (2011). Article  Google Scholar  * Vogeli, C. _et al_. Data


Withholding and the Next Generation of Scientists: Results of a National Survey. _Acad. Med_. 81, (2006). * Kim, Y. & Zhang, P. Understanding data sharing behaviors of STEM researchers:


The roles of attitudes, norms, and data repositories. _Libr. Inf. Sci. Res._ 37, 189–200 (2015). Article  Google Scholar  * Browne, I. & Misra, J. The Intersection of Gender and Race in


the Labor Market. _Annu. Rev. Sociol._ 29, 487–513 (2003). Article  Google Scholar  * Fiske, S. T., Cuddy, A. J. C., Glick, P. & Xu, J. A model of (often mixed) stereotype content:


Competence and warmth respectively follow from perceived status and competition. _Journal of Personality and Social Psychology_ 82, 878–902 (2002). Article  PubMed  Google Scholar  * Eagly,


A. H. & Crowley, M. Gender and helping behavior: A meta-analytic review of the social psychological literature. _Psychol. Bull._ 100, 283–308 (1986). Article  Google Scholar  *


Dufwenberg, M. & Muren, A. Generosity, anonymity, gender. _J. Econ. Behav. Organ._ 61, 42–49 (2006). Article  Google Scholar  * Weber, M., Koehler, C. & Schnauber-Stockmann, A. Why


Should I Help You? Man Up! Bystanders’ Gender Stereotypic Perceptions of a Cyberbullying Incident. _Deviant Behav._ 40, 585–601 (2019). Article  Google Scholar  * Erlandsson, A. _et al_.


Moral preferences in helping dilemmas expressed by matching and forced choice. _Judgm. Decis. Mak._ 15, 452–475 (2020). Article  Google Scholar  * Bursell, M. The Multiple Burdens of


Foreign-Named Men—Evidence from a Field Experiment on Gendered Ethnic Hiring Discrimination in Sweden. _Eur. Sociol. Rev._ 30, 399–409 (2014). Article  Google Scholar  * Arai, M., Bursell,


M. & Nekby, L. The Reverse Gender Gap in Ethnic Discrimination: Employer Stereotypes of Men and Women with Arabic Names. _Int. Migr. Rev._ 50, 385–412 (2016). Article  Google Scholar  *


Carol, S., Eich, D., Keller, M., Steiner, F. & Storz, K. Who can ride along? Discrimination in a German carpooling market. _Popul. Space Place_ 25, e2249 (2019). Article  Google Scholar


  * Dahl, M. & Krog, N. Experimental Evidence of Discrimination in the Labour Market: Intersections between Ethnicity, Gender, and Socio-Economic Status. _Eur. Sociol. Rev._ 34, 402–417


(2018). Article  Google Scholar  * Flage, A. Ethnic and gender discrimination in the rental housing market: Evidence from a meta-analysis of correspondence tests, 2006–2017. _J. Hous. Econ._


41, 251–273 (2018). Article  ADS  Google Scholar  * Midtbøen, A. H. Discrimination of the Second Generation: Evidence from a Field Experiment in Norway. _J. Int. Migr. Integr._ 17, 253–272


(2016). Article  Google Scholar  * Simonovits, B., Shvets, I. & Taylor, H. Discrimination in the sharing economy: evidence from a Hungarian field experiment. _Corvinus J. Sociol. Soc.


Policy_ 9, 55–79 (2018). Article  Google Scholar  * Sidanius, J. & Pratto, F. _Social dominance: An intergroup theory of social hierarchy and oppression. Social dominance: An intergroup


theory of social hierarchy and oppression_. (Cambridge University Press, 1999). * Ert, E., Fleischer, A. & Magen, N. Trust and reputation in the sharing economy: The role of personal


photos in Airbnb. _Tour. Manag._ 55, 62–73 (2016). Article  Google Scholar  * Eagly, A. H. & Kite, M. E. Are stereotypes of nationalities applied to both women and men? _J. Pers. Soc.


Psychol._ 53, 451–462 (1987). Article  Google Scholar  * Guo, E., Aloe, J., & Hao, K. The US crackdown on Chinese economic espionage is a mess. We have the data to show it. _MIT


Technology Review_ (2021). * Lu, Y., Kaushal, N., Huang, X. & Gaddis, S. M. Priming COVID-19 salience increases prejudice and discriminatory intent against Asians and Hispanics. _Proc.


Natl. Acad. Sci._ 118, e2105125118 (2021). Article  CAS  PubMed  PubMed Central  Google Scholar  * Cao, A., Lindo, J. M. & Zhong, J. _Can Social Media Rhetoric Incite Hate Incidents?


Evidence from Trump’s“ Chinese Virus” Tweets_. (2022). * Ryan, T. J. & Krupnikov, Y. Split Feelings: Understanding Implicit and Explicit Political Persuasion. _Am. Polit. Sci. Rev._ 115,


1424–1441 (2021). Article  Google Scholar  * Quintana, D. S. A synthetic dataset primer for the biobehavioural sciences to promote reproducibility and hypothesis generation. _Elife_ 9,


e53275 (2020). Article  PubMed  PubMed Central  Google Scholar  * Acciai, C., Jesper, W. S. & Mathias, W. N. Estimating social bias in data sharing behaviours: an open science


experiment. _Open Science Framework_ https://doi.org/10.17605/OSF.IO/PJC9G (2023). Download references ACKNOWLEDGEMENTS This study was funded by Carlsbergfondet (the Carlsberg foundation) –


Award # CF19-0566. [P.I. M.W.N] The funder had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript. AUTHOR INFORMATION AUTHORS AND


AFFILIATIONS * Department of Sociology, University of Copenhagen, Øster Farimagsgade 5, 1353, Copenhagen, Denmark Claudia Acciai & Mathias W. Nielsen * Danish Centre for Studies in


Research and Research Policy, Department of Political Science, Aarhus University, Bartholins Allé 7, 8000, Aarhus C, Denmark Jesper W. Schneider Authors * Claudia Acciai View author


publications You can also search for this author inPubMed Google Scholar * Jesper W. Schneider View author publications You can also search for this author inPubMed Google Scholar * Mathias


W. Nielsen View author publications You can also search for this author inPubMed Google Scholar CONTRIBUTIONS C.A. and M.W.N. designed research with input from J.W.S.; C.A. and M.W.N.


performed research; J.W.S. contributed supplementary data. C.A. and M.W.N. analyzed data; C.A. and M.W.N. and wrote the paper with input from J.W.S. CORRESPONDING AUTHORS Correspondence to


Claudia Acciai or Mathias W. Nielsen. ETHICS DECLARATIONS COMPETING INTERESTS The authors declare no competing interests. ADDITIONAL INFORMATION PUBLISHER’S NOTE Springer Nature remains


neutral with regard to jurisdictional claims in published maps and institutional affiliations. SUPPLEMENTARY INFORMATION SUPPLEMENTARY INFORMATION RIGHTS AND PERMISSIONS OPEN ACCESS This


article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as


you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party


material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s


Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder.


To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/. Reprints and permissions ABOUT THIS ARTICLE CITE THIS ARTICLE Acciai, C., Schneider, J.W. & Nielsen,


M.W. Estimating social bias in data sharing behaviours: an open science experiment. _Sci Data_ 10, 233 (2023). https://doi.org/10.1038/s41597-023-02129-8 Download citation * Received: 01


February 2023 * Accepted: 31 March 2023 * Published: 21 April 2023 * DOI: https://doi.org/10.1038/s41597-023-02129-8 SHARE THIS ARTICLE Anyone you share the following link with will be able


to read this content: Get shareable link Sorry, a shareable link is not currently available for this article. Copy to clipboard Provided by the Springer Nature SharedIt content-sharing


initiative