Meta-Psychology
https://conferences.lnu.se/index.php/metapsychology
<p>Meta-Psychology publishes theoretical and empirical contributions that advance psychology as a science through critical discourse related to individual articles, research lines, research areas, or psychological science as a field.</p>Linnaeus University Pressen-USMeta-Psychology2003-2714Response to responsible research assessment I and II from the perspective of the DGPs working group on open science in clinical psychology
https://conferences.lnu.se/index.php/metapsychology/article/view/3794
<p>We comment on the papers by Schönbrodt et al. (2022) and Gärtner et al. (2022) on responsible research assessment from the perspective of clinical psychology and psychotherapy research. Schönbrodt et al. (2022) propose four principles to guide hiring and promotion in psychology: (1) In addition to publications in scientific journals, data sets and the development of research software should be considered. (2) Quantitative metrics can be useful, but they should be valid and applied responsibly. (3) Methodological rigor, research impact, and work quantity should be considered as three separate dimensions for evaluating research contributions. (4) The quality of work should be prioritized over the number of citations or the quantity of research output. From the perspective of clinical psychology, we endorse the initiative to update current practice by establishing a matrix for comprehensive, transparent and fair evaluation criteria. In the following, we will both comment on and complement these criteria from a clinical-psychological perspective.</p>Jakob Fink-LamotteKevin HilbertDorothée BentzSimon BlackwellJan R. BoehnkeJuliane BurghardtBarbara CludiusJohannes C. EhrenthalMoritz ElsaesserAnke HaberkampTanja HechlerAnja KräplinChristian ParetLars SchulzeSarah WilkerHelen Niemeyer
Copyright (c) 2024 Jakob Fink-Lamotte, Kevin Hilbert, Dorothée Bentz, Simon Blackwell, Jan R. Boehnke, Juliane Burghardt, Barbara Cludius, Johannes C. Ehrenthal, Moritz Elsaesser, Anke Haberkamp, Tanja Hechler, Anja Kräplin, Christian Paret, Lars Schulze, Sarah Wilker, Helen Niemeyer
https://creativecommons.org/licenses/by/4.0/
2024-03-172024-03-17810.15626/MP.2023.3794Assessing rigor and impact of research software for hiring and promotion in psychology: A comment on Gärtner et al. (2022)
https://conferences.lnu.se/index.php/metapsychology/article/view/3715
<p>Based on four principles of a more responsible research assessment in academic hiring and promotion processes, Gärtner et al. (2022) suggested an evaluation scheme for published manuscripts, reusable data sets, and research software. This commentary responds to the proposed indicators for the evaluation of research software contributions in academic hiring and promotion processes. Acknowledging the significance of research software as a critical component of modern science, we propose that an evaluation scheme must emphasize the two major dimensions of rigor and impact. Generally, we believe that research software should be recognized as valuable scientific output in academic hiring and promotion, with the hope that this incentivizes the development of more open and better research software.</p>Andreas Markus BrandmaierMaximilian ErnstAaron Peikert
Copyright (c) 2024 Andreas M. Brandmaier, Maximilian Ernst, Aaron Peikert
https://creativecommons.org/licenses/by/4.0/
2024-07-152024-07-15810.15626/MP.2023.3715Responsible Research Assessment requires structural more than procedural reforms
https://conferences.lnu.se/index.php/metapsychology/article/view/3734
<p>In their target articles, Schönbrodt et al. (2022) and Gärtner et al. (2022) propose new metrics and their practical implementation to improve responsible research assessment. Generally, I welcome the inclusion of open science and scientific rigor into evaluating job candidates. However, the proposed reform mainly focuses on the first stage of selecting candidates who then continue towards a second stage of in-depth evaluation of research quality. Yet, this second selection stage is underdeveloped but likely more critical concerning responsible research assessment and hiring decisions. I argue that an adequate assessment of research quality at this second stage requires the representation of specific knowledge in the subfield of a discipline that the candidate should be hired for by the hiring committee. This is rarely achieved given the current structural organization of departments, especially in German-speaking countries, and potentially explains the reliance on suboptimal indicators such as h-index and Journal Impact factor. Therefore, I argue that responsible research assessment requires structural reform to ensure that institutions have several researchers in permanent positions with specific knowledge in different subfields to provide an adequate and responsible assessment of research quality by hiring committees at all evaluation stages.</p>Gidon T. Frischkorn
Copyright (c) 2024 Gidon T. Frischkorn
https://creativecommons.org/licenses/by/4.0/
2024-03-172024-03-17810.15626/MP.2023.3734Responsible Research Assessment Should Prioritize Theory Development and Testing Over Ticking Open Science Boxes
https://conferences.lnu.se/index.php/metapsychology/article/view/3735
<p>We appreciate the initiative to seek for ways to improve academic assessment by broadening the range of relevant research contributions and by considering a candidate’s scientific rigor. Evaluating a candidate's ability to contribute to science is a complex process that cannot be captured through one metric alone. While the proposed changes have some advantages, such as an increased focus on quality over quantity, the proposal's focus on adherence to open science practices is not sufficient, as it undervalues theory building and formal modelling: A narrow focus on open science conventions is neither a sufficient nor valid indicator for a “good scientist” and may even encourage researchers to choose easy, pre-registerable studies rather than engage in time-intensive theory building. Further, when in a first step only a minimum standard for following easily achievable open science goals is set, most applicants will soon pass this threshold. At this point, one may ask if the additional benefit of such a low bar outweighs the potential costs of such an endeavour. We conclude that a reformed assessment system should put at least equal emphasis on theory building and adherence to open science principles and should not completely disregard traditional performance metrices.</p>Hannah DamesPhilipp MusfeldVencislav PopovKlaus OberauerGidon T. Frischkorn
Copyright (c) 2024 Hannah Dames, Philipp Musfeld, Vencislav Popov, Klaus Oberauer, Gidon T. Frischkorn
https://creativecommons.org/licenses/by/4.0/
2024-03-172024-03-17810.15626/MP.2023.3735Valuing Preprints Must be Part of Responsible Research Assessment
https://conferences.lnu.se/index.php/metapsychology/article/view/3758
<p>Comments on papers by Schönbrodt et al. (2022) and Gärtner et al. (2022) proposing reforms to the research assessment process. Given the prominent role of preprints in contemporary scientific practice, they must be an accepted and central component of research assessment.</p>Moin Syed
Copyright (c) 2024 Moin Syed
https://creativecommons.org/licenses/by/4.0/
2024-03-172024-03-17810.15626/MP.2023.3758Indicators for teaching assessment
https://conferences.lnu.se/index.php/metapsychology/article/view/3763
<p>This commentary on Schönbrodt et al. (2022) and Gärtner et al. (2022) aims at complementing the ideas regarding an implementation of DORA for the domain of teaching. As there is neither a comprehensive assessment system based on empirical data nor a competence model for teaching competencies available, yet, we describe some pragmatic ideas for indicators of good teaching and formulate desiderates for future research programs and validation.</p>Miriam HansenJulia BeitnerHolger HorzMartin Schultze
Copyright (c) 2024 Miriam Hansen, Julia Beitner, Holger Horz, Martin Schultze
https://creativecommons.org/licenses/by/4.0/
2024-03-172024-03-17810.15626/MP.2023.3763Research assessment using a narrow definition of “research quality” is an act of gatekeeping: A comment on Gärtner et al. (2022)
https://conferences.lnu.se/index.php/metapsychology/article/view/3764
<p>Gärtner et al. (2022) propose a system for quantitatively scoring the methodological rigour of papers during the hiring and promotion of psychology researchers, with the aim of advantaging researchers who conduct open, reproducible work. However, the quality criteria proposed for assessing methodological rigour are drawn from a narrow post-positivist paradigm of quantitative, confirmatory research conducted from an epistemology of scientific realism. This means that research conducted from a variety of other approaches, including constructivist, qualitative research, becomes structurally disadvantaged under the new system. The implications of this for particular fields, demographics of researcher, and the future of the discipline of psychology are discussed.</p>Tom Hostler
Copyright (c) 2024 Tom Hostler
https://creativecommons.org/licenses/by/4.0/
2024-03-172024-03-17810.15626/MP.2023.3764Comment on "Responsible Research Assessment: Implementing DORA for hiring and promotion in psychology”
https://conferences.lnu.se/index.php/metapsychology/article/view/3779
<p>In target papers, Schönbrodt et al. (2022), and Gärtner et al. (2022) proposed to broaden the range of the considered research contributions, namely (i) bringing strong empirical evidence, (ii) building open databases, (iii) building and maintaining packages, where each dimension being scored independently in marking scheme. Using simulations, we show that the current proposal places a significant weight on software development, potentially at the expense of other academic activities – a weight that should be explicit to committees before they make use of the proposed marking scheme. Following Gärtner et al. (2022) recommendations, we promote the use of flexible weights which more closely match an institution’s specific needs by the weighting of the relevant dimensions. We propose a Shinyapp that implement the marking scheme with adaptative weights to both help the hiring committee define and foresee the consequences of weights’ choices and increase the transparency and understandability of the procedure.</p>Victor AugerNele Claes
Copyright (c) 2024 Victor Auger, Nele Claes
https://creativecommons.org/licenses/by/4.0/
2024-03-172024-03-17810.15626/MP.2023.3779Responsible Research is also concerned with generalizability: Recognizing efforts to reflect upon and increase generalizability in hiring and promotion decisions in psychology
https://conferences.lnu.se/index.php/metapsychology/article/view/3695
<p>We concur with the authors of the two target articles that Open Science practices can help combat the ongoing reproducibility and replicability crisis in psychological science and should hence be acknowledged as responsible research practices in hiring and promotion decisions. However, we emphasize that another crisis is equally threatening the credibility of psychological science in Germany: The sampling or generalizability crisis. We suggest that scientists’ efforts to contextualize their research, reflect upon, and increase its generalizability should be incentivized as responsible research practices in hiring and promotion decisions. To that end, we present concrete suggestions for how efforts to combat the additional generalizability crisis could be operationalized within Gärtner et al. (2022) evaluation scheme. Tackling the replicability and the generalizability crises in tandem will advance the credibility and quality of psychological science and teaching in Germany.</p>Roman StengelinManuel BohnAlejandro Sánchez-AmaroDaniel HaunMaleen ThieleMoritz DaumElisa FelscheFrankie FongAnja GampeMarta Giner TorrénsSebastian GrueneisenDavid HardeckerLisa HornKarri NeldnerSarah Pope-CaldwellNils Schuhmacher
Copyright (c) 2024 Roman Stengelin, Manuel Bohn, Alejandro Sánchez-Amaro, Daniel Haun, Maleen Thiele, Moritz Daum, Elisa Felsche, Frankie Fong, Anja Gampe, Marta Giner Torréns, Sebastian Grueneisen, David Hardecker, Lisa Horn, Karri Neldner, Sarah Pope-Caldwell, Nils Schuhmacher
https://creativecommons.org/licenses/by/4.0/
2024-03-172024-03-17810.15626/MP.2023.3695Responsible research assessment in the area of quantitative methods research: A comment on Gärtner et al.
https://conferences.lnu.se/index.php/metapsychology/article/view/3796
<p>In this commentary, we discuss the proposed criteria in Gärtner et al. (2022) for hiring or promoting quantitative methods researchers. We argue that the criteria do not reflect aspects that are relevant to quantitative methods researchers and typical publications they produce. We introduce a new set of criteria that can be used to evaluate the performance of quantitative methods researchers in a more valid fashion. We discuss the necessity to balance scientific expertise and open science commitment in such ranking schemes.</p>Holger BrandtMirka HenningerEsther UlitzschKristian KleinkeThomas Schäfer
Copyright (c) 2024 Holger Brandt, Mirka Henninger, Esther Ulitzsch, Kristian Kleinke, Thomas Schäfer
https://creativecommons.org/licenses/by/4.0/
2024-03-172024-03-17810.15626/MP.2023.3796Responsible assessment of what research? Beware of epistemic diversity!
https://conferences.lnu.se/index.php/metapsychology/article/view/3797
<p>Schönbrodt et al. (2022) and Gärtner et al. (2022) aim to outline in the target articles why and how research assessment could be improved in psychological science in accordance with DORA, resulting in a focus on abandoning the impact factor as an indicator for research quality and aligning assessment with methodological rigor and open science practices. However, I argue that their attempt is guided by a rather narrow statistical and quantitative understanding of knowledge production in psychological science. Consequently, the authors neglect the epistemic diversity within psychological science, leading to the potential danger of committing epistemic injustice. Hence, the criteria they introduce for research assessment might be appropriate for some approaches to knowledge production; it could, however, neglect or systematically disadvantage others. Furthermore, I claim that the authors lack some epistemic (intellectual) humility about their proposal. Further information is required regarding when and for which approaches their proposal is appropriate and, maybe even more importantly, when and where it is not. Similarly, a lot of the proposed improvements of the reform movement, like the one introduced in the target articles, are probably nothing more than trial and error due to a lack of investigation of their epistemic usefulness and understanding of underlying mechanisms and theories. Finally, I argue that with more awareness about epistemic diversity in psychological science in combination with more epistemic (intellectual) humility, the danger of epistemic injustice could be attenuated.</p>Sven Ulpts
Copyright (c) 2024 Sven Ulpts
https://creativecommons.org/licenses/by/4.0/
2024-03-172024-03-17810.15626/MP.2023.3797A broader view of research contributions: Necessary adjustments to DORA for hiring and promotion in psychology.
https://conferences.lnu.se/index.php/metapsychology/article/view/3652
<p>Recently Schönbrodt et al. (2022) released recommendations for improving how psychologists could be evaluated for recruitment, retention, and promotion. Specifically, they provided four principles of responsible research assessment in response to current methods that rely heavily on bibliometric indices of journal quality and research impact. They build their case for these principles on the San Francisco Declaration on Research Assessment (DORA) perspective that decries reliance on invalid quantitative metrics of research quality and productivity in hiring and promotion. The paper makes clear the tension panels have to address in evaluating applications—too little time to do an in-depth evaluation of an individual’s career and contribution, so reliance on easy to understand, but perhaps invalid, metrics. This dilemma requires an alternative mechanism rather than simply a rejection of metrics. To that end, the authors are to be congratulated for operationalising what those alternatives might look like. Nonetheless, the details embedded in the principles seem overly narrow and restrictive.</p>Gavin Brown
Copyright (c) 2024 Gavin Brown
https://creativecommons.org/licenses/by/4.0/
2024-03-172024-03-17810.15626/MP.2022.3652Commentary: “Responsible Research Assessment: Implementing DORA for hiring and promotion in psychology”
https://conferences.lnu.se/index.php/metapsychology/article/view/3655
<p>A commentary on: Gärtner et al., 2022; Schönbrodt et al., 2022.</p>Alejandro Sandoval-Lentisco
Copyright (c) 2024 Alejandro Sandoval-Lentisco
https://creativecommons.org/licenses/by/4.0/
2024-03-172024-03-17810.15626/MP.2022.3655Interdisciplinary Value
https://conferences.lnu.se/index.php/metapsychology/article/view/3679
<p>This is a commentary on interdisciplinary value in the special issue "Responsible Research Assessment: Implementing DORA for hiring and promotion in psychology."</p>Veli-Matti Karhulahti
Copyright (c) 2024 Veli-Matti Karhulahti
https://creativecommons.org/licenses/by/4.0/
2024-03-172024-03-17810.15626/MP.2023.3679Comment on: Responsible Research Assessment I and Responsible Research Assessment II
https://conferences.lnu.se/index.php/metapsychology/article/view/3685
<p>A long-term personnel policy in filling professorships, aimed at remedying deficits in psychological research, should be able to significantly improve the scientific quality of psychology: “The main reason is that the hiring and promotion of such researchers is most likely to contribute to the emergence of a credible scientific knowledge base“ (Gärtner et al., in press). </p>Erich H. Witte
Copyright (c) 2024 Erich H. Witte
https://creativecommons.org/licenses/by/4.0/
2024-03-172024-03-17810.15626/MP.2023.3685Re-analysis of a meta-analysis about tryptophan and depression
https://conferences.lnu.se/index.php/metapsychology/article/view/3716
<p style="font-weight: normal; line-height: 200%; margin-bottom: 0cm;" align="left"><span style="font-size: small;">This is a reanalysis of a meta-analysis about L-tryptophan blood levels and depression, which became part of the controversy around a recent umbrella review about the role of serotonin in depression. The reanalysis revealed major methodological limitations, raising doubts on the conclusions in the original publication that levels of tryptophan are lowered among depressed compared to non-depressed individuals. The data is also compatible with a null effect and no firm conclusion should be made.</span></p>Martin Plöderl
Copyright (c) 2024 Martin Plöderl
https://creativecommons.org/licenses/by/4.0/
2024-05-032024-05-03810.15626/MP.2023.3716The Untrustworthy Evidence in Dishonesty Research
https://conferences.lnu.se/index.php/metapsychology/article/view/3987
<p>Replicable and reliable research is essential for cumulative science and its applications in practice. This article examines the quality of research on dishonesty using a sample of 286 hand-coded test statistics from 99 articles. Z-curve analysis indicates a low expected replication rate, a high proportion of missing studies, and an inflated false discovery risk. Test of insufficient variance (TIVA) finds that 11/61 articles with multiple test statistics contain results that are ``too-good-to-be-true''. Sensitivity analysis confirms the robustness of the findings. In conclusion, caution is advised when relying on or applying the existing literature on dishonesty.</p>František Bartoš
Copyright (c) 2024 František Bartoš
https://creativecommons.org/licenses/by/4.0/
2024-04-192024-04-19810.15626/MP.2023.3987How should we investigate variation in the relation between social media and well-being?
https://conferences.lnu.se/index.php/metapsychology/article/view/3322
<p>Most researchers studying the relation between social media use and well-being find small to no associations, yet policymakers and public stakeholders keep asking for more evidence. One way the field is reacting is by inspecting the variation around average relations—with the goal of describing individual social media users. Here, we argue that this approach produces findings that are not as informative as they could be. Our analysis begins by describing how the field got to this point. Then, we explain the problems with the current approach of studying variation and how it loses sight of one of the most important goals of a quantitative social science: generalizing from a sample to a population. We propose a principled approach to quantify, interpret, and explain variation in average relations by: (1) conducting model comparisons, (2) defining a region of practical equivalence and testing the theoretical distribution of relations against that region, (3) defining a smallest effect size of interest and comparing it against the theoretical distribution. We close with recommendations to either study moderators as systematic factors that explain variation or to commit to a person-specific approach and conduct N=1 studies and qualitative research.</p>Niklas JohannesPhilipp K. MasurMatti VuorreAndrew K. Przybylski
Copyright (c) 2024 Niklas Johannes, Philipp K. Masur, Matti Vuorre, Andrew K. Przybylski
https://creativecommons.org/licenses/by/4.0/
2024-07-012024-07-01810.15626/MP.2022.3322Investigating Heterogeneity in (Social) Media Effects: Experience-Based Recommendations
https://conferences.lnu.se/index.php/metapsychology/article/view/3649
<p><span id="page3R_mcid13" class="markedContent">We recently introduced a new, unified approach to investigate the effects of social media use on well-being. Using experience sampling methods among sizeable samples of respondents, our unified approach combines the strengths of nomothetic methods of analysis (e.g., mean comparisons, regression models), which are suited to understand group averages and generalize to populations, with idiographic methods of analysis (e.g., N=1 time series analyses), which are suitable to assess the effects of social media use on each single person (i.e., person-specific effects). Our approach challenges existing knowledge of media effects based on the nomothetic-only approach. As with many innovations, our approach has raised questions. In this article, we discuss our experience with our unified media effects approach that we have been building since 2018. We will explain what our approach exactly entails and what it requires. For example, how many observations are needed per person? Which methods did we employ to assess the meaningfulness of variation around average effects? How can we generalize our findings to our target populations? And how can our person-specific results aid policy decisions? Finally, we hope to answer questions of colleagues who are interested in replicating, extending, or building on our work.<br /></span></p>Patti ValkenburgIne BeyensLoes Keijsers
Copyright (c) 2024 Patti Valkenburg, Ine Beyens, Loes Keijsers
https://creativecommons.org/licenses/by/4.0/
2024-07-012024-07-01810.15626/MP.2022.3649The many faces of early life adversity - Content overlap in validated assessment instruments as well as in fear and reward learning research
https://conferences.lnu.se/index.php/metapsychology/article/view/3958
<p>The precise assessment of childhood adversity is crucial for understanding the impact of aversive events on mental and physical development. However, the plethora of assessment tools currently used in the literature with unknown overlap in childhood adversity types covered hamper comparability and cumulative knowledge generation. In this study, we conducted two separate item-level content analyses of in total 35 questionnaires aiming to assess childhood adversity. These include 13 questionnaires that were recently recommended based on strong psychometric properties as well as additional 25 questionnaires that were identified through a systematic literature search. The latter provides important insights into the actual use of childhood adversity questionnaires in a specific, exemplary research field (i.e., the association between childhood adversity and threat and reward learning). Of note, only 3 of the recommended questionnaires were employed in this research field. Both item-wise content analysis illustrate substantial heterogeneity in the adversity types assessed across these questionnaires and hence highlight limited overlap in content (i.e., adversity types) covered by different questionnaires. Furthermore, we observed considerable differences in structural properties across all included questionnaires such as the number of items, age ranges assessed as well as the specific response formats (e.g., binary vs. continuous assessments, self vs. caregiver). We discuss implications for the interpretation, comparability and integration of the results from the existing literature and derive specific recommendations for future research. In sum, the substantial heterogeneity in the assessment and operationalization of childhood adversity emphasizes the urgent need for theoretical and methodological solutions to promote comparability, replicability of childhood adversity assessment and foster cumulative knowledge generation in research on the association of childhood adversity and physical as well as psychological health.</p>Alina KoppoldJulia RugeTobias HeckerTina Lonsdorf
Copyright (c) 2024 Alina Koppold, Julia Ruge, Tobias Hecker, Tina B. Lonsdorf
https://creativecommons.org/licenses/by/4.0/
2024-08-072024-08-07810.15626/MP.2023.3958ReproduceMe: Lessons from a pilot project on computational reproducibility
https://conferences.lnu.se/index.php/metapsychology/article/view/4021
<p><span style="font-weight: 400;">If a scientific paper is computationally reproducible, the analyses it reports can be repeated independently by others. At the present time most papers are not reproducible. However, the tools to enable computational reproducibility are now widely available, using free and open source software. We conducted a pilot study in which we offered ‘reproducibility as a service’ within a UK psychology department for a period of 6 months. Our rationale was that most researchers lack either the time or expertise to make their own work reproducible, but might be willing to allow this to be done by an independent team. Ten papers were converted into reproducible format using <em>R markdown</em>, such that all analyses were conducted by a single script that could download raw data from online platforms as required, generate figures, and produce a pdf of the final manuscript. For some studies this involved reproducing analyses originally conducted using commercial software. The project was an overall success, with strong support from the contributing authors who saw clear benefit from this work, including greater transparency and openness, and ease of use for the reader. Here we describe our framework for reproducibility, summarise the specific lessons learned during the project, and discuss the future of computational reproducibility. Our view is that computationally reproducible manuscripts embody many of the core principles of open science, and should become the default format for scientific communication.</span></p>Daniel H. BakerMareike BergKirralise J. HansfordBartholomew P.A. QuinnFederico G. SegalaErin L. Warden-English
Copyright (c) 2024 Daniel H. Baker, Mareike Berg, Kirralise J. Hansford, Bartholomew P.A. Quinn, Federico G. Segala, Erin L. Warden-English
https://creativecommons.org/licenses/by/4.0/
2024-09-062024-09-06810.15626/MP.2023.4021Testing the Usability of the Psychological Research Preregistration-Quantitative (PRP-QUANT) Template
https://conferences.lnu.se/index.php/metapsychology/article/view/4039
<p>The Psychological Research Preregistration-Quantitative (PRP-QUANT) Template provides researchers with a comprehensive list of elements to consider when planning a psychological study. We assessed its usability and researchers’ intention to use it. We conducted a usability test (study 1) and surveyed researchers who submitted or reviewed a preregistration created with the template (study 2, authors: <em>N</em> = 19, reviewers: <em>N</em> = 29) regarding their impression of the template. For the usability test, we recruited participants via the mailing lists of the German Psychological Society, the American Psychological Association, and the British Psychological Society, and social media. Participants answered selected template and web probing items and provided an overall rating (<em>N</em> = 88). Based on the Unified Theory of Acceptance and Use of Technology (UTAUT), we expected that the intention to use the template is influenced by performance expectancy (moderated by age), effort expectancy (moderated by age and experience), and social influence (moderated by age, experience, and voluntariness, <em>N</em> = 60). The results suggest that the PRP-QUANT Template is suitable for different research areas within psychology, is evaluated as effective, and perceived positively. Performance expectancy and all predictors combined significantly predicted researchers’ intention to use the template. A plain language summary in English and German is available in Text S1 in the supplemental material.</p>Lisa SpitzerMichael BosnjakStefanie Mueller
Copyright (c) 2024 Lisa Spitzer, Michael Bosnjak, Stefanie Mueller
https://creativecommons.org/licenses/by/4.0/
2024-12-202024-12-20810.15626/MP.2023.4039Evaluating the Replicability of Social Priming Studies
https://conferences.lnu.se/index.php/metapsychology/article/view/3308
<p>To assess the replicability of social priming findings we reviewed the extant close replication attempts in the field. In total, we found 70 close replications, that replicated 49 unique findings. Ninety-four percent of the replications had effect sizes smaller than the effect they replicated and only 17% of the replications reported a significant p-value in the original direction. The strongest predictor of replication success was whether or not the replication team included at least one of the authors of the original paper. Twelve of the 18 replications with at least one original author produced a significant effect in the original direction and the meta-analytic average of these studies suggest a significant priming effect (d = 0.40, 95% CI[0.23; 0.58]). In stark contrast, none of the 52 replications by independent research teams produced a significant effect in the original direction and the meta-analytic average was virtually zero (d = 0.002, 95% CI[-0.03; 0.03]). We argue that these results have shifted the burden of proof back onto advocates of social priming. Successful replications from independent research teams will likely be required to convince sceptics that social priming exists at all.</p>Erik Mac GiollaSimon KarlssonDavid A. NeequayeMagnus Bergquist
Copyright (c) 2024 Erik Mac Giolla, Simon Karlsson, David A. Neequaye, Magnus Bergquist
https://creativecommons.org/licenses/by/4.0/
2024-11-122024-11-12810.15626/MP.2022.3308Bayesian Evaluation of Replication Studies
https://conferences.lnu.se/index.php/metapsychology/article/view/2554
<p>In this paper a method is proposed to determine whether the result from an original study is corroborated in a replication study. The paper is illustrated using two replication studies and the corresponding original studies from the Reproducibility Project: Psychology by the Open Science Collaboration. This method emphasizes the need to determine what one wants to replicate from the original paper. This can be done by translating the research hypotheses formulated in the introduction into informative hypotheses, or, by translating the results into interval hypotheses. The Bayes factor will be used to determine whether the hypotheses resulting from the original study are corroborated by the replication study. Our method to assess the successfulness of replication will better fit the needs and desires of researchers in fields that use replication studies.</p>Hidde Jelmer LeplaaCharlotte RietbergenHerbert Hoijtink
Copyright (c) 2024 Hidde Jelmer Leplaa, Charlotte Rietbergen, Herbert Hoijtink
https://creativecommons.org/licenses/by/4.0/
2024-12-202024-12-20810.15626/MP.2020.2554Beyond a Dream: The Practical Foundations of Disconnected Psychology
https://conferences.lnu.se/index.php/metapsychology/article/view/2740
<p><em>Disconnected</em> psychology is a form of psychological science in which researchers ground their work upon the main principles of psychological method but are detached from a “field” consisting of other psychologists that comprises <em>connected</em> psychology. It has previously been proposed that combining the two forms of psychology would result in the most significant advancement of psychological knowledge (Krpan, 2020). However, disconnected psychology may seem to be an “abstract utopia”, given that it has not been previously detailed how to put it into practice. The present article therefore sets the practical foundations of disconnected psychology. In this regard, I first describe a hypothetical disconnected psychologist and discuss relevant methodological and epistemological implications. I then propose how this variant of psychology could be integrated with the current academic system (i.e., with connected psychology). Overall, the present article transforms disconnected psychology from a hazy dream into substance that could eventually maximize psychological knowledge, even if implementing it would require a radical transformation of psychological science. </p>Dario Krpan
Copyright (c) 2024 Dario Krpan
https://creativecommons.org/licenses/by/4.0/
2024-04-192024-04-19810.15626/MP.2020.2740The Devil is Mainly in the Nuisance Parameters: Performance of Structural Fit Indices Under Misspecified Structural Models in SEM
https://conferences.lnu.se/index.php/metapsychology/article/view/2757
<p>To provide researchers with a means of assessing the fit of the structural component of structural equation models, structural fit indices- modifications of the composite fit indices, RMSEA, SRMR, and CFI- have recently been developed. We investigated the performance of four of these structural fit indices- RMSEA-P, RMSEAs, SRMRs, and CFIs-, when paired with widely accepted cutoff values, in the service of detecting structural misspecification. In particular, by way of simulation study, for each of seven fit indices- 3 composite and 4 structural-, and the traditional chi-square test of perfect composite fit, we estimated the following rates: a) Type I error rate (i.e., the probability of (incorrect) rejection of a correctly specified structural component), under each of four degrees of misspecification in the measurement component; and b) Power (i.e., the probability of (correct) rejection of an incorrectly specified structural model), under each condition formed of the pairing of one of three degrees of structural misspecification with one of four degrees of measurement component misspecification. In addition to sample size, the impacts of two model features, incidental to model misspecification- number of manifest variables per latent variable and magnitude of factor loading- were investigated. The results suggested that, although the structural fit indices performed relatively better than the composite fit indices, none of the goodness-of-fit index with a fixed cutoff value pairings was capable of delivering an entirely satisfactory Type I error rate/Power balance, [RMSEAs, .05] failing entirely in this regard. Of the remaining pairings; a) RMSEA-P and CFIs suffered from a severely inflated Type I error rate; b) despite the fact that they were designed to pick up on structural features of candidate models, all pairings- and especially, RMSEA-P and CFIs-manifested sensitivities to model features, incidental to structural misspecification; and c) although, in the main, behaving in a sensible fashion, SRMRs was only sensitive to structural misspecification when it occurred at a relatively high degree.</p>Moritz HeeneMichael MaraunNadine Juliana GlushkoSunthud Pornprasertmanit
Copyright (c) 2024 Moritz Heene, Michael Maraun, Nadine Juliana Glushko, Sunthud Pornprasertmanit
https://creativecommons.org/licenses/by/4.0/
2024-12-202024-12-20810.15626/MP.2021.2757Preregistration specificity and adherence: A review of preregistered gambling studies and cross-disciplinary comparison
https://conferences.lnu.se/index.php/metapsychology/article/view/2909
<p>Study preregistration is one of several “open science” practices (e.g., open data, preprints) that researchers use to improve the transparency and rigour of their research. As more researchers adopt preregistration as a regular practice, examining the nature and content of preregistrations can help identify the strengths and weaknesses of current practices. The value of preregistration, in part, relates to the specificity of the study plan and the extent to which investigators adhere to this plan. We identified 53 preregistrations from the gambling studies field meeting our predefined eligibility criteria and scored their level of specificity using a 23-item protocol developed to measure the extent to which a clear and exhaustive preregistration plan restricts various researcher degrees of freedom (RDoF; i.e., the many methodological choices available to researchers when collecting and analysing data, and when reporting their findings). We also scored studies on a 32-item protocol that measured adherence to the preregistered plan in the study manuscript. We found gambling preregistrations had low specificity levels on most RDoF. However, a comparison with a sample of cross-disciplinary preregistrations (N = 52; Bakker et al., 2020) indicated that gambling preregistrations scored higher on 12 (of 29) items. Thirteen (65%) of the 20 associated published articles or preprints deviated from the protocol without declaring as much (the mean number of undeclared deviations per article was 2.25, SD = 2.34). Overall, while we found improvements in specificity and adherence over time (2017-2020), our findings suggest the purported benefits of preregistration—including increasing transparency and reducing RDoF—are not fully achieved by current practices. Using our findings, we provide 10 practical recommendations that can be used to support and refine preregistration practices.</p>Robert HeireneDebi LaPlanteEric LouderbackBrittany KeenMarjan BakkerAnastasia SerafimovskaSally Gainsbury
Copyright (c) 2024 Robert Heirene, Debi LaPlante, Eric Louderback, Brittany Keen, Marjan Bakker, Anastasia Serafimovska, Sally Gainsbury
https://creativecommons.org/licenses/by/4.0/
2024-07-012024-07-01810.15626/MP.2021.2909Distinguishing Between Models and Hypotheses: Implications for Significance Testing
https://conferences.lnu.se/index.php/metapsychology/article/view/2957
<p>In the debate about the merits or demerits of null hypothesis significance testing (NHST), authorities on both sides assume that the <em>p</em> value that a researcher computes is based on the null hypothesis or test hypothesis. If the assumption is true, it suggests that there are proper uses for NHST, such as distinguishing between competing directional hypotheses. And once it is admitted that there are proper uses for NHST, it makes sense to educate substantive researchers about how to use NHST properly and avoid using it improperly. From this perspective, the conclusion would be that researchers in the business and social sciences could benefit from better education pertaining to NHST. In contrast, my goal is to demonstrate that the <em>p</em> value that a researcher computes is not based on a hypothesis, but on a model in which the hypothesis is embedded. In turn, the distinction between hypotheses and models indicates that NHST cannot soundly be used to distinguish between competing directional hypotheses or to draw any conclusions about directional hypotheses whatsoever. Therefore, it is not clear that better education is likely to prove satisfactory. It is the temptation issue, not the education issue, that deserves to be in the forefront of NHST discussions.</p>David Trafimow
Copyright (c) 2024 David Trafimow
https://creativecommons.org/licenses/by/4.0/
2024-11-112024-11-11810.15626/MP.2021.2957Measurements of Susceptibility to Anchoring are Unreliable
https://conferences.lnu.se/index.php/metapsychology/article/view/3236
<p>Theories on anchoring effects—the assimilation of numerical estimates toward previously considered numbers—have been used to derive hypotheses that susceptibility to anchoring is correlated with certain personality traits. Thus, for the last decade, a considerable amount of research has investigated relationships between people’s susceptibility to anchoring and personality traits (e.g., intelligence, the Big Five, narcissism). However, many of the findings are contradictory. We suspect that this inconsistency is grounded in imprecise measurements. Unfortunately, few reports have disclosed estimates of the susceptibility scores’ reliability (e.g., Cronbach’s alpha). We created a large and open data set of anchoring susceptibility scores and conducted a meta-analysis to test how extensive the <em>reliability problem</em> is. Results suggest that the reliability of most tasks is very low. In the few cases in which the reliability is acceptable, the validity of anchoring scores is questionable. We discuss requirements for further attempts to solve the reliability problem.</p>Lukas RöselerLucia WeberKatharina A. C. HelgerthElena StichMiriam GüntherPaulina TegethoffFelix Stefan WagnerAstrid Schütz
Copyright (c) 2024 Lukas Röseler, Lucia Weber, Katharina A. C. Helgerth, Elena Stich, Miriam Günther, Paulina Tegethoff, Felix Stefan Wagner, Astrid Schütz
https://creativecommons.org/licenses/by/4.0/
2024-12-202024-12-20810.15626/MP.2022.3236How Close to the Mark Might Published Heritability Estimates Be?
https://conferences.lnu.se/index.php/metapsychology/article/view/1479
<p>The behavioural scientist who requires an estimate of narrow heritability, h<sup>2</sup>, will conduct a twin study, and input the resulting estimated covariance matrices into a particular mode of estimation, the latter derived under supposition of the standard biometric model (SBM). It is known that the standard biometric model can be expected to misrepresent the phenotypic (genetic) architecture of human traits. The impact of this misrepresentation on the accuracy of h<sup>2</sup> estimation is unknown. We aimed to shed some light on this general issue, by undertaking three simulation studies. In each, we investigated the parameter recovery performance of five modes- Falconer’s coefficient and the SEM models, ACDE, ADE, ACE, and AE- when they encountered a constructed, non-SBM, architecture, under a particular informational input. In study 1, the architecture was single-locus with dominance effects and genetic-environment covariance, and the input was a set of population covariance matrices yielded under the four twin designs, monozygotic-reared together, monozygotic-reared apart, dizygotic-reared together, and dizygotic-reared apart; in study 2, the architecture was identical to that of study 1, but the informational input was monozygotic-reared together and dizygotic-reared together; and in study 3, the architecture was multi-locus with dominance effects, genetic-environment covariance, and epistatic interactions. The informational input was the same as in study 1. The results suggest that conclusions regarding the coverage of h<sup>2</sup> must be drawn conditional on a) the general class of generating architecture in play; b) specifics of the architecture’s parametric instantiations; c) the informational input into a mode of estimation; and d) the particular mode of estimation<br />employed. The results showed that the more complicated the generating architecture, the poorer a mode’s h<sup>2</sup> recovery performance. Random forest analyses furthermore revealed that, depending on the genetic architecture, h<sup>2</sup>, the dominance and locus additive parameter, and proportions of alleles were involved in complex interaction effects impacting on h<sup>2</sup> parameter recovery performance of a mode of estimation. Data and materials: <a href="https://osf.io/aq9sx/">https://osf.io/aq9sx/</a></p>Michael MaraunMoritz HeenePhilipp Sckopke
Copyright (c) 2024 Michael Maraun, Moritz Heene, Philipp Sckopke
https://creativecommons.org/licenses/by/4.0/
2024-05-222024-05-22810.15626/MP.2018.1479Knowing What We're Talking About
https://conferences.lnu.se/index.php/metapsychology/article/view/3638
<p><span style="font-weight: 400;">A theory crisis and measurement crisis have been argued to be root causes of psychology's replication crisis. In both, the lack of conceptual clarification and the jingle-jangle jungle at the construct definition level as well the measurement level play a central role. We introduce a conceptual tool that can address these issues: Decentralized Construct Taxonomy specifications (DCTs). These consist of comprehensive specifications of construct definitions, corresponding instructions for quantitative and qualitative research, and unique identifiers. We discuss how researchers can develop DCT specifications as well as how DCT specifications can be used in research, practice, and theory development. Finally, we discuss the implications and potential for future developments to answer the call for conceptual clarification and epistemic iteration. This contributes to the move towards a psychological science that progresses in a cumulative fashion through discussion and comparison.</span></p>Gjalt-Jorn PetersRik Crutzen
Copyright (c) 2024 Gjalt-Jorn Peters, Rik Crutzen
https://creativecommons.org/licenses/by/4.0/
2024-04-192024-04-19810.15626/MP.2022.3638Facets of Trust in Science: Researchers can be perceived as ethical and competent despite inconsistent research results
https://conferences.lnu.se/index.php/metapsychology/article/view/4230
<p>The public perception of science and scientists themselves has become a much-debated topic in recent years. In this article, we contribute to a more nuanced understanding of the public’s trust in science by focussing on the practices of science, which are often not known by the public. Building on previous research by Ebersole, Axt and Nosek (2016), we conducted a preregistered, quota-sampled survey in Austria (<em>N</em> = 564), where we presented participants with different scenarios about scientific practices. Thereby, we disentangled the perception of scientists–i.e., how competent and ethical they are being perceived–from the confidence in their scientific findings–i.e., how correct their results are being perceived. For instance, when “a researcher X conducted a study with an interesting finding, which he then publishes”, this researcher was–in our study– perceived as averagely competent and ethical, and the findings were perceived as neither correct nor incorrect (but somewhere in between). However, if “another researcher Y tried to replicate X’s finding, but failed - and X then criticized Y’s methodology and dismissed the new study”, researcher X was perceived as less competent, less ethical and the original results were perceived as less correct by participants. Importantly, if researcher X “acknowledges Y’s methodology” or “investigates the difference between the original study and the failed replication”, ratings for X’s competence and ethical behavior were <em>higher </em>than for how correct his results were being perceived. Moreover, the highest competence and ethics ratings were obtained, when researcher X was described to share the methods and data online for transparency. Psychological dispositions of the participants, such as political orientation or motivation for cognition, did not seem to affect these ratings to a large degree. These results are discussed in the light of Mertonian norms of science, which highlight cooperativeness and disinterestedness.</p>Hilmar BrohmerSimon CehMarta CzerwonkaSandra GrinschglGabriela HoferIzabela LebudaHildrun Walter
Copyright (c) 2024 Hilmar Brohmer, Simon Ceh, Marta Czerwonka, Sandra Grinschgl, Gabriela Hofer, Izabela Lebuda, Hildrun Walter
https://creativecommons.org/licenses/by/4.0/
2024-12-202024-12-20810.15626/MP.2024.4230Associations between Goal Orientation and Self-Regulated Learning Strategies are Stable across Course Types, Underrepresented Minority Status, and Gender
https://conferences.lnu.se/index.php/metapsychology/article/view/2918
<p>In this pre-registered replication of findings from Muis and Franco [2009; Contemporary Educational Psychology, 34(4), 306-318], college students (N = 978) from across the United States and Canada were surveyed regarding their goal orientations and learning strategies. A structural equation modelling approach was used to assess the associations between goal orientations and learning strategies. Six of the eight significant associations (75%) found by Muis and Franco replicated successfully in the current study. Mastery approach goals positively predicted endorsement of all learning strategies (Rehearsal, Critical Thinking, Metacognitive Self-Regulation and Elaboration). Performance avoidance goals negatively predicted critical thinking, while positively predicting metacognitive self-regulation and rehearsal. Evidence for moderation by assignment type was found. No evidence of the moderation of these associations by gender, underrepresented minority status, or course type (STEM, Humanities, or Social Sciences) was found. The reliability of common scales used in educational research and issues concerning the replication of studies using structural equation modeling are discussed.</p>Brendan SchuetzeVeronica Yan
Copyright (c) 2024 Brendan Schuetze, Veronica Yan
https://creativecommons.org/licenses/by/4.0/
2024-04-192024-04-19810.15626/MP.2021.2918The Effect of Variety on Perceived Quantity
https://conferences.lnu.se/index.php/metapsychology/article/view/2639
<p>Redden and Hoch (2009) found that variety in a set of items robustly decreased the perceived quantity of the sum of these items across multiple studies. For example, a set of multicolored M&M’s was estimated to contain fewer M&M’s than an equally large set of single-colored M&M’s (e.g., Redden & Hoch, 2009, Study 3). We conducted six close replication studies of the studies reported by Redden and Hoch and did not find this effect in any of them. A meta-analysis of the four original studies and 6 replication studies (N = 1,383) revealed no evidence for the phenomenon that variety reduces perceived quantity.</p>Lukas RöselerGeorg FelserJana AsbergerAstrid Schütz
Copyright (c) 2024 Lukas Röseler, Georg Felser, Jana Asberger, Astrid Schütz
https://creativecommons.org/licenses/by/4.0/
2024-08-272024-08-27810.15626/MP.2020.2639