<?xml version="1.0" encoding="utf-8" standalone="yes" ?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
  <channel>
    <title>metascience | Pablo Bernabeu</title>
    <link>https://pablobernabeu.github.io/tags/metascience/</link>
      <atom:link href="https://pablobernabeu.github.io/tags/metascience/index.xml" rel="self" type="application/rss+xml" />
    <description>metascience</description>
    <generator>Source Themes Academic (https://sourcethemes.com/academic/)</generator><language>en-uk</language><copyright>Pablo Bernabeu, 2015—2026. Licence: [CC BY 4.0](https://creativecommons.org/licenses/by/4.0/). Email: pcbernabeu@gmail.com. Cookies only used by third-party systems such as [Disqus](https://help.disqus.com/en/articles/1717155-use-of-cookies).</copyright><lastBuildDate>Sun, 01 Feb 2026 00:00:00 +0000</lastBuildDate>
    <image>
      <url>https://pablobernabeu.github.io/img/default_preview_image.png</url>
      <title>metascience</title>
      <link>https://pablobernabeu.github.io/tags/metascience/</link>
    </image>
    
    <item>
      <title>Unlock the Lab: Your guide to reading science like a scientist</title>
      <link>https://pablobernabeu.github.io/applications-and-dashboards/unlock-the-lab/</link>
      <pubDate>Sun, 01 Feb 2026 00:00:00 +0000</pubDate>
      <guid>https://pablobernabeu.github.io/applications-and-dashboards/unlock-the-lab/</guid>
      <description>&lt;p&gt;&lt;a href=&#34;https://unlock-the-lab.web.app&#34;&gt;&lt;strong&gt;Unlock the Lab&lt;/strong&gt;&lt;/a&gt; is an educational web application designed to develop science literacy by guiding participants through the evaluation of research quality using evidence-based criteria. Rather than passively consuming information, participants actively engage with 48 fictional research scenarios, rating study quality and predicting how their peers will rate the same studies. This peer-anchored design fosters reflective thinking and helps participants calibrate their own judgements against a broader community standard.&lt;/p&gt;
&lt;p&gt;The application is suitable for use in university workshops, open science training events and self-directed learning. It requires no login or prior knowledge, and its browser-based format makes it accessible from any device.&lt;/p&gt;
&lt;div style=&#34;width: 100%; height: 800px; margin: 2em 0; border-radius: 8px; overflow: hidden; border: 1px solid #ddd;&#34;&gt;
  &lt;iframe src=&#34;https://unlock-the-lab.web.app&#34;
      style=&#34;width: 100%; height: 100%; border: none;&#34;
      title=&#34;Unlock the Lab web application&#34;&gt;
  &lt;/iframe&gt;
&lt;/div&gt;
&lt;h2 id=&#34;educational-objectives&#34;&gt;Educational Objectives&lt;/h2&gt;
&lt;p&gt;The application targets several interconnected competencies in scientific reasoning. Participants learn to evaluate research quality using a structured rubric that covers methodology, sample size, data transparency, pre-registration, and publication practices. They also learn to recognise misleading framing, such as sensationalised headlines and clickbait abstracts that misrepresent underlying findings. The scenarios help users identify barriers to knowledge access, including paywalled journals and predatory publishing. A core objective is to practise objective assessment by decoupling conclusions from title framing and focusing instead on the evidence presented. Finally, participants build calibrated consensus skills by comparing their personal ratings against the community average for each study. These objectives are embedded in both the educational content and the task design, ensuring that learning occurs through active participation rather than passive instruction.&lt;/p&gt;
&lt;img src=&#34;img/screenshot-rubric.png&#34; alt=&#34;Evaluation rubric in Unlock the Lab&#34; style=&#34;max-width:100%; border-radius:6px; margin: 1em 0;&#34;&gt;
&lt;p style=&#34;font-size:0.85em; color:#555; margin-top:-0.5em;&#34;&gt;&lt;em&gt;Evaluation rubric presented before the study scenarios and accessible throughout the experience&lt;/em&gt;&lt;/p&gt;
&lt;h2 id=&#34;application-structure&#34;&gt;Application Structure&lt;/h2&gt;
&lt;p&gt;The workshop unfolds across three main phases. The first is an &lt;strong&gt;educational introduction&lt;/strong&gt; in which participants read background material on how to assess research, covering key concepts in study design, transparency and publication ethics. A glossary of 21 scientific terms with accessible definitions is available throughout and can be consulted at any point. The second phase is &lt;strong&gt;scenario evaluation&lt;/strong&gt;: participants work through 48 fictional research scenarios one at a time, providing two ratings for each study on a 1–7 scale — a prediction of the peer consensus score followed by the participant&#39;s own rating. The scenarios span a range of disciplines and vary in quality, methodology and framing. The third phase, &lt;strong&gt;results and reflection&lt;/strong&gt;, invites participants to view their leaderboard position and explore the &lt;a href=&#34;https://unlock-the-lab.web.app/dashboard.html&#34;&gt;live analytics dashboard&lt;/a&gt; to see how their ratings compare with the community as a whole.&lt;/p&gt;
&lt;img src=&#34;img/screenshot-scenario.png&#34; alt=&#34;Research scenario evaluation screen&#34; style=&#34;max-width:100%; border-radius:6px; margin: 1em 0;&#34;&gt;
&lt;p style=&#34;font-size:0.85em; color:#555; margin-top:-0.5em;&#34;&gt;&lt;em&gt;Example research scenario with dual rating interface&lt;/em&gt;&lt;/p&gt;
&lt;h2 id=&#34;scoring-system&#34;&gt;Scoring System&lt;/h2&gt;
&lt;p&gt;Performance is measured by prediction accuracy rather than by agreeing with any predetermined correct answer. Each study is scored as:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;score = 100 − |predicted_rating − actual_peer_average| × 12&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;The multiplier of 12 is a deliberate design choice. Because ratings are on a 1–7 scale, the maximum possible error is 6. Multiplying by 12 means a worst-case prediction still yields a score of 28 (100 − 6 × 12), ensuring participants are never completely penalised for a single poor estimate. A harsher multiplier of roughly 17 would reduce a maximum-error prediction to zero; 12 was chosen as a more forgiving constant that keeps participants engaged throughout the task.&lt;/p&gt;
&lt;p&gt;Scores are capped between 0 and 100. The aggregate score is the sum across all 48 studies, giving a maximum of 4800. This design rewards participants who understand how their peers reason about research quality, rather than those who simply hold strong opinions.&lt;/p&gt;
&lt;h2 id=&#34;leaderboard&#34;&gt;Leaderboard&lt;/h2&gt;
&lt;p&gt;A real-time leaderboard ranks participants by their aggregate prediction score. Two views are available: the &lt;strong&gt;top 200 of the last 24 hours&lt;/strong&gt; and the &lt;strong&gt;all-time top 200&lt;/strong&gt;. Participants are identified by automatically assigned anonymous usernames (e.g., &amp;quot;Cheerful Penguin&amp;quot;), ensuring data privacy while still enabling a competitive and engaging ranking experience.&lt;/p&gt;
&lt;img src=&#34;img/screenshot-leaderboard.png&#34; alt=&#34;Leaderboard in Unlock the Lab&#34; style=&#34;max-width:100%; border-radius:6px; margin: 1em 0;&#34;&gt;
&lt;p style=&#34;font-size:0.85em; color:#555; margin-top:-0.5em;&#34;&gt;&lt;em&gt;Real-time leaderboard showing prediction accuracy rankings&lt;/em&gt;&lt;/p&gt;
&lt;h2 id=&#34;analytics-dashboard&#34;&gt;Analytics Dashboard&lt;/h2&gt;
&lt;p&gt;A publicly accessible &lt;a href=&#34;https://unlock-the-lab.web.app/dashboard.html&#34;&gt;live analytics dashboard&lt;/a&gt; provides visualisations of the aggregate data collected across all participants. In addition to the leaderboard shown above, the dashboard includes a criterion importance chart showing the degree of importance that participants assigned to each evaluation criterion (title, access, source, theory, methods and data, and conclusion), and a study-level bar chart of mean quality ratings with 95% confidence intervals across all 48 studies. The dashboard is intended both for participants reviewing their own results and for facilitators and researchers interested in population-level patterns.&lt;/p&gt;
&lt;img src=&#34;img/screenshot-dashboard-1.png&#34; alt=&#34;Top Predictors leaderboard and Criterion Importance chart&#34; style=&#34;max-width:100%; border-radius:6px; margin: 1em 0;&#34;&gt;
&lt;p style=&#34;font-size:0.85em; color:#555; margin-top:-0.5em;&#34;&gt;&lt;em&gt;Dashboard section showing the Criterion Importance chart, which displays the average token allocation per evaluation criterion across participants&lt;/em&gt;&lt;/p&gt;
&lt;img src=&#34;img/screenshot-dashboard-2.png&#34; alt=&#34;Average Quality Ratings by Study chart&#34; style=&#34;max-width:100%; border-radius:6px; margin: 1em 0;&#34;&gt;
&lt;p style=&#34;font-size:0.85em; color:#555; margin-top:-0.5em;&#34;&gt;&lt;em&gt;Dashboard section showing mean quality ratings (1–7 scale) with 95% confidence intervals for each of the 48 studies; bars are colour-coded and clickable for detailed study information&lt;/em&gt;&lt;/p&gt;
&lt;h2 id=&#34;broader-themes-for-discussion&#34;&gt;Broader Themes for Discussion&lt;/h2&gt;
&lt;p&gt;The scenarios in Unlock the Lab are fictional, but the dynamics they expose are real. Several broader themes naturally arise when the application is used in a workshop or classroom setting.&lt;/p&gt;
&lt;p&gt;A question that surfaces in most discussions of science communication is whether scientists are actually rewarded for communicating their work to non-specialist audiences — and the answer depends heavily on institutional context. In many academic systems, promotion and tenure are tied almost exclusively to publication metrics and grant income, leaving little professional incentive for scientists to invest in public engagement. Yet demand for accessible science has grown, particularly in the wake of high-profile controversies over vaccine safety, climate data and pandemic modelling. Some funders now require evidence of public engagement as a condition of grant awards, and frameworks such as the Research Excellence Framework in the United Kingdom have begun to recognise impact beyond academia. Despite these positive developments, a significant gap persists between the stated importance of public outreach and the professional rewards allocated to it.&lt;/p&gt;
&lt;p&gt;Closely related to the question of outreach is a broader set of pressures that shape what research gets produced and how. The quantity and quality of research are governed by incentive structures operating at the level of individuals, institutions and journals. The pressure to publish frequently — sometimes described as publish-or-perish — has been associated with a range of questionable research practices, including selective reporting, inflated effect sizes and the suppression of null results in what is sometimes called the file-drawer problem. Journal impact factors, though widely criticised as crude proxies for article quality, continue to influence hiring and promotion decisions in ways that reward prestige over reproducibility. Funding bodies, which typically favour novelty over replication, have contributed to a research landscape in which confirmation is systematically undervalued. These pressures operate subtly: few researchers consciously intend to distort the scientific record, yet the cumulative effect of individually rational decisions can be a literature that overstates certainty and understates uncertainty. Recognising how these dynamics operate is itself a form of science literacy, and one that the evaluation scenarios in Unlock the Lab are designed to activate.&lt;/p&gt;
&lt;p&gt;It is tempting to trace the open science movement to a gradual, principled awakening, but the reality is considerably less flattering. The push towards greater transparency in research did not emerge from idealism alone. Some of its most consequential catalysts were scandals. The case of Diederik Stapel, the Dutch social psychologist whose fabrication of data across dozens of studies was uncovered in 2011, became one of the most widely discussed episodes of scientific fraud in recent memory. Stapel had built a prolific career on results that were, in some cases, entirely invented, and his exposure prompted sustained reflection both about individual responsibility and about the structural conditions that had made long-term detection so difficult.&lt;/p&gt;
&lt;div&gt;
  &lt;div style=&#34;position: relative; padding-top: 56.25%; margin-top: 3%; margin-bottom: 4%;&#34;&gt;
  &lt;iframe src=&#34;https://www.youtube.com/embed/WUGh2VWR4JA&#34; frameborder=&#34;0&#34; allowfullscreen
      style=&#34;position: absolute; top: 0; left: 0; width: 100%; height: 100%;&#34;&gt;&lt;/iframe&gt;
  &lt;/div&gt;
&lt;/div&gt;
&lt;p&gt;The inquiries that followed identified several systemic vulnerabilities. The absence of raw data sharing, the reluctance of journals to publish replications and the deference that junior researchers typically accord senior colleagues all featured prominently. Similar cases — including those of Marc Hauser in evolutionary psychology, Dirk Smeesters in consumer behaviour and Jens Förster in social cognition — reinforced the argument that individual fraud was symptomatic of broader cultural problems in research. Out of this period of reckoning grew a set of reforms that are now, to varying degrees, embedded in scientific practice: pre-registration of hypotheses and analysis plans, open data and materials repositories, registered reports, large-scale replication efforts such as the Many Labs project, and the emergence of meta-scientific fields dedicated to studying science itself (&lt;a href=&#34;https://doi.org/10.1007/s11948-020-00182-9&#34;&gt;Bruton et al., 2020&lt;/a&gt;; &lt;a href=&#34;https://doi.org/10.1371/journal.pone.0263023&#34;&gt;Gopalakrishna et al., 2022&lt;/a&gt;; &lt;a href=&#34;https://doi.org/10.1002/job.2623&#34;&gt;Kepes et al., 2022&lt;/a&gt;; &lt;a href=&#34;https://doi.org/10.1016/j.rmal.2023.100064&#34;&gt;Larsson et al., 2023&lt;/a&gt;; &lt;a href=&#34;https://doi.org/10.3102/0013189X211001356&#34;&gt;Makel et al., 2021&lt;/a&gt;; &lt;a href=&#34;https://doi.org/10.1007/s11948-021-00314-9&#34;&gt;Xie et al., 2021&lt;/a&gt;). These changes were not imposed from above alone; many were championed by researchers who recognised that the credibility of their own work depended on the credibility of the field as a whole.&lt;/p&gt;
&lt;p&gt;Generative artificial intelligence has since added a further layer of complexity to these debates. The rapid uptake of large language models and other generative AI tools has introduced an entirely new set of questions about transparency and attribution in research. At the level of individual tasks, these tools can assist with drafting, literature synthesis, code generation and data analysis, offering genuine efficiency gains particularly for researchers working in their second or third language. At a systemic level, however, their integration into the research workflow raises concerns that overlap directly with the themes of Unlock the Lab. AI-generated text can produce plausible-sounding but inaccurate citations — a phenomenon sometimes called hallucination — and the persuasive fluency of generated prose can make methodological weaknesses harder to detect. Several journals now require explicit disclosure of AI use in submitted manuscripts, though policies vary widely and enforcement is difficult. Alongside these content-level concerns, large-scale bibliometric analysis of 41.3 million papers has revealed a structural paradox: scientists who engage in AI-augmented research publish more frequently and receive more citations, yet collective AI adoption narrows the range of scientific topics studied and reduces engagement among scientists, with AI-augmented work gravitating towards data-rich areas and appearing to automate established fields rather than pioneer new ones (&lt;a href=&#34;https://doi.org/10.1038/s41586-025-09922-y&#34;&gt;Hao et al., 2026&lt;/a&gt;). Perhaps most relevant to science communication is the question of trust: as AI lowers the cost of producing formally credible-looking content, the skills involved in evaluating sources, identifying clickbait and distinguishing rigorous research from superficially persuasive claims become more rather than less important.&lt;/p&gt;
&lt;h2 id=&#34;technology-stack&#34;&gt;Technology Stack&lt;/h2&gt;
&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Layer&lt;/th&gt;
&lt;th&gt;Technologies&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Frontend&lt;/td&gt;
&lt;td&gt;HTML5, CSS3, JavaScript (ES6+)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Visualisation&lt;/td&gt;
&lt;td&gt;&lt;a href=&#34;https://www.chartjs.org/&#34;&gt;Chart.js&lt;/a&gt; 4.4.0&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Build tooling&lt;/td&gt;
&lt;td&gt;&lt;a href=&#34;https://vite.dev/&#34;&gt;Vite&lt;/a&gt; 5.4&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Database&lt;/td&gt;
&lt;td&gt;Firebase Realtime Database&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Authentication&lt;/td&gt;
&lt;td&gt;Firebase Authentication (anonymous)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Hosting&lt;/td&gt;
&lt;td&gt;Firebase Hosting&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;
&lt;h2 id=&#34;source-code-and-contributions&#34;&gt;Source Code and Contributions&lt;/h2&gt;
&lt;p&gt;The &lt;a href=&#34;https://github.com/pablobernabeu/Unlock_the_Lab&#34;&gt;source code is available on GitHub&lt;/a&gt; under a &lt;a href=&#34;https://creativecommons.org/licenses/by/4.0/&#34;&gt;Creative Commons Attribution 4.0 International&lt;/a&gt; licence. The application can be extended or adapted via pull requests. Feature requests, bug reports and other suggestions can be submitted as &lt;a href=&#34;https://github.com/pablobernabeu/Unlock_the_Lab/issues&#34;&gt;issues&lt;/a&gt;.&lt;/p&gt;
&lt;h2 id=&#34;references&#34;&gt;References&lt;/h2&gt;
&lt;p&gt;Bruton, S. V., Medlin, M., Brown, M., &amp;amp; Sacco, D. F. (2020). Personal motivations and systemic incentives: Scientists on questionable research practices. &lt;em&gt;Science and Engineering Ethics, 26&lt;/em&gt;(3), 1531–1547. &lt;a href=&#34;https://doi.org/10.1007/s11948-020-00182-9&#34;&gt;https://doi.org/10.1007/s11948-020-00182-9&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;Gopalakrishna, G., Ter Riet, G., Vink, G., Stoop, I., Wicherts, J. M., &amp;amp; Bouter, L. M. (2022). Prevalence of questionable research practices, research misconduct and their potential explanatory factors: A survey among academic researchers in The Netherlands. &lt;em&gt;PloS ONE, 17&lt;/em&gt;(2), e0263023. &lt;a href=&#34;https://doi.org/10.1371/journal.pone.0263023&#34;&gt;https://doi.org/10.1371/journal.pone.0263023&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;Hao, Q., Xu, F., Li, Y., &amp;amp; Evans, J. (2026). Artificial intelligence tools expand scientists’ impact but contract science’s focus. &lt;em&gt;Nature, 679&lt;/em&gt;, 1237–1243. &lt;a href=&#34;https://doi.org/10.1038/s41586-025-09922-y&#34;&gt;https://doi.org/10.1038/s41586-025-09922-y&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;Kepes, S., Keener, S. K., McDaniel, M. A., &amp;amp; Hartman, N. S. (2022). Questionable research practices among researchers in the most research‐productive management programs. &lt;em&gt;Journal of Organizational Behavior, 43&lt;/em&gt;(7), 1190–1208. &lt;a href=&#34;https://doi.org/10.1002/job.2623&#34;&gt;https://doi.org/10.1002/job.2623&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;Larsson, T., Plonsky, L., Sterling, S., Kytö, M., Yaw, K., &amp;amp; Wood, M. (2023). On the frequency, prevalence, and perceived severity of questionable research practices. &lt;em&gt;Research Methods in Applied Linguistics, 2&lt;/em&gt;(3), 100064. &lt;a href=&#34;https://doi.org/10.1016/j.rmal.2023.100064&#34;&gt;https://doi.org/10.1016/j.rmal.2023.100064&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;Makel, M. C., Hodges, J., Cook, B. G., &amp;amp; Plucker, J. A. (2021). Both questionable and open research practices are prevalent in education research. &lt;em&gt;Educational Researcher, 50&lt;/em&gt;(8), 493–504. &lt;a href=&#34;https://doi.org/10.3102/0013189X211001356&#34;&gt;https://doi.org/10.3102/0013189X211001356&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;Xie, Y., Wang, K., &amp;amp; Kong, Y. (2021). Prevalence of research misconduct and questionable research practices: A systematic review and meta-analysis. &lt;em&gt;Science and Engineering Ethics, 27&lt;/em&gt;(4), 41. &lt;a href=&#34;https://doi.org/10.1007/s11948-021-00314-9&#34;&gt;https://doi.org/10.1007/s11948-021-00314-9&lt;/a&gt;&lt;/p&gt;
&lt;a href=&#39;https://unlock-the-lab.web.app/&#39; target=&#39;_blank&#39; style=&#34;display:inline-block; margin-top: 0.8rem;&#34;&gt;
      &lt;button style = &#34;background-color: white; color: black; border: 2px solid #196F27; border-radius: 12px;&#34;&gt;
      &lt;h3 style = &#34;margin-top: 7px !important; margin-left: 9px !important; margin-right: 9px !important;&#34;&gt;
      &lt;span style=&#34;color:#DBE6DA;&#34;&gt;&lt;/span&gt; Access web application
      &lt;/h3&gt;&lt;/button&gt;
      &lt;/a&gt;
</description>
    </item>
    
  </channel>
</rss>
