Presentation for EA Anywhere, 5 Nov. 2023
Youtube video: The Unjournal: Bridging the gap between EA and academia
David Reinstein - Founder and Co-Director
Unjournal.org: Our full explanation and progress
These slides are linked at bit.ly/ujpresents; see speakers’ notes for more details
Also see
Intro - The Unjournal Presentation for March 25 Event
What is The Unjournal?
The Unjournal is not a journal.
We coordinate and fund the public evaluation of research projects in any format.
We’re building an open, sustainable system
for evaluation, feedback, ratings, and assessment.
Our initial focus is quantitative work that informs global priorities,
especially in economics, policy, and other social sciences.
Links: Unjournal.org, An Introduction to The Unjournal
Output: unjournal.pubpub.org
“Academic peer review” (background)
In Economics:
Main ingredients
Research Submission/Identification and Selection
Paid Evaluators (AKA 'reviewers')
Eliciting Quantifiable and Comparable Metrics
Public Evaluation
Linking, Not Publishing
Financial Prizes
Transparency
Our Theory of Change
To believe The Unjournal has value one must believe...
EA Research
Academic Research
Commission Direct Public Evaluation
The Unjournal Collaboration
Our Approach: leveraging problem synergy
Solution: Direct public evaluation of (global-priorities-relevant) research
→ The Unjournal is funding, organizing, and scaling this up. See our full ToC here.
Impactful Research and Evaluation:
What is it and why does it matter?
See also:
Unjournal.org “What is global-priorities-relevant research?” and links within �
EA Forum: Why scientific research is less effective in producing value than it could be: a mapping
How do we ‘have a positive impact’?
How can research support impact?
Research needs to produce true, useful information which enables better decision-making, driving choices and behavior that lead to better outcomes.
This can occur through:
→ We focus on global-priorities-relevant research.
Brief… some of Unjournal’s paths to impact
==
Prioritizing research: existing frameworks
But cause prioritization ≠ research prioritization!
CF: GPI framework
Moral theory, decision theory, epistemology
E.g., Behavioral research on altruism,
game theory & peace building
E.g., development economics, monitoring & evaluation, cost-benefit and predictive modeling
Increasingly relevant to UJ focus
What global-priorities-relevant research does Unjournal (currently) cover?
Not: Philosophy, computer science (AI interpretability), animal behavior, pure math
Not: Pure theory, research inputs, shallow reviews, informal discussion
4. Causes/outcomes including… →
Our “Field Specialist groups” �Updated 3-Nov-23, see “our team” for more
GH&D
Dev Econ
Economics/welfare
Psychology and attitudes
Innovation, meta-science
Catastrophic risks & AI gov.
Environmental Economics
Building:
Animal Welfare,
Social impact of tech., Macro/growth/finance,
LT trends and demographics
Why should research be evaluated?
The value of journals
Journals are not really publishers.
Journals are evaluators.
Journals offer quality control, credibility, prestige.
We have arXiv (and RePEc etc.) for that.
The value of evaluation
Rigor & Quality Control. Researchers should receive feedback and be held accountable to high standards of logic and evidence.
The value of evaluation
Credibility, domain, usefulness. Research-users want to know how much to trust research, update their beliefs, & adjust their decisions in different contexts.
(Without validating it all themselves)
The value of evaluation
Prioritizing research. We need a way to choose which researchers and organizations should receive more funding.
The state of the art in EA research evaluation
Weak underbelly of “EA/GPI/adjacent research”?
Bates:
… A cost of $86m to mitigate approximately 40% of the impact of a full-scale nuclear war between the US and a peer country seems prima facie absurd, and the level of exploration of such an important parameter is simply not in line with best practice in a cost-effectiveness analysis (especially since this is the parameter on which we might expect the authors to be least expert). … these issues could potentially reverse the authors’ conclusions, and should have been substantially defended in the text.
Authors:
We agree that this estimate from the published work is likely low and have since updated our view on cost upwards. The nuclear war probability utilized does not include other sources of nuclear risk such as accidental detonation of nuclear weapons leading to escalation, intentional attack, or dyads involving China.
Other ‘wins, rethinks, and potential’
Apr. 2022 Evidence Action’s Dispensers for Safe Water program “… a remarkable new investment of up to $64.7 million.” “recommended by GiveWell… and funded by Open Philanthropy”
“Underpinned by rigorous research by Nobel Laureate Michael Kremer and colleagues…”
a recent meta-analysis by Michael Kremer … shows that water treatment reduces the odds of mortality of children under five, from all causes, by around 25%.
Release cites:
Why not just use academic publishing?
Problems with academic publishing:
Why are we still doing this?
How do we solve these problems?
Towards a New Equilibrium: The Unjournal
Commission evaluations which are:
A New Equilibrium: The Unjournal
1. Rents & barriers to research access.
2. Static, limited formats: the PDF prison.
3. Gaming the system: wasted research & review effort.
4. Encourages academic flexing.
A New Equilibrium: The Unjournal
1. Rents & barriers to research access. Completely free to access.
2. Static, limited formats: the PDF prison.
3. Gaming the system: wasted research & review effort.
4. Encourages academic flexing
A New Equilibrium: The Unjournal
1. Rents & barriers to research access. Completely free to access.
2. Static, limited formats: the PDF prison. Open to any format.
3. Gaming the system: wasted research & review effort.
4. Encourages academic flexing.
A New Equilibrium: The Unjournal
1. Rents & barriers to research access. Completely free to access.
2. Static, limited formats: the PDF prison. Open to any format.
3. Gaming the system: wasted research & review effort. Paid, quantified, public evaluations.
4. Encourages academic flexing.
A New Equilibrium: The Unjournal
1. Rents & barriers to research access. Completely free to access.
2. Static, limited formats: the PDF prison. Open to any format.
3. Gaming the system: wasted research & review effort. Paid, quantified, public evaluations.
4. Encourages academic flexing. Directly rated on credibility, relevance & impact.
Progress, Challenges, and Roadmap
The Unjournal’s paths to impact
See here for an explanation in context
The academic collective action problem
Overcoming academic inertia is hard. Image source.
Overcoming the academic collective action problem
Our advantages:
Make ourselves impossible to ignore.
UJ evaluations should be public before traditional journals do their reviews.
Our progress: evaluation
(Pilot) output hosted at unjournal.pubpub.org:
Systems for prioritization, evaluation, aggregation…
Our workflow …
…
… workflow
5. Source evaluators with relevant complementary expertise
6. Evaluation & rating process
7. Author response
8. Eval. Manager summary+
9. “Publish” package, link to bibliometrics+
We built a team
Building, improving, and grounding/benchmarking
Our roadmap ahead
Challenges and pivotal choices, heavies
Not there yet: Get some ‘crowned heads of academia’ on board. Commitments from open science/open access orgs. Get submissions of “dynamic docs” (code/data).
Bigger fish: Getting UJ evaluations to ‘count’ in academia and research orgs.
Unanticipated challenge: “Hidden most recent version of paper”/author engagement
Big questions: Evaluation criteria and aggregation. Useful outputs for research-users. Quantification/quantified uncertainty. ‘Performance’ incentives for evaluators; evaluation manager oversight.
How you can get involved
Questions?
Bonus slides
How?
4. Financial prizes for strongest work
Replication crisis, p-hacks, fraud/ error → Transparent ‘research pipeline’ formats
🔎: “Rejected after third revision”
Traditional binary (0/1) ‘publish or reject’ process
Wastes resources
#sneakypubsorgoodresearch ???
.
�→ Evaluate & rate, don’t accept/reject
Global priorities/EA research orgs need:
How?
Key Hurdles…
… And about 60 volunteers for referee pool
Protocol for choosing/communicating work to evaluate
Guideline/form for evaluators (HERE)
Academic publishers
Extract rents and discourage innovation
But there is a coordination problem in ‘escaping’ this.
Funders like Open Phil and EA-affiliated researchers are not stuck, we can facilitate an exit.
Enable new formats and approaches
More readable, reliable, and replicable research formats, e.g., dynamic documents
and allow research projects to continue to improve without “paper bloat”