Publications

The following are selected publications from our researchers.


Cooperation, conflict, and transformative AI

Multi-agent systems

Show all

DiGiovanni, Anthony. Beginner’s guide to reducing s-risks. CLR Website, September 2023.
Links | BibTeX
Kokotajlo, Daniel. Persuasion Tools: AI takeover without AGI or agency?. LessWrong, November 2020.
Links | BibTeX
Althaus, David; Kokotajlo, Daniel. Incentivizing forecasting via social media. Effective Altruism Forum, December 2020.
Links | BibTeX
Sotala, Kaj. Sequence on non-agent and multiagent models of mind. LessWrong, January 2019.
Links | BibTeX
Oesterheld, Caspar. Moral realism and AI alignment. LessWrong, September 2018.
Links | BibTeX
Gloor, Lukas. Suffering-Focused AI Safety: In Favor of “Fail-Safe” Measures. CLR Website, June 2016.
Links | BibTeX
Gloor, Lukas. Room for Other Things: How to adjust if EA seems overwhelming. Effective Altruism Forum, March 2015.
Links | BibTeX

Show all

DiGiovanni, Anthony. Beginner’s guide to reducing s-risks. CLR Website, September 2023.
Links | BibTeX
Kokotajlo, Daniel. Persuasion Tools: AI takeover without AGI or agency?. LessWrong, November 2020.
Links | BibTeX
Althaus, David; Kokotajlo, Daniel. Incentivizing forecasting via social media. Effective Altruism Forum, December 2020.
Links | BibTeX
Sotala, Kaj. Sequence on non-agent and multiagent models of mind. LessWrong, January 2019.
Links | BibTeX
Oesterheld, Caspar. Moral realism and AI alignment. LessWrong, September 2018.
Links | BibTeX
Gloor, Lukas. Suffering-Focused AI Safety: In Favor of “Fail-Safe” Measures. CLR Website, June 2016.
Links | BibTeX
Gloor, Lukas. Room for Other Things: How to adjust if EA seems overwhelming. Effective Altruism Forum, March 2015.
Links | BibTeX

Strategic considerations

Show all

DiGiovanni, Anthony. Beginner’s guide to reducing s-risks. CLR Website, September 2023.
Links | BibTeX
Kokotajlo, Daniel. Persuasion Tools: AI takeover without AGI or agency?. LessWrong, November 2020.
Links | BibTeX
Althaus, David; Kokotajlo, Daniel. Incentivizing forecasting via social media. Effective Altruism Forum, December 2020.
Links | BibTeX
Sotala, Kaj. Sequence on non-agent and multiagent models of mind. LessWrong, January 2019.
Links | BibTeX
Oesterheld, Caspar. Moral realism and AI alignment. LessWrong, September 2018.
Links | BibTeX
Gloor, Lukas. Suffering-Focused AI Safety: In Favor of “Fail-Safe” Measures. CLR Website, June 2016.
Links | BibTeX
Gloor, Lukas. Room for Other Things: How to adjust if EA seems overwhelming. Effective Altruism Forum, March 2015.
Links | BibTeX

Show all

DiGiovanni, Anthony. Beginner’s guide to reducing s-risks. CLR Website, September 2023.
Links | BibTeX
Kokotajlo, Daniel. Persuasion Tools: AI takeover without AGI or agency?. LessWrong, November 2020.
Links | BibTeX
Althaus, David; Kokotajlo, Daniel. Incentivizing forecasting via social media. Effective Altruism Forum, December 2020.
Links | BibTeX
Sotala, Kaj. Sequence on non-agent and multiagent models of mind. LessWrong, January 2019.
Links | BibTeX
Oesterheld, Caspar. Moral realism and AI alignment. LessWrong, September 2018.
Links | BibTeX
Gloor, Lukas. Suffering-Focused AI Safety: In Favor of “Fail-Safe” Measures. CLR Website, June 2016.
Links | BibTeX
Gloor, Lukas. Room for Other Things: How to adjust if EA seems overwhelming. Effective Altruism Forum, March 2015.
Links | BibTeX

Decision theory

Show all

Sorry, no publications matched your criteria.

Malevolence

Show all

DiGiovanni, Anthony. Beginner’s guide to reducing s-risks. CLR Website, September 2023.
Links | BibTeX
Kokotajlo, Daniel. Persuasion Tools: AI takeover without AGI or agency?. LessWrong, November 2020.
Links | BibTeX
Althaus, David; Kokotajlo, Daniel. Incentivizing forecasting via social media. Effective Altruism Forum, December 2020.
Links | BibTeX
Sotala, Kaj. Sequence on non-agent and multiagent models of mind. LessWrong, January 2019.
Links | BibTeX
Oesterheld, Caspar. Moral realism and AI alignment. LessWrong, September 2018.
Links | BibTeX
Gloor, Lukas. Suffering-Focused AI Safety: In Favor of “Fail-Safe” Measures. CLR Website, June 2016.
Links | BibTeX
Gloor, Lukas. Room for Other Things: How to adjust if EA seems overwhelming. Effective Altruism Forum, March 2015.
Links | BibTeX

Ethics & meta-ethics

Show all

DiGiovanni, Anthony. Beginner’s guide to reducing s-risks. CLR Website, September 2023.
Links | BibTeX
Kokotajlo, Daniel. Persuasion Tools: AI takeover without AGI or agency?. LessWrong, November 2020.
Links | BibTeX
Althaus, David; Kokotajlo, Daniel. Incentivizing forecasting via social media. Effective Altruism Forum, December 2020.
Links | BibTeX
Sotala, Kaj. Sequence on non-agent and multiagent models of mind. LessWrong, January 2019.
Links | BibTeX
Oesterheld, Caspar. Moral realism and AI alignment. LessWrong, September 2018.
Links | BibTeX
Gloor, Lukas. Suffering-Focused AI Safety: In Favor of “Fail-Safe” Measures. CLR Website, June 2016.
Links | BibTeX
Gloor, Lukas. Room for Other Things: How to adjust if EA seems overwhelming. Effective Altruism Forum, March 2015.
Links | BibTeX

Prioritization & macrostrategy

Show all

DiGiovanni, Anthony. Beginner’s guide to reducing s-risks. CLR Website, September 2023.
Links | BibTeX
Kokotajlo, Daniel. Persuasion Tools: AI takeover without AGI or agency?. LessWrong, November 2020.
Links | BibTeX
Althaus, David; Kokotajlo, Daniel. Incentivizing forecasting via social media. Effective Altruism Forum, December 2020.
Links | BibTeX
Sotala, Kaj. Sequence on non-agent and multiagent models of mind. LessWrong, January 2019.
Links | BibTeX
Oesterheld, Caspar. Moral realism and AI alignment. LessWrong, September 2018.
Links | BibTeX
Gloor, Lukas. Suffering-Focused AI Safety: In Favor of “Fail-Safe” Measures. CLR Website, June 2016.
Links | BibTeX
Gloor, Lukas. Room for Other Things: How to adjust if EA seems overwhelming. Effective Altruism Forum, March 2015.
Links | BibTeX

AI Forecasting

Show all

DiGiovanni, Anthony. Beginner’s guide to reducing s-risks. CLR Website, September 2023.
Links | BibTeX
Kokotajlo, Daniel. Persuasion Tools: AI takeover without AGI or agency?. LessWrong, November 2020.
Links | BibTeX
Althaus, David; Kokotajlo, Daniel. Incentivizing forecasting via social media. Effective Altruism Forum, December 2020.
Links | BibTeX
Sotala, Kaj. Sequence on non-agent and multiagent models of mind. LessWrong, January 2019.
Links | BibTeX
Oesterheld, Caspar. Moral realism and AI alignment. LessWrong, September 2018.
Links | BibTeX
Gloor, Lukas. Suffering-Focused AI Safety: In Favor of “Fail-Safe” Measures. CLR Website, June 2016.
Links | BibTeX
Gloor, Lukas. Room for Other Things: How to adjust if EA seems overwhelming. Effective Altruism Forum, March 2015.
Links | BibTeX

Other

DiGiovanni, Anthony. Beginner’s guide to reducing s-risks. CLR Website, September 2023.
Links | BibTeX
Kokotajlo, Daniel. Persuasion Tools: AI takeover without AGI or agency?. LessWrong, November 2020.
Links | BibTeX
Althaus, David; Kokotajlo, Daniel. Incentivizing forecasting via social media. Effective Altruism Forum, December 2020.
Links | BibTeX
Sotala, Kaj. Sequence on non-agent and multiagent models of mind. LessWrong, January 2019.
Links | BibTeX
Oesterheld, Caspar. Moral realism and AI alignment. LessWrong, September 2018.
Links | BibTeX
Gloor, Lukas. Suffering-Focused AI Safety: In Favor of “Fail-Safe” Measures. CLR Website, June 2016.
Links | BibTeX
Gloor, Lukas. Room for Other Things: How to adjust if EA seems overwhelming. Effective Altruism Forum, March 2015.
Links | BibTeX