GoodAI and the AI Roadmap Institute have organized two workshops where we have started outlining various issues and suggested the first steps to tackle the AI race. These will make a good start for participants:
Below is a list of resources that can help you get a better idea of the existing research landscape on this topic. We encourage you to share any articles you find interesting.
Amodei, D., Olah, C., Steinhardt, J., Christiano, P., Schulman, J., & Mané, D. (2016). Concrete Problems in AI Safety.
Armstrong, S., Bostrom, N., & Shulman, C. (2016). Racing to the precipice: a model of artificial intelligence development. AI & SOCIETY, 31(2), 201–206.
Baum, S. D. (2016). On the promotion of safe and socially beneficial artificial intelligence. AI and Society, (2011), 1–9.
Baum, S. D. (2017). A Survey of Artificial General Intelligence Projects for Ethics, Risk, and Policy Executive Summary. Global Catastrophic Risk Institute Working Paper 17-1.
Bostrom, N. (2017). Strategic Implications of Openness in AI Development. Global Policy, 8(2), 135–148.
Conn, A. (2017). Can AI Remain Safe as Companies Race to Develop It?
Marcus, G. (2017). Deep Learning: A Critical Appraisal (pg 17)
Orseau, L., & Armstrong, S. (2016). Safely Interruptible Agents.
Ramamoorthy, A., Yampolskiy, R. (2018). Beyond MAD?: The Race for Artificial General Intelligence. ITU Journal: ICT Discoveries, Special Issue No. 1, 2 Feb. 2018.
Tomasik, B. (2013). International Cooperation vs. AI Arms Race. Foundational Research Institute.
Tomasik, B. (2013). Differential Intellectual Progress as a Positive-Sum Project. Foundational Research Institute.
Tomasik, B. (2013). Gains from Trade through Compromise. Foundational Research Institute.
Reading Guide for the Global Politics of Artificial Intelligence. Compiled by Matthijs Maas, Carrick Flynn, Sophie-Charlotte Fischer on behalf of Allan Dafoe.
AI Arms Race
Geist, E. M. (2016). It’s already too late to stop the AI arms race — We must manage it instead. Bulletin of the Atomic Scientists, 72(5), 318–321.
Shulman, C., (2009) Arms Control and Intelligence Explosions. Paper presented at the 7th
European Conference on Computing and Philosophy (ECAP), Bellaterra, Spain, July 2–4.
Shulman, C., & Armstrong, S. (2011) Arms races and intelligence explosions . Singularity Hypotheses.
Critch, A,. (2017) Toward negotiable reinforcement learning: shifting priorities in Pareto optimal sequential decision-making. Cornell University Library.
Grace, K., Salvatier, J., Dafoe, A., Zhang, B., Evans, O,. (2017) When Will AI Exceed Human Performance? Evidence from AI Experts. Cornell University Library.