
First, do no harm.
1,500+ Posts…
Free knowledge sharing for Safe AI. Not for profit. Linkouts to sources provided. Ads are likely to appear on link-outs (zero benefit to this journal publisher)
ANTHROPIC. Constitutional AI: Harmlessness from AI Feedback. Dec 15, 2022
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER. ANTHROPIC. Constitutional AI: Harmlessness from AI Feedback. Dec 15, 2022 Abstract As AI systems become more capable, we would like to enlist their help to supervise other AIs. We experiment with methods [...]
GPAI. The Global Partnership on Artificial Intelligence
The Global Partnership on Artificial Intelligence (GPAI) is where governments and leading AI experts work together on values-based pathways for AI. It is a multi-stakeholder initiative which aims to bridge the gap between theory and practice. TRANSCRIPT. Artificial intelligence is [...]
Edouard Harris – New Research: Advanced AI may tend to seek power *by default*. 22 OCT 2023.
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER. Power Seeking (AI) - Lesswrong Power Seeking is a property that agents might have, where they attempt to gain more general ability to control their environment. It's particularly relevant to AIs, and [...]
MICROSOFT RESEARCH. Responsible, Equitable, and Ethical AI panel discussion. 18 OCT 2022.
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER. Responsible, Equitable, and Ethical AI panel discussion Event:Research Summit 2022 Track:Precision Health – From Discovery to Delivery Date:October 19, 2022 This timely panel discussion did not air during Microsoft Research Summit [...]
ANTHROPIC. Predictability and Surprise in Large Generative Models. 03 OCT 2023
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER. Predictability and Surprise in Large Generative Models DEEP GANGULI∗, DANNY HERNANDEZ∗, LIANE LOVITT∗, NOVA DASSARMA†, TOM HENIGHAN†, ANDY JONES†, NICHOLAS JOSEPH†, JACKSON KERNION†, BEN MANN†, AMANDA ASKELL, YUNTAO BAI, ANNA CHEN, TOM [...]
TECHTALK. AI scientists are studying the “emergent” abilities of Large Language Models (LLM). August 22, 2022.
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER. TECHTALK. AI scientists are studying the “emergent” abilities of large language models. August 22, 2022. By Ben Dickson Large language models (LLMs) have become the center of attention and hype because of [...]
GROKKING. NEW PHENOMENA. A Mechanistic Interpretability Analysis of Grokking. Models trained on small algorithmic tasks like modular addition will initially memorise the training data, but after a long time will suddenly learn to generalise to unseen data.
Grokking is the mysterious phenomenon of explosive machine learning. Learn more: QUICK STUDY on Twitter. A Mechanistic Interpretability Analysis of Grokking by Neel Nanda, Tom Lieberum. 15th Aug 2022 Introduction Grokking is a recent phenomena discovered by OpenAI researchers, that in my opinion is one of the most fascinating mysteries [...]
2022 Expert Survey on Progress in AI. 03 AUGUST 2022
2022 Expert Survey on Progress in AI. 03 AUGUST 2022 48% of respondents gave at least 10% chance of an extremely bad outcome We contacted approximately 4271 researchers who published at the conferences NeurIPS or ICML in 2021. These people were selected by taking all of the authors at those [...]
ROUGH NOTE. Interpretability vs Neuroscience by Christopher Olah. March 12th, 2021
Interpretability vs Neuroscience by Christopher Olah Six major advantages which make artificial neural networks much easier to study than biological ones. Posted on March 12th, 2021 This article is a rough note. Writing rough notes allows me share more content, since polishing takes lots of time. While I hope it's [...]
George Carlin. “Never underestimate the power of stupid people in large groups.”
George Carlin. "Never underestimate the power of stupid people in large groups."
Demis Hassabis: DeepMind – AI, Superintelligence & the Future of Humanity | Lex Fridman Podcast #299
DEMIS. I think there's going to be some amazing advances in biology built on top of things like alpha fold. We're already seeing that with the community doing that after we've open sourced it and released it um and uh you know i also i [...]
U.S. DEPARTMENT OF DEFENSE RESPONSIBLE ARTIFICIAL INTELLIGENCE STRATEGY AND IMPLEMENTATION PATHWAY
U.S. DEPARTMENT OF DEFENSE RESPONSIBLE ARTIFICIAL INTELLIGENCE STRATEGY AND IMPLEMENTATION PATHWAY - UNCLASSIFIED Prepared by the DoD Responsible AI Working Council in accordance with the memorandum issued by Deputy Secretary of Defense Kathleen Hicks on May 26, 2021, Implementing Responsible Artificial Intelligence in the Department of Defense. June 2022 [...]
We Can Make COVID-19 the Last Pandemic | Bill Gates | TED | 22 APRIL 2022
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER. We Can Make COVID-19 the Last Pandemic | Bill Gates | TED | 22 APRIL 2022 "The IMF estimates that COVID has cost nearly 14 trillion dollars. And [...]
THE WHITE HOUSE. Cancer Moonshot
THE WHITE HOUSE. Cancer Moonshot CANCER MOONSHOT Share Your Actions Share Your Stories & Ideas Your Stories Events and Webinars Fact Sheets Fact Sheet: President Biden Reignites Cancer Moonshot to End Cancer as We Know It Biden-Harris Administration Sets Goal of Reducing Cancer [...]
Future of Life Institute. Slaughterbots – if human: kill()
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER. This could be our future, unless we agree to prohibit lethal autonomous weapons. To learn what Slaughterbots are, how they pose a risk, and what solutions are available, visit [...]
Final Report. National Security Commission on Artificial Intelligence (NSCAI)
Final Report. National Security Commission on Artificial Intelligence (NSCAI) Executive Summary The Final Report presents the NSCAI’s strategy for winning the artificial intelligence era. The 16 chapters explain the steps the United States must take to responsibly use AI for national security and defense, defend [...]
NATIONAL ARTIFICIAL INTELLIGENCE INITIATIVE OVERSEEING AND IMPLEMENTING THE UNITED STATES NATIONAL AI STRATEGY
NATIONAL ARTIFICIAL INTELLIGENCE INITIATIVE OVERSEEING AND IMPLEMENTING THE UNITED STATES NATIONAL AI STRATEGY The National Artificial Intelligence Initiative (NAII) was established by the National Artificial Intelligence Initiative Act of 2020 (NAIIA) (DIVISION E, SEC. 5001) – bipartisan legislation enacted on January 1, 2021. The main purposes of the initiative [...]
Kubeflow. MLOps open-source tool by Google on top of Kubernetes
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER. Kubeflow. MLOps open-source tool by Google on top of Kubernetes Almost immediately after Kubernetes established itself as the standard for working with a cluster of containers, Google created [...]
Zoom In: An Introduction to Circuits. Published by OpenAI. March 10, 2020
Zoom In: An Introduction to Circuits By studying the connections between neurons, we can find meaningful algorithms in the weights of neural networks - OpenAI. March 10, 2020 Zoom In: An Introduction to Circuits By studying the connections between neurons, we can find meaningful algorithms in the weights of neural [...]
STUART RUSSELL. Human Compatible: Artificial Intelligence and the Problem of Control. 2019.
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER. STUART RUSSELL. Human Compatible: Artificial Intelligence and the Problem of Control. 2019. Amazon. "The most important book on AI this year." --The Guardian "Mr. Russell's exciting book goes deep, while sparkling with [...]
The Bitter Lesson. Rich Sutton. March 13, 2019.
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER. The Bitter Lesson. Rich Sutton. March 13, 2019. The biggest lesson that can be read from 70 years of AI research is that general methods that leverage computation are ultimately the most [...]
Will MacAskill | Effective Altruism | What are the most important moral problems of our time?
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER. Will MacAskill | Effective Altruism | What are the most important moral problems of our time? FOR EDUCATIONAL AND KNOWLEDGE [...]
Strategic Implications of Openness in AI Development (2107). Nick Bostrom. Future of Humanity Institute, University of Oxford.
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER. Strategic Implications of Openness in AI Development (2107) Nick Bostrom. Future of Humanity Institute, University of Oxford. Policy Implications The global desirability of openness in AI development – sharing e.g. source code, [...]