
First, do no harm.
1,500+ Posts…
Free knowledge sharing for Safe AI. Not for profit. Linkouts to sources provided. Ads are likely to appear on link-outs (zero benefit to this journal publisher)
ANTHROPIC. Predictability and Surprise in Large Generative Models. 03 OCT 2023
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER. Predictability and Surprise in Large Generative Models DEEP GANGULI∗, DANNY HERNANDEZ∗, LIANE LOVITT∗, NOVA DASSARMA†, TOM HENIGHAN†, ANDY JONES†, NICHOLAS JOSEPH†, JACKSON KERNION†, BEN MANN†, AMANDA ASKELL, YUNTAO BAI, ANNA CHEN, TOM [...]
TECHTALK. AI scientists are studying the “emergent” abilities of Large Language Models (LLM). August 22, 2022.
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER. TECHTALK. AI scientists are studying the “emergent” abilities of large language models. August 22, 2022. By Ben Dickson Large language models (LLMs) have become the center of attention and hype because of [...]
GROKKING. NEW PHENOMENA. A Mechanistic Interpretability Analysis of Grokking. Models trained on small algorithmic tasks like modular addition will initially memorise the training data, but after a long time will suddenly learn to generalise to unseen data.
Grokking is the mysterious phenomenon of explosive machine learning. Learn more: QUICK STUDY on Twitter. A Mechanistic Interpretability Analysis of Grokking by Neel Nanda, Tom Lieberum. 15th Aug 2022 Introduction Grokking is a recent phenomena discovered by OpenAI researchers, that in my opinion is one of the most fascinating mysteries [...]
2022 Expert Survey on Progress in AI. 03 AUGUST 2022
2022 Expert Survey on Progress in AI. 03 AUGUST 2022 48% of respondents gave at least 10% chance of an extremely bad outcome We contacted approximately 4271 researchers who published at the conferences NeurIPS or ICML in 2021. These people were selected by taking all of the authors at those [...]
ROUGH NOTE. Interpretability vs Neuroscience by Christopher Olah. March 12th, 2021
Interpretability vs Neuroscience by Christopher Olah Six major advantages which make artificial neural networks much easier to study than biological ones. Posted on March 12th, 2021 This article is a rough note. Writing rough notes allows me share more content, since polishing takes lots of time. While I hope it's [...]
George Carlin. “Never underestimate the power of stupid people in large groups.”
George Carlin. "Never underestimate the power of stupid people in large groups."
Demis Hassabis: DeepMind – AI, Superintelligence & the Future of Humanity | Lex Fridman Podcast #299
DEMIS. I think there's going to be some amazing advances in biology built on top of things like alpha fold. We're already seeing that with the community doing that after we've open sourced it and released it um and uh you know i also i [...]
U.S. DEPARTMENT OF DEFENSE RESPONSIBLE ARTIFICIAL INTELLIGENCE STRATEGY AND IMPLEMENTATION PATHWAY
U.S. DEPARTMENT OF DEFENSE RESPONSIBLE ARTIFICIAL INTELLIGENCE STRATEGY AND IMPLEMENTATION PATHWAY - UNCLASSIFIED Prepared by the DoD Responsible AI Working Council in accordance with the memorandum issued by Deputy Secretary of Defense Kathleen Hicks on May 26, 2021, Implementing Responsible Artificial Intelligence in the Department of Defense. June 2022 [...]
We Can Make COVID-19 the Last Pandemic | Bill Gates | TED | 22 APRIL 2022
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER. We Can Make COVID-19 the Last Pandemic | Bill Gates | TED | 22 APRIL 2022 "The IMF estimates that COVID has cost nearly 14 trillion dollars. And [...]
THE WHITE HOUSE. Cancer Moonshot
THE WHITE HOUSE. Cancer Moonshot CANCER MOONSHOT Share Your Actions Share Your Stories & Ideas Your Stories Events and Webinars Fact Sheets Fact Sheet: President Biden Reignites Cancer Moonshot to End Cancer as We Know It Biden-Harris Administration Sets Goal of Reducing Cancer [...]
Future of Life Institute. Slaughterbots – if human: kill()
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER. This could be our future, unless we agree to prohibit lethal autonomous weapons. To learn what Slaughterbots are, how they pose a risk, and what solutions are available, visit [...]
Final Report. National Security Commission on Artificial Intelligence (NSCAI)
Final Report. National Security Commission on Artificial Intelligence (NSCAI) Executive Summary The Final Report presents the NSCAI’s strategy for winning the artificial intelligence era. The 16 chapters explain the steps the United States must take to responsibly use AI for national security and defense, defend [...]
NATIONAL ARTIFICIAL INTELLIGENCE INITIATIVE OVERSEEING AND IMPLEMENTING THE UNITED STATES NATIONAL AI STRATEGY
NATIONAL ARTIFICIAL INTELLIGENCE INITIATIVE OVERSEEING AND IMPLEMENTING THE UNITED STATES NATIONAL AI STRATEGY The National Artificial Intelligence Initiative (NAII) was established by the National Artificial Intelligence Initiative Act of 2020 (NAIIA) (DIVISION E, SEC. 5001) – bipartisan legislation enacted on January 1, 2021. The main purposes of the initiative [...]
Kubeflow. MLOps open-source tool by Google on top of Kubernetes
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER. Kubeflow. MLOps open-source tool by Google on top of Kubernetes Almost immediately after Kubernetes established itself as the standard for working with a cluster of containers, Google created [...]
Zoom In: An Introduction to Circuits. Published by OpenAI. March 10, 2020
Zoom In: An Introduction to Circuits By studying the connections between neurons, we can find meaningful algorithms in the weights of neural networks - OpenAI. March 10, 2020 Zoom In: An Introduction to Circuits By studying the connections between neurons, we can find meaningful algorithms in the weights of neural [...]
STUART RUSSELL. Human Compatible: Artificial Intelligence and the Problem of Control. 2019.
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER. STUART RUSSELL. Human Compatible: Artificial Intelligence and the Problem of Control. 2019. Amazon. "The most important book on AI this year." --The Guardian "Mr. Russell's exciting book goes deep, while sparkling with [...]
The Bitter Lesson. Rich Sutton. March 13, 2019.
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER. The Bitter Lesson. Rich Sutton. March 13, 2019. The biggest lesson that can be read from 70 years of AI research is that general methods that leverage computation are ultimately the most [...]
Will MacAskill | Effective Altruism | What are the most important moral problems of our time?
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER. Will MacAskill | Effective Altruism | What are the most important moral problems of our time? FOR EDUCATIONAL AND KNOWLEDGE [...]
Strategic Implications of Openness in AI Development (2107). Nick Bostrom. Future of Humanity Institute, University of Oxford.
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER. Strategic Implications of Openness in AI Development (2107) Nick Bostrom. Future of Humanity Institute, University of Oxford. Policy Implications The global desirability of openness in AI development – sharing e.g. source code, [...]
Artificial Superintelligence: A Futuristic Approach by Roman V. Yampolskiy (2016)
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER. AT AMAZON. Artificial Superintelligence: A Futuristic Approach 1st Edition by Roman V. Yampolskiy (2016) A day does not go by without a news article reporting some amazing breakthrough in artificial intelligence [...]
TED | What happens when our computers get smarter than we are? | Nick Bostrom | 27 APRIL 2015
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. TED | What happens when our computers get smarter than we are? | Nick Bostrom | 27 APRIL 2015 TED | How civilization could destroy itself -- and 4 ways we could prevent [...]
VANITY FAIR. Elon Musk: Artificial Intelligence Could Wipe Out Humanity. 09 Oct 2014.
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER. VANITY FAIR. Elon Musk: Artificial Intelligence Could Wipe Out Humanity. 09 Oct 2014. FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE [...]
Stephen Hawking: ‘Transcendence looks at the implications of artificial intelligence – but are we taking AI seriously enough?’ 01 MAY 2014
FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER. Stephen Hawking: 'Transcendence looks at the implications of artificial intelligence - but are we taking AI seriously enough?' - THE INDEPENDENT - 01 MAY 2014 FOR EDUCATIONAL PURPOSES Success in creating AI [...]