Are SAE features from the Base Model still meaningful to LLaVA? — AI Alignment Forum

[ad_1] Shan Chen, Jack Gallifant, Kuleen Sasse, Danielle Bitterman[1]Please read this as a work in progress where we are colleagues sharing this in a lab (https://www.bittermanlab.org) meeting to help/motivate potential parallel research.TL;DR:Recent work has evaluated the generalizability of Sparse Autoencoder (SAE) features; this study examines their effectiveness in multimodal settings.We evaluate feature extraction using a CIFAR-100-inspired explainable classification task, analyzing the impact of pooling strategies, binarization, and layer selection on

How Formula 1® uses generative AI to accelerate race-day issue resolution

[ad_1] Formula 1® (F1) races are high-stakes affairs where operational efficiency is paramount. During these live events, F1 IT engineers must triage critical issues across its services, such as network degradation to one of its APIs. This impacts downstream services that consume data from the API, including products such as F1 TV, which offer live and on-demand coverage of every race as well as real-time telemetry. Determining the root cause

Abstract Mathematical Concepts vs. Abstractions Over Real-World Systems — AI Alignment Forum

[ad_1] Consider concepts such as "a vector", "a game-theoretic agent", or "a market". Intuitively, those are "purely theoretical" abstractions: they don't refer to any specific real-world system. Those abstractions would be useful even in universes very different from ours, and reasoning about them doesn't necessarily involve reasoning about our world.Consider concepts such as "a tree", "my friend Alice", or "human governments". Intuitively, those are "real-world" abstractions. While "a tree" bundles

AGI Safety & Alignment @ Google DeepMind is hiring — AI Alignment Forum

[ad_1] The AGI Safety & Alignment Team (ASAT) at Google DeepMind (GDM) is hiring! Please apply to the Research Scientist and Research Engineer roles. Strong software engineers with some ML background should also apply (to the Research Engineer role). Our initial batch of hiring will focus more on hiring engineers, but we expect to continue to use the applications we receive for future hiring this year, which we expect will be more

Automation and intelligent assistance (2025 guide)

[ad_1] Imagine having a personal assistant who can not only schedule your appointments and send emails but also proactively anticipate your needs, learn your preferences, and complete complex tasks on your behalf. That’s the promise of AI agents — intelligent software entities designed to operate autonomously and achieve specific goals.What are AI agents?In simple terms, an AI agent is a computer program that can perceive its environment, make decisions,

Using Amazon Rekognition to improve bicycle safety

[ad_1] Cycling is a fun way to stay fit, enjoy nature, and connect with friends and acquaintances. However, riding is becoming increasingly dangerous, especially in situations where cyclists and cars share the road. According to the NHTSA, in the United States an average of 883 people on bicycles are killed in traffic crashes, with an average of about 45,000 injury-only crashes reported annually. While total bicycle fatalities only account for

How Trump’s ‘drill, baby, drill’ pledge is affecting other countries

[ad_1] Navin Singh KhadkaEnvironment Correspondent, BBC World ServiceGetty ImagesTrump has said the US's oil and gas will be sold all over the worldThe UN climate summit in the United Arab Emirates in 2023 ended with a call to "transition away from fossil fuels". It was applauded as a historic milestone in global climate action.Barely a year later, however, there are fears that the global commitment may be losing momentum, as

Gauging Interest for a Learning-Theoretic Agenda Mentorship Programme — AI Alignment Forum

[ad_1] I'm planning to organize a mentorship programme for people who want to become researchers working on the Learning-Theoretic Agenda (LTA). I'm still figuring out the detailed plan, the logistics and the funding, but here's an outline of how it would looks like. To express interest, submit this form.I believe that the risk of a global catastrophe due to unaligned artificial superintelligence is the most pressing problem of our time.

A short course on AGI safety from the GDM Alignment team — AI Alignment Forum

[ad_1] We are excited to release a short course on AGI safety for students, researchers and professionals interested in this topic. The course offers a concise and accessible introduction to AI alignment, consisting of short recorded talks and exercises (75 minutes total) with an accompanying slide deck and exercise workbook. It covers alignment problems we can expect as AI capabilities advance, and our current approach to these problems (on technical and governance

How to talk to AIs like ChatGPT?

[ad_1] Large Language Models (LLMs) like ChatGPT have become a staple in machine learning and artificial intelligence discussions, marking a profound shift in these fields. Long overshadowed by more mainstream technologies, LLMs have emerged as pivotal elements in everyday digital interactions for millions worldwide.As a researcher with over four years of experience in this domain, I have engaged deeply with these tools, gaining valuable insights, especially recently. Exploring the