conference
★ 0.71 CFP closes May 4, 2026
Neural Information Processing Systems 2026 held across three satellite locations: Sydney, Atlanta, and Paris. Community-driven event created by and for the community. Paper abstract deadline May 4, full paper deadline May 6, workshop application deadline June 6, author notifications September 24. Safety-related workshops to be announced.
#interpretability#evals#alignment#ml-research#safety-research major-conferencemulti-location
Salience signals
{
"type_weight": 1,
"source_trust": 0.85,
"topic_relevance": 0.75,
"time_proximity": 0.17081081081081081,
"community_signal": 0.8,
"speaker_org_signal": 0.75,
"is_deadline_open": 1,
"source_count": 1
} workshop
★ 0.95 CFP closes May 8, 2026
Annual mechanistic interpretability workshop at ICML bringing together researchers from academia, industry, and independent research to discuss advances in understanding neural networks' internal mechanisms through analysis of weights and activations. Addresses fundamental scientific challenge as neural networks grow in influence and capability. Follows successful workshops at ICML 2024 and NeurIPS 2025.
#interpretability#alignment#circuit-tracing#sparse-autoencoders#mechanistic-interpretability#instrument-science ICMLworkshopmechanistic-interpretabilityICML-workshopthird-iteration
Salience signals
{
"type_weight": 0.85,
"source_trust": 0.9,
"topic_relevance": 0.95,
"time_proximity": 0.7685534591194969,
"community_signal": 0.85,
"speaker_org_signal": 0.85,
"is_deadline_open": 1,
"source_count": 1
} workshop
★ 0.81 CFP closes May 8, 2026
Annual workshop on pluralistic AI alignment focusing on aligning with the diversity of human values. Welcomes research spanning philosophy, machine learning, human-computer interaction, social sciences, policy studies, and applications. Topics include handling annotation disagreements, consensus methods, ethical considerations, and case studies. Papers 4-8 pages, non-archival, submitted via OpenReview. Authors must agree to serve as reviewers.
#alignment#governance#pluralistic-alignment#pluralistic-values ICML-workshopinterdisciplinary
Salience signals
{
"type_weight": 0.85,
"source_trust": 0.85,
"topic_relevance": 0.8,
"time_proximity": 0.7635220125786164,
"community_signal": 0.75,
"speaker_org_signal": 0.75,
"is_deadline_open": 1,
"source_count": 1
} workshop
★ 0.92 Apps close May 10, 2026
5-week project-based course for engineers and early researchers to work with an AI safety expert on a contribution to AI safety research or engineering. Includes mentorship, regular check-ins, and a published write-up. Covers alignment, mechanistic interpretability, evaluations, red-teaming, AI control, and scalable oversight.
#alignment#interpretability#evals
Salience signals
{
"type_weight": 0.85,
"source_trust": 0.9,
"topic_relevance": 0.9,
"time_proximity": 0.8,
"community_signal": 0.8,
"speaker_org_signal": 0.85,
"is_deadline_open": 1,
"source_count": 1
} conference
★ 0.79 Reg closes May 13, 2026
Third annual Technical AI Safety conference, free one-day event at Oxford. Welcomes researchers and professionals from all backgrounds regardless of prior research experience. Organized by Oxford Martin AI Governance Initiative and Noeon Research. Registration available through Luma. Video content from past conferences available.
#alignment#governance#safety-research#evals#interpretability conferencetechnicalOxfordfreeone-day
Salience signals
{
"type_weight": 1,
"source_trust": 0.9,
"topic_relevance": 0.9,
"time_proximity": 0.6857142857142857,
"community_signal": 0.8,
"speaker_org_signal": 0.85,
"is_deadline_open": 0,
"source_count": 1
} fellowship
★ 0.87 Apps close May 18, 2026
Three-week ML upskilling bootcamp for AI safety focusing on interpretability and reinforcement learning, based on ARENA curriculum. Run by Cambridge Boston Alignment Initiative. Provides housing, meals, 24/7 office access, teaching assistants, and travel support. Prerequisites include Python familiarity and comfort with multivariable calculus and linear algebra.
#interpretability#alignment#control#mechanistic-interpretability bootcampARENA-curriculum
Salience signals
{
"type_weight": 0.8,
"source_trust": 0.85,
"topic_relevance": 0.9,
"time_proximity": 0.8,
"community_signal": 0.75,
"speaker_org_signal": 0.8,
"is_deadline_open": 1,
"source_count": 1
} conference
★ 0.82 Apps close May 20, 2026
EA Global London 2026 conference organized by Centre for Effective Altruism. Three-day gathering with heavy AI safety attendance. Applications open with default ticket price £500, reduced-price and free tickets available. In scope for social event reasons given in the AAE codex.
#alignment#governance effective-altruism
Salience signals
{
"type_weight": 1,
"source_trust": 0.8,
"topic_relevance": 0.7,
"time_proximity": 0.979874213836478,
"community_signal": 0.8,
"speaker_org_signal": 0.7,
"is_deadline_open": 1,
"source_count": 1
} hackathon
★ 0.72 Reg closes May 21, 2026
Three-day hybrid hackathon organized by Apart Research focused on secure program synthesis. Part of Apart Research's monthly AI safety sprint series with 55+ sprints completed and 6,000+ participants globally.
#alignment#control#safety-research#evals#security#automated-research#safety-applications hybridsprint
Salience signals
{
"type_weight": 0.65,
"source_trust": 0.85,
"topic_relevance": 0.7,
"time_proximity": 0.9142857142857143,
"community_signal": 0.65,
"speaker_org_signal": 0.75,
"is_deadline_open": 1,
"source_count": 1
} fellowship
★ 0.68 Early-bird ends May 22, 2026
Nine-week intensive fellowship at London Initiative for Safe AI (LISA). Provides £6,000-£8,000 stipend, travel to London, £2,000 housing allowance for non-London residents, meals and compute resources. 70-90% receive extensions for continued support up to 6 months. Open to anyone committed to ensuring AI develops safely, diverse backgrounds (ML, philosophy, policy, physics, biology). 7 completed cohorts with 129 alumni now at Anthropic, Google DeepMind, UK AISI, and founding ventures.
#alignment#interpretability#governance#evals#safety-research#technical-safety Londonstipendquarterly
Salience signals
{
"type_weight": 0.8,
"source_trust": 0.75,
"topic_relevance": 0.85,
"time_proximity": 0.8238993710691824,
"community_signal": 0.7,
"speaker_org_signal": 0.75,
"is_deadline_open": 0,
"source_count": 1
} conference
★ 0.83 Early-bird ends May 24, 2026
International Conference on Machine Learning 2026. July 6: Expo/Tutorial Day, July 7-9: Main Conference, July 10-11: Workshops. Features safety-related workshops including Mechanistic Interpretability and Pluralistic Alignment. Emphasis on community standards, peer review ethics, and research ethics. Early registration deadline May 24.
#alignment#interpretability#evals#ml-research major-conference
Salience signals
{
"type_weight": 1,
"source_trust": 0.85,
"topic_relevance": 0.75,
"time_proximity": 0.7886792452830189,
"community_signal": 0.8,
"speaker_org_signal": 0.75,
"is_deadline_open": 1,
"source_count": 1
} hackathon
★ 0.85 Apps close May 26, 2026
Five-day intensive programme for AI safety founders going from idea to funded. Successful pitches receive £50k in equity-free seed funding. Part of BlueDot Impact's incubator and rapid-funding initiatives supporting concrete AI safety work.
#alignment#governance startupincubator
Salience signals
{
"type_weight": 0.65,
"source_trust": 0.9,
"topic_relevance": 0.85,
"time_proximity": 0.9647798742138365,
"community_signal": 0.75,
"speaker_org_signal": 0.75,
"is_deadline_open": 1,
"source_count": 1
} conference
★ 0.93 Apps close Jun 1, 2026
Five-day multi-track conference bringing together researchers in theoretical AI alignment. Unconference format where attendees propose and lead sessions. Covers mathematical approaches including Singular Learning Theory, Agent Foundations, Causal Incentives, Computational Mechanics, Safety-by-Debate, and Scalable Oversight. Free attendance, limited accommodations, financial support available for travel/lodging. Organized by Iliad umbrella organization.
#alignment#theory#interpretability#agent-foundations#formal-foundations#theoretical-alignment#theoretical-foundations conferenceunconferencetheoreticalfreemathematical
Salience signals
{
"type_weight": 1,
"source_trust": 0.85,
"topic_relevance": 0.95,
"time_proximity": 0.6477987421383647,
"community_signal": 0.8,
"speaker_org_signal": 0.85,
"is_deadline_open": 1,
"source_count": 1
} fellowship
★ 0.87 Apps close Jun 1, 2026
10-week fellowship for researchers and entrepreneurs working on mitigating risks from frontier AI, combining technical and governance approaches. Three focus areas: technical AI safety, AI governance, and technical AI governance. Provides competitive stipend, meals, transportation, visa, and lodging. Open to talented individuals worldwide at any career stage. Expression of interest form available.
#alignment#governance#evals#safety-research#technical-safety Cambridge-UKstipend
Salience signals
{
"type_weight": 0.8,
"source_trust": 0.85,
"topic_relevance": 0.9,
"time_proximity": 0.7886792452830189,
"community_signal": 0.75,
"speaker_org_signal": 0.8,
"is_deadline_open": 1,
"source_count": 1
} conference
★ 0.65 Reg closes Jun 4, 2026
Foresight Institute's flagship event gathering leading scientists, entrepreneurs, funders, and policymakers to explore frontiers of science and technology. Includes AI safety track alongside other frontier science topics. 40-year-old organization focused on transformative technology.
#governance#alignment#ai-safety#frontier-tech#frontier-technology#policy networkingflagship
Salience signals
{
"type_weight": 1,
"source_trust": 0.75,
"topic_relevance": 0.7,
"time_proximity": 0.9446540880503145,
"community_signal": 0.65,
"speaker_org_signal": 0.7,
"is_deadline_open": 0,
"source_count": 1
} hackathon
★ 0.76 Reg closes Jun 6, 2026
One-day hackathon focused on creating AI risk content and educational materials. Part of BlueDot Impact's community-building initiatives to improve public understanding of AI safety challenges.
#alignment#governance content-creation
Salience signals
{
"type_weight": 0.65,
"source_trust": 0.85,
"topic_relevance": 0.75,
"time_proximity": 0.939622641509434,
"community_signal": 0.7,
"speaker_org_signal": 0.7,
"is_deadline_open": 1,
"source_count": 1
} conference
★ 0.67 Reg closes Jun 7, 2026
Two-day conference addressing artificial intelligence implications for international peace and security. Organized by United Nations Institute for Disarmament Research. Focuses on AI governance, security implications, and ethical frameworks for AI development in international context.
#governance#policy#ai-ethics#international-policy UNinternationalgovernancehybridGenevainternational-policy
Salience signals
{
"type_weight": 1,
"source_trust": 0.85,
"topic_relevance": 0.7,
"time_proximity": 0.879245283018868,
"community_signal": 0.6,
"speaker_org_signal": 0.8,
"is_deadline_open": 0,
"source_count": 1
} conference
★ 0.82 CFP closes Jun 15, 2026
Two-day interdisciplinary gathering bringing together researchers, policymakers, and practitioners from academia, government, industry, and civil society. Grounded in the International AI Safety Report, focusing on technical challenges, governance, risk assessment, safety evaluation and testing frameworks. Features keynotes, panels, workshops, breakout sessions, lightning talks, and networking. Organized by Gradient Institute with RAND, Timaeus, Good Ancestors, University of Sydney, and CSIRO Data61.
#evals#governance#evaluation#measurement-science#alignment interdisciplinary
Salience signals
{
"type_weight": 1,
"source_trust": 0.75,
"topic_relevance": 0.85,
"time_proximity": 0.7836477987421384,
"community_signal": 0.65,
"speaker_org_signal": 0.75,
"is_deadline_open": 1,
"source_count": 1
} workshop
★ 0.75 Reg closes Jul 5, 2026
Part of FAR.AI's ongoing Alignment Workshop series, deepening collective understanding of potential risks from Artificial General Intelligence (AGI) and exploring mitigation strategies. Hosted in Seoul during ICML 2026 week.
#alignment#governance#control#interpretability#technical-safety
Salience signals
{
"type_weight": 0.85,
"source_trust": 0.9,
"topic_relevance": 0.85,
"time_proximity": 0.7886792452830189,
"community_signal": 0.75,
"speaker_org_signal": 0.85,
"is_deadline_open": 0,
"source_count": 1
} fellowship
★ 0.87 Apps close Jul 6, 2026
Three-week ML upskilling bootcamp for AI safety focusing on interpretability and reinforcement learning, based on ARENA curriculum. Run by Cambridge Boston Alignment Initiative. Hosted by Collider in Manhattan. Provides housing, meals, teaching assistants, and travel support.
#interpretability#alignment#control#mechanistic-interpretability bootcampARENA-curriculum
Salience signals
{
"type_weight": 0.8,
"source_trust": 0.85,
"topic_relevance": 0.9,
"time_proximity": 0.7886792452830189,
"community_signal": 0.75,
"speaker_org_signal": 0.8,
"is_deadline_open": 1,
"source_count": 1
} conference
★ 0.80 CFP closes Jul 15, 2026
Summit bringing together academic leaders, entrepreneurs, AI experts, venture capitalists, and policymakers to discuss the future of AI and Agentic AI. Call for Papers and Startup Spotlight applications open. In-person and livestream available.
#alignment#evals#governance
Salience signals
{
"type_weight": 1,
"source_trust": 0.85,
"topic_relevance": 0.75,
"time_proximity": 0.6477987421383647,
"community_signal": 0.75,
"speaker_org_signal": 0.8,
"is_deadline_open": 1,
"source_count": 1
} workshop
★ 0.60 Reg closes Jul 17, 2026
Two-day technical workshop hosted by Foresight Institute focusing on secure and sovereign AI development. Part of Foresight's broader work on transformative technology and AI safety.
#governance#evals#alignment#safety-research#security#control technicalBerlin
Salience signals
{
"type_weight": 0.85,
"source_trust": 0.75,
"topic_relevance": 0.75,
"time_proximity": 0.7283018867924529,
"community_signal": 0.6,
"speaker_org_signal": 0.7,
"is_deadline_open": 0,
"source_count": 1
} fellowship
★ 0.77 Early-bird ends Jul 25, 2026
Fellowship running September 14, 2026 through February 5, 2027, supporting researchers pursuing rigorous, high-impact research on safety and alignment of advanced AI systems. Fellows work in peer group setting with mentorship from OpenAI staff. Workspace offered at Constellation in Berkeley, remote participation permitted. Provides monthly stipends, computational resources, ongoing guidance. Requires meaningful research deliverables (papers, benchmarks, datasets). Welcomes diverse academic backgrounds. Partnering with Constellation.
#alignment#safety-research#governance#evals#control#safety-evals#robustness#oversight#safety-evaluation remote-allowedAPI-credits
Salience signals
{
"type_weight": 0.8,
"source_trust": 0.9,
"topic_relevance": 0.95,
"time_proximity": 0.43647798742138366,
"community_signal": 0.85,
"speaker_org_signal": 0.95,
"is_deadline_open": 0,
"source_count": 1
} fellowship
★ 0.83 Apps close Aug 10, 2026
Three-week ML upskilling bootcamp for AI safety focusing on interpretability and reinforcement learning, based on ARENA curriculum. Run by Cambridge Boston Alignment Initiative. Provides housing, meals, 24/7 office access, teaching assistants, and travel support.
#interpretability#alignment#control#mechanistic-interpretability bootcampARENA-curriculum
Salience signals
{
"type_weight": 0.8,
"source_trust": 0.85,
"topic_relevance": 0.9,
"time_proximity": 0.6125786163522012,
"community_signal": 0.75,
"speaker_org_signal": 0.8,
"is_deadline_open": 1,
"source_count": 1
} fellowship
★ 0.74 Apps close Sep 26, 2026
Astra is a fully funded, in-person fellowship program operating from Constellation's Berkeley research center. Fellows work on technical, governance, strategy, and field-building projects with senior mentors. Benefits include $8,400 monthly stipend, ~$15K/month research budget for empirical fellows, visa support, workspace access, weekly mentorship, and placement services. Extension period available through June 31.
#alignment#control#evals#governance#safety-research#interpretability#strategy fellowshipempiricalgovernancestrategyhigh-stipendcompute-budget
Salience signals
{
"type_weight": 0.8,
"source_trust": 0.85,
"topic_relevance": 0.95,
"time_proximity": 0.43647798742138366,
"community_signal": 0.85,
"speaker_org_signal": 0.85,
"is_deadline_open": 0,
"source_count": 1
} fellowship
★ 0.96 Apps closed May 3, 2026
Four-month Anthropic Fellows Program providing funding, compute (~$15k/month), and close mentorship from Anthropic researchers. Weekly stipend of $3,850 USD / £2,310 / $4,300 CAD. Focus on scalable oversight, adversarial robustness, interpretability, AI welfare, and safety evaluations. Over 40% of first cohort fellows joined Anthropic full-time.
#alignment#interpretability#control#evals#adversarial-robustness fellowshipresearchAnthropic
Salience signals
{
"type_weight": 0.8,
"source_trust": 0.95,
"topic_relevance": 0.95,
"time_proximity": 0.7182389937106919,
"community_signal": 0.85,
"speaker_org_signal": 0.95,
"is_deadline_open": 1,
"source_count": 1
} fellowship
★ 0.93 Apps closed Apr 12, 2026
Nine-week AI safety research fellowship for 30 fellows with $10,000 stipend, housing in Harvard dorms, 24/7 office access in Harvard Square. Weekly 1-2 hour individual mentorship from researchers at Harvard, MIT, Northeastern. Up to $10,000 in compute credits per fellow, conference submission support, weekly speaker events, networking, workshops. Rolling application with 4-stage process: form, 15-min interview, mentor-specific tasks, mentor interview. International students with OPT/CPT eligible, full in-person participation required (18+ only).
#alignment#interpretability#governance#evals#biosecurity#safety-research fellowshipresearchCambridgeHarvardstipendhousingcompute-credits
Salience signals
{
"type_weight": 0.8,
"source_trust": 0.85,
"topic_relevance": 0.95,
"time_proximity": 0.9245283018867925,
"community_signal": 0.75,
"speaker_org_signal": 0.85,
"is_deadline_open": 1,
"source_count": 1
} workshop
★ 0.88 Apps closed May 1, 2026
Workshop on secure and verifiable AI development, bringing together researchers, builders, and funders across ML, hardware security, systems, cryptography, and computer security. Focuses on verification techniques for AI safety. Colocated with IEEE Security and Privacy conference. Organized by FAR.AI.
#evals#alignment#safety-research#security workshopverificationcryptographyhardware-security
Salience signals
{
"type_weight": 0.85,
"source_trust": 0.9,
"topic_relevance": 0.85,
"time_proximity": 0.9428571428571428,
"community_signal": 0.65,
"speaker_org_signal": 0.85,
"is_deadline_open": 1,
"source_count": 1
} fellowship
★ 0.87 Apps closed Jan 18, 2026
12-week research fellowship designed to train and support emerging researchers working on AI alignment, interpretability, governance, and safety. Fellows receive mentorship from leading researchers, research management support, and join focused community. In-person cohorts in Berkeley and London. Top performers can extend for additional 6-12 months with continued funding. Applications closed (deadline Jan 18), collecting expressions of interest for future cohorts.
#alignment#interpretability#governance#theory#control#safety-research#technical-safety fellowshipmentorshipMATSresearchhigh-prestigestipend
Salience signals
{
"type_weight": 0.8,
"source_trust": 0.95,
"topic_relevance": 0.95,
"time_proximity": 0.9647798742138365,
"community_signal": 0.9,
"speaker_org_signal": 0.9,
"is_deadline_open": 0,
"source_count": 1
} fellowship
★ 0.85 Apps closed Mar 24, 2026
The CAIS AI and Society Fellowship is a fully-funded three-month research program for scholars in economics, law, international relations, and adjacent disciplines to investigate how advanced AI may reshape social, economic, geopolitical, and legal systems. Fellows receive $25,000 stipend, covered travel to San Francisco, daily meals, and work with significant autonomy defining their own research directions at CAIS offices. Features regular guest speakers from Stanford, law schools, and international affairs experts.
#governance#policy fellowshipgovernancepolicyresearch
Salience signals
{
"type_weight": 0.8,
"source_trust": 0.9,
"topic_relevance": 0.85,
"time_proximity": 0.959748427672956,
"community_signal": 0.8,
"speaker_org_signal": 0.9,
"is_deadline_open": 0,
"source_count": 2
} fellowship
★ 0.81 Apps closed Apr 1, 2026
ARENA's eighth cohort, a 4-5 week in-person bootcamp for alignment research engineering. Fully funded including travel, visas, accommodation, and meals. Participants build skills in mechanistic interpretability, RL, and transformers. Applications closed. Operated by London Initiative for Safe AI and funded by Coefficient Giving.
#alignment#interpretability#control#technical-safety#evals bootcamptechnicalintensive
Salience signals
{
"type_weight": 0.8,
"source_trust": 0.9,
"topic_relevance": 0.9,
"time_proximity": 1,
"community_signal": 0.8,
"speaker_org_signal": 0.85,
"is_deadline_open": 0,
"source_count": 1
} fellowship
★ 0.79 Apps closed Mar 1, 2026
3-month fellowship to launch or accelerate impactful careers in American AI governance and policy. Fellows conduct independent research projects under expert mentorship while building professional networks and developing policy expertise. Focus areas include public policy, political science, engineering, economics, biosecurity, cybersecurity, China studies, and risk management. Prioritizes bipartisan engagement, rigorous analysis, and practical policy relevance. $21,000 stipend plus travel support, weekday lunches, and DC office space. US work authorization required.
#governance#policy fellowshippolicygovernanceDC
Salience signals
{
"type_weight": 0.8,
"source_trust": 0.9,
"topic_relevance": 0.9,
"time_proximity": 0.89937106918239,
"community_signal": 0.8,
"speaker_org_signal": 0.85,
"is_deadline_open": 0,
"source_count": 1
} fellowship
★ 0.79 Apps closed Jan 4, 2026
3-month fellowship for conducting independent research on AI governance topics. Fellows receive mentorship from field experts, participate in seminars and Q&A sessions, and build professional networks. Research outputs may include reports, white papers, journal articles, op-eds, or blog posts. £12,000 stipend plus travel support and weekday lunches. Open to candidates from government, academia, industry, or civil society with expertise in policy, political science, computer science, economics, or risk management. Visa sponsorship available.
#governance#policy fellowshipresearchgovernanceLondon
Salience signals
{
"type_weight": 0.8,
"source_trust": 0.9,
"topic_relevance": 0.9,
"time_proximity": 0.89937106918239,
"community_signal": 0.8,
"speaker_org_signal": 0.85,
"is_deadline_open": 0,
"source_count": 1
} fellowship
★ 0.78 Apps closed Apr 15, 2026
4-month empirical AI safety research fellowship with Anthropic. Fellows work on scalable oversight, adversarial robustness, AI control, model organisms, mechanistic interpretability, AI security, and model welfare. Includes $3,850/week stipend, ~$15k/month compute, and close mentorship from Anthropic researchers. Over 80% of first cohort fellows produced papers.
#alignment#interpretability#control#evals#adversarial-robustness fellowshipresearchAnthropicapplications-open-may-cohort
Salience signals
{
"type_weight": 0.8,
"source_trust": 0.95,
"topic_relevance": 0.95,
"time_proximity": 0.3666666666666667,
"community_signal": 0.9,
"speaker_org_signal": 0.95,
"is_deadline_open": 0,
"source_count": 1
} Foresight Institute flagship event gathering leading scientists, entrepreneurs, funders, and policymakers to explore frontiers of science and technology including AI safety. 40-year-old organization focused on transformative technology with dedicated AI safety track.
#alignment#governance frontier-sciencemulti-track
Salience signals
{
"type_weight": 1,
"source_trust": 0.75,
"topic_relevance": 0.7,
"time_proximity": 0.9446540880503145,
"community_signal": 0.65,
"speaker_org_signal": 0.7,
"is_deadline_open": 1,
"source_count": 1
} Second Workshop on Technical AI Governance Research at ICML 2026, focusing on technical approaches to AI governance, policy, and regulation. Part of the main conference workshop track.
#governance#policy ICMLworkshopgovernancetechnical-governance
Salience signals
{
"type_weight": 0.85,
"source_trust": 0.85,
"topic_relevance": 0.9,
"time_proximity": 0.7383647798742139,
"community_signal": 0.75,
"speaker_org_signal": 0.7,
"is_deadline_open": 0,
"source_count": 1
} Workshop at ICML 2026 focused on identifying, diagnosing, and fixing failure modes in agentic AI systems. Covers reproducible triggers for failures, diagnostic tracing methods, and verified repair approaches. Highly relevant to AI safety and robustness.
#evals#alignment ICMLworkshopfailure-modesagentsdiagnostics
Salience signals
{
"type_weight": 0.85,
"source_trust": 0.85,
"topic_relevance": 0.9,
"time_proximity": 0.7383647798742139,
"community_signal": 0.75,
"speaker_org_signal": 0.7,
"is_deadline_open": 0,
"source_count": 1
} Second Workshop on Agents in the Wild focusing on safety and security of AI agents deployed in real-world environments. Addresses challenges in ensuring safe and secure operation of autonomous agents. Part of ICML 2026 workshop track.
#alignment#evals ICMLworkshopagentssafetysecurity
Salience signals
{
"type_weight": 0.85,
"source_trust": 0.85,
"topic_relevance": 0.9,
"time_proximity": 0.7333333333333334,
"community_signal": 0.75,
"speaker_org_signal": 0.7,
"is_deadline_open": 0,
"source_count": 1
} Mentorship for Alignment Researchers at Cambridge AI Safety Hub. Teams of 2-4 participants paired with experienced mentors from AI labs, think tanks, and academia to produce published AI safety research. One-week in-person kick-off (July 13-19 or July 20-26) followed by 8-10 week part-time remote phase. Provides $2k+ compute budgets and research management support. Applications closed (Stage 1 deadline was May 3).
#alignment#interpretability#evals#governance mentorship
Salience signals
{
"type_weight": 0.8,
"source_trust": 0.8,
"topic_relevance": 0.9,
"time_proximity": 0.7534591194968554,
"community_signal": 0.7,
"speaker_org_signal": 0.75,
"is_deadline_open": 0,
"source_count": 1
} Weekly AI safety evaluations paper reading club hosted by BlueDot Impact. Meets every Tuesday at 4:00 PM UTC to discuss evaluation methodologies, safety benchmarks, and measurement frameworks. Open to all interested in AI safety evals research.
#evals#alignment weeklypaper-discussion
Salience signals
{
"type_weight": 0.45,
"source_trust": 0.85,
"topic_relevance": 0.9,
"time_proximity": 0.4285714285714286,
"community_signal": 0.7,
"speaker_org_signal": 0.7,
"is_deadline_open": 1,
"source_count": 1
} Weekend AI safety hackathon focused on Global South participation and perspectives. Hybrid format allowing both online and in-person participation. Organized by Apart Research as part of their 55+ research sprints series with 6,000+ participants across 200+ global locations.
#alignment#safety-research#evals#governance
Salience signals
{
"type_weight": 0.65,
"source_trust": 0.85,
"topic_relevance": 0.8,
"time_proximity": 0.8490566037735849,
"community_signal": 0.7,
"speaker_org_signal": 0.7,
"is_deadline_open": 0,
"source_count": 1
} fellowship
★ 0.65 Apps closed Jan 15, 2026
World's first academic programme dedicated to AI evaluation combining technical depth with policy and governance perspectives. 150 hours total: 90 hours online (lectures, networking, activities), 20 hours hands-on courses, 40 hours in-person capstone week in Valencia. Cohort of 40 top global participants. Fully funded scholarships available. Graduates receive 15 ECTS Expert Diploma from ValgrAI. Faculty from Cambridge, Stanford, Princeton, EU AI Office, UK AI Safety Institute, FAR AI, Apollo Research. Funded by Coefficient Giving.
#evals#safety-research#governance#policy fellowshipevalsacademichybriddiplomafundedprestigious
Salience signals
{
"type_weight": 0.8,
"source_trust": 0.85,
"topic_relevance": 0.95,
"time_proximity": 0,
"community_signal": 0.75,
"speaker_org_signal": 0.9,
"is_deadline_open": 0,
"source_count": 1
} workshop
★ 0.64 CFP closed Mar 12, 2026
Two-day workshop focused on tensions between model developers and evaluation researchers, surfacing practical insights from across the evaluation ecosystem. Organized by EvalEval Coalition, hosted by Hugging Face, University of Edinburgh, and EleutherAI. Accepts full papers (6-8 pages), short papers (up to 4 pages), and tiny papers (up to 2 pages). Two-way anonymized review process.
#evals#safety-research#measurement ACL-workshoptwo-day
Salience signals
{
"type_weight": 0.85,
"source_trust": 0.75,
"topic_relevance": 0.8,
"time_proximity": 0.8037735849056604,
"community_signal": 0.6,
"speaker_org_signal": 0.7,
"is_deadline_open": 0,
"source_count": 1
} Part-time remote research program pairing aspiring researchers with 130+ experienced mentors from Google DeepMind, RAND, Apollo Research, MATS, UK AISI, and others for three-month projects. Participants dedicate 5-40 hours weekly tailored to individual availability. Research areas include AI safety, AI policy, AI security, interpretability, biosecurity, and societal impacts. Open to undergraduate, graduate/PhD students, and professionals at different experience levels. Applications closed (mentee deadline was Jan 14).
#alignment#governance#evals#safety-research#interpretability#technical-safety remotementorship
Salience signals
{
"type_weight": 0.8,
"source_trust": 0.85,
"topic_relevance": 0.9,
"time_proximity": 0,
"community_signal": 0.75,
"speaker_org_signal": 0.85,
"is_deadline_open": 0,
"source_count": 1
} conference
★ 0.58 CFP closed Apr 30, 2026
9th annual Cognitive Computational Neuroscience conference, a forum for discussion among researchers in cognitive science, neuroscience, and artificial intelligence focused on understanding computations underlying complex behavior. Single-track format with keynote speakers including Brenden Lake (Princeton), Ila Fiete (MIT), Kenji Doya (OIST), Doris Tsao (Berkeley), and Alona Fyshe (Alberta). AAE attendees follow for predictive-coding, metacognition, and signal-detection-theory measurement work applicable to LLMs.
#interpretability#cognitive-science#control#mechanistic-interpretability#formal-foundations#instrument-science#computational-neuroscience#measurement-science cog-neuromeasurement-sciencecognitive-science
Salience signals
{
"type_weight": 1,
"source_trust": 0.8,
"topic_relevance": 0.7,
"time_proximity": 0.6477987421383647,
"community_signal": 0.5,
"speaker_org_signal": 0.65,
"is_deadline_open": 0,
"source_count": 1
} Two-day conference on global cooperation for cybersecurity resilience and stability. Organized by United Nations Institute for Disarmament Research. Addresses international frameworks for cyber governance and security cooperation.
#governance#cyber-security UNdisarmament
Salience signals
{
"type_weight": 1,
"source_trust": 0.85,
"topic_relevance": 0.6,
"time_proximity": 0.4,
"community_signal": 0.5,
"speaker_org_signal": 0.75,
"is_deadline_open": 0,
"source_count": 1
} Two-week intensive summer school at Columbia University covering machine learning topics including mechanistic interpretability, alignment/safety, RAG & agents, and LLM systems. Approximately 200 PhD students participate alongside faculty and industry speakers. In-scope due to dedicated alignment and mechanistic interpretability tracks.
#interpretability#alignment
Salience signals
{
"type_weight": 0.35,
"source_trust": 0.75,
"topic_relevance": 0.7,
"time_proximity": 0.869182389937107,
"community_signal": 0.5,
"speaker_org_signal": 0.6,
"is_deadline_open": 0,
"source_count": 1
} A free, accessible workshop hosted by AI Safety Awareness Group Oakland exploring AI's trajectory and societal impact. No technical background required. Features live demonstrations of current AI systems, interactive forecasting activities, and discussions about AI's implications for work, relationships, and society over the next 1-5 years.
#governance#alignment
Salience signals
{
"type_weight": 0.45,
"source_trust": 0.7,
"topic_relevance": 0.7,
"time_proximity": 0.5714285714285714,
"community_signal": 0.6,
"speaker_org_signal": 0.5,
"is_deadline_open": 0,
"source_count": 1
}