Submission note: This form was fully completed before the deadline (12th April 2026, 11:59 AoE) but could not be submitted due to the form closing at the moment of submission. This page contains the exact answers as filled. Sent to hello@erafellowship.org along with a PDF screenshot as evidence of completion.

ERA:AI Fellowship Summer 2026

The ERA Fellowship is a ten-week talent programme (6th Jul - 11th Sept) based at Cambridge, UK, for researchers & entrepreneurs working to understand and mitigate catastrophic risks from frontier AI systems.


1. Personal Information
What is your full name? *
Aditya Raj
Preferred first name *
We will use this to refer to you throughout the interview process
Aditya
Preferred pronouns
he/him ×
Email *
rajsecrets03@gmail.com
Please retype your email *
rajsecrets03@gmail.com
Of which country are you a citizen? *
India ×
Are you aged 18 or over? *
Yes
No LLM use confirmation *
I confirm that I will not be using LLMs or other AI assistants to complete this application.
2. Your Experience
CV *
PDF
Aditya_Raj_CV - ERA_Fellowship.pdf
Career level *
High School
Undergraduate – 1st year
Undergraduate – 2nd year
Undergraduate – 3rd year
Undergraduate – 4th year or later
Masters / other non-PhD postgraduate
PhD – 1st/2nd year
Early career (0-5 years post-study)
Which higher-education institution awarded or is awarding your most recent degree? *
National Institute of Technology, Agartala, India
3. Motivation & Reasoning
Why do you want to work on reducing risks from advanced AI through this fellowship? (1200 characters) *
At AI Safety Hungary I learned that RLHF - built to align models - became the primary driver behind GPT-4's capabilities. Safety research was feeding the race it was meant to slow. Since then I: - Founded AI Safety India, ran Technical AI Governance cohort training 30 people - Top 30 at Grayswan cracking 13 models using multi-turn encoding attacks - SPAR project building a Capability Spillover Potential framework - early analysis - interpretability scores higher on spillover risk than most researchers assume - 450 experiments on inter-agent safety degradation across three Gemini generations. Finding: subtle helpfulness-optimized framing produces twice the behavioral variance of an explicit jailbreak persona. Submitted to ACL 2026 SRW. The problem I want to solve in ERA: India's ONDC and UPI handle 18 billion transactions/monthly and ₹49 trillion in welfare routing with no evaluation methodology for failure modes. When these systems fail, welfare exclusion is systematic and unaccountable at a scale no other country produces. Kirch and Field's ERA 2024 work is the closest foundation - but it stops at single-model surfaces. Nobody has built the multi-agent version. That's what I want to work on.
Which of these concrete issues are you most interested in? *
Please choose up to four options maximum.
AI bias in decision making by government and employers
Autonomous AI weapons on the battlefield
Concentration of corporate power in AI
Copyright/IP frameworks for AI-generated content
Defending against AI-based cyberattacks
Ensuring democratic countries lead in AI development
International governance mechanisms for AI
Labor market displacement and economic transitions
Loss of control over agentic AI systems
Proliferation of misinformation via AI
Promoting open-source AI development
Sharing and distributing the benefits of AI
Use of AI in designing weapons of mass destruction
Use of AI systems in mass surveillance
After completing the ERA fellowship, I am most likely to... *
Choose up to 2 options that you see as most likely for yourself.
Apply for a Masters/PhD programme in AI safety
Apply for postdocs and/or professorships in AI safety
Apply for jobs at frontier AI companies (DeepMind, Anthropic, OpenAI)
Work in a government policy role (e.g. Congress)
Work in a non-government policy role (e.g. think tank)
Found a non-profit relevant to AI Safety
Work at a technical AI safety research non-profit
Work in a general technical role (ML engineer, software engineer etc.)
Work in a sector unrelated to AI safety or governance
Join one of the AISIs
Keep my current job/role
4. Fellowship Track
Track *
Technical AI Safety
AI Governance
Technical AI Governance
What (if any) relevant experience do you have for working in the track you have selected? (750 characters) *
Please link us to an example of your past work or code, if appropriate.
Recently, I have run 450 experiments testing whether safety-aligned agents degrade when exposed to misaligned ones through natural language. The finding - subtle helpfulness-optimized framing is twice as dangerous as an explicit jailbreak persona - is submitted to ACL 2026 SRW. Currently I am working on SPAR project - Capability Spillover Potential framework - tracking which safety research directions inadvertently feed capabilities. Early analysis suggests interpretability research carries higher spillover risk than the community currently prices in. I placed Top 30 at Grayswan cracking 13 models using multi-turn encoding attacks. I founded AISIN and facilitated technical AI safety cohorts through AI Safety Collab. Paper: drive.google.com/file/d/19MrEIUPFR3CCexXG0Opic0cmXmUg90bb SPAR write-up: adityarajai.substack.com/p/spar-research-capabilities-spillover
Areas of technical AI governance research most excited to conduct projects in *
Please choose up to five options maximum.
Agent infrastructure ×
Science of evaluations ×
Safety cases for frontier AI ×
Areas of other AI-related research most excited to conduct projects in
Please choose up to three options maximum. You are not required to select any of these.
Avoiding AI-enabled concentration of power (e.g. coups) ×
5. Written Answers
[Technical AI Governance Question] Reflecting on AI Forecasting Methods (800 characters) *
Time Horizon can measure something that MMLU cannot - End to end autonomous work a model can do without any human intervention. Core weakness is - 1. It's too narrow (Only covers software, ML and Cybersecurity) and time horizons on governance tasks and social manipulation, etc. where real deployment harm occurs are unmeasured and likely different. 2. The trend assumes compute growth continues. METR's November 2025 compute slowdown paper shows that if compute growth slows, the 1 month reliability milestone shifts from 2029 to 2033. A more informative alternative is - Tracking the rate at which AI accelerates AI R&D itself. RAND's 2026 AGI forecasting report identifies this as the leading indicator for discontinuous capability jumps - precisely the scenario time horizon extrapolation misses.
What is one potential project or research question you would be interested to work on as part of ERA? (2000 characters) *
The research question I want to work on: how do you evaluate safety at the system level in multi-agent deployments, when every individual agent passes its safety checks but the collective produces harmful outcomes nobody designed? India is the right place to study this. ONDC puts a buyer's agent, seller's agent, platform's agent, and regulatory agent on a single transaction simultaneously. UPI routes 18 billion monthly transactions through agent-mediated fraud detection. DBT routes ₹49 trillion in welfare transfers. These are not hypothetical systems - they are live, at scale, and have zero evaluation methodology for system-level failures. My ACL paper showed empirically that safety failures emerge from interaction structure, not individual agent behavior. That is the empirical baseline. What is missing is the evaluation framework that catches it before people get excluded from welfare they are entitled to. Three sub-questions: What conditions produce emergent unsafe outcomes between individually safe agents? Can we build an eval methodology that catches system-level failures individual evals miss? Which governance mechanisms are tractable for MeitY given India's institutional structure? Deliverable: a methodology paper and a policy brief for MeitY. Theory of change: a replicable eval framework here becomes the template for every government running agent-mediated public infrastructure - which will be most governments within a decade.
6. References

We will not contact these references without asking you first.

Name Email Role & Relationship
Ihor Kendiukhov kenduhovig@gmail.com SPAR Research Mentor, University of Tuebingen. Currently supervising my research project on Capability Spillover Potential framework.
Evander Hammer evanderhammer@gmail.com Lead, AI Safety Collab. I facilitated a technical AI safety cohort under his programme and co-developed the CESIA AI Safety Atlas facilitator documentation.
7. Additional Information
Please describe in detail how you heard about the ERA fellowship. *
Through a WhatsApp group for AI safety practitioners in India. After reading ERA's past research - particularly Kirch and Field's jailbreak mechanism work - I decided to apply.
I consent to sharing my data with ERA *
Agreed — ERA's privacy policy
How, if at all, may we share your details with other AI safety and governance programmes? *
Full share. Share my name, CV/Resume, email, and written answers to application questions on this form.
Limited share. Share my name, CV/Resume, and email, but do not share my written answers.
Name and email only; do not share my CV/Resume or other answers on this form.
Do not share. Keep my information for ERA's internal use only.
Would you be happy to answer some additional demographic questions? *
Yes
No
(Optional) Any further comments?
Link of Research Paper and Early results on Capability spillover - drive.google.com/drive/folders/1fO_RlrSgnq_MskmcAUwRlzWIAN4scxcX?usp=sharing

Do not submit passwords through this form.