AI Offense-Defense Dynamics Lead Researcher jobs in United States
cer-icon
Apply on Employer Site
company-logo

Future of Life Institute (FLI) · 1 month ago

AI Offense-Defense Dynamics Lead Researcher

Future of Life Organizations is focused on addressing the risks posed by advanced AI systems. As an AI Offense-Defense Dynamics Lead Researcher, you will lead research to understand the dynamics of AI technologies and their societal implications, developing frameworks that provide actionable insights for stakeholders to manage AI risks effectively.

Non Profit
badNo H1Bnote

Responsibilities

Develop quantitative system dynamics models capturing the interrelationships between technological, social, and institutional factors that influence AI risk landscapes
Design detailed analytical models and simulations to identify critical leverage points where policy interventions could shift offense-defense balances toward safer outcomes
Expand and operationalize our current offense/defense dynamics taxonomy and nascent framework, developing metrics and models to predict whether specific AI system features favor offensive or defensive applications
Build empirically-informed analytical frameworks using documented cases of AI misuse and beneficial deployed uses to validate theoretical models
Research how specific technical characteristics (capabilities breadth/depth, accessibility, adaptability, etc.) interact with sociotechnical contexts to determine offense-defense balances
Build public understanding of offense-defense dynamics through blog posts, articles, conference talks, and media engagement
Create tools and methodologies to assess new AI models upon release for their likely offense-defense implications
Draft evidence-based guidance for AI governance that accounts for complex interdependencies between technological capabilities and deployment contexts
Translate research findings into actionable guidance for key stakeholders including policymakers, AI developers, security professionals, and standards organizations

Qualification

Complex systems modelingRisk assessment methodologiesDual-use technologiesModern AI systems analysisSecurity mindsetAI governanceOperational risk managementSystems dynamics modelingNetwork theoryComplexity scienceExceptional analytical thinkingInterdisciplinary researchPublications in technology risk

Required

A M.Sc. or higher in either Computer Science, Cybersecurity, Criminology, Security Studies, AI Policy, Risk Management, or a related field
Demonstrated experience with complex systems modeling, risk assessment methodologies, or security analysis
Strong understanding of dual-use technologies and the factors that influence whether capabilities favor offensive or defensive applications
Deep understanding of modern AI systems, including large language models, multimodal models, and autonomous agents, with ability to analyze their technical architectures and capability profiles
Experience in any of the following: Security mindset, Security studies research, Cybersecurity, Safety engineering, AI governance, Operational risk management, Systems dynamics modeling, Network theory, Complexity science, Adversarial analysis, or Technical standards development
Ability to develop both qualitative frameworks and quantitative models that capture sociotechnical interactions, and comfort creating semi-quantitative semi-empirical models also grounded in logic
Record of relevant publications or research contributions related to technology risk, governance, or security
Exceptional analytical thinking with ability to identify non-obvious path dependencies and feedback loops in complex systems

Preferred

PhD in a relevant field
Experience with system dynamics modeling, hypergraph techniques, or other complex network analysis methods
Skills in developing interactive tools or dashboards for risk visualization and communication
Background in interdisciplinary research bridging technical and social science domains
Demonstrated aptitude in top-down techniques and first-principles thinking
Experience with the quantification of qualitative risk factors or developing proxy metrics for complex phenomena
Background in compiling and analyzing incident databases or case studies for pattern recognition
Familiarity with empirical approaches to technology assessment and impact prediction
Knowledge of international relations theory as it applies to technology proliferation dynamics

Benefits

Good benefits if U.S. employee

Company

Future of Life Institute (FLI)

twittertwittertwitter
company-logo
The Future of Life Institute (FLI) is an independent nonprofit that works to reduce extreme, large-scale risks from transformative technologies, as well as steer the development and use of these technologies to benefit life.

Funding

Current Stage
Early Stage
Total Funding
$1.67M
Key Investors
Open Philanthropy
2021-11-15Grant· $0.48M
2015-08-01Grant· $1.19M

Leadership Team

leader-logo
Viktoriya Krakovna
Co-founder
linkedin
Company data provided by crunchbase