Artwork

Content provided by The Nonlinear Fund. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by The Nonlinear Fund or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://player.fm/legal.
Player FM - Podcast App
Go offline with the Player FM app!

AF - Timaeus is hiring! by Jesse Hoogland

4:04
 
Share
 

Manage episode 428611924 series 3314709
Content provided by The Nonlinear Fund. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by The Nonlinear Fund or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://player.fm/legal.
Welcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: Timaeus is hiring!, published by Jesse Hoogland on July 12, 2024 on The AI Alignment Forum. TLDR: We're hiring two research assistants to work on advancing developmental interpretability and other applications of singular learning theory to alignment. About Us Timaeus's mission is to empower humanity by making breakthrough scientific progress on alignment. Our research focuses on applications of singular learning theory to foundational problems within alignment, such as interpretability (via "developmental interpretability"), out-of-distribution generalization (via "structural generalization"), and inductive biases (via "geometry of program synthesis"). Our team spans Melbourne, the Bay Area, London, and Amsterdam, collaborating remotely to tackle some of the most pressing challenges in AI safety. For more information on our research and the position, see our Manifund application, this update from a few months ago, our previous hiring call and this advice for applicants. Position Details Title: Research Assistant Location: Remote Duration: 6-month contract with potential for extension. Compensation: Starting at $35 USD per hour as a contractor (no benefits). Start Date: Starting as soon as possible. Key Responsibilities Conduct experiments using PyTorch/JAX on language models ranging from small toy systems to billion-parameter models. Collaborate closely with a team of 2-4 researchers. Document and present research findings. Contribute to research papers, reports, and presentations. Maintain detailed research logs. Assist with the development and maintenance of codebases and repositories. Projects As a research assistant, you would likely work on one of the following two projects/research directions (this is subject to change): Devinterp of language models: (1) Continue scaling up techniques like local learning coefficient (LLC) estimation to larger models to study the development of LLMs in the 1-10B range. (2) Work on validating the next generations of SLT-derived techniques such as restricted LLC estimation and certain kinds of weight- and data-correlational analysis. This builds towards a suite of SLT-derived tools for automatically identifying and analyzing structure in neural networks. SLT of safety fine-tuning: Investigate the use of restricted LLCs as a tool for measuring (1) reversibility of safety fine-tuning and (2) susceptibility to jailbreaking. Having now validated many of our predictions around SLT, we are now working hard to make contact with real-world safety questions as quickly as possible. See our recent Manifund application for a more in-depth description of this research. Qualifications Strong Python programming skills, especially with PyTorch and/or JAX. Strong ML engineering skills (you should have completed at least the equivalent of a course like ARENA). Excellent communication skills. Ability to work independently in a remote setting. Passion for AI safety and alignment research. A Bachelor's degree or higher in a technical field is highly desirable. Full-time availability. Bonus: Familiarity with using LLMs in your workflow is not necessary but a major plus. Knowledge of SLT and Developmental Interpretability is not required, but is a plus. Application Process Interested candidates should submit their applications by July 31st. Promising candidates will be invited for an interview consisting of: 1. A 30-minute background interview, and 2. A 30-minute research-coding interview to assess problem-solving skills in a realistic setting (i.e., you will be allowed expected to use LLMs and whatever else you can come up with). Apply Now To apply, please submit your resume, write a brief statement of interest, and answer a few quick questions here. Join us in shaping the future of AI alignment research. Apply now to be part of the Timaeus team! Than...
  continue reading

2434 episodes

Artwork
iconShare
 
Manage episode 428611924 series 3314709
Content provided by The Nonlinear Fund. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by The Nonlinear Fund or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://player.fm/legal.
Welcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: Timaeus is hiring!, published by Jesse Hoogland on July 12, 2024 on The AI Alignment Forum. TLDR: We're hiring two research assistants to work on advancing developmental interpretability and other applications of singular learning theory to alignment. About Us Timaeus's mission is to empower humanity by making breakthrough scientific progress on alignment. Our research focuses on applications of singular learning theory to foundational problems within alignment, such as interpretability (via "developmental interpretability"), out-of-distribution generalization (via "structural generalization"), and inductive biases (via "geometry of program synthesis"). Our team spans Melbourne, the Bay Area, London, and Amsterdam, collaborating remotely to tackle some of the most pressing challenges in AI safety. For more information on our research and the position, see our Manifund application, this update from a few months ago, our previous hiring call and this advice for applicants. Position Details Title: Research Assistant Location: Remote Duration: 6-month contract with potential for extension. Compensation: Starting at $35 USD per hour as a contractor (no benefits). Start Date: Starting as soon as possible. Key Responsibilities Conduct experiments using PyTorch/JAX on language models ranging from small toy systems to billion-parameter models. Collaborate closely with a team of 2-4 researchers. Document and present research findings. Contribute to research papers, reports, and presentations. Maintain detailed research logs. Assist with the development and maintenance of codebases and repositories. Projects As a research assistant, you would likely work on one of the following two projects/research directions (this is subject to change): Devinterp of language models: (1) Continue scaling up techniques like local learning coefficient (LLC) estimation to larger models to study the development of LLMs in the 1-10B range. (2) Work on validating the next generations of SLT-derived techniques such as restricted LLC estimation and certain kinds of weight- and data-correlational analysis. This builds towards a suite of SLT-derived tools for automatically identifying and analyzing structure in neural networks. SLT of safety fine-tuning: Investigate the use of restricted LLCs as a tool for measuring (1) reversibility of safety fine-tuning and (2) susceptibility to jailbreaking. Having now validated many of our predictions around SLT, we are now working hard to make contact with real-world safety questions as quickly as possible. See our recent Manifund application for a more in-depth description of this research. Qualifications Strong Python programming skills, especially with PyTorch and/or JAX. Strong ML engineering skills (you should have completed at least the equivalent of a course like ARENA). Excellent communication skills. Ability to work independently in a remote setting. Passion for AI safety and alignment research. A Bachelor's degree or higher in a technical field is highly desirable. Full-time availability. Bonus: Familiarity with using LLMs in your workflow is not necessary but a major plus. Knowledge of SLT and Developmental Interpretability is not required, but is a plus. Application Process Interested candidates should submit their applications by July 31st. Promising candidates will be invited for an interview consisting of: 1. A 30-minute background interview, and 2. A 30-minute research-coding interview to assess problem-solving skills in a realistic setting (i.e., you will be allowed expected to use LLMs and whatever else you can come up with). Apply Now To apply, please submit your resume, write a brief statement of interest, and answer a few quick questions here. Join us in shaping the future of AI alignment research. Apply now to be part of the Timaeus team! Than...
  continue reading

2434 episodes

All episodes

×
 
Loading …

Welcome to Player FM!

Player FM is scanning the web for high-quality podcasts for you to enjoy right now. It's the best podcast app and works on Android, iPhone, and the web. Signup to sync subscriptions across devices.

 

Quick Reference Guide