mirror of
https://github.com/jackyzha0/quartz.git
synced 2025-12-20 11:24:05 -06:00
initial posting
This commit is contained in:
parent
6aa1314e13
commit
40af1da766
69
content/careers/Applied ML Engineer.md
Normal file
69
content/careers/Applied ML Engineer.md
Normal file
@ -0,0 +1,69 @@
|
||||
---
|
||||
title: Applied ML Engineer
|
||||
date: 02.13.25
|
||||
tags:
|
||||
- positions
|
||||
- product
|
||||
- dev
|
||||
- announcements
|
||||
---
|
||||
|
||||
(NYC, Full-Time)
|
||||
|
||||
## About the Role
|
||||
We're searching for an applied machine learning engineer excited to work on the ML side of [Honcho](https://honcho.dev). You'll work alongside our interdisciplinary team to transform novel ideas into production systems that help LLMs understand and align with individual users.
|
||||
|
||||
This role requires a strong engineer who can rapidly prototype and ship ML systems. The pace of the LLM space is staggering - we need someone with a hacker mentality who is excited about diving into papers/codebases, implementing novel methods at breakneck speed, and figuring out what actually works. Our team is small and fast-moving, so you'll have the freedom to experiment widely and ship impactful features quickly.
|
||||
|
||||
|
||||
## About You
|
||||
- 2-3 years applied LLM experience or equivalent
|
||||
- Proficiency with a popular Python ML library (e.g PyTorch, TF, JAX, HF transformers, etc)
|
||||
- Experience building LLM systems
|
||||
- Experience with post-training methods & implementing LLM papers
|
||||
- Comfortable in Unix environment + attendant command line tools (Git, Docker, etc)
|
||||
- Up to date on OS AI community & technologies
|
||||
- High cultural alignment with Plastic Labs' ethos
|
||||
- In NYC or willing to move to NYC
|
||||
- Complementary interest or experience specific to reinforcement learning, representation engineering, control vectors, prompt optimization, sparse auto-encoders, agentic frameworks, emergent behaviors, theory of mind, identity a plus
|
||||
- Complementary background in cognitive sciences (cs, linguistics, neuroscience, philosophy, & psychology) or other adjacent interdisciplinary fields a plus
|
||||
|
||||
## How to Apply
|
||||
Please send the following to research@plasticlabs.ai:
|
||||
- **Resume/CV** in whatever form it exists (PDF, LinkedIn, website, etc)
|
||||
- **Portfolio** of notable work (GitHub, pubs, ArXiv, blog, X, etc)
|
||||
- **Statement** of alignment specific to Plastic Labs--how do you identify with our mission, how can you contribute, etc? (points for brief, substantive, heterodox)
|
||||
|
||||
Applications without these 3 items won't be considered, but be sure to optimize for speed over perfection. If relevant, be sure to credit the LLM you used.
|
||||
|
||||
And it can't hurt to [join Discord](https://discord.gg/plasticlabs) and introduce yourself or engage with [our GitHub](https://github.com/plastic-labs).
|
||||
|
||||
## Research We're Excited About
|
||||
[s1: Simple test-time scaling](https://arxiv.org/abs/2501.19393)
|
||||
[Neural Networks Are Elastic Origami!](https://youtu.be/l3O2J3LMxqI?si=bhodv2c7GG75N2Ku)
|
||||
[Titans: Learning to Memorize at Test Time](https://arxiv.org/abs/2501.00663v1)
|
||||
[Mind Your Theory: Theory of Mind Goes Deeper Than Reasoning](https://arxiv.org/abs/2412.13631)
|
||||
[Generative Agent Simulations of 1,000 People](https://arxiv.org/abs/2411.10109)
|
||||
[DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning](https://arxiv.org/abs/2501.12948)
|
||||
[Multiagent Finetuning: Self Improvement with Diverse Reasoning Chains](https://arxiv.org/abs/2501.05707)
|
||||
[Prompt Programming for Large Language Models: Beyond the Few-Shot Paradigm](https://arxiv.org/pdf/2102.07350)
|
||||
[Theory of Mind May Have Spontaneously Emerged in Large Language Models](https://arxiv.org/pdf/2302.02083v3)
|
||||
[Think Twice: Perspective-Taking Improved Large Language Models' Theory-of-Mind Capabilities](https://arxiv.org/pdf/2311.10227)
|
||||
[Representation Engineering: A Top-Down Approach to AI Transparency](https://arxiv.org/abs/2310.01405)
|
||||
[Theia Vogel's post on Representation Engineering Mistral 7B an Acid Trip](https://vgel.me/posts/representation-engineering/)
|
||||
[A Roadmap to Pluralistic Alignment](https://arxiv.org/abs/2402.05070)
|
||||
[Open-Endedness is Essential for Artificial Superhuman Intelligence](https://arxiv.org/pdf/2406.04268)
|
||||
[Simulators](https://generative.ink/posts/simulators/)
|
||||
[Extended Mind Transformers](https://arxiv.org/pdf/2406.02332)
|
||||
[Violation of Expectation via Metacognitive Prompting Reduces Theory of Mind Prediction Error in Large Language Models](https://arxiv.org/abs/2310.06983)
|
||||
[Constitutional AI: Harmlessness from AI Feedback](https://arxiv.org/pdf/2212.08073)
|
||||
[Claude's Character](https://www.anthropic.com/research/claude-character)
|
||||
[Language Models Represent Space and Time](https://arxiv.org/pdf/2310.02207)
|
||||
[Generative Agents: Interactive Simulacra of Human Behavior](https://arxiv.org/abs/2304.03442)
|
||||
[Meta-Rewarding Language Models: Self-Improving Alignment with LLM-as-a-Meta-Judge](https://arxiv.org/abs/2407.19594)
|
||||
[Cyborgism](https://www.lesswrong.com/posts/bxt7uCiHam4QXrQAA/cyborgism)
|
||||
[Spontaneous Reward Hacking in Iterative Self-Refinement](https://arxiv.org/abs/2407.04549)
|
||||
[... accompanying twitter thread](https://x.com/JanePan_/status/1813208688343052639)
|
||||
|
||||
|
||||
(Back to [[Work at Plastic]])
|
||||
Loading…
Reference in New Issue
Block a user