Agency science in LLMs
I am currently looking for collaborators (or students) for several ongoing projects:
Mechanistic interpretability in DeepRL and LLMs
Agency foundation research on a variety of topics. See here for several topics on game theory, RL/IRL, and conceptual research.
Funding may be available from AI safety/alignment organizations. Contact me to chat more: cm5635(at)nyu(dot).edu(ignore), or find me on the AI alignment slack channel.