Chenyang An

AI Research Scientist @ Miromind

prof_pic.jpg

New York City

email: cya.portfolio at gmail dot com

I am an LLM researcher and AI agent engineer focused on post-training, data curation, and reasoning in both natural and formal languages. I’m an AI Research Scientist at Miromind. Before that, I was an Applied Scientist at Amazon AWS, and have interned at Microsoft Research, Scale AI, and Amazon.

My PhD research centers on improving LLM reasoning through post-training. I have published at ACL, ICLR, and EMNLP, with work on curating higher-quality reasoning data via trial-and-error trajectories, studying LLM generalization through internal representations, and designing diversity-aware reward mechanisms for mathematical reasoning.

I build AI agents across low-latency, medium-latency, and long-horizon settings, focusing on reasoning, verification, and trading.

Long-horizon AI systems: I developed QED, an open-source agent for mathematical discovery that solved a nontrivial open PDE research problem by automatically searching for a valid solution and proving correctness end-to-end (QED is currently solving other open math research questions, more to be announced). I also built supporting tools for mathematical workflows, including Proofread, an AI agent for proofreading technical papers and books, and pdf-to-Lean, which converts mathematical documents into formal proofs.

Medium-latency AI systems: I work with my Amazon colleagues to build agentic systems that translate natural language into formal specifications and executable logic, and use symbolic verification methods to ensure correctness, reduce hallucination, and improve logical consistency in LLM outputs.

Low-latency AI systems: I build open source AI agents for real-time financial decision-making, including systems that process live news, generate signal, and execute trades within 1.5–2 seconds across stocks and options.

[Resume]

If you are interested in any of the topics above, feel free to drop me an email!

news

Apr 10, 2026 I open-sourced QED, a multi-agent pipeline that transforms mathematical problem statements into rigorous proofs. QED solved a research-level open problem in PDEs, with the proof verified by domain experts from three institutions and incorporated into their mathematical work.
Feb 25, 2026 I’m happy to release an agent pipeline based on Claude Code that help proofread latex source code of paper and books! Check https://github.com/chenyang-an/proofread for details!
Aug 30, 2025 I’m excited to share that I will join Amazon AWS Automated Reasoning Group as an Applied Scientist!
Apr 01, 2025 I’m excited to share that our paper, “The Price of Format: Diversity Collapse in LLMs”, has been accepted to Empirical Methods in Natural Language Processing 2025 (EMNLP)! In this work, we find out that structured templates in instruction-tuned LLMs cause diversity collapse—limiting open-ended generation—even under high-temperature sampling, and systematically evaluated this effect across tasks to show the trade-off between alignment, task performance, and output diversity.
Apr 01, 2025 I’m excited to share that I’ll be joining Amazon AWS Neurosymbolic as an Applied Scientist Intern, where I’ll be working on LLM reasoning in both natural and formal language!

latest posts

selected publications

  1. Arxiv
    The Price of Format: Diversity Collapse in LLMs
    May 2025
  2. Arxiv
    Linear Correlation in LM’s Compositional Generalization and Hallucination
    Feb 2025
  3. Arxiv
    Next-Token Prediction Task Assumes Optimal Data Ordering for LLM Training in Proof Generation
    Oct 2024
  4. ICLR
    Correlation and Navigation in the Vocabulary Key Representation Space of Language Models
    Jan 2025
  5. ACL
    Learn from Failure: Fine-Tuning LLMs with Trial-and-Error Data for Intuitionistic Propositional Logic Proving
    May 2024