Portrait of Junyang Lin

Junyang Lin

Independent researcher

I work on language models, multimodal systems, and agentic learning.

My recent work focuses on how models reason, use tools, learn across modalities, and improve through post-training and interaction with environments.

About

About

I am an individual researcher working on large language models, multimodal systems, and agentic learning.

I care about a simple question: how do we move from models that generate convincing outputs to systems that can reason, act, and improve through interaction with the world? That question sits behind my interest in long-horizon planning, tool use, multimodal representation, and training methods that make model behavior more grounded.

My recent work spans agentic coding, multimodal modeling, reinforcement-learning methods, and efficient attention design. Across these projects, I keep returning to the same underlying theme: useful intelligence depends not only on model scale, but also on how architectures, environments, and training pipelines are designed together.

If you want a quick overview, start with the selected papers below or the essays in the blog.

Research Themes

Research

01

How can models plan and act instead of only answer?

I care about agents that can decide when to think, when to call tools, and how to revise a plan after seeing feedback from an environment. The core shift is from static reasoning to systems that sustain useful action over longer horizons.

02

How do language and vision become one working system?

Multimodal modeling matters when a system needs to read, inspect, and respond to the world rather than stay inside text alone. I am interested in model families that share representations across modalities and make interaction more grounded.

03

What training stack actually supports stronger post-training?

Better model behavior depends on more than prompts or scale. I am interested in reinforcement-learning algorithms, attention mechanisms, and infrastructure choices that make large-scale post-training stable, efficient, and aligned with real capability gains.

Selected Work

Selected work

Full publication list
2025 · Technical Report

Qwen3 Technical Report

System-level report on a Qwen generation shaped by large-scale language modeling and post-training.

2025 · Technical Report

Qwen3-VL Technical Report

Report on a multimodal Qwen model family designed to operate across language and vision.

2025 · Agentic Systems

Qwen3-Coder: Agentic Coding in the World

Work on coding agents that interact with tools and external environments instead of generating code in isolation.

2025 · Reinforcement Learning

Group Sequence Policy Optimization

Research on sequence-level policy optimization for reinforcement-learning settings in large model training.

2025 · Model Architecture

Gated Attention for Large Language Models

Study of attention designs aimed at non-linearity, sparsity, and attention-sink-free behavior in language models.

2025 · Technical Report

Qwen-Image Technical Report

Report on a Qwen image-generation effort, extending the model family beyond text and toward visual synthesis.

Writing

Writing

All essays

Contact

Contact

If your interests overlap with agentic systems, multimodal learning, reinforcement-learning infrastructure, or the broader question of how foundation models become more grounded and useful, feel free to reach out.