Xiaoze Liu Xiaoze Liu
Blog
LinkedIn (opens new window)
Scholar (opens new window)
GitHub (opens new window)
Twitter (opens new window)
Blog
LinkedIn (opens new window)
Scholar (opens new window)
GitHub (opens new window)
Twitter (opens new window)
  • Posts

    • Blog
    • On cross-model-family collaboration
      • Training-level: Mutual Reinforcement Learning
      • Merging-level: The Trojan in the Vocabulary
      • Agentic-pipeline-level: The Vision Wormhole
      • What ties them together
    • Mutual Reinforcement Learning: Sharing experience across heterogeneous LLM policies
      • The hard problem
      • What we built (the substrate)
      • Three probes against GRPO
      • The trade-off, sketched
      • What we found
      • Where this sits
      • BibTeX
    • A note on copyright protection in the LLM era
      • What got me started
      • SHIELD: detect the problem, then defend at inference
      • Why fixing open-weight models is nasty
      • SUV's reframing
      • Why the two papers together
      • References
      • BibTeX
    • Both sides of LLM peer review at ICML 2026
      • Why I went in pro-LLM
      • What LLMs were actually useful for
      • Then I got a paper rejected by what was clearly the other use case
      • The asymmetry I keep coming back to
      • Mean Reviewer
      • Where I land now
    • The Vision Wormhole: Latent-space communication for heterogeneous multi-agent systems
      • Latent communication is having a moment, but it stays inside one family
      • What we do
      • What we found
      • Where this sits
      • What I'm less sure about
      • BibTeX
    • Don't beg. Earn it. On a line from Eureka Seven
      • The Linguistic Anatomy of the Phrase
      • A Structural Echo
      • The Evolution of "Winning"
      • A Philosophy for Action
    • The Trojan in the Vocabulary: How a single token can sabotage LLM composition
      • Why I went looking for this
      • The attack, in one paragraph
      • What surprised me in the results
      • What I want you to take away
      • Where this sits
      • Limitations and ethics
      • BibTeX

# Blog

Occasional notes on research, ideas, and things I find worth writing down.

  • On cross-model-family collaboration
  • Mutual Reinforcement Learning: Sharing experience across heterogeneous LLM policies
  • A note on copyright protection in the LLM era
  • Both sides of LLM peer review at ICML 2026
  • The Vision Wormhole: Latent-space communication for heterogeneous multi-agent systems
  • Don't beg. Earn it. On a line from Eureka Seven
  • The Trojan in the Vocabulary: How a single token can sabotage LLM composition

On cross-model-family collaboration →