-
Beyond A*: Better Planning with Transformers via Search Dynamics Bootstrapping
Paper • 2402.14083 • Published • 47 -
Linear Transformers are Versatile In-Context Learners
Paper • 2402.14180 • Published • 7 -
Training-Free Long-Context Scaling of Large Language Models
Paper • 2402.17463 • Published • 24 -
The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits
Paper • 2402.17764 • Published • 628
Yang Lee
innovation64
AI & ML interests
AGI
Recent Activity
upvoted a paper about 11 hours ago
ClawBench: Can AI Agents Complete Everyday Online Tasks? upvoted a paper about 11 hours ago
SkillClaw: Let Skills Evolve Collectively with Agentic Evolver upvoted a paper about 13 hours ago
Matrix-Game 3.0: Real-Time and Streaming Interactive World Model with Long-Horizon Memory