## 🔍 Key Insights AI engineering is rapidly advancing along two parallel tracks: **standardization of Agent architectures** and **fine-grained evaluation of model capabilities**. Frameworks like **OpenClaw** and **Learn Claude Code** continue to strengthen the foundational practices of Agent engineering, while CMU’s **DIAGRAMMA benchmark** has, for the first time, quantitatively exposed systemic weaknesses in mainstream models’ scientific chart comprehension—**top-performing models such as GPT-4o achieve only 59.64% accuracy** [4]. Concurrently, **Kimi’s Attention Residuals** and **Beihang University’s InCoder-32B** deliver critical breakthroughs—one in low-level architecture design, the other in industrial-scale code modeling [7][8]. ## 🚀 Key Updates - **In-depth analysis of the OpenClaw Agent framework’s Workspace** [5]: A systematic breakdown of core configuration files (e.g., `AGENTS.md`, `SOUL.md`) clarifies their responsibilities and establishes a practical, “truly usable” configuration paradigm for Agent engineering - **Launch of the Learn Claude Code tutorial** [2]: Focuses on real-world AI Agent engineering implementation, delivering a complete methodology—from design principles to reusable architectural patterns - **Release of CMU’s DIAGRAMMA benchmark results** [4]: GPT-4o, Claude, and Gemini all fall short, revealing a fundamental bottleneck in scientific chart understanding; the highest score achieved is just 59.64% - **Kimi introduces Attention Residuals**, a novel architecture [7]: Replaces conventional residual connections with depth-wise attention mechanisms, enabling on-demand cross-layer information retrieval and aggregation - **Beihang University releases InCoder-32B**, an industrial-grade code foundation model [8]: The first 32B-parameter code model tailored for chip design, GPU optimization, and similar domains—trained on 2.5 million simulation-and-verification data samples - **daVinci-Env open-sources the OpenSWE training framework** [9]: The largest transparent SWE Agent training environment to date, featuring 45,320 executable Docker environments and over 128,000 open-source code repositories - **Peking University’s Peng Yuxin team proposes TARA** [10]: Infuses biological taxonomy tree priors into multimodal large models to resolve logical consistency and zero-shot generalization challenges in hierarchical recognition - **Top 10 Agent Skills for frontend, product, and UI practitioners** [3]: Curated list of highly reliable skill tools from OpenAI, Anthropic, Vercel, and others—with scenario-based selection guidance ## 🔗 Sources [1] What You Didn’t Know About Agents: Principles, Architectures, and Engineering Practices — Tw93 — https://www.bestblogs.dev/article/58852dc5 [2] Learn Claude Code Tutorial: A Practical Guide to AI Agent Engineering — https://www.bestblogs.dev/status/2035338785668653363 [3] Top 10 Recommended Agent Skills for Frontend, Product, and UI Practitioners — https://www.bestblogs.dev/status/2035316234271764654 [4] AI Models Fail to Interpret Basic Charts from High-School Textbooks: CMU’s DIAGRAMMA Benchmark Reveals Critical Deficiencies — https://www.bestblogs.dev/status/2035315182755578061 [5] A Deep Dive into OpenClaw🦞: Crossing the Threshold from “Functional” to “Truly Usable”—Workspace Explained in Detail — https://www.bestblogs.dev/article/0