All pages (in chronological order)


(논문 요약) Cognitive Behaviors that Enable Self-Improving Reasoners, or, Four Habits of Highly Effective STaRs

(논문 요약) Open-Reasoner-Zero: An Open Source Approach to Scaling Up Reinforcement Learning on the Base Model

(논문 요약) The Danger of Overthinking: Examining the Reasoning-Action Dilemma in Agentic Tasks

(논문 요약) NATURALREASONING: Reasoning in the Wild with 2.8M Challenging Questions

(논문 요약) Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention

(논문 요약) SWE-Lancer: Can Frontier LLMs Earn $1 Million from Real-World Freelance Software Engineering?

(블로그 요약) R1-reproduce

(개념 요약) Tokenization

(잡생각) 지능과 환경

(논문 요약) LIMO: Less Is More for Reasoning

(논문 요약) Gold-medalist Performance in Solving Olympiad Geometry with AlphaGeometry2

(논문 요약) SFT Memorizes, RL Generalizes: A Comparative Study of Foundation Model Post-training

(블로그 요약) Qwen2.5-VL: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning

(블로그 요약) DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning

(블로그 요약) Sky-T1: Train your own O1 preview model within $450

(논문 요약) rStar-Math: Small LLMs Can Master Math Reasoning with Self-Evolved Deep Thinking

(논문 요약) Can LLMs Design Good Questions Based on Context?

(잡생각) 탁월함

(논문 요약) Do NOT Think That Much for 2+3=? On the Overthinking of o1-Like LLMs

(논문 요약) DeepSeek-V3 Technical Report

(논문 요약) Large Concept Models: Language Modeling in a Sentence Representation Space

(논문 요약) THEAGENTCOMPANY: BENCHMARKING LLM AGENTS ON CONSEQUENTIAL REAL WORLD TASKS

(잡생각) 역사의 굴레

(잡생각) 인간과 기계

(잡생각) 프로 스포츠

(논문 요약) Byte Latent Transformer: Patches Scale Better Than Tokens

(논문 요약) CUT YOUR LOSSES IN LARGE-VOCABULARY LANGUAGE MODELS

(논문 요약) Training Large Language Models to Reason in a Continuous Latent Space

(잡생각) 지식

(논문 요약) Phi-4 Technical Report

(논문 요약) DOES RLHF SCALE? EXPLORING THE IMPACTS FROM DATA, MODEL, AND METHOD

(논문 요약) Clio: Privacy-Preserving Insights into Real-World AI Use

(논문 요약) PROCEDURAL KNOWLEDGE IN PRETRAINING DRIVES REASONING IN LARGE LANGUAGE MODELS

(논문 요약) Reverse Thinking Makes LLMs Stronger Reasoners

(논문 요약) OpenAI o1 System Card

(논문 요약) TULU 3: Pushing Frontiers in Open Language Model Post-Training

(논문 요약) Mixture-of-Transformers: A Sparse and Scalable Architecture for Multi-Modal Foundation Models

(논문 요약) LLaVA-o1: Let Vision Language Models Reason Step-by-Step

(논문 요약) Hunyuan-Large: An Open-Source MoE Model with 52 Billion Activated Parameters by Tencent

(논문 요약) Scaling Laws for Precision

(논문 요약) OPENCODER: THE OPEN COOKBOOK FOR TOP-TIER CODE LARGE LANGUAGE MODELS

(논문 요약) LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding

(잡생각) 목표의 최적화

(잡생각) 가치에 관하여

(논문 요약) Executable Code Actions Elicit Better LLM Agents

(논문 요약) ARITHMETIC WITHOUT ALGORITHMS: LANGUAGE MODELS SOLVE MATH WITH A BAG OF HEURISTICS

(논문 요약) AFLOW: AUTOMATING AGENTIC WORKFLOW GENERATION

(논문 요약) TOKEN MERGING: YOUR VIT BUT FASTER

(논문 요약) A COMPARATIVE STUDY ON REASONING PATTERNS OF OPENAI’S O1 MODEL

(모델 요약) GRANITE 3.0 LANGUAGE MODELS

(모델 요약) Aya Expanse: Connecting Our World

(논문 요약) CodeRetriever: Large-scale Contrastive Pre-training for Code Search

(모델 요약) Lightweight Llama Models

(논문 요약) BitNet: Scaling 1-bit Transformers for Large Language Models

(논문 요약) SPIRIT LM: Interleaved Spoken and Written Language Model

(논문 요약) Janus: Decoupling Visual Encoding for Unified Multimodal Understanding and Generation

(논문 요약) THINKING LLMS: GENERAL INSTRUCTION FOLLOWING WITH THOUGHT GENERATION

(논문 요약) Matryoshka Representation Learning

(논문 요약) jina-embeddings-v3: Multilingual Embeddings With Task LoRA

(논문 요약) DIFFERENTIAL TRANSFORMER

(논문 요약) Astute RAG: Overcoming Imperfect Retrieval Augmentation and Knowledge Conflicts for Large Language Models

(논문 요약) GSM-Symbolic: Understanding the Limitations of Mathematical Reasoning in Large Language Models

(논문 요약) TOOLGEN: UNIFIED TOOL RETRIEVAL AND CALLING VIA GENERATION

(논문 요약) MLE-BENCH: EVALUATING MACHINE LEARNING AGENTS ON MACHINE LEARNING ENGINEERING

(잡생각) 스토리

(논문 요약) Movie Gen: A Cast of Media Foundation Models

(논문 요약) Scaling Proprioceptive-Visual Learning with Heterogeneous Pre-trained Transformers

(논문 요약) One Initialization to Rule them All: Fine-tuning via Explained Variance Adaptation

(논문 요약) Not All LLM Reasoners Are Created Equal

(논문 요약) Archon: An Architecture Search Framework for Inference-Time Techniques

(잡생각) 피라미드 구조의 조직

(논문 요약) Molmo and PixMo: Open Weights and Open Data for State-of-the-Art Multimodal Models

(논문 요약) Logic-of-Thought: Injecting Logic into Contexts for Full Reasoning in Large Language Models

(논문 요약) LOTUS: Diffusion-based Visual Foundation Model for High-quality Dense Prediction

(논문 요약) Retrieval Augmented Generation (RAG) and Beyond: A Comprehensive Survey on How to Make your LLMs use External Data More Wisely

(논문 요약) GPT o1 검증

(Quant) 자산 배분

(Quant) 포트폴리오 구성

(Quant) 종목 선정

(잡생각) 자본과 미디어

(모델 요약) Llama 3.2: Revolutionizing edge AI and vision with open, customizable models

(논문 요약) Training Language Models to Self-Correct via Reinforcement Learning

(논문 요약) TORA: A TOOL-INTEGRATED REASONING AGENT FOR MATHEMATICAL PROBLEM SOLVING

(잡생각) 미래의 직업

(논문 요약) Qwen2.5-Coder Technical Report

(논문 요약) QWEN2 TECHNICAL REPORT

(잡생각) 종과 개체

(책 요약) Quantitative Finance with Python

(잡생각) Intelligence

(논문 요약) LLAMA-OMNI: SEAMLESS SPEECH INTERACTION WITH LARGE LANGUAGE MODELS

(논문 요약) AGENT WORKFLOW MEMORY

(논문 요약) Knowing When to Ask - Bridging Large Language Models and Data

(잡생각) statistical models

(잡생각) 좌절 관리 알고리즘

(논문 요약) Automatic Data Curation for Self-Supervised Learning: A Clustering-Based Approach

(논문 요약) Can LLMs Generate Novel Research Ideas? A Large-Scale Human Study with 100+ NLP Researchers

(논문 요약) Are Emergent Abilities of Large Language Models a Mirage?

(논문 요약) The Effects of Generative AI on High Skilled Work: Evidence from Three Field Experiments with Software Developers

(논문 요약) Strategic Chain-of-Thought: Guiding Accurate Reasoning in LLMs through Strategy Elicitation

(논문 요약) Open Mixture-of-Experts Language Models

(논문 요약) DeepSeek-V2: A Strong, Economical, and Efficient Mixture-of-Experts Language Model

(모델 요약) Yi-Coder

(모델 요약) Reflection Llama

(논문 요약) DIFFUSION MODELS ARE REAL-TIME GAME ENGINES

(논문 요약) Transfusion: Predict the Next Token and Diffuse Images with One Multi-Modal Model

(논문 요약) Smaller, Weaker, Yet Better: Training LLM Reasoners via Compute-Optimal Sampling

(논문 요약) To Code, or Not To Code? Exploring Impact of Code in Pre-training

(논문 요약) Speculative RAG: Enhancing Retrieval Augmented Generation through Drafting

(논문 요약) Scaling Laws for Data Filtering—Data Curation cannot be Compute Agnostic

(논문 요약) LLM Pruning and Distillation in Practice: The Minitron Approach

(논문 요약) Automated Design of Agentic Systems

(논문 요약) xGen-MM (BLIP-3): A Family of Open Large Multimodal Models

(논문 요약) MUTUAL REASONING MAKES SMALLER LLMS STRONGER PROBLEM-SOLVERS

(논문 요약) MEDICAL GRAPH RAG: TOWARDS SAFE MEDICAL LARGE LANGUAGE MODEL VIA GRAPH RETRIEVALAUGMENTED GENERATION

(논문 요약) Can Generalist Foundation Models Outcompete Special-Purpose Tuning? Case Study in Medicine

(논문 요약) Distributed Inference and Fine-tuning of Large Language Models Over The Internet

(논문 요약) Self-Taught Evaluators

(논문 요약) Scaling Exponents Across Parameterizations and Optimizers

(논문 요약) Large Language Monkeys: Scaling Inference Compute with Repeated Sampling

(논문 요약) REACT: SYNERGIZING REASONING AND ACTING IN LANGUAGE MODELS

(논문 요약) Stretching Each Dollar: Diffusion Training from Scratch on a Micro-Budget

(논문 요약) MindSearch: Mimicking Human Minds Elicits Deep AI Searcher

(논문 요약) Apple Intelligence Foundation Language Models

(논문 요약) SAM 2: Segment Anything in Images and Videos

(논문 요약) Segment Anything

(논문 요약) META-REWARDING LANGUAGE MODELS: Self-Improving Alignment with LLM-as-a-Meta-Judge

(논문 요약) LazyLLM: DYNAMIC TOKEN PRUNING FOR EFFICIENT LONG CONTEXT LLM INFERENCE

(논문 요약) MINT-1T: Scaling Open-Source Multimodal Data by 10x: A Multimodal Dataset with One Trillion Tokens

(논문 요약) Solving olympiad geometry without human demonstrations

(논문 요약) LEAN-GitHub: Compiling GitHub LEAN repositories for a versatile LEAN prover

(논문 요약) Weak-to-Strong Reasoning

(논문 요약) PROVER-VERIFIER GAMES IMPROVE LEGIBILITY OF LLM OUTPUTS

(논문 요약) The Llama 3 Herd of Models

(논문 요약) From GaLore to WeLore: How Low-Rank Weights Non-uniformly Emerge from Low-Rank Gradients

(논문 요약) Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients

(논문 요약) DataComp-LM: In search of the next generation of training sets for language models

(논문 요약) Florence-2: Advancing a Unified Representation for a Variety of Vision Tasks

(논문 요약) FlashAttention-3: Fast and Accurate Attention with Asynchrony and Low-precision

(논문 요약) RouteLLM: Learning to Route LLMs with Preference Data

(논문 요약) INTERNET OF AGENTS: WEAVING A WEB OF HETEROGENEOUS AGENTS FOR COLLABORATIVE INTELLIGENCE

(논문 요약) Planetarium: A Rigorous Benchmark for Translating Text to Structured Planning Languages

(논문 요약) DiPaCo: Distributed Path Composition

(논문 요약) Data curation via joint example selection further accelerates multimodal learning

(논문 요약) MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases

(논문 요약) APIGen: Automated PIpeline for Generating Verifiable and Diverse Function-Calling Datasets

(논문 요약) Searching for Best Practices in Retrieval-Augmented Generation

(논문 요약) AGENTLESS: Demystifying LLM-based Software Engineering Agents

(논문 요약) InternLM-XComposer-2.5: A Versatile Large Vision Language Model Supporting Long-Contextual Input and Output

(논문 요약) Gemma 2: Improving Open Language Models at a Practical Size

(논문 요약) Connecting the Dots: LLMs can Infer and Verbalize Latent Structure from Disparate Training Data

(논문 요약) WorkBench: a Benchmark Dataset for Agents in a Realistic Workplace Setting

(논문 요약) Meta Large Language Model Compiler: Foundation Models of Compiler Optimization

(논문 요약) TREE SEARCH FOR LANGUAGE MODEL AGENTS

(논문 요약) Debug like a Human: A Large Language Model Debugger via Verifying Runtime Execution Step by Step

(논문 요약) DeepSeek-Coder-V2: Breaking the Barrier of Closed-Source Models in Code Intelligence

(논문 요약) AGENTGYM: Evolving Large Language Model-based Agents across Diverse Environments

(논문 요약) MAGPIE: Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing

(논문 요약) Mixture-of-Agents Enhances Large Language Model Capabilities

(논문 요약) OpenVLA: An Open-Source Vision-Language-Action Model

(논문 요약) Show, Don’t Tell: Aligning Language Models with Demonstrated Feedback

(논문 요약) Towards Scalable Automated Alignment of LLMs: A Survey

(논문 요약) Buffer of Thoughts: Thought-Augmented Reasoning with Large Language Models

(논문 요약) SimPO: Simple Preference Optimization with a Reference-Free Reward

(논문 요약) Faithful Logical Reasoning via Symbolic Chain-of-Thought

(논문들 요약) Large Language Model Tuning

(논문 요약) Extreme Compression of Large Language Models via Additive Quantization

(논문 요약) QuIP#: Even Better LLM Quantization with Hadamard Incoherence and Lattice Codebooks

(논문 요약) DeepSeek-Prover: Advancing Theorem Proving in LLMs through Large-Scale Synthetic Data

(논문 요약) Layer-Condensed KV Cache for Efficient Inference of Large Language Models

(논문 요약) Scaling Monosemanticity: Extracting Interpretable Features from Claude 3 Sonnet

(코드 탐색) llama.cpp

(논문 요약) Granite Code Models: A Family of Open Foundation Models for Code Intelligence

(논문 요약) Does Fine-Tuning LLMs on New Knowledge Encourage Hallucinations?

(논문 요약) Chameleon: Mixed-Modal Early-Fusion Foundation Models

(논문 요약) What matters when building vision-language models?

(코드 실행) Alphacodium

(논문 요약) Code Generation with AlphaCodium: From Prompt Engineering to Flow Engineering

(논문 요약) Better & Faster Large Language Models via Multi-token Prediction

(논문 요약) LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale

(논문 요약) OBELICS: An Open Web-Scale Filtered Dataset of Interleaved Image-Text Documents

(논문 요약) SWE-AGENT: AGENT-COMPUTER INTERFACES ENABLE AUTOMATED SOFTWARE ENGINEERING

(논문 요약) Chat Vector: A Simple Approach to Equip LLMs with Instruction Following and Model Alignment in New Languages

(논문 요약) AgentCoder: Multiagent-Code Generation with Iterative Testing and Optimisation

(논문 요약) NExT: Teaching Large Language Models to Reason about Code Execution

(논문 요약) Make Your LLM Fully Utilize the Context

(논문 요약) Phi-3 Technical Report:A Highly Capable Language Model Locally on Your Phone

(데이터 요약) common crawl filtered data

(논문 요약) One Embedder, Any Task: Instruction-Finetuned Text Embeddings

(모델 요약) LLaMa3

(논문 요약) CodeGemma: Open Code Models Based on Gemma

(논문 요약) RecurrentGemma: Moving Past Transformers for Efficient Open Language Models

(논문 요약) Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention

(논문 요약) Towards General Computer Control: A Multimodal Agent for Red Dead Redemption II as a Case Study

(논문 요약) Integrating Code Generation with Execution and Refinement

(논문 요약) Gecko: Versatile Text Embeddings Distilled from Large Language Models

(논문 요약) Gorilla: Large Language Model Connected with Massive APIs

(논문 요약) The Unreasonable Ineffectiveness of the Deeper Layers

(논문 요약) SWE-BENCH: CAN LANGUAGE MODELS RESOLVE REAL-WORLD GITHUB ISSUES?

(논문 요약) GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection

(논문 요약) RAFT: Adapting Language Model to Domain Specific RAG

(논문 요약) LONG-FORM FACTUALITY IN LARGE LANGUAGE MODELS

(논문 요약) MTEB: Massive Text Embedding Benchmark

(논문 요약) Chart-based Reasoning: Transferring Capabilities from LLMs to VLMs

(논문 요약) Generative Representational Instruction Tuning

(논문 요약) Text Embeddings by Weakly-Supervised Contrastive Pre-training

(논문 요약) Branch-Train-MiX: Mixing Expert LLMs into a Mixture-of-Experts LLM

(논문 요약) Direct Preference Optimization: Your Language Model is Secretly a Reward Model

(논문 요약) RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Horizon Generation

(논문 요약) Quiet-STaR: Language Models Can Teach Themselves to Think Before Speaking

(논문 요약) STaR: Self-Taught Reasoner Bootstrapping Reasoning With Reasoning

(논문 요약) SceneScript: Reconstructing Scenes With An Autoregressive Structured Language Model

(논문 요약) DEMYSTIFYING EMBEDDING SPACES USING LARGE LANGUAGE MODELS

(논문 요약) Stop Regressing: Training Value Functions via Classification for Scalable Deep RL

(논문 요약) Scaling Instructable Agents Across Many Simulated Worlds

(논문 요약) Simple and Scalable Strategies to Continually Pre-train Large Language Models

(논문 요약) Efficient Tool Use with Chain-of-Abstraction Reasoning

(논문 요약) Beyond A*: Better Planning with Transformers via Search Dynamics Bootstrapping

(논문 요약) InternLM-Math: Open Math Large Language Models Toward Verifiable Reasoning

(논문 요약) Grounded Language-Image Pre-training

(코드 실행) AnyMAL: An Efficient and Scalable Any-Modality Augmented Language Model

(논문 요약) Textbooks Are All You Need

(논문 요약) Generating Diverse High-Fidelity Images with VQ-VAE-2

(논문 요약) Neural Discrete Representation Learning

(논문 요약) ROFORMER: ENHANCED TRANSFORMER WITH ROTARY POSITION EMBEDDING

FED 금리 수익률 곡선

Undistort Image with OpenCV-Python

Top-view perspective transform with OpenCV-Python

Rotation matrix with Quaternion

Visualize a point cloud with Open3D