- formatting
- images
- links
- math
- code
- blockquotes
- external-services
•
•
•
•
•
•
-
LoRA on the Go: Instance-level Dynamic LoRA Selection and Merging
-
Loong: Synthesize Long Chain-of-Thoughts at Scale through Verifiers
-
LLM$ imes$MapReduce-V3: Enabling Interactive In-Depth Survey Generation through a MCP-Driven Hierarchically Modular Agent System
-
LLMs Encode How Difficult Problems Are
-
LLM-guided Hierarchical Retrieval
-
LLM-ERM: Sample-Efficient Program Learning via LLM-Guided Search
-
LLM-empowered knowledge graph construction: A survey
-
LLM-as-a-Judge: Toward World Models for Slate Recommendation Systems
-
Llama Guard: LLM-based Input-Output Safeguard for Human-AI Conversations
-
LLaMA-Adapter: Efficient Fine-tuning of Language Models with Zero-init Attention
-
Llama 2: Open Foundation and Fine-Tuned Chat Models
-
LiveThinking: Enabling Real-Time Efficient Reasoning for AI-Powered Livestreaming via Reinforcement Learning
-
LiveCodeBench: Holistic and Contamination Free Evaluation of Large Language Models for Code
-
LIMO: Less is More for Reasoning
-
Limits of trust in medical AI
-
LIMI: Less is More for Agency
-
LIME: Making LLM Data More Efficient with Linguistic Metadata Embeddings
-
Let's Verify Step by Step
-
Less LLM, More Documents: Searching for Improved RAG
-
Less is More Tokens: Efficient Math Reasoning via Difficulty-Aware Chain-of-Thought Distillation
-
Learning When to Plan: Efficiently Allocating Test-Time Compute for LLM Agents
-
Learning to Reason: Training LLMs with GPT-OSS or DeepSeek R1 Reasoning Traces
-
Learning to Focus: Focal Attention for Selective and Scalable Transformers
-
Learning on the Job: An Experience-Driven Self-Evolving Agent for Long-Horizon Tasks
-
Learning Fine-Grained Bimanual Manipulation with Low-Cost Hardware
-
Latent Traits and Cross-Task Transfer: Deconstructing Dataset Interactions in LLM Fine-tuning
-
Latent learning: episodic memory complements parametric learning by enabling flexible reuse of experiences
-
Larger Datasets Can Be Repeated More: A Theoretical Analysis of Multi-Epoch Scaling in Linear Regression
-
Large Language Monkeys: Scaling Inference Compute with Repeated Sampling
-
Large Language Models Meet Virtual Cell: A Survey
-
Large Language Model Sourcing: A Survey
-
Language Self-Play For Data-Free Training
-
Language models as tools for investigating the distinction between possible and impossible natural languages
-
KTO: Model Alignment as Prospect Theoretic Optimization
-
Knowledge-tuning Large Language Models with Structured Medical Knowledge Bases for Reliable Response Generation in Chinese
-
Kimi Linear: An Expressive, Efficient Attention Architecture
-
Kimi k1.5: Scaling Reinforcement Learning with LLMs
-
KCM: KAN-Based Collaboration Models Enhance Pretrained Large Models
-
KAN: Kolmogorov-Arnold Networks
-
Judging LLM-as-a-Judge with MT-Bench and Chatbot Arena
-
Jailbroken: How Does LLM Safety Training Fail?
-
Jailbreaking Black Box Large Language Models in Twenty Queries
-
iTransformer: Inverted Transformers Are Effective for Time Series Forecasting
-
Is Your Code Generated by ChatGPT Really Correct? Rigorous Evaluation of Large Language Models for Code Generation
-
Is ChatGPT a General-Purpose Natural Language Processing Task Solver?
-
Introduction to Machine Learning
-
Internalizing World Models via Self-Play Finetuning for Agentic RL
-
Inter-Agent Trust Models: A Comparative Study of Brief, Claim, Proof, Stake, Reputation and Constraint in Agentic Web Protocol Design-A2A, AP2, ERC-8004, and Beyond
-
Intelligence per Watt: Measuring Intelligence Efficiency of Local AI
-
InstructBLIP: Towards General-purpose Vision-Language Models with Instruction Tuning