Top Important LLM Papers for the Week from 05/02 to 11/02
Stay Updated with Recent Large Language Models Research
Large language models (LLMs) have advanced rapidly in recent years. As new generations of models are developed, researchers and engineers need to stay informed on the latest progress. This article summarizes some of the most important LLM papers published during the Second Week of February 2024.
The papers cover various topics shaping the next generation of language models, from model optimization and scaling to reasoning, benchmarking, and enhancing performance. Keeping up with novel LLM research across these domains will help guide continued progress toward models that are more capable, robust, and aligned with human values.
Table of Contents:
LLM Progress & Benchmarking
LLM Reasoning
LLM Training & Evaluation
Transformers & Attention Based Models
My E-book: Data Science Portfolio for Success Is Out!
I recently published my first e-book Data Science Portfolio for Success which is a practical guide on how to build your data science portfolio. The book covers the following topics: The Importance of Having a Portfolio as a Data Scientist How to Build a Data Science Portfolio That Will Land You a Job?
1. LLM Progress & Benchmarking
StepCoder: Improve Code Generation with Reinforcement Learning from Compiler Feedback
PokéLLMon: A Human-Parity Agent for Pokémon Battles with Large Language Models
TravelPlanner: A Benchmark for Real-World Planning with Language Agents
Nomic Embed: Training a Reproducible Long Context Text Embedder
OpenMoE: An Early Effort on Open Mixture-of-Experts Language Models
Can Mamba Learn How to Learn? A Comparative Study on In-Context Learning Tasks
Scaling Laws for Downstream Task Performance of Large Language Models
CodeIt: Self-Improving Language Models with Prioritized Hindsight Replay
Tag-LLM: Repurposing General-Purpose LLMs for Specialized Domains
Memory Consolidation Enables Long-Context Video Understanding
Driving Everywhere with Large Language Model Policy Adaptation
WebLINX: Real-World Website Navigation with Multi-Turn Dialogue
SPHINX-X: Scaling Data and Parameters for a Family of Multi-modal Large Language Models
2. LLM Reasoning
DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models
Self-Discover: Large Language Models Self-Compose Reasoning Structures
3. LLM Training, Inference, Evaluation & OptimizationÂ
Specialized Language Models with Cheap Inference from Limited Domain Data
Rethinking Interpretability in the Era of Large Language Models
Rethinking Optimization and Architecture for Tiny Language Models
LiPO: Listwise Preference Optimization through Learning-to-Rank
Shortened LLaMA: A Simple Depth Pruning for Large Language Models
BiLLM: Pushing the Limit of Post-Training Quantization for LLMs
Hydrogen: High-Throughput LLM Inference with Shared Prefixes
Offline Actor-Critic Reinforcement Learning Scales to Large Models
4. Transformers & Attention Based Models
Repeat After Me: Transformers are Better than State Space Models at Copying
Progressive Gradient Flow for Robust N:M Sparsity Training in Transformers
The Hedgehog & the Porcupine: Expressive Linear Attentions with Softmax Mimicry
5. LLM Fine-TuningÂ
Are you looking to start a career in data science and AI and do not know how? I offer data science mentoring sessions and long-term career mentoring:
Mentoring sessions: https://lnkd.in/dXeg3KPW
Long-term mentoring: https://lnkd.in/dtdUYBrM