Top Important LLM Papers for the Week from 18/12 to 24/12
Stay Updated with Recent Large Language Models Research
Every week, several top-tier academic conferences and journals showcased innovative research in computer vision, presenting exciting breakthroughs in various subfields such as image recognition, vision model optimization, generative adversarial networks (GANs), image segmentation, video analysis, and more.
This article provides a comprehensive overview of the most significant papers published in the third week of December 2023, highlighting the latest research and advancements in computer vision. Whether you’re a researcher, practitioner, or enthusiast, this article will provide valuable insights into the state-of-the-art techniques and tools in computer vision.
Table of Contents:
LLM Progress & Benchmarking
LLM Fine Tuning
LLM Reasoning
LLM Training & Evaluation
Responsible AI & LLM Ethics
Transformers & Attention Models
My E-book: Data Science Portfolio for Success Is Out!
I recently published my first e-book Data Science Portfolio for Success which is a practical guide on how to build your data science portfolio. The book covers the following topics: The Importance of Having a Portfolio as a Data Scientist How to Build a Data Science Portfolio That Will Land You a Job?
1. LLM Progress & Benchmarking
Amphion: An Open-Source Audio, Music, and Speech Generation Toolkit
Extending Context Window of Large Language Models via Semantic Compression
M3DBench: Let’s Instruct Large Models with Multi-modal 3D Prompts
Mini-GPTs: Efficient Large Language Models through Contextual Pruning
LLM in a flash: Efficient Large Language Model Inference with Limited Memory
PowerInfer: Fast Large Language Model Serving with a Consumer-grade GPU
Topic-VQ-VAE: Leveraging Latent Codebooks for Flexible Topic-Guided Document Generation
Faithful Persona-based Conversational Dataset Generation with Large Language Models
2. LLM Fine Tuning
3. LLM Reasoning
4. LLM Training & Evaluation
Weak-to-Strong Generalization: Eliciting Strong Capabilities With Weak Supervision
Self-Evaluation Improves Selective Generation in Large Language Models
G-LLaVA: Solving Geometric Problem with Multi-Modal Large Language Model
Catwalk: A Unified Language Model Evaluation Framework for Many Datasets
5. Transformers & Attention Models
Weight subcloning: direct initialization of transformers using larger pretrained ones
Cached Transformers: Improving Transformers with Differentiable Memory Cache
Are you looking to start a career in data science and AI and do not know how? I offer data science mentoring sessions and long-term career mentoring:
Mentoring sessions: https://lnkd.in/dXeg3KPW
Long-term mentoring: https://lnkd.in/dtdUYBrM