Skip to content

weAIDB/awesome-data-llm

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 

Repository files navigation

A Survey of LLM × DATA

A collection of papers and projects related to LLMs and corresponding data-centric methods. arXiv

Other publicly-available materials: [Slide]

If you find our survey useful, please cite the paper:

@article{LLMDATASurvey,
    title={A Survey of LLM × DATA},
    author={Xuanhe Zhou, Junxuan He, Wei Zhou, Haodong Chen, Zirui Tang, Haoyu Zhao, Xin Tong, Guoliang Li, Youmin Chen, Jun Zhou, Zhaojun Sun, Binyuan Hui, Shuo Wang, Conghui He, Zhiyuan Liu, Jingren Zhou, Fan Wu},
    year={2025},
    journal={arXiv preprint arXiv:2505.18458},
    url={https://arxiv.org/abs/2505.18458}
}
@article{tangllmasanalyst,
    title={LLM/Agent-as-Data-Analyst: A Survey},
    author={Zirui Tang, Weizheng Wang, Zihang Zhou, Yang Jiao, Bangrui Xu, Boyu Niu, Xuanhe Zhou, Guoliang Li, Yeye He, Wei Zhou, Yitong Song, Cheng Tan, Bin Wang, Conghui He, Xiaoyang Wang, Fan Wu},
    year={2025},
    journal={arXiv preprint arXiv:2509.23988},
    url={https://arxiv.org/abs/2509.23988}
}

🌤 The IaaS Concept of DATA4LLM

The IaaS concept for LLM data (phonetically echoing Infrastructure as a Service) defines the characteristics of high-quality datasets along four key dimensions: (1) Inclusiveness ensures broad coverage across domains, tasks, sources, languages, styles, and modalities. (2) Abundance emphasizes sufficient and well-balanced data volume to support scaling, fine-tuning, and continual learning without overfitting. (3) Articulation requires clear, coherent, and instructive content with step-by-step reasoning to enhance model understanding and task performance. (4) Sanitization involves rigorous filtering to remove private, toxic, unethical, and misleading content, ensuring data safety, neutrality, and compliance. arXiv

Cover

🌟 LLM/Agent-as-Data-Analyst

We observe the evolution of LLM/Agent-as-Data-Analyst techniques follows a five-dimension trajectory: (1) Data Modality (homogeneous → heterogeneous); (2) Analysis Functionality (literal → semantic); (3) Knowledge Scope (closed-world →open-world); (4) Tool Integration (tool-coupled → tool-assisted); (5) Development Autonomy (manual → fully autonomous).

Cover2

Table of Contents

Datasets

  1. CommonCrawl: A massive web crawl dataset covering diverse languages and domains; widely used for LLM pretraining. [Source]

  2. The Stack: A large-scale dataset of permissively licensed source code in multiple programming languages; used for code LLMs. [HuggingFace]

  3. RedPajama: A replication of LLaMA’s training data recipe with open datasets; spans web, books, arXiv, and more. [Github]

  4. SlimPajama-627B-DC: A deduplicated and filtered subset of RedPajama (627B tokens); optimized for clean and efficient training. [HuggingFace]

  5. Alpaca-CoT: Instruction-following dataset enhanced with Chain-of-Thought (CoT) reasoning prompts; used for dialogue fine-tuning. [Github]

  6. LLaVA-Pretrain: A multimodal dataset with image-text pairs for training visual language models like LLaVA. [HuggingFace]

  7. Wikipedia: Structured and encyclopedic content; a foundational source for general-purpose language models. [HuggingFace]

  8. C4: A cleaned version of CommonCrawl data, widely used in models like T5 for high-quality web text. [HuggingFace]

  9. BookCorpus: Contains free fiction books; often used to teach models long-form language understanding. [HuggingFace]

  10. Arxiv: Scientific paper corpus from arXiv, covering physics, math, CS, and more; useful for academic language modeling. [HuggingFace]

  11. PubMed: Biomedical literature dataset from the PubMed database; key resource for medical domain models. [Source]

  12. StackExchange: Community Q&A data covering domains like programming, math, philosophy, etc.; useful for QA and dialogue tasks. [Source]

  13. OpenWebText2: A high-quality open-source web text dataset based on URLs commonly cited on Reddit; GPT-style training corpus. [Source]

  14. OpenWebMath: A dataset of math questions and answers; designed to improve mathematical reasoning in LLMs. [HuggingFace]

  15. Falcon-RefinedWeb: Filtered web data used in training Falcon models; emphasizes data quality through rigorous preprocessing. [HuggingFace]

  16. CCI 3.0: A large-scale multi-domain Chinese web corpus, suitable for training high-quality Chinese LLMs. [HuggingFace]

  17. OmniCorpus: A unified multimodal dataset (text, image, audio) designed for general-purpose AI training. [Github]

  18. WanJuan3.0: A diverse and large-scale Chinese dataset including news, fiction, QA, and more; released by OpenDataLab. [Source]

0 Data Characteristics across LLM Stages

⬆️top

Data for Pretraining

  1. OBELICS: An Open Web-Scale Filtered Dataset of Interleaved Image-Text Documents
    Hugo Laurençon, Lucile Saulnier, Léo Tronchon, et al. NeurIPS 2023. [Paper]
  2. Aligning Books and Movies: Towards Story-like Visual Explanations by Watching Movies and Reading Books
    Yukun Zhu, Ryan Kiros, Richard Zemel, et al. ICCV 2015. [Paper]

Data for Continual Pre-training

  1. MedicalGPT: Training Medical GPT Model
    Ming Xu. [Github]
  2. BBT-Fin: Comprehensive Construction of Chinese Financial Domain Pre-trained Language Model, Corpus and Benchmark
    Dakuan Lu, Hengkui Wu, Jiaqing Liang, et al. arXiv 2023. [Paper]

Data for Supervised Fine-Tuning (SFT)

General Instruction Following

  1. Free dolly: Introducing the world’s first truly open instruction-tuned llm
    Mike Conover, Matt Hayes, Ankit Mathur, et al. 2023. [Source]

Specific Domain Usage

  1. MedicalGPT: Training Medical GPT Model [Github]
  2. DISC-LawLLM: Fine-tuning Large Language Models for Intelligent Legal Services
    Shengbin Yue, Wei Chen, Siyuan Wang, et al. arXiv 2023. [Paper]

Data for Reinforcement Learning (RL)

RLHF

  1. MedicalGPT: Training Medical GPT Model [Github]
  2. UltraFeedback: Boosting Language Models with Scaled AI Feedback
    Ganqu Cui, Lifan Yuan, Ning Ding, et al. ICML 2024. [Paper]

RoRL

  1. DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning
    DeepSeek-AI. arXiv 2025. [Paper]
  2. Kimi k1.5: Scaling Reinforcement Learning with LLMs
    Kimi Team. arXiv 2025. [Paper]

Data for Retrieval-Augmented Generation (RAG)

  1. DH-RAG: A Dynamic Historical Context-Powered Retrieval-Augmented Generation Method for Multi-Turn Dialogue
    Feiyuan Zhang, Dezhi Zhu, James Ming, et al. arXiv 2025. [Paper]
  2. Medical Graph RAG: Towards Safe Medical Large Language Model via Graph Retrieval-Augmented Generation
    Junde Wu, Jiayuan Zhu, Yunli Qi, et al. arXiv 2024. [Paper]
  3. ERAGent: Enhancing Retrieval-Augmented Language Models with Improved Accuracy, Efficiency, and Personalization
    Yunxiao Shi, Xing Zi, Zijing Shi, et al. arXiv 2024. [Paper]
  4. PersonaRAG: Enhancing Retrieval-Augmented Generation Systems with User-Centric Agents
    Saber Zerhoudi, Michael Granitzer. arXiv 2024. [Paper]
  5. DISC-LawLLM: Fine-tuning Large Language Models for Intelligent Legal Services [Paper]

Data for LLM Evaluation

  1. MMMU: A Massive Multi-discipline Multimodal Understanding and Reasoning Benchmark for Expert AGI
    Xiang Yue, Yuansheng Ni, Kai Zhang, et al. CVPR 2024. [Paper]
  2. LexEval: A Comprehensive Chinese Legal Benchmark for Evaluating Large Language Models
    Haitao Li, You Chen, Qingyao Ai, et al. NeurIPS 2024. [Paper]
  3. What disease does this patient have? a large-scale open domain question answering dataset from medical exams
    Di Jin, Eileen Pan, Nassim Oufattole, et al. AAAI 2021. [Paper]
  4. Evaluating Large Language Models Trained on Code
    Mark Chen, Jerry Tworek, Heewoo Jun, et al. arXiv 2021. [Paper]

Data for LLM Agents

  1. STeCa: Step-level Trajectory Calibration for LLM Agent Learning
    Hanlin Wang, Jian Wang, Chak Tou Leong, Wenjie Li. arXiv 2025. [Paper]
  2. Large Language Model-Based Agents for Software Engineering: A Survey
    Junwei Liu, Kaixin Wang, Yixuan Chen, et al. arXiv 2024. [Paper]
  3. Advancing LLM Reasoning Generalists with Preference Trees
    Lifan Yuan, Ganqu Cui, Hanbin Wang, et al. arXiv 2024. [Paper]
  4. Tool Learning in the Wild: Empowering Language Models as Automatic Tool Agents
    Zhengliang Shi, Shen Gao, Lingyong Yan, et al. arXiv 2024. [Paper]
  5. Enhancing Chat Language Models by Scaling High-quality Instructional Conversations
    Ning Ding, Yulin Chen, Bokai Xu, et al. EMNLP 2023. [Paper]

1 Data Processing for LLM

⬆️top

1.1 Data Acquisition

Data Sources

Public Data
  1. Project Gutenberg: A large collection of free eBooks from the public domain; supports training language models on long-form literary text. [Source]
  2. Open Library: A global catalog of books with metadata and some open-access content; useful for multilingual and knowledge-enhanced language modeling. [Source]
  3. GitHub: The world’s largest open-source code hosting platform; supports training models for code generation and understanding. [Source]
  4. GitLab: A DevOps platform for hosting both private and open-source projects; provides high-quality programming and documentation data. [Source]
  5. Bitbucket: A source code hosting platform by Atlassian; suitable for mining enterprise-level software development data. [Source]
  6. CulturaX: A Cleaned, Enormous, and Multilingual Dataset for Large Language Models in 167 Languages
    Thuat Nguyen, Chien Van Nguyen, Viet Dac Lai, et al. LREC-COLING 2024. [Paper]
  7. The Stack: 3 TB of permissively licensed source code
    Denis Kocetkov, Raymond Li, Loubna Ben Allal, et al. arXiv 2022. [Paper]
  8. mT5: A Massively Multilingual Pre-trained Text-to-Text Transformer
    Linting Xue, Noah Constant, Adam Roberts, et al. NAACL 2021. [Paper]
  9. Exploring the limits of transfer learning with a unified text-to-text transformer
    Colin Raffel, Noam Shazeer, Adam Roberts, et al. JMLR 2020. [Paper]
  10. CodeSearchNet Challenge: Evaluating the State of Semantic Code Search
    Hamel Husain, Ho-Hsiang Wu, Tiferet Gazit, et al. arXiv 2019. [Paper]
  11. Aligning Books and Movies: Towards Story-like Visual Explanations by Watching Movies and Reading Books [Paper]

Data Acquisition Methods

Website Crawling
  1. Beautiful Soup: A Python-based library for parsing HTML and XML documents; supports extracting structured information from static web pages. [Source]
  2. Selenium: A browser automation tool that enables interaction with dynamic web pages; suitable for scraping JavaScript-heavy content. [Github]
  3. Playwright: A browser automation framework developed by Microsoft; supports multi-browser environments and is ideal for high-quality, concurrent web scraping tasks. [Source]
  4. Puppeteer: A Node.js library that provides a high-level API to control headless Chrome or Chromium; useful for scraping complex pages, taking screenshots, or generating PDFs. [Source]
  5. An Empirical Comparison of Web Content Extraction Algorithms
    Janek Bevendorff, Sanket Gupta, Johannes Kiesel, Benno Stein. SIGIR 2023. [Paper]
  6. Trafilatura: A Web Scraping Library and Command-Line Tool for Text Discovery and Extraction
    Adrien Barbaresi. ACL 2021 Demo. [Paper]
  7. Fact or Fiction: Content Classification for Digital Libraries
    Aidan Finn, N. Kushmerick, Barry Smyth. DELOS Workshops / Conferences 2001. [Paper]
Layout Analysis
  1. PaddleOCR: An open-source Optical Character Recognition (OCR) toolkit based on the PaddlePaddle deep learning framework; supports multilingual text detection and recognition, ideal for extracting text from images and document layout analysis. [Github]
  2. YOLOv10: Real-Time End-to-End Object Detection
    Ao Wang, Hui Chen, Lihao Liu, et al. NeurIPS 2024. [Paper]
  3. UMIE: Unified Multimodal Information Extraction with Instruction Tuning
    Lin Sun, Kai Zhang, Qingyuan Li, Renze Lou. AAAI 2024. [Paper]
  4. ChatEL: Entity linking with chatbots
    Yifan Ding, Qingkai Zeng, Tim Weninger. LREC | COLING 2024. [Paper]
  5. Vary: Scaling up the Vision Vocabulary for Large Vision-Language Models
    Haoran Wei, Lingyu Kong, Jinyue Chen, et al. ECCV 2024. [Paper]
  6. General OCR Theory: Towards OCR - 2.0 via a Unified End - to - end Model
    Haoran Wei, Chenglong Liu, Jinyue Chen, et al. arXiv 2024. [Paper]
  7. Focus Anywhere for Fine-grained Multi-page Document Understanding
    Chenglong Liu, Haoran Wei, Jinyue Chen, et al. arXiv 2024. [Paper]
  8. MinerU: An Open-Source Solution for Precise Document Content Extraction
    Bin Wang, Chao Xu, Xiaomeng Zhao, et al. arXiv 2024. [Paper]
  9. WebIE: Faithful and Robust Information Extraction on the Web
    Chenxi Whitehouse, Clara Vania, Alham Fikri Aji, et al. ACL 2023. [Paper]
  10. ReFinED: An Efficient Zero-shot-capable Approach to End-to-End Entity Linking
    Tom Ayoola, Shubhi Tyagi, Joseph Fisher, et al. NAACL 2022 Industry Track. [Paper]
  11. Alignment-Augmented Consistent Translation for Multilingual Open Information Extraction
    Keshav Kolluru, Muqeeth Mohammed, Shubham Mittal, et al. ACL 2022. [Paper]
  12. LayoutLMv3: Pre-training for Document AI with Unified Text and Image Masking
    Yupan Huang, Tengchao Lv, Lei Cui, et al. ACM Multimedia 2022. [Paper]
  13. Learning Transferable Visual Models From Natural Language Supervision
    Alec Radford, Jong Wook Kim, Chris Hallacy, et al. ICML 2021. [Paper]
  14. Tesseract: an open-source optical character recognition engine
    Anthony Kay. Linux Journal, Volume 2007. [Paper]

1.2 Data Deduplication

⬆️top

  1. Analysis of the Reasoning with Redundant Information Provided Ability of Large Language Models
    Wenbei Xie. arXiv 2023. [Paper]
  2. Scaling Laws and Interpretability of Learning from Repeated Data
    Danny Hernandez, Tom Brown, Tom Conerly, et al. arXiv 2022. [Paper]

Exact Substring Matching

  1. BaichuanSEED: Sharing the Potential of ExtensivE Data Collection and Deduplication by Introducing a Competitive Large Language Model Baseline
    Guosheng Dong, Da Pan, Yiding Sun, et al. arXiv 2024. [Paper]
  2. Deduplicating Training Data Makes Language Models Better
    Katherine Lee, Daphne Ippolito, Andrew Nystrom, et al. ACL 2022. [Paper]
  3. Suffix arrays: a new method for on-line string searches
    Udi Manber, Gene Myers. SIAM Journal on Computing 1993. [Paper]

Approximate Hashing-based Deduplication

  1. BaichuanSEED: Sharing the Potential of ExtensivE Data Collection and Deduplication by Introducing a Competitive Large Language Model Baseline [Paper]
  2. LSHBloom: Memory-efficient, Extreme-scale Document Deduplication
    Arham Khan, Robert Underwood, Carlo Siebenschuh, et al. arXiv 2024. [Paper]
  3. SimiSketch: Efficiently Estimating Similarity of streaming Multisets
    Fenghao Dong, Yang He, Yutong Liang, et al. arXiv 2024. [Paper]
  4. DotHash: Estimating Set Similarity Metrics for Link Prediction and Document Deduplication
    Igor Nunes, Mike Heddes, Pere Vergés, et al. KDD 2023. [Paper]
  5. Formalizing BPE Tokenization
    Martin Berglund (Umeå University), Brink van der Merwe (Stellenbosch University). NCMA 2023. [Paper]
  6. SlimPajama-DC: Understanding Data Combinations for LLM Training
    Zhiqiang Shen, Tianhua Tao, Liqun Ma, et al. arXiv 2023. [Paper]
  7. Deduplicating Training Data Makes Language Models Better [Paper]
  8. Noise-Robust De-Duplication at Scale
    Emily Silcock, Luca D'Amico-Wong, Jinglin Yang, Melissa Dell. arXiv 2022. [Paper]
  9. In Defense of Minhash over Simhash
    Anshumali Shrivastava, Ping Li. AISTATS 2014. [Paper]
  10. Similarity estimation techniques from rounding algorithms
    Moses S. Charikar. STOC 2002. [Paper]
  11. On the Resemblance and Containment of Documents
    A. Broder. Compression and Complexity of SEQUENCES 1997. [Paper]

Approximate Frequency-based Down-Weighting

  1. SoftDedup: an Efficient Data Reweighting Method for Speeding Up Language Model Pre-training
    Nan He, Weichen Xiong, Hanwen Liu, et al. ACL 2024. [Paper]

Embedding-Based Clustering

  1. FairDeDup: Detecting and Mitigating Vision-Language Fairness Disparities in Semantic Dataset Deduplication
    Eric Slyman, Stefan Lee, Scott Cohen, Kushal Kafle. CVPR 2024. [Paper]
  2. Effective Pruning of Web-Scale Datasets Based on Complexity of Concept Clusters
    Amro Abbas, Evgenia Rusak, Kushal Tirumala, et al. ICLR 2024. [Paper]
  3. D4: Improving LLM Pretraining via Document De-Duplication and Diversification
    Kushal Tirumala, Daniel Simig, Armen Aghajanyan, Ari Morcos. NeurIPS 2023. [Paper]
  4. SemDeDup: Data-efficient learning at web-scale through semantic deduplication
    Amro Abbas, Kushal Tirumala, Dániel Simig, et al. ICLR 2023. [Paper]
  5. OPT: Open Pre-trained Transformer Language Models
    Susan Zhang, Stephen Roller, Naman Goyal, et al. arXiv 2022. [Paper]
  6. Learning Transferable Visual Models From Natural Language Supervision [Paper]
  7. OpenCLIP
    Gabriel Ilharco, Mitchell Wortsman, Ross Wightman, et al. 2021. [Paper]
  8. LAION-400M: Open Dataset of CLIP-Filtered 400 Million Image-Text Pairs
    Christoph Schuhmann, Richard Vencu, Romain Beaumont, et al. NeurIPS 2021. [Paper]

Non-Text Data Deduplication

  1. DataComp: In search of the next generation of multimodal datasets
    Samir Yitzhak Gadre, Gabriel Ilharco, Alex Fang, et al. NeurIPS 2023. [Paper]
  2. SemDeDup: Data-efficient learning at web-scale through semantic deduplication [Paper]
  3. Learning Transferable Visual Models From Natural Language Supervision [Paper]
  4. Contrastive Learning with Large Memory Bank and Negative Embedding Subtraction for Accurate Copy Detection
    Shuhei Yokoo. arXiv 2021. [Paper]

1.3 Data Filtering

⬆️top

Sample-level Filtering

(1) Statistical Evaluation
  1. Perplexed by Perplexity: Perplexity-Based Data Pruning With Small Reference Models
    Zachary Ankner, Cody Blakeney, Kartik Sreenivasan, et al. ICLR 2025. [Paper]
  2. Data-efficient Fine-tuning for LLM-based Recommendation
    Xinyu Lin, Wenjie Wang, Yongqi Li, et al. SIGIR 2024. [Paper]
  3. SHED: Shapley-Based Automated Dataset Refinement for Instruction Fine-Tuning
    Yexiao He, Ziyao Wang, Zheyu Shen, et al. NeurIPS 2024. [Paper]
  4. SmallToLarge (S2L): Scalable Data Selection for Fine-tuning Large Language Models by Summarizing Training Trajectories of Small Models
    Yu Yang, Siddhartha Mishra, Jeffrey Chiang, et al. NeurIPS 2024. [Paper]
  5. Effective Pruning of Web-Scale Datasets Based on Complexity of Concept Clusters [Paper]
  6. WizardLM: Empowering Large Pre-Trained Language Models to Follow Complex Instructions
    Can Xu, Qingfeng Sun, Kai Zheng, et al. ICLR 2024. [Paper]
  7. Superfiltering: Weak-to-Strong Data Filtering for Fast Instruction-Tuning
    Ming Li, Yong Zhang, Shwai He, et al. ACL 2024. [Paper]
  8. Smaller Language Models are capable of selecting Instruction-Tuning Training Data for Larger Language Models
    Dheeraj Mekala, Alex Nguyen, Jingbo Shang. ACL 2024. [Paper]
  9. Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research
    Luca Soldaini, Rodney Kinney, Akshita Bhagia, et al. ACL 2024. [Paper]
  10. From Quantity to Quality: Boosting LLM Performance with Self-Guided Data Selection for Instruction Tuning
    Ming Li, Yong Zhang, Zhitao Li, et al. NAACL 2024. [Paper]
  11. Improving Pretraining Data Using Perplexity Correlations
    Tristan Thrush, Christopher Potts, Tatsunori Hashimoto. arXiv 2024. [Paper]
  12. Introducing MPT-7B: A New Standard for Open-Source, Commercially Usable LLMs
    The Mosaic Research Team. 2023. [Paper]
  13. Instruction Tuning with GPT-4
    Baolin Peng, Chunyuan Li, Pengcheng He, et al. arXiv 2023. [Paper]
  14. DINOv2: Learning Robust Visual Features without Supervision
    Maxime Oquab, Timothée Darcet, Théo Moutakanni, et al. arXiv 2023. [Paper]
  15. The Pile: An 800GB Dataset of Diverse Text for Language Modeling
    Leo Gao, Stella Biderman, Sid Black, et al. arXiv 2021. [Paper]
  16. Language Models are Unsupervised Multitask Learners
    Alec Radford, Jeffrey Wu, Rewon Child, et al. OpenAI blog 2019. [Paper]
  17. Bag of Tricks for Efficient Text Classification
    Armand Joulin, Edouard Grave, Piotr Bojanowski, Tomas Mikolov. EACL 2017. [Paper]
  18. The Shapley Value: Essays in Honor of Lloyd S. Shapley
    A. E. Roth, Ed. Cambridge: Cambridge University Press, 1988. [Source]
(2) Model Scoring
  1. SEAL: Safety-enhanced Aligned LLM Fine-tuning via Bilevel Data Selection
    Han Shen, Pin-Yu Chen, Payel Das, Tianyi Chen. ICLR 2025. [Paper]
  2. SCAR: Data Selection via Style-Consistency-Aware Response Ranking for Efficient Instruction Tuning of Large Language Models
    Zhuang Li, Yuncheng Hua, Thuy-Trang Vu, et al. ACL 2025. [Paper] [Github]
  3. QuRating: Selecting High-Quality Data for Training Language Models
    Alexander Wettig, Aatmik Gupta, Saumya Malik, Danqi Chen. ICML 2024. [Paper]
  4. What Makes Good Data for Alignment? A Comprehensive Study of Automatic Data Selection in Instruction Tuning
    Wei Liu, Weihao Zeng, Keqing He, et al. ICLR 2024. [Paper]
  5. LAB: Large-Scale Alignment for ChatBots
    Shivchander Sudalairaj, Abhishek Bhandwaldar, Aldo Pareja, et al. arXiv 2024. [Paper]
  6. Biases in Large Language Models: Origins, Inventory, and Discussion
    Roberto Navigli, Simone Conia, Björn Ross. ACM JDIQ, 2023. [Paper]
(3) Hybrid Methods
  1. Emergent and predictable memorization in large language models
    Stella Biderman, USVSN Sai Prashanth, Lintang Sutawika, et al. NeurIPS 2023. [Paper]
  2. When Less is More: Investigating Data Pruning for Pretraining LLMs at Scale
    Max Marion, Ahmet Üstün, Luiza Pozzobon, et al. arXiv 2023. [Paper]
  3. Instruction Mining: Instruction Data Selection for Tuning Large Language Models
    Yihan Cao, Yanbin Kang, Chi Wang, Lichao Sun. arXiv 2023. [Paper]
  4. Llama 2: Open Foundation and Fine-Tuned Chat Models
    Hugo Touvron, Louis Martin, Kevin Stone, et al. arXiv 2023. [Paper]
  5. MoDS: Model-oriented Data Selection for Instruction Tuning
    Qianlong Du, Chengqing Zong, Jiajun Zhang. arXiv 2023. [Paper]
  6. Economic Hyperparameter Optimization With Blended Search Strategy
    Chi Wang, Qingyun Wu, Silu Huang, Amin Saied. ICLR 2021. [Paper]
  7. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
    Jacob Devlin, Ming-Wei Chang, Kenton Lee, et al. NAACL 2019. [Paper]
  8. Active Learning for Convolutional Neural Networks: A Core-Set Approach
    Ozan Sener, Silvio Savarese. ICLR 2018. [Paper]

Content-level Filtering

  1. spaCy: An industrial-strength Natural Language Processing (NLP) library that supports tokenization, part-of-speech tagging, named entity recognition, dependency parsing, and more; well-suited for fast and accurate text processing and information extraction. [Source]
  2. CogVideoX: Text-to-Video Diffusion Models with An Expert Transformer
    Zhuoyi Yang, Jiayan Teng, Wendi Zheng, et al. ICLR 2025. [Paper]
  3. HunyuanVideo: A Systematic Framework For Large Video Generative Models
    Weijie Kong, Qi Tian, Zijian Zhang, et al. arXiv 2025. [Paper]
  4. Wan: Open and Advanced Large-Scale Video Generative Models
    Team Wan et al. arXiv 2025. [Paper]
  5. Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding
    Hang Zhang, Xin Li, Lidong Bing. EMNLP 2023 (System Demonstrations). [Paper]
  6. Analyzing Leakage of Personally Identifiable Information in Language Models
    Nils Lukas, Ahmed Salem, Robert Sim, et al. IEEE S&P 2023. [Paper]
  7. DeID-GPT: Zero-shot Medical Text De-Identification by GPT-4
    Zhengliang Liu, Yue Huang, Xiaowei Yu, et al. arXiv 2023. [Paper]
  8. Baichuan 2: Open Large-scale Language Models
    Aiyuan Yang, Bin Xiao, Bingning Wang, et al. arXiv 2023. [Paper]
  9. Exploring Video Quality Assessment on User Generated Contents from Aesthetic and Technical Perspectives
    Haoning Wu, Erli Zhang, Liang Liao, et al. arXiv 2022. [Paper]
  10. YOLOX: Exceeding YOLO Series in 2021
    Zheng Ge, Songtao Liu, Feng Wang, et al. arXiv 2021. [Paper]
  11. LAION-400M: Open Dataset of CLIP-Filtered 400 Million Image-Text Pairs [Paper]
  12. FLAIR: An Easy-to-Use Framework for State-of-the-Art NLP
    Alan Akbik, Tanja Bergmann, Duncan Blythe, et al. NAACL 2019 Demos. [Paper]

1.4 Data Selection

⬆️top

  1. A Survey on Data Selection for Language Models
    Alon Albalak, Yanai Elazar, Sang Michael Xie, et al. arXiv 2024. [Paper]

  2. A Survey on Data Selection for LLM Instruction Tuning
    Jiahao Wang, Bolin Zhang, Qianlong Du, et al. arXiv 2024. [Paper]

Similarity-based Data Selection

  1. spaCy: [Source]
  2. Enabling On-Device Large Language Model Personalization with Self-Supervised Data Selection and Synthesis
    Ruiyang Qin, Jun Xia, Zhenge Jia, et al. DAC 2024. [Paper]
  3. CoLoR-Filter: Conditional Loss Reduction Filtering for Targeted Language Model Pre-training
    David Brandfonbrener, Hanlin Zhang, Andreas Kirsch, et al. NeurIPS 2024. [Paper]
  4. Efficient Continual Pre-training for Building Domain Specific Large Language Models
    Yong Xie, Karan Aggarwal, Aitzaz Ahmad. Findings of ACL 2024. [Paper]
  5. Data Selection for Language Models via Importance Resampling
    Sang Michael Xie, Shibani Santurkar, Tengyu Ma, Percy Liang. NeurIPS 2023. [Paper]

Optimization-based Data Selection

  1. DSDM: model-aware dataset selection with datamodels
    Logan Engstrom, Axel Feldmann, Aleksander Mądry. ICML 2024. [Paper]
  2. LESS: Selecting Influential Data for Targeted Instruction Tuning
    Mengzhou Xia, Sadhika Malladi, Suchin Gururangan, et al. ICML 2024. [Paper]
  3. TSDS: Data Selection for Task-Specific Model Finetuning
    Zifan Liu, Amin Karbasi, Theodoros Rekatsinas. arXiv 2024. [Paper]
  4. Datamodels: Understanding Predictions with Data and Data with Predictions
    Andrew Ilyas, Sung Min Park, Logan Engstrom, et al. ICML 2022. [Paper]

Model-based Data Selection

  1. Autonomous Data Selection with Language Models for Mathematical Texts
    Yifan Zhang, Yifan Luo, Yang Yuan, et al. ICLR 2024. [Paper]

1.5 Data Mixing

⬆️top

  1. Mixtera: A Data Plane for Foundation Model Training Maximilian Böther, Xiaozhe Yao, Tolga Kerimoglu, Dan Graur, Viktor Gsteiger, Ana Klimovic. arXiv 2025. [Paper]
  2. Scalable Data Ablation Approximations for Language Models through Modular Training and Merging
    Clara Na, Ian Magnusson, Ananya Harsh Jha, et al. EMNLP 2024. [Paper]
  3. Mixture-of-LoRAs: An Efficient Multitask Tuning for Large Language Models
    Wenfeng Feng, Chuzhan Hao, Yuewei Zhang, et al. COLING 2024. [Paper]

Heuristic Optimization

  1. BiMix: Bivariate Data Mixing Law for Language Model Pretraining
    Ce Ge, Zhijian Ma, Daoyuan Chen, et al. arXiv 2024. [Paper]
  2. Maximize Your Data's Potential: Enhancing LLM Accuracy with Two-Phase Pretraining
    Steven Feng, Shrimai Prabhumoye, Kezhi Kong, et al. arXiv 2024. [Paper]
  3. SlimPajama-DC: Understanding Data Combinations for LLM Training [Paper]
  4. Evaluating Large Language Models Trained on Code [Paper]
  5. Exploring the limits of transfer learning with a unified text-to-text transformer [Paper]
  6. CommonsenseQA: A Question Answering Challenge Targeting Commonsense Knowledge
    Alon Talmor, Jonathan Herzig, Nicholas Lourie, et al. NAACL 2019. [Paper]
  7. A mathematical theory of communication
    C. E. Shannon. The Bell system technical journal 1948. [Paper]

Bilevel Optimization

  1. ScaleBiO: Scalable Bilevel Optimization for LLM Data Reweighting
    Rui Pan, Jipeng Zhang, Xingyuan Pan, et al. ACL 2025. [Paper]
  2. DoGE: Domain Reweighting with Generalization Estimation
    Simin Fan, Matteo Pagliardini, Martin Jaggi. ICML 2024. [Paper]
  3. An overview of bilevel optimization
    Benoît Colson, Patrice Marcotte, Gilles Savard. AOR 2007. [Paper]

Distributionally Robust Optimization

  1. Task-level Distributionally Robust Optimization for Large Language Model-based Dense Retrieval
    Guangyuan Ma, Yongliang Ma, Xing Wu, et al. AAAI 2025. [Paper]
  2. DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining
    Sang Michael Xie, Hieu Pham, Xuanyi Dong, et al. NeurIPS 2023. [Paper]
  3. Qwen Technical Report
    Jinze Bai, Shuai Bai, Yunfei Chu, Zeyu Cui, et al. arXiv 2023. [Paper]

Model-Based Optimization

  1. RegMix: Data Mixture as Regression for Language Model Pre-training
    Qian Liu, Xiaosen Zheng, Niklas Muennighoff, et al. ICLR 2025. [Paper]
  2. Data Mixing Laws: Optimizing Data Mixtures by Predicting Language Modeling Performance
    Jiasheng Ye, Peiju Liu, Tianxiang Sun, et al. ICLR 2025. [Paper]
  3. CMR Scaling Law: Predicting Critical Mixture Ratios for Continual Pre-training of Language Models
    Jiawei Gu, Zacc Yang, Chuanghao Ding, et al. EMNLP 2024. [Paper]
  4. TinyLlama: An Open-Source Small Language Model
    Peiyuan Zhang, Guangtao Zeng, Tianduo Wang, Wei Lu. arXiv 2024. [Paper]
  5. BiMix: Bivariate Data Mixing Law for Language Model Pretraining [Paper]
  6. D-CPT Law: Domain-specific Continual Pre-Training Scaling Law for Large Language Models
    Haoran Que, Jiaheng Liu, Ge Zhang, et al. arXiv 2024. [Paper]
  7. Data Proportion Detection for Optimized Data Management for Large Language Models
    Hao Liang, Keshi Zhao, Yajie Yang, et al. arXiv 2024. [Paper]
  8. DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining [Paper]
  9. Training compute-optimal large language models
    Jordan Hoffmann, Sebastian Borgeaud, Arthur Mensch, et al. NeurIPS 2022. [Paper]
  10. LightGBM: a highly efficient gradient boosting decision tree
    Guolin Ke, Qi Meng, Thomas Finley, et al. NeurIPS 2017. [Paper]

1.6 Data Distillation and Synthesis

⬆️top

  1. How to Synthesize Text Data without Model Collapse?
    Xuekai Zhu, Daixuan Cheng, Hengli Li, et al. ICML 2025. [Paper]
  2. Differentially Private Synthetic Data via Foundation Model APIs 2: Text
    Chulin Xie, Zinan Lin, Arturs Backurs, et al. ICML 2024. [Paper]
  3. LLM See, LLM Do: Leveraging Active Inheritance to Target Non-Differentiable Objectives
    Luísa Shimabucoro, Sebastian Ruder, Julia Kreutzer, et al. EMNLP 2024. [Paper]
  4. WizardLM: Empowering Large Pre-Trained Language Models to Follow Complex Instructions [Paper]
  5. Augmenting Math Word Problems via Iterative Question Composing
    Haoxiong Liu, Yifan Zhang, Yifan Luo, et al. arXiv 2024. [Paper]

Knowledge Distillation

  1. Multistage Collaborative Knowledge Distillation from a Large Language Model for Semi-Supervised Sequence Generation
    Jiachen Zhao, Wenlong Zhao, Andrew Drozdov, et al. ACL 2024. [Paper]
  2. PaD: Program-aided Distillation Can Teach Small Models Reasoning Better than Chain-of-thought Fine-tuning
    Xuekai Zhu, Biqing Qi, Kaiyan Zhang, et al. NAACL 2024. [Paper]
  3. Knowledge Distillation Using Frontier Open-source LLMs: Generalizability and the Role of Synthetic Data
    Anup Shirgaonkar, Nikhil Pandey, Nazmiye Ceren Abay, et al. arXiv 2024. [Paper]
  4. Training Verifiers to Solve Math Word Problems
    Karl Cobbe, Vineet Kosaraju, Mohammad Bavarian, et al. arXiv 2021. [Paper]
  5. Dialogue chain-of-thought distillation for commonsense-aware conversational agents
    Hyungjoo Chae, Yongho Song, Kai Tzu-iunn Ong, et al. arXiv 2023. [Paper]
  6. MCC-KD: Multi-CoT consistent knowledge distillation
    Hongzhan Chen, Siyue Wu, Xiaojun Quan, et al. arXiv 2023. [Paper]
  7. Large language models are reasoning teachers
    Namgyu Ho, Laura Schmid, Se-Young Yun. arXiv 2023. [Paper]
  8. Leveraging training data in few-shot prompting for numerical reasoning
    Zhanming Jie, Wei Lu. arXiv 2023. [Paper]
  9. Knowledge-augmented reasoning distillation for small language models in knowledge-intensive tasks
    Minki Kang, Seanie Lee, Jinheon Baek, et al. NeurIPS 2023. [Paper]
  10. Symbolic chain-of-thought distillation: Small models can also "think" step-by-step
    Liunian Harold Li, Jack Hessel, Youngjae Yu, et al. arXiv 2024. [Paper]
  11. Explanations from large language models make small reasoners better
    Shiyang Li, Jianshu Chen, Yelong Shen, et al. arXiv 2022. [Paper]
  12. Distilling reasoning capabilities into smaller language models
    Kumar Shridhar, Alessandro Stolfo, Mrinmaya Sachan. arXiv 2023. [Paper]
  13. SCOTT: Self-consistent chain-of-thought distillation
    Peifeng Wang, Zhengyang Wang, Zheng Li, et al. arXiv 2023. [Paper]
  14. Democratizing reasoning ability: Tailored learning from large language model
    Zhaoyang Wang, Shaohan Huang, Yuxuan Liu, et al. arXiv 2023. [Paper]

Pre-training Data Augmentation

  1. BERT-Tiny-Chinese: A lightweight Chinese BERT pre-trained model released by CKIP Lab, with a small number of parameters; suitable for use as an encoder in pre-training data augmentation tasks to enhance efficiency for compact models. [Source]
  2. Case2Code: Scalable Synthetic Data for Code Generation
    Yunfan Shao, Linyang Li, Yichuan Ma, et al. COLING 2025. [Paper]
  3. Advancing Mathematical Reasoning in Language Models: The Impact of Problem-Solving Data, Data Synthesis Methods, and Training Stages
    Zui Chen, Tianqiao Liu, Mi Tian, et al. ICLR 2025. [Paper]
  4. JiuZhang3.0: Efficiently Improving Mathematical Reasoning by Training Small Data Synthesis Models
    Kun Zhou, Beichen Zhang, Jiapeng Wang, et al. arXiv 2024. [Paper]
  5. Florence-2: Advancing a Unified Representation for a Variety of Vision Tasks
    Bin Xiao, Haiping Wu, Weijian Xu, et al. CVPR 2024. [Paper]
  6. DiffuseMix: Label-Preserving Data Augmentation with Diffusion Models
    Khawar Islam, Muhammad Zaigham Zaheer, Arif Mahmood, et al. CVPR 2024. [Paper]
  7. Magicoder: Empowering Code Generation with OSS-Instruct
    Yuxiang Wei, Zhe Wang, Jiawei Liu, et al. ICML 2024. [Paper]
  8. Instruction Pre-Training: Language Models are Supervised Multitask Learners
    Daixuan Cheng, Yuxian Gu, Shaohan Huang, et al. EMNLP 2024. [Paper]
  9. Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research [Paper]
  10. Rephrasing the Web: A Recipe for Compute and Data-Efficient Language Modeling
    Pratyush Maini, Skyler Seto, Richard Bai, et al. ACL 2024. [Paper]
  11. VeCLIP: Improving CLIP Training via Visual-Enriched Captions
    Zhengfeng Lai, Haotian Zhang, Bowen Zhang, et al. ECCV 2024. [Paper]
  12. Diffusion Models and Representation Learning: A Survey
    Michael Fuest, Pingchuan Ma, Ming Gui, et al. arXiv 2024. [Paper]
  13. CtrlSynth: Controllable Image Text Synthesis for Data-Efficient Multimodal Learning
    Qingqing Cao, Mahyar Najibi, Sachin Mehta. arXiv 2024. [Paper]
  14. Qwen2 Technical Report
    An Yang, Baosong Yang, Binyuan Hui, et al. arXiv 2024. [Paper]
  15. TinyLlama: An Open-Source Small Language Model [Paper]
  16. On the Diversity of Synthetic Data and its Impact on Training Large Language Models
    Hao Chen, Abdul Waheed, Xiang Li, et al. arXiv 2024. [Paper]
  17. Towards Effective and Efficient Continual Pre-training of Large Language Models
    Jie Chen, Zhipeng Chen, Jiapeng Wang, et al. arXiv 2024. [Paper]
  18. Improving CLIP Training with Language Rewrites
    Lijie Fan, Dilip Krishnan, Phillip Isola, et al. NeurIPS 2023. [Paper]
  19. Effective Data Augmentation With Diffusion Models
    Brandon Trabucco, Kyle Doherty, Max Gurinas, et al. arXiv 2023. [Paper]
  20. Mistral 7B
    Albert Q. Jiang, Alexandre Sablayrolles, Arthur Mensch, et al. arXiv 2023. [Paper]
  21. Llama 2: Open Foundation and Fine-Tuned Chat Models [Paper]
  22. SDXL: Improving Latent Diffusion Models for High-Resolution Image Synthesis
    Dustin Podell, Zion English, Kyle Lacey, et al. arXiv 2023. [Paper]
  23. Documenting Large Webtext Corpora: A Case Study on the Colossal Clean Crawled Corpus
    Jesse Dodge, Maarten Sap, Ana Marasović, et al. EMNLP 2021. [Paper]
  24. The Pile: An 800GB Dataset of Diverse Text for Language Modeling [Paper]
  25. First Steps of an Approach to the ARC Challenge based on Descriptive Grid Models and the Minimum Description Length Principle
    Sébastien Ferré (Univ Rennes, CNRS, IRISA). arXiv 2021. [Paper]
  26. TinyBERT: Distilling BERT for Natural Language Understanding
    Xiaoqi Jiao, Yichun Yin, Lifeng Shang, et al. Findings of EMNLP 2020. [Paper]
  27. HellaSwag: Can a Machine Really Finish Your Sentence?
    Rowan Zellers, Ari Holtzman, Yonatan Bisk, et al. ACL 2019. [Paper]

SFT Data Augmentation

  1. Key-Point-Driven Data Synthesis with its Enhancement on Mathematical Reasoning
    Yiming Huang, Xiao Liu, Yeyun Gong, et al. arXiv 2024. [Paper]
  2. Augmenting Math Word Problems via Iterative Question Composing [Paper]
  3. AgentInstruct: Toward Generative Teaching with Agentic Flows
    Arindam Mitra, Luciano Del Corro, Guoqing Zheng, et al. arXiv 2024. [Paper]
  4. Synthetic Data (Almost) from Scratch: Generalized Instruction Tuning for Language Models
    Haoran Li, Qingxiu Dong, Zhengyang Tang, et al. arXiv 2024. [Paper]
  5. Self-Instruct: Aligning Language Models with Self-Generated Instructions
    Yizhong Wang, Yeganeh Kordi, Swaroop Mishra, et al. ACL 2023. [Paper]

SFT Reasoning Data Augmentation

  1. DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning [Paper]
  2. LIMO: Less is More for Reasoning
    Yixin Ye, Zhen Huang, Yang Xiao, et al. arXiv 2025. [Paper]
  3. LLMs Can Easily Learn to Reason from Demonstrations: Structure, Not Content, Is What Matters!
    Dacheng Li, Shiyi Cao, Tyler Griggs, et al. arXiv 2025. [Paper]
  4. Satori: Reinforcement Learning with Chain-of-Action-Thought Enhances LLM Reasoning via Autoregressive Search
    Maohao Shen, Guangtao Zeng, Zhenting Qi, et al. arXiv 2025. [Paper]
  5. Advancing Language Model Reasoning through Reinforcement Learning and Inference Scaling
    Zhenyu Hou, Xin Lv, Rui Lu, et al. arXiv 2025. [Paper]
  6. MUSTARD: Mastering Uniform Synthesis of Theorem and Proof Data
    Yinya Huang, Xiaohan Lin, Zhengying Liu, et al. ICLR 2024. [Paper]
  7. Math-Shepherd: Verify and Reinforce LLMs Step-by-step without Human Annotations
    Peiyi Wang, Lei Li, Zhihong Shao, et al. ACL 2024. [Paper]
  8. NuminaMath: The largest public dataset in AI4Maths with 860k pairs of competition math problems and solutions
    Jia Li, Edward Beeching, Lewis Tunstall, et al. 2024. [Paper]
  9. QwQ: Reflect Deeply on the Boundaries of the Unknown
    Qwen Team. 2024. [Source]
  10. Let's Verify Step by Step
    Hunter Lightman, Vineet Kosaraju, Yura Burda, et al. arXiv 2023. [Paper]

Reinforcement Learning

  1. Judging LLM-as-a-Judge with MT-Bench and Chatbot Arena
    Lianmin Zheng, Wei-Lin Chiang, Ying Sheng, et al. NeurIPS 2023. [Paper]
  2. Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback
    Yuntao Bai, Andy Jones, Kamal Ndousse, et al. arXiv 2022. [Paper]

Retrieval-Augmentation Generation

  1. Mitigating the Privacy Issues in Retrieval-Augmented Generation (RAG) via Pure Synthetic Data
    Shenglai Zeng, Jiankun Zhang, Pengfei He, et al. arXiv 2024. [Paper]

1.7 End-to-End Data Processing Pipelines

⬆️top

1.7.1 Typical data processing frameworks

  1. Mixtera: A Data Plane for Foundation Model Training
    Maximilian Böther, Xiaozhe Yao, Tolga Kerimoglu, et al. arXiv 2025. [Paper]
  2. Data-Juicer: A One-Stop Data Processing System for Large Language Models
    Daoyuan Chen, Yilun Huang, Zhijian Ma, et al. SIGMOD 2024. [Paper]
  3. An Integrated Data Processing Framework for Pretraining Foundation Models
    Yiding Sun, Feng Wang, Yutao Zhu, et al. SIGIR 2024. [Paper]
  4. Dataverse: Open-Source ETL (Extract, Transform, Load) Pipeline for Large Language Models
    Hyunbyung Park, Sukyung Lee, Gyoungjin Gim, et al. arXiv 2024. [Paper]

1.7.2 Typical data pipelines

  1. Common Crawl: A large-scale publicly accessible web crawl dataset that provides massive raw webpages and metadata. It serves as a crucial raw data source in typical pretraining data pipelines, where it undergoes multiple processing steps such as cleaning, deduplication, and formatting to produce high-quality corpora for downstream model training. [Source]
  2. The RefinedWeb dataset for falcon LLM: outperforming curated corpora with web data only
    Guilherme Penedo, Quentin Malartic, Daniel Hesslow, et al. NeurIPS 2023. [Paper]
  3. Trafilatura: A Web Scraping Library and Command-Line Tool for Text Discovery and Extraction [Paper]
  4. Scaling Language Models: Methods, Analysis & Insights from Training Gopher
    Jack W. Rae, Sebastian Borgeaud, Trevor Cai, et al. arXiv 2021. [Paper]
  5. CCNet: Extracting High Quality Monolingual Datasets from Web Crawl Data
    Guillaume Wenzek, Marie - Anne Lachaux, Alexis Conneau, et al. LREC 2020. [Paper]
  6. Exploring the limits of transfer learning with a unified text-to-text transformer [Paper]
  7. Bag of Tricks for Efficient Text Classification [Paper]

1.7.3 Orchestration of data pipelines

  1. Data-Juicer Sandbox: A Feedback-Driven Suite for Multimodal Data-Model Co-development [Paper]

2 Data Storage for LLM

⬆️top

2.1 Data Formats

Training Data Format

  1. TFRecord: A binary data storage format recommended by TensorFlow, suitable for efficient storage and reading of large-scale training data. [Source]
  2. MindRecord: An efficient data storage format used by MindSpore, supporting multi-platform data management. [Source]
  3. tf.data.Dataset: An abstract interface in TensorFlow representing collections of training data, enabling flexible data manipulation. [Source]
  4. COCO JSON: COCO JSON format uses structured JSON to store images and their corresponding labels, widely used in computer vision datasets. [Source]

Model Data Format

  1. PyTorch-specific formats (.pt, .pth): PyTorch’s .pt and .pth formats are used to save model parameters and architecture, supporting model storage and loading. [Source]
  2. TensorFlow(SavedModel, .ckpt): TensorFlow’s SavedModel and checkpoint formats save complete model information, facilitating model reproduction and deployment. [Source]
  3. Hugging Face Transformers library: Hugging Face offers a unified model format interface to facilitate saving and usage of various pretrained models. [Source]
  4. Pickle (.pkl): Pickle format is used for serializing models and data, suitable for quick saving and loading. [Source]
  5. ONNX: An open cross-platform model format supporting model conversion and deployment across different frameworks. [Source]
  6. An Empirical Study of Safetensors' Usage Trends and Developers' Perceptions
    Beatrice Casey, Kaia Damian, Andrew Cotaj, et al. arXiv 2025. [Paper]

2.2 Data Distribution

⬆️top

  1. DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning [Paper]
  2. CC-GPX: Extracting High-Quality Annotated Geospatial Data from Common Crawl
    Ilya Ilyankou, Meihui Wang, Stefano Cavazzi, et al. SIGSPATIAL 2024. [Paper]

Distributed Storage Systems

  1. JuiceFS: A high-performance cloud-native distributed file system designed for efficient storage and access of large-scale data. [Github]
  2. 3FS: A distributed file system designed for deep learning and large-scale data processing, emphasizing high throughput and reliability. [Github]
  3. S3: A widely used cloud storage service offering secure, scalable, and highly available object storage solutions. [Source]
  4. Hdfs architecture guide. Hadoop apache project
    D. Borthakur et al. Hadoop apache project, 53(1-13):2, 2008. [Source]

Heterogeneous Storage Systems

  1. ProTrain: Efficient LLM Training via Memory-Aware Techniques
    Hanmei Yang, Jin Zhou, Yao Fu, et al. arXiv 2024. [Paper]
  2. ZeRO-infinity: breaking the GPU memory wall for extreme scale deep learning
    Samyam Rajbhandari, Olatunji Ruwase, Jeff Rasley, et al. SC 2021. [Paper]
  3. ZeRO-Offload: Democratizing Billion-Scale Model Training
    Jie Ren, Samyam Rajbhandari, Reza Yazdani Aminabadi, et al. USENIX ATC 2021. [Paper]
  4. ZeRO: memory optimizations toward training trillion parameter models
    Samyam Rajbhandari, Jeff Rasley, Olatunji Ruwase, et al. SC 2020. [Paper]
  5. vDNN: virtualized deep neural networks for scalable, memory-efficient neural network design
    Minsoo Rhu, Natalia Gimelshein, Jason Clemons, et al. MICRO-49 2016. [Paper]

2.3 Data Organization

⬆️top

  1. Survey of Hallucination in Natural Language Generation
    Ziwei Ji, Nayeon Lee, Rita Frieske, et al. ACM Computing Surveys (2022). [Paper]
  2. Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks
    Patrick Lewis, Ethan Perez, Aleksandra Piktus, et al. NeurIPS 2020. [Paper]

Vector-Based Organization

  1. STELLA: A large-scale Chinese vector database supporting efficient vector search and semantic retrieval applications. [Source]
  2. Milvus: An open-source vector database focused on large-scale, high-performance similarity search and analysis. [Source]
  3. Weaviate: Weaviate offers a cloud-native vector search engine supporting intelligent search and knowledge graph construction for multimodal data. [Source]
  4. LanceDB: An efficient vector database designed for large-scale machine learning and recommendation systems. [Source]
  5. Mix-of-Granularity: Optimize the Chunking Granularity for Retrieval-Augmented Generation
    Zijie Zhong, Hanwen Liu, Xiaoya Cui, et al. COLING 2025. [Paper]
  6. Dense X Retrieval: What Retrieval Granularity Should We Use?
    Tong Chen, Hongwei Wang, Sihao Chen, et al. EMNLP 2024. [Paper]
  7. Scalable and Domain-General Abstractive Proposition Segmentation
    Mohammad Javad Hosseini, Yang Gao, Tim Baumgärtner, et al. Findings of EMNLP 2024. [Paper]
  8. A Hierarchical Context Augmentation Method to Improve Retrieval-Augmented LLMs on Scientific Papers
    Tian-Yi Che, Xian-Ling Mao, Tian Lan, et al. KDD 2024. [Paper]
  9. M3-Embedding: Multi-Linguality, Multi-Functionality, Multi-Granularity Text Embeddings Through Self-Knowledge Distillation
    Jianlyu Chen, Shitao Xiao, Peitian Zhang, et al. Findings of ACL 2024. [Paper]
  10. Thread: A Logic-Based Data Organization Paradigm for How-To Question Answering with Retrieval Augmented Generation
    Kaikai An, Fangkai Yang, Liqun Li, et al. arXiv 2024. [Paper]
  11. GleanVec: Accelerating Vector Search with Minimalist Nonlinear Dimensionality Reduction
    Mariano Tepper, Ishwar Singh Bhati, Cecilia Aguerrebere, et al. arXiv 2024. [Paper]
  12. The Faiss Library
    Matthijs Douze, Alexandr Guzhva, Chengqi Deng, et al. arXiv 2024. [Paper]
  13. Similarity Search in the Blink of an Eye with Compressed Indices
    Cecilia Aguerrebere, Ishwar Singh Bhati, Mark Hildebrand, et al. VLDB Endowment 2023. [Paper]
  14. LeanVec: Searching Vectors Faster by Making Them Fit
    Mariano Tepper, Ishwar Singh Bhati, Cecilia Aguerrebere, et al. arXiv 2023. [Paper]
  15. Towards General Text Embeddings with Multi-stage Contrastive Learning
    Zehan Li, Xin Zhang, Yanzhao Zhang, et al. arXiv 2023. [Paper]

Graph-Based Organization

  1. ArangoDB: A multi-model database that supports graph, document, and key-value data, suitable for handling complex relational queries. [Source]
  2. MiniRAG: Towards Extremely Simple Retrieval-Augmented Generation
    Tianyu Fan, Jingyuan Wang, Xubin Ren, et al. arXiv 2025. [Paper]
  3. From Local to Global: A Graph RAG Approach to Query-Focused Summarization
    Darren Edge, Ha Trinh, Newman Cheng, et al. arXiv 2024. [Paper]
  4. LightRAG: Simple and Fast Retrieval-Augmented Generation
    Zirui Guo, Lianghao Xia, Yanhua Yu, et al. arXiv 2024. [Paper]
  5. Graph Databases Assessment: JanusGraph, Neo4j, and TigerGraph
    Jéssica Monteiro, et al. Perspectives and Trends in Education and Technology 2023. [Paper]
  6. Empirical Evaluation of a Cloud-Based Graph Database: the Case of Neptune
    Ghislain Auguste Atemezing. KGSWC 2021. [Paper]

2.4 Data Movement

⬆️top

Caching Data

  1. CacheLib: An open-source, high-performance embedded caching library developed by Meta to accelerate data access and increase system throughput. [Source]
  2. Tectonic-Shift: A Composite Storage Fabric for Large-Scale ML Training
    Mark Zhao, Satadru Pan, Niket Agarwal, et al. USENIX ATC 2023. [Paper]
  3. Fluid: Dataset Abstraction and Elastic Acceleration for Cloud-native Deep Learning Training Jobs
    Rong Gu, Kai Zhang, Zhihao Xu, et al. ICDE 2022. [Paper]
  4. Quiver: An Informed Storage Cache for Deep Learning
    Abhishek Kumar, Muthian Sivathanu. USENIX FAST 2020. [Paper]

Data/Operator Offloading

  1. cedar: Optimized and Unified Machine Learning Input Data Pipelines
    Mark Zhao, et al. Proceedings of the VLDB Endowment, Volume 18, Issue 2, 2025. [Paper]
  2. Pecan: cost-efficient ML data preprocessing with automatic transformation ordering and hybrid placement
    Dan Graur, Oto Mraz, Muyu Li, et al. USENIX ATC 2024. [Paper]
  3. tf.data service: A Case for Disaggregating ML Input Data Processing
    Andrew Audibert, Yang Chen, Dan Graur, et al. SoCC 2023. [Paper]
  4. Cachew: Machine Learning Input Data Processing as a Service
    Dan Graur, Damien Aymon, Dan Kluser, et al. USENIX ATC 2022. [Paper]
  5. Borg: the next generation
    Muhammad Tirmazi, Adam Barker, Nan Deng, et al. EuroSys 2020. [Paper]

Overlapping of storage and computing

  1. Optimizing RLHF Training for Large Language Models with Stage Fusion
    Yinmin Zhong, Zili Zhang, Bingyang Wu, et al. NSDI 2025. [Paper]
  2. SiloD: A Co-design of Caching and Scheduling for Deep Learning Clusters
    Hanyu Zhao, Zhenhua Han, Zhi Yang, et al. EuroSys 2023. [Paper]
  3. Optimization by Simulated Annealing
    S. Kirkpatrick, C. D. Gelatt, Jr., M. P. Vecchi. Science, 220(4598):671–680, 1983. [Paper]

2.5 Data Fault Tolerance

⬆️top

Checkpoints

  1. PaddleNLP: PaddleNLP supports checkpoint saving and resuming during training, enabling fault tolerance and recovery for long-running training tasks. [Source]
  2. MegaScale: Scaling Large Language Model Training to More Than 10,000 GPUs
    Ziheng Jiang, Haibin Lin, Yinmin Zhong, et al. USENIX NSDI 2024. [Paper]
  3. ByteCheckpoint: A Unified Checkpointing System for Large Foundation Model Development
    Borui Wan, Mingji Han, Yiyao Sheng, et al. arXiv 2024. [Paper]
  4. GEMINI: Fast Failure Recovery in Distributed Training with In-Memory Checkpoints
    Zhuang Wang, Zhen Jia, Shuai Zheng, et al. SOSP 2023. [Paper]
  5. CheckFreq: Frequent, Fine-Grained DNN Checkpointing
    Jayashree Mohan, Amar Phanishayee, Vijay Chidambaram. USENIX FAST 2021. [Paper]

Redundant Computations

  1. ReCycle: Resilient Training of Large DNNs using Pipeline Adaptation
    Swapnil Gandhi, Mark Zhao, Athinagoras Skiadopoulos, et al. SOSP 2024. [Paper]
  2. Bamboo: Making Preemptible Instances Resilient for Affordable Training of Large DNNs
    John Thorpe, Pengzhan Zhao, Jonathan Eyolfson, et al. NSDI 2023 . [Paper]
  3. Oobleck: Resilient Distributed Training of Large Models Using Pipeline Templates
    Insu Jang, Zhenning Yang, Zhen Zhang, et al. SOSP 2023. [Paper]

2.6 KV Cache

⬆️top

Cache Space Management

  1. Efficient Memory Management for Large Language Model Serving with PagedAttention
    Woosuk Kwon, Zhuohan Li, Siyuan Zhuang, et al. SOSP 2023. [Paper]
  2. VTensor: Using Virtual Tensors to Build a Layout-oblivious AI Programming Framework
    Feng Yu, Jiacheng Zhao, Huimin Cui, et al. PACT 2020. [Paper]

KV Placement

  1. Cost-Efficient Large Language Model Serving for Multi-turn Conversations with CachedAttention
    Bin Gao, Zhuomin He, Puru Sharma, et al. USENIX ATC 2024. [Paper]
  2. RAGCache: Efficient Knowledge Caching for Retrieval-Augmented Generation
    Chao Jin, Zili Zhang, Xuanlin Jiang, et al. arXiv 2024. [Paper]

KV Shrinking

  1. Adaptive KV-Cache Compression without Manually Setting Budget
    Chenxia Tang, Jianchun Liu, Hongli Xu, et al. arXiv 2025. [Paper]
  2. Fast State Restoration in LLM Serving with HCache
    Shiwei Gao, Youmin Chen, Jiwu Shu. EuroSys 2025. [Paper]
  3. CacheGen: KV Cache Compression and Streaming for Fast Large Language Model Serving
    Yuhan Liu, Hanchen Li, Yihua Cheng, et al. SIGCOMM 2024. [Paper]
  4. MiniCache: KV Cache Compression in Depth Dimension for Large Language Models
    Akide Liu, Jing Liu, Zizheng Pan, et al. NeurIPS 2024. [Paper]
  5. Animating rotation with quaternion curves
    Ken Shoemake. ACM SIGGRAPH Computer Graphics, Volume 19, Issue 3. 1985. [Paper]

KV Indexing

  1. ChunkAttention: Efficient Self-Attention with Prefix-Aware KV Cache and Two-Phase Partition
    Lu Ye, Ze Tao, Yong Huang, et al. ACL 2024. [Paper]
  2. BatchLLM: Optimizing Large Batched LLM Inference with Global Prefix Sharing and Throughput-oriented Token Batching
    Zhen Zheng, Xin Ji, Taosong Fang, et al. arXiv 2024. [Paper]

3 Data Serving for LLM

⬆️top

3.1 Data Shuffling

Data Shuffling for Training

  1. Mixtera: A Data Plane for Foundation Model Training
    Maximilian Böther, Xiaozhe Yao, Tolga Kerimoglu, et al. arXiv 2025. [Paper]
  2. Velocitune: A Velocity-based Dynamic Domain Reweighting Method for Continual Pre-training
    Zheheng Luo, Xin Zhang, Xiao Liu, et al. ACL 2025. [Paper]
  3. How Abilities in Large Language Models are Affected by Supervised Fine-tuning Data Composition
    Guanting Dong, Hongyi Yuan, Keming Lu, et al. ACL 2024. [Paper]
  4. Mixture-of-Skills: Learning to Optimize Data Usage for Fine-Tuning Large Language Models
    Minghao Wu, Thuy-Trang Vu, Lizhen Qu, et al. EMNLP 2024. [Paper]
  5. Strategic Data Ordering: Enhancing Large Language Model Performance through Curriculum Learning
    Jisu Kim, Juhwan Lee. arXiv 2024. [Paper]
  6. NLU on Data Diets: Dynamic Data Subset Selection for NLP Classification Tasks
    Jean-michel Attendu, Jean-philippe Corbeil. SustaiNLP @ ACL 2023. [Paper]
  7. Efficient Online Data Mixing For Language Model Pre-Training
    Alon Albalak, Liangming Pan, Colin Raffel, et al. arXiv 2023. [Paper]
  8. Data Pruning via Moving-one-Sample-out
    Haoru Tan, Sitong Wu, Fei Du, et al. NeurIPS 2023. [Paper]
  9. BERT on a Data Diet: Finding Important Examples by Gradient-Based Pruning
    Mohsen Fayyaz, Ehsan Aghazadeh, Ali Modarressi, et al. ENLSP @ NeurIPS2022. [Paper]
  10. Scaling Laws for Neural Language Models
    Jared Kaplan, Sam McCandlish, Tom Henighan, et al. arXiv 2020. [Paper]
  11. Why there are complementary learning systems in the hippocampus and neocortex: insights from the successes and failures of connectionist models of learning and memory
    James L. McClelland, Bruce L. McNaughton, Randall C. O’Reilly. Psychological Review 1995. [Paper]
  12. Catastrophic Interference in Connectionist Networks: The Sequential Learning Problem
    M. McCloskey, N. J. Cohen. Psychology of Learning and Motivation 1989. [Paper]

Data Selection for RAG

  1. Cohere rerank: Cohere's rerank model reorders initial retrieval results to improve relevance to the query, making it a key component for building high-quality RAG systems. [Source]
  2. ASRank: Zero-Shot Re-Ranking with Answer Scent for Document Retrieval
    Abdelrahman Abdallah, Jamshid Mozafari, Bhawna Piryani, et al. NAACL 2025. [Paper]
  3. MAIN-RAG: Multi-Agent Filtering Retrieval-Augmented Generation
    Chia-Yuan Chang, Zhimeng Jiang, Vineeth Rakesh, et al. arXiv 2025. [Paper]
  4. ARAGOG: Advanced RAG Output Grading
    Matouš Eibich, Shivay Nagpal, Alexander Fred-Ojala. arXiv 2024. [Paper]
  5. Large Language Model Is Not a Good Few-shot Information Extractor, but a Good Reranker for Hard Samples!
    Yubo Ma, Yixin Cao, YongChing Hong, et al. Findings of EMNLP 2023. [Paper]
  6. Chatlaw: A Multi-Agent Collaborative Legal Assistant with Knowledge Graph Enhanced Mixture-of-Experts Large Language Model
    Jiaxi Cui, Munan Ning, Zongjian Li, et al. arXiv 2023. [Paper]
  7. RankVicuna: Zero-Shot Listwise Document Reranking with Open-Source Large Language Models
    Ronak Pradeep, Sahel Sharifymoghaddam, Jimmy Lin. arXiv 2023. [Paper]

3.2 Data Compression

⬆️top

RAG Knowledge Compression

  1. Context Embeddings for Efficient Answer Generation in RAG
    David Rau, Shuai Wang, Hervé Déjean, et al. WSDM 2025. [Paper]
  2. xRAG: Extreme Context Compression for Retrieval-augmented Generation with One Token
    Xin Cheng, Xun Wang, Xingxing Zhang, et al. NeurIPS 2024. [Paper]
  3. RECOMP: Improving Retrieval-Augmented LMs with Context Compression and Selective Augmentation
    Fangyuan Xu, Weijia Shi, Eunsol Choi. ICLR 2024. [Paper]
  4. Compressing Long Context for Enhancing RAG with AMR-based Concept Distillation
    Kaize Shi, Xueyao Sun, Qing Li, et al. arXiv 2024. [Paper]
  5. Familiarity-Aware Evidence Compression for Retrieval-Augmented Generation
    Dongwon Jung, Qin Liu, Tenghao Huang, et al. arXiv 2024. [Paper]

Prompt Compression

  1. LongLLMLingua: Accelerating and Enhancing LLMs in Long Context Scenarios via Prompt Compression
    Huiqiang Jiang, Qianhui Wu, Xufang Luo, et al. ACL 2024. [Paper]
  2. LLMLingua-2: Data Distillation for Efficient and Faithful Task-Agnostic Prompt Compression
    Zhuoshi Pan, Qianhui Wu, Huiqiang Jiang, et al. Findings of ACL 2024. [Paper]
  3. LLMLingua: Compressing Prompts for Accelerated Inference of Large Language Models
    Huiqiang Jiang, Qianhui Wu, Chin-Yew Lin, et al. EMNLP 2023. [Paper]
  4. Learning to Compress Prompts with Gist Tokens
    Jesse Mu, Xiang Lisa Li, Noah Goodman. NeurIPS 2023. [Paper]
  5. Adapting Language Models to Compress Contexts
    Alexis Chevalier, Alexander Wettig, Anirudh Ajith, et al. EMNLP 2023. [Paper]

3.3 Data Packing

⬆️top

Short Sequence Insertion

  1. Fewer Truncations Improve Language Modeling
    Hantian Ding, Zijian Wang, Giovanni Paolini, et al. ICML 2024. [Paper]
  2. Bucket Pre-training is All You Need
    Hongtao Liu, Qiyao Peng, Qing Yang, et al. arXiv 2024. [Paper]

Sequence Combination Optimization

  1. Dataset Decomposition: Faster LLM Training with Variable Sequence Length Curriculum
    Hadi Pouransari, Chun-Liang Li, Jen-Hao Rick Chang, et al. NeurIPS 2024. [Paper]
  2. Efficient Sequence Packing without Cross-contamination: Accelerating Large Language Models without Impacting Performance
    Mario Michael Krell, Matej Kosec, Sergio P. Perez, et al. arXiv 2021. [Paper]

Semantic-Based Packing

  1. Structured Packing in LLM Training Improves Long Context Utilization
    Konrad Staniszewski, Szymon Tworkowski, Sebastian Jaszczur, et al. AAAI 2025. [Paper]
  2. In-context Pretraining: Language Modeling Beyond Document Boundaries
    Weijia Shi, Sewon Min, Maria Lomeli, et al. ICLR 2024. [Paper]

3.4 Data Provenance

⬆️top

  1. A comprehensive survey on data provenance: : State-of-the-art approaches and their deployments for IoT security enforcement
    Md Morshed Alam, Weichao Wang. Journal of Computer Security, Volume 29, Issue 4. 2021. [Paper]

Embedding Markers

  1. Bileve: Securing Text Provenance in Large Language Models Against Spoofing with Bi-level Signature
    Tong Zhou, Xuandong Zhao, Xiaolin Xu, et al. NeurIPS 2024. [Paper]
  2. Undetectable Watermarks for Language Models
    Miranda Christ, et al. in Proceedings of the 37th Annual Conference on Learning Theory (COLT 2024). [Paper]
  3. An Unforgeable Publicly Verifiable Watermark for Large Language Models
    Aiwei Liu, Leyi Pan, Xuming Hu, et al. ICLR 2024. [Paper]
  4. A Watermark for Large Language Models
    John Kirchenbauer, Jonas Geiping, Yuxin Wen, et al. ICML 2023. [Paper]
  5. Publicly-Detectable Watermarking for Language Models
    Jaiden Fairoze, Sanjam Garg, Somesh Jha, et al. arXiv 2023. [Paper]

Statistical Provenance

  1. A Watermark for Large Language Models [Paper]

4 LLM for Data Management

⬆️top

4.1 LLM for Data Manipulation

4.1.1 LLM for Data Cleaning

Data Standardization
  1. Exploring the Feasibility of Automated Data Standardization using Large Language Models for Seamless Positioning

    Lee, Max JL, et al. 2024 14th International Conference on Indoor Positioning and Indoor Navigation (IPIN). IEEE, 2024. [Paper]

  2. Language Models Enable Simple Systems for Generating Structured Views of Heterogeneous Data Lakes
    Simran Arora, et al. Proceedings of the VLDB Endowment, Volume 17, Issue 2, 2024. [Paper]

  3. CleanAgent: Automating Data Standardization with LLM-based Agents
    Danrui Qi, Jiannan Wang. arXiv 2024. [Paper]

  4. AutoDCWorkflow: LLM-based Data Cleaning Workflow Auto-Generation and Benchmark
    Lan Li, Liri Fang, Vetle I. Torvik. arXiv 2024. [Paper]

  5. LLMs with User-defined Prompts as Generic Data Operators for Reliable Data Processing
    Luyi Ma, et al. 1st IEEE International Workshop on Data Engineering and Modeling for AI (DEMAI), IEEE BigData 2023. [Paper]

  6. Large language models as data preprocessors.

    Zhang, Haochen, et al. arXiv 2023. [Paper]

Data Error Processing
  1. Exploring LLM Agents for Cleaning Tabular Machine Learning Datasets
    Tommaso Bendinelli, Artur Dox, Christian Holz. ICLR 2025 Workshop on Foundation Models in the Wild. [Paper]

  2. Exploring LLM Agents for Cleaning Tabular Machine Learning Datasets

    Bendinelli, Tommaso, Artur Dox, and Christian Holz. arXiv 2025. [Paper]

  3. ZeroED: Hybrid Zero-shot Error Detection through Large Language Model Reasoning

    Ni, Wei, et al. arXiv 2025. [Paper]

  4. GIDCL: A Graph-Enhanced Interpretable Data Cleaning Framework with Large Language Models
    Mengyi Yan, et al. Proceedings of the ACM on Management of Data, Volume 2, Issue 6, 2024. [Paper]

  5. Multi-News+: Cost-efficient Dataset Cleansing via LLM-based Data Annotation
    Juhwan Choi, Jungmin Yun, Kyohoon Jin, et al. EMNLP 2024. [Paper]

  6. Data Cleaning Using Large Language Models
    Shuo Zhang, Zezhou Huang, Eugene Wu. arXiv 2024. [Paper]

  7. LLMClean: Context-Aware Tabular Data Cleaning via LLM-Generated OFDs
    Fabian Biester, Mohamed Abdelaal, Daniel Del Gaudio. arXiv 2024. [Paper]

  8. Anomaly Detection of Tabular Data Using LLMs

    Li, Aodong, et al. arXiv 2024. [Paper]

  9. Cleaning Semi-Structured Errors in Open Data Using Large Language Models

    M. Mondal, J. Audiffren, L. Dolamic, et al, 2024 11th IEEE Swiss Conference on Data Science (SDS). [Paper]

  10. IterClean: An Iterative Data Cleaning Framework with Large Language Models
    Wei Ni, et al. Proceedings of the ACM Turing Award Celebration Conference - China 2024. [Paper]

Data Imputation
  1. Does Prompt Design Impact Quality of Data Imputation by LLMs?

    Srinivasan, Shreenidhi, and Lydia Manikonda. arXiv 2025. [Paper]

  2. On LLM-Enhanced Mixed-Type Data Imputation with High-Order Message Passing

    Wang, Jianwei, et al. arXiv 2025. [Paper]

  3. RetClean: Retrieval-Based Data Cleaning Using LLMs and Data Lakes
    Zan Ahmad Naeem, et al. VLDB Endowment 2024. [Paper]

  4. Data Augmentation using Large Language Models: Data Perspectives, Learning Paradigms and Challenges

    Ding, Bosheng, et al. arXiv 2024. [Paper]

  5. A Context-Aware Approach for Enhancing Data Imputation with Pre-trained Language Models

    Hayat, Ahatsham, and Mohammad Rashedul Hasan. arXiv 2024. [Paper]

4.1.2 LLM for Data Integration

Entity Matching
  1. A Deep Dive Into Cross-Dataset Entity Matching with Large and Small Language Models

    Zhang, Zeyu, et al. International Conference on Extending Database Technology (EDBT) 2025. [Paper]

  2. Large Language Models for Data Discovery and Integration: Challenges and Opportunities

    Freire, Juliana, et al. IEEE Data Eng. Bull. 49(1): 3-31 (2025). [Paper]

  3. Entity matching using large language models
    Ralph Peeters, Christian Bizer. EDBT 2025. [Paper]

  4. Match, Compare, or Select? An Investigation of Large Language Models for Entity Matching
    Tianshu Wang, Hongyu Lin, Xiaoyang Chen, et al. COLING 2025. [Paper]

  5. Cost-Effective In-Context Learning for Entity Resolution: A Design Space Exploration
    Meihao Fan, Xiaoyue Han, Ju Fan, et al. ICDE 2024. [Paper]

  6. KcMF: A Knowledge-compliant Framework for Schema and Entity Matching with Fine-tuning-free LLMs
    Yongqin Xu, Huan Li, Ke Chen, Lidan Shou. arXiv 2024. [Paper]

  7. Jellyfish: A Large Language Model for Data Preprocessing
    Haochen Zhang, Yuyang Dong, Chuan Xiao, et al. EMNLP 2024. [Paper]

  8. Fine-tuning Large Language Models for Entity Matching

    Steiner, Aaron, Ralph Peeters, et al. arXiv 2024. [Paper]

Schema Matching
  1. SCHEMORA: Schema Matching via Multi-stage Recommendation and Metadata Enrichment using Off-the-Shelf LLMs
    Osman Erman Gungor, Derak Paulsen, William Kang. arXiv 2025. [Paper]

  2. Towards Scalable Schema Mapping using Large Language Models

    Buss, Christopher, et al. arXiv 2025. [Paper]

  3. Knowledge Graph-based Retrieval-Augmented Generation for Schema Matching
    Chuangtao Ma, Sriom Chakrabarti, Arijit Khan, et al. arXiv 2025. [Paper]

  4. Interactive Data Harmonization with LLM Agents
    Aécio Santos, Eduardo H. M. Pena, Roque Lopez, et al. arXiv 2025. [Paper]

  5. Schema Matching with Large Language Models: an Experimental Study
    Marcel Parciak, Brecht Vandevoort, Frank Neven, et al. TaDA 2024 Workshop, collocated with VLDB 2024. [Paper]

  6. Magneto: Combining Small and Large Language Models for Schema Matching
    Yurong Liu, Eduardo Pena, Aecio Santos, et al. VLDB Endowment 2024. [Paper]

  7. Agent-OM: Leveraging LLM Agents for Ontology Matching Zhangcheng Qiang, et al. Proceedings of the VLDB Endowment, Volume 18, Issue 3, 2024. [Paper]

  8. Matchmaker: Self-Improving Large Language Model Programs for Schema Matching

    Seedat, Nabeel, and Mihaela van der Schaar. arXiv 2024. [Paper]

  9. TableGPT2: A Large Multimodal Model with Tabular Data Integration

    Su, Aofeng, et al. arXiv 2024. [Paper]

4.1.3 LLM for Data Discovery

  1. ArcheType: A Novel Framework for Open-Source Column Type Annotation using Large Language Models
    Benjamin Feuer, Yurong Liu, Chinmay Hegde, et al. VLDB 2024. [Paper]
Data Profiling
  1. Flexible metadata harvesting for ecology using large language models
    Zehao Lu, Thijs L van der Plas, Parinaz Rashidi, et al. arXiv 2025. [Paper]

  2. Pneuma: Leveraging LLMs for Tabular Data Representation and Retrieval in an End-to-End System
    Muhammad Imam Luthfi Balaka, David Alexander, Qiming Wang, et al. SIGMOD 2025. [Paper]

  3. AutoDDG: Automated Dataset Description Generation using Large Language Models
    Haoxiang Zhang, Yurong Liu, Wei-Lun (Allen) Hung, et al. arXiv 2025. [Paper]

  4. LEDD: Large Language Model-Empowered Data Discovery in Data Lakes
    Qi An, Chihua Ying, Yuqing Zhu, et al. arXiv 2025. [Paper]

  5. LLM-Aided Customizable Profiling of Code Data Based On Programming Language Concepts

    Thorat, Pankaj, et al. arXiv 2025. [Paper]

  6. Cocoon: Semantic Table Profiling Using Large Language Models

    Huang, Zezhou, et al. Proceedings of the 2024 Workshop on Human-In-the-Loop Data Analytics. 2024. [Paper]

Data Annotation
  1. LLMs as Data Annotators: How Close Are We to Human Performance
    Haq, Muhammad Uzair Ul, Davide Rigoni, et al. arXiv 2025. [Paper]

  2. Birdie: Natural Language-Driven Table Discovery Using Differentiable Search Index
    Yuxiang Guo, Zhonghao Hu, Yuren Mao, et al. VLDB 2025. [Paper]

  3. Mind the Data Gap: Bridging LLMs to Enterprise Data Integration
    Moe Kayali, Fabian Wenz, Nesime Tatbul, et al. CIDR 2025. [Paper]

  4. Evaluating Knowledge Generation and Self-Refinement Strategies for LLM-based Column Type Annotation
    Keti Korini, Christian Bizer. arXiv 2025. [Paper]

  5. Columbo: Expanding Abbreviated Column Names for Tabular Data Using Large Language Models
    Ting Cai, Stephen Sheen, AnHai Doan. arXiv 2025. [Paper]

  6. An LLM Agent-Based Complex Semantic Table Annotation Approach
    Yilin Geng, Shujing Wang, Chuan Wang, et al. arXiv 2025. [Paper]

  7. Open-Source LLMs for Text Annotation: A Practical Guide for Model Setting and Fine-Tuning

    Alizadeh, Meysam, et al. Journal of Computational Social Science 8.1 (2025): 1-25. [Paper]

  8. Prompt Candidates, then Distill: A Teacher-Student Framework for LLM-driven Data Annotation

    Xia, Mingxuan, et al. arXiv 2025. [Paper]

  9. Evaluating how LLM annotations represent diverse views on contentious topics

    Brown, Megan A., et al. arXiv 2025. [Paper]

  10. CHORUS: Foundation Models for Unified Data Discovery and Exploration
    Moe Kayali, et al. Proceedings of the VLDB Endowment, Volume 17, Issue 8, 2024. [Paper]

  11. RACOON: An LLM-based Framework for Retrieval-Augmented Column Type Annotation with a Knowledge Graph
    Lindsey Linxi Wei, Guorui Xiao, Magdalena Balazinska. arXiv 2024. [Paper]

  12. AutoLabel: Automated Textual Data Annotation Method Based on Active Learning and Large Language Model
    Ming, Xuran, et al. International Conference on Knowledge Science, Engineering and Management. 2024. [Paper]

  13. The Promises and Pitfalls of LLM Annotations in Dataset Labeling: a Case Study on Media Bias Detection

    Horych, Tomas, et al. arXiv 2024. [Paper]

  14. Large Language Models as Annotators: Enhancing Generalization of NLP Models at Minimal Cost

    Bansal, Parikshit, and Amit Sharma. arXiv 2023. [Paper]

4.2 LLM for Data System Optimization

⬆️top

4.2.1 LLM for Configuration Tuning

  1. ELMo-Tune-V2: LLM-Assisted Full-Cycle Auto-Tuning to Optimize LSM-Based Key-Value Stores
    Viraj Thakkar, Qi Lin, Kenanya Keandra Adriel Prasetyo, et al. arXiv 2025. [Paper]
  2. MLETune: Streamlining Database Knob Tuning via Multi-LLMs Experts Guided Deep Reinforcement Learning
    Wenlong Dong, Wei Liu, Rui Xi, et al. ICPADS 2024. [Paper]
Tuning Task-Aware Prompt Engineering
  1. λ-Tune: Harnessing Large Language Models for Automated Database System Tuning
    Victor Giannankouris, Immanuel Trummer. SIGMOD 2025. [Paper]
  2. LLMIdxAdvis: Resource-Efficient Index Advisor Utilizing Large Language Model
    Xinxin Zhao, Haoyang Li, Jing Zhang, et al. arXiv 2025. [Paper]
  3. LATuner: An LLM-Enhanced Database Tuning System Based on Adaptive Surrogate Model
    Chongjiong Fan, Zhicheng Pan, Wenwen Sun, et al. ECML PKDD 2024. [Paper]
  4. Is Large Language Model Good at Database Knob Tuning? A Comprehensive Experimental Evaluation
    Yiyan Li, Haoyang Li, Zhao Pu, et al. arXiv 2024. [Paper]
RAG Based Tuning Experience Enrichment
  1. Automatic Database Configuration Debugging using Retrieval-Augmented Language Models
    Sibei Chen, Ju Fan, Bin Wu, et al. Proceedings of the ACM on Management of Data, Volume 3, Issue 1, 2025. [Paper]
  2. GPTuner: A Manual-Reading Database Tuning System via GPT-Guided Bayesian Optimization
    Jiale Lao, Yibo Wang, Yufei Li, et al. VLDB 2024. [Paper]
Training Enhanced Tuning Goal Alignment
  1. E2ETune: End-to-End Knob Tuning via Fine-tuned Generative Language Model
    Xinmei Huang, Haoyang Li, Jing Zhang, et al. VLDB 2025. [Paper]
  2. DB-GPT: Large Language Model Meets Database
    Xuanhe Zhou, Zhaoyan Sun, Guoliang Li. Data Science and Engineering 2024. [Paper]
  3. HEBO: Heteroscedastic Evolutionary Bayesian Optimisation
    Alexander I. Cowen-Rivers, Wenlong Lyu, Zhi Wang, et al. NeurIPS 2020. [Paper]

4.2.2 LLM for Query Optimization

Optimization-Aware Prompt Engineering
  1. E3-Rewrite: Learning to Rewrite SQL for Executability, Equivalence,and Efficiency
    Dongjie Xu, Yue Cui, Weijie Shi, et al. arXiv 2025. [Paper]
  2. LLM4Hint: Leveraging Large Language Models for Hint Recommendation in Offline Query Optimization
    Suchen Liu, Jun Gao, Yinjun Han, et al. arXiv 2025. [Paper]
  3. QUITE: A Query Rewrite System Beyond Rules with LLM Agents
    Yuyang Song, Hanxu Yan, Jiale Lao, et al. arXiv 2025. [Paper]
  4. Can Large Language Models Be Query Optimizer for Relational Databases?
    Jie Tan, Kangfei Zhao, Rui Li, et al. arXiv 2025. [Paper]
  5. A Query Optimization Method Utilizing Large Language Models
    Zhiming Yao, Haoyang Li, Jing Zhang, et al. arXiv 2025. [Paper]
  6. Query Rewriting via LLMs
    Sriram Dharwada, Himanshu Devrani, Jayant Haritsa, et al. arXiv 2025. [Paper]
  7. DB-GPT: Large Language Model Meets Database [Paper]
  8. LLM-R2: A Large Language Model Enhanced Rule-Based Rewrite System for Boosting Query Efficiency
    Zhaodonghui Li, Haitao Yuan, Huiming Wang, et al. VLDB 2024. [Paper]
  9. The Unreasonable Effectiveness of LLMs for Query Optimization
    Peter Akioyamen, Zixuan Yi, Ryan Marcus. ML for Systems Workshop at NeurIPS 2024. [Paper]
  10. R-Bot: An LLM-based Query Rewrite System
    Zhaoyan Sun, Xuanhe Zhou, Guoliang Li. arXiv 2024. [Paper]
  11. Query Rewriting via Large Language Models
    Jie Liu, Barzan Mozafari. arXiv 2024. [Paper]

4.2.3 LLM for Anomaly Diagnosis

Manually Crafted Prompts for Anomaly Diagnosis
  1. DBG-PT: A Large Language Model Assisted Query Performance Regression Debugger
    Victor Giannakouris, Immanuel Trummer. Proceedings of the VLDB Endowment, Volume 17, Issue 12, 2024. [Paper]
RAG Based Diagnosis Experience Enrichment
  1. DBAIOps: A Reasoning LLM-Enhanced Database Operation and Maintenance System using Knowledge Graphs
    Wei Zhou, Peng Sun, Xuanhe Zhou, et al. arXiv 2025. [Paper]
  2. Query Performance Explanation through Large Language Model for HTAP Systems
    Haibo Xiu, Li Zhang, Tieying Zhang, et al. ICDE 2025. [Paper]
  3. D-Bot: Database Diagnosis System using Large Language Models
    Xuanhe Zhou, Guoliang Li, Zhaoyan Sun, et al. Proceedings of the VLDB Endowment, Volume 17, Issue 10. 2024. [Paper]
  4. LLM As DBA
    Xuanhe Zhou, Guoliang Li, Zhiyuan Liu. arXiv 2023. [Paper]
Multi-Agent Mechanism for Collaborative Diagnosis
  1. GaussMaster: An LLM-based Database Copilot System
    Wei Zhou, Ji Sun, Xuanhe Zhou, et al. arXiv 2025. [Paper]
  2. D-Bot: Database Diagnosis System using Large Language Models [Paper]
  3. Panda: Performance Debugging for Databases using LLM Agents
    Vikramank Singh, Kapil Eknath Vaidya, Vinayshekhar Bannihatti Kumar, et al. CIDR 2024. [Paper]
  4. LLM As DBA [Paper]
Localized LLM Enhancement via Specialized FineTuning
  1. D-Bot: Database Diagnosis System using Large Language Models [Paper]
  2. LLM for Data Management
    Guoliang Li, Xuanhe Zhou, Xinyang Zhao. PVLDB 17(12). 2024. [Paper]
  3. LLM-Enhanced Data Management
    Xuanhe Zhou, Xinyang Zhao, Guoliang Li. arXiv 2024. [Paper]

5 LLM as Data Analyst

5.1 LLM for Structured Data Analysis

5.1.1 Relational Data

  1. A relational model of data for large shared data banks. [Paper]

  2. Multilinear tensor regression for longitudinal relational data [Paper]

  3. Probabilistic classification and clustering in relational data [Paper]

  4. Outlier detection in relational data: A case study in geographical information systems [Paper]

NL2SQL
  1. Finsql: Model-agnostic llms-based text-to-sql framework for financial analysis [Paper]

  2. Pet-sql: A prompt-enhanced two-round refinement of text-to-sql with cross-consistency [Paper]

  3. Chess: Contextual harnessing for efficient sql synthesis [Paper]

  4. Codes: Towards building open-source language models for text-to-sql [Paper]

  5. Combining small language models and large language models for zero-shot nl2sql [Paper]

  6. Cracking SQL Barriers: An llm-based dialect translation system [Paper]

  7. Cracksql: A hybrid sql dialect translation system powered by large language models [Paper]

  8. Din-sql: Decomposed in-context learning of text-to-sql with self-correction [Paper]

  9. Opensearch-sql: Enhancing text-to-sql with dynamic few-shot and consistency alignment [Paper]

  10. Bridging the semantic gap between text and table: A case study on nl2sql [Paper]

  11. The dawn of natural language to sql: Are we fully ready? [Paper]

  12. A Survey of Text-to-SQL in the Era of LLMs: Where Are We, and Where Are We Going? [Paper]

  13. Natural Language to SQL: State of the Art and Open Problems [Paper]

  14. A survey on employing large language models for text-to-sql tasks [Paper]

NL2Code
  1. Natural language to code generation in interactive data science notebooks [Paper]

  2. Palm: Scaling language modeling with pathways [Paper]

  3. Contextualized data-wrangling code generation in computational notebooks [Paper]

  4. Data interpreter: An llm agent for data science [Paper]

  5. Collaboration between intelligent agents and large language models: A novel approach for enhancing code generation capability [Paper]

  6. BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. [Paper]

LLM for Semantic Analysis.

Multi-Step QA.

  1. Tat-llm: A specialized language model for discrete reasoning over financial tab- ular and textual data [Paper]

  2. S3HQA: A three-stage approach for multi-hop text-table hybrid question answering [Paper]

  3. Plugging schema graph into multi-table qa: A human-guided framework for reducing llm reliance. [Paper]

  4. TaPERA: Enhancing faithfulness and interpretability in long-form table QA by content planning and execution-based reasoni [Paper]

  5. Reactable: Enhancing react for table question answering [Paper]

  6. Chain-of-table: Evolving tables in the reasoning chain for table understanding [Paper]

End-to-End QA

  1. Table-gpt: Table-tuned gpt for diverse table tasks [Paper]

  2. Tablegpt2: A large multimodal model with tabular data integration [Paper]

  3. Cabinet: Content relevance based noise reduction for table question answering [Paper]

  4. Tablemaster: A recipe to advance table understanding with language models [Paper]

  5. Mmqa: Evaluating llms with multi-table multi-hop complex questions. [Paper]

  6. Multimodal table understanding [Paper]

  7. Improved baselines with visual instruction tuning [Paper]

  8. Tabpedia: Towards comprehensive visual table understanding with concept synergy [Paper]

  9. Judging llm-as-a-judge with mt-bench and chatbot arena. [Paper]

LLM for Time Series Analysis.
  1. Time series databases and influxdb [Paper]

  2. Towards cross-modality modeling for time series analytics: A survey in the llm era [Paper]

  3. A comparison of arima and lstm in forecasting time series [Paper]

  4. Association between forecasting models’ precision and nonlinear patterns of daily river flow time series [Paper]

  5. The performance of lstm and bilstm in forecasting time series [Paper]

  6. Hmckrautoencoder: An interpretable deep learning framework for time series analysis. [Paper]

TS2NL.

  1. Can large language models be anomaly detectors for time series? [Paper]

  2. Timerag: Boosting llm time series forecasting via retrieval-augmented generation. [Paper]

  3. Dynamic time warping algorithm review. [Paper]

  4. Temporal data meets llm–explainable financial time series forecasting. [Paper]

  5. Exploring large language models for climate forecasting [Paper]

  6. Timecap: Learning to contextualize, augment, and predict time series events with large language model agents [Paper]

  7. Explainable multi-modal time series prediction with llm-in-the-loop [Paper]

  8. From news to forecast: Integrating event analysis in llm-based time series forecasting with reflection [Paper]

Alignment.

  1. Time-llm: Time series forecasting by reprogramming large language models [Paper]

  2. Seed: A structural encoder for embedding-driven decoding in time series prediction with llms [Paper]

  3. Timecma: Towards llm-empowered multivariate time series forecasting via cross-modality alignment [Paper]

  4. Calf: Aligning llms for time series forecasting via cross-modal fine-tuning [Paper]

  5. S2IP-LLM: Semantic space informed prompt learning with LLM for time series forecasting [Paper]

  6. Llm4ts: Aligning pre-trained llms as data-efficient time-series forecasters [Paper]

  7. Large language models are few-shot multivariate time series classifiers. [Paper]

5.1.2 Graph Data Analysis

  1. A comparison of current graph database models [Paper]

Natural Language To Graph Analysis Query.

  1. Nat-nl2gql: A novel multi-agent framework for translating natural language to graph query language [Paper]

  2. r3-NL2GQL: A model coordination and knowledge graph alignment approach for NL2GQL [Paper]

  3. Aligning large language models to a domain-specific graph database for nl2gql [Paper]

  4. Graph learning in the era of llms: A survey from the perspective of data, models, and tasks [Paper]

  5. Leveraging biomolecule and natural language through multi-modal learning: A survey [Paper]

LLM-based Semantic Analysis.

  • Retrieval-Then-Reasoning.
  1. Subgraph retrieval enhanced model for multi-hop knowledge base question answering [Paper]

  2. Unikgqa: Unified retrieval and reasoning for solving multi-hop question answering over knowledge graph [Paper]

  3. G-retriever: Retrieval-augmented generation for textual graph understanding and question answering [Paper]

  • Execution-Then-Reasoning
  1. Interactive-kbqa: Multi-turn inter-actions for knowledge base question answering with large language models [Paper]

  2. Mcts-kbqa: Monte carlo tree search for knowledge base question answering [Paper]

  3. Flexkbqa: A flexible llm-powered framework for few-shot knowledge base question answering [Paper]https://ojs.aaai.org/index.php/AAAI/article/view/29823

Graph Task Based Fine-tuning Methods.

  1. Language is all a graph needs [Paper]

  2. Instruct-graph: Boosting large language models via graph-centric instruction tuning and preference alignment [Paper]

  3. Direct preference optimization: Your language model is secretly a reward model [Paper]

  4. Graphgpt: Graph instruction tuning for large language models [Paper]

  5. Inductive representation learning on large graphs [Paper]

  6. Semi-supervised classification with graph convolutional networks. [Paper]

  7. Glam: Fine-tuning large language models for domain knowledge graph alignment via neighborhood partitioning and generative sub-graph encoding [Paper]

  • Agent Based Methods.
  1. Structgpt: A general framework for large language model to reason over structured data [Paper]

  2. Kbqa-o1: Agentic knowledge base question answering with monte carlo tree search. [Paper]

  3. Call me when necessary: Llms can efficiently and faithfully reason over structured environments [Paper]

5.1.3 Structured Data Generation for LLM

  1. Compositional Semantic Parsing on Semi-Structured Tables [Paper]

  2. Spider: A large-scale human-labeled dataset for complex and cross-domain semantic parsing and text-to-sql task [Paper]

Relational Data Generation.
  1. REaLTabFormer: Generating Realistic Relational and Tabular Data using Transformers [Paper]

  2. Relational data generation with graph neural networks and latent diffusion models [Paper]

  3. Synthetic data generation of many-to-many datasets via random graph generation. [Paper]

  4. Mixed-type tabular data synthesis with score-based diffusion in latent space [Paper]

  5. Syntaxsqlnet: Syntax tree networks for complex and cross-domain text-to-sql task [Paper]

  6. Codes: Towards building open-source language models for text-to-sql [Paper]

  7. Itf-gan: Synthetic time series dataset generation and manipulation by interpretable features [Paper]

  8. ChatTS: Aligning Time Series with LLMs via Synthetic Data for Enhanced Understanding and Reasoning [Paper]

Graph Data Generation.
  1. A framework for large-scale synthetic graph dataset generation [Paper]

  2. A temporal knowledge graph generation dataset supervised distantly by large language models [Paper]

5.2 LLM for Semi-Structured Data Analysis

5.2.1 Markup Language

Markup Extraction.

  1. Language models enable simple systems for generating structured views of heterogeneous data lakes [Paper]

  2. Webformer: The web-page transformer for structure information extraction [Paper]

Markup Query.

  1. XPath Agent: An Efficient XPath Programming Agent Based on LLM for Web Crawler [Paper]

  2. Bridging the gap: Enabling natural language queries for nosql databases through text-to-nosql translation [Paper]

Markup Understanding.

  1. Dom-lm: Learning generalizable representations for html documents [Paper]

  2. Markuplm: Pre-training of text and markup language for visually-rich document understanding [Paper]

  3. Hierarchical multimodal pre-training for visually rich webpage understanding [Paper]

5.2.2 Semi-Structured Table

Table Representation.

  1. Tuta: Tree-based transformers for generally structured table pre-training [Paper]

  2. ST-Raptor: LLM-Powered Semi-Structured Table Question Answering [Paper]

  3. Reasoning and Retrieval for Complex Semi-structured Tables via Reinforced Relational Data Transformation [Paper]

  4. Auto-Tables: Synthesizing Multi-Step Transformations to Relationalize Tables without Using Examples [Paper]

  5. Can an LLM find its way around a Spreadsheet? [Paper]

Table Prompting.

  1. SpreadsheetLLM: encoding spreadsheets for large language models [Paper]

  2. HySem: A context length optimized LLM pipeline for unstructured tabular extraction [Paper]

Table Querying.

  1. SpreadsheetLLM: encoding spreadsheets for large language models [Paper]

  2. ST-Raptor: LLM-Powered Semi-Structured Table Question Answering [Paper]

5.3 LLM for Unstructured Data Analysis

5.3.1 Chart Analysis

Traditional Approaches

  1. DVQA: Understanding Data Visualizations viaQuestion Answering [Paper]

Chart Captioning

  1. Describing Complex Charts in Natural Language A Caption Generation System [Paper]

  2. An Architecture for Data-to-Text Systems[Paper]

  3. Chartthinker: A contextual chain-of-thought approach to optimized chart summarization[Paper]

  4. Chart-to-Text: Generating Natural Language Descriptions for Charts by Adapting the Transformer Model[Paper]

  5. FigCaps-HF: A Figure-to-Caption Generative Framework and Benchmark with Human Feedback[Paper]

  6. Unichart: A universal vision-language pretrained model for chart comprehension and reasoning[Paper]

Chart Question Answering

  1. ChartLlama: A Multimodal LLM for Chart Undestanding and Generation [Paper]

  2. ChartBench: A Benchmark for Complex Visual easoning in Charts [Paper]

  3. Evochart: A benchmark and a self-training approach towards real-world chart understanding[Paper]

  4. Chartinsights: Evaluating multimodal large language models for low-level chart question answering[Paper]

  5. Vizability: Enhancing chart accessibility with llm-based conversational interaction[Paper]

  6. Charts-of-Thought: Enhancing LLM Visualization Literacy Through Structured Data Extraction[Paper]

  7. ChartMoE: Mixture of Diversely Aligned Expert Connector for Chart Understanding[Paper]

  8. ChartGemma: Visual Instruction-tuning for Cart Reasoning in the Wild [Paper]

  9. mPLUG-Owl: Modularization Empowers Large Laguage Models with Multimodality [Paper]

Chart-to-Code

  1. ChartMimic: Evaluating LMM's Cross-Modal Reasoning Capability via Chart-to-Code Generation[Paper]

  2. Text2Chart31: Instruction Tuning for Chart Generation with Automatic Feedback[Paper]

  3. Breaking the SFT Plateau: Multimodal Structured Reinforcement Learning for Chart-to-Code Generation[Paper]

5.3.2 Video Analysis

Temporally-Anchored Approaches
  1. Timemarker: A versatile video-llm for long and short video understanding with superior temporal localization ability [Paper]

  2. Seq2time: Sequential knowledge transfer for video llm temporal grounding [Paper]

  3. Tempme: Video temporal token merging for efficient text-video retrieval [Paper]

  4. Video token merging for long-form video understanding [Paper]

  5. Grounded-videollm: Sharpening fine-grained temporal grounding in video large language models [Paper]

Instruction-Aware Relative Temporal Localization
  1. From image to video, what do we need in multimodal llms? [Paper]

  2. LLMs meet long video: Advancing long video comprehension with an interactive visual adapter in llms [Paper]

Video Emotional Analysis
  1. Predicting Team Well-Being through Face Video Analysis with AI [Paper]

  2. AI based multimodal emotion and behavior analysis of interviewee [Paper]

Object Detection
  1. Videorefer suite: Advancing spatial-temporal object understanding with video llm [Paper]

  2. Video summarisation with incident and context information using generative ai [Paper]

  3. Abnormal event detection in surveillance videos through LSTM auto-encoding and local minima assistance [Paper]

Gesture and Behavior Detection
  1. Utilizing multimodal large language models for video analysis of posture in studying collaborative learning: A case study [Paper]

  2. Artificial intelligence–powered 3D analysis of video-based caregiver-child interactions [Paper]

Video Data for LLM
  1. VideoITG: Multimodal Video Understanding with Instructed Temporal Grounding [Paper]

  2. Nuwa-infinity: Autoregressive over autoregressive generation for infinite visual synthesis [Paper]

  3. Text2video-zero: Text-to-image diffusion models are zero-shot video generators [Paper]

  4. Align your latents: High-resolution video synthesis with latent diffusion models [Paper]

  5. Sadtalker: Learning realistic 3d motion coefficients for stylized audio-driven single image talking face animation [Paper]

  6. DreamTalk: When Emotional Talking Head Generation Meets Diffusion Probabilistic Models [Paper]

  7. Disco: Disentangled control for realistic human dance generation [Paper]

  8. Imagen video: High definition video generation with diffusion models [Paper]

  9. Make-a-video: Text-to-video generation without text-video data [Paper]

5.3.3 Document Analysis

  1. M3DocRAG: Multi-modal Retrieval is What You Need for Multi-page Multi-document Understanding [Paper]

  2. SV-RAG: LoRA-Contextualizing Adaptation of MLLMs for Long Document Understanding [Paper]

  3. VisFocus: Prompt-Guided Vision Encoders for OCR-Free Dense Document Understanding [Paper]

  4. DocFormer: End-to-End Transformer for Document Understanding [Paper]

  5. VLCDoC: Vision-Language contrastive pre-training model for cross-Modal document classification [Paper]

  6. Efficient End-to-End Visual Document Understanding with Rationale Distillation [Paper]

  7. LayoutLMv3: Pre-training for Document AI with Unified Text and Image Masking [Paper]

  8. MMP: Towards Robust Multi-Modal Learning with Masked Modality Projection [Paper]

  9. Unifying Layout Generation with a Decoupled Diffusion Model [Paper]

  10. LayoutDM: Discrete Diffusion Model for Controllable Layout Generation [Paper]

  11. FUNSD: A Dataset for Form Understanding in Noisy Scanned Documents [Paper]

  12. ViLT: Vision-and-Language Transformer Without Convolution or Region Supervision [Paper]

  13. VisDoM: Multi-Document QA with Visually Rich Elements Using Multimodal Retrieval-Augmented Generation [Paper]

  14. Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks [Paper]

  15. MissModal: Increasing Robustness to Missing Modality in Multimodal Sentiment Analysis [Paper]

  16. CREPE: Coordinate-Aware End-to-End Document Parser [Paper]

  17. LTSim: Layout Transportation-based Similarity Measure for Evaluating Layout Generation [Paper]

  18. AesthetiQ: Enhancing Graphic Layout Design via Aesthetic-Aware Preference Alignment of Multi-modal Large Language Models [Paper]

  19. Automatic generation of scientific papers for data augmentation in document layout analysis [Paper]

  20. PosterLlama: Bridging Design Ability of Language Model to Content-Aware Layout Generation [Paper]

  21. LayoutCoT: Chain-of-Thought Prompting for Layout Generation [Paper]

  22. SciPostLayout: A Dataset for Layout Analysis and Layout Generation of Scientific Posters [Paper]

  23. OmniParser: A Unified Framework for Text Spotting, Key Information Extraction and Table Recognition [Paper]

  24. DLAFormer: An End-to-End Transformer For Document Layout Analysis [Paper]

  25. DocLLM: A layout-aware generative language model for multimodal document understanding [Paper]

  26. LayoutLM: Pre-training of Text and Layout for Document Image Understanding [Paper]

  27. LayoutLMv2: Multi-modal Pre-training for Visually-Rich Document Understanding [Paper]

  28. Corrective Retrieval Augmented Generation [Paper]

  29. RAFT: Adapting Language Model to Domain Specific RAG [Paper]

  30. VASCAR: Content-Aware Layout Generation via Visual-Spatial Self-Correction [Paper]

5.3.4 Program Analysis

  1. Automatic Semantic Augmentation of Language Model Prompts (for Code Summarization) [Paper]

  2. Teaching Large Language Models to Self-Debug [Paper]

  3. Syntax-directed variational autoencoder for structured data [Paper]

  4. Large Language Models are Few-Shot Summarizers: Multi-Intent Comment Generation via In-Context Learning [Paper]

  5. FT2Ra: A Fine-Tuning-Inspired Approach to Retrieval-Augmented Code Completion [Paper]

  6. Composing graphical models with neural networks for structured representations and fast inference [Paper]

  7. Kimina-Prover Preview: Towards Large Formal Reasoning Models with Reinforcement Learning [Paper]

  8. REPOFUSE: Repository-Level Code Completion with Fused Dual Context [Paper]

  9. Goedel-Prover: A Frontier Model for Open-Source Automated Theorem Proving [Paper]

  10. Software Vulnerability Detection with GPT and In-Context Learning [Paper]

  11. Pre-training by Predicting Program Dependencies for Vulnerability Analysis Tasks [Paper]

  12. WizardCoder: Empowering Code Large Language Models with Evol-Instruct [Paper]

  13. SCLA: Automated Smart Contract Summarization via LLMs and Semantic Augmentation [Paper]

  14. Self-Instruct: Aligning Language Models with Self-Generated Instructions [Paper]

  15. Magicoder: Empowering Code Generation with OSS-Instruct [Paper]

  16. Repoformer: Selective Retrieval for Repository-Level Code Completion [Paper]

  17. DeepSeek-Prover: Advancing Theorem Proving in LLMs through Large-Scale Synthetic Data [Paper]

  18. Vulnerability Detection by Learning From Syntax-Based Execution Paths of Code [Paper]

  19. RepoCoder: Repository-Level Code Completion Through Iterative Retrieval and Generation [Paper]

  20. Large Language Model for Vulnerability Detection: Emerging Results and Future Directions [Paper]

5.3.5 3D Model Analysis

3D-Language Fusion
  1. 3d-llm: Injecting the 3d world into large language models [Paper]

  2. 3ur-llm: An end-to-end multimodal large language model for 3d scene understanding [Paper]

  3. Towards 3d molecule-text interpretation in language models [Paper]

  4. Proteinchat: Towards achieving chatgpt-like functionalities on protein 3d structures [Paper]

  5. Protchatgpt: Towards understanding proteins with large language models [Paper]

3D-Derived Task Enhancement
  1. Do Large Language Models Truly Understand Geometric Structures? [Paper]

  2. 3DSMILES-GPT: 3D molecular pocket-based generation with token-only large language model [Paper]

  3. SMILES, a chemical language and information system. 1. Introduction to methodology and encoding rules [Paper]

  4. ProtChat: An AI Multi-Agent for Automated Protein Analysis Leveraging GPT-4 and Protein Language Model [Paper]

  5. A multimodal protein representation framework for quantifying transferability across biochemical downstream tasks [Paper]

Cross-modal Capability Refinement
  1. Self-supervised image-based 3d model retrieval [Paper]

  2. Llmi3d: Empowering llm with 3d perception from a single 2d image [Paper]

3-D data for LLM
  1. Fantasia3d: Disentangling geometry and appearance for high-quality text-to-3d content creation [Paper]

  2. Sweetdreamer: Aligning geometric priors in 2d diffusion for consistent text-to-3d [Paper]

  3. Richdreamer: A generalizable normal-depth diffusion model for detail richness in text-to-3d [Paper]

  4. Zero-1-to-3: Zero-shot one image to 3d object [Paper]

  5. Hunyuan3d 1.0: A unified framework for text-to-3d and image-to-3d generation [Paper]

  6. Craftsman3d: High-fidelity mesh generation with 3d native generation and interactive geometry refiner [Paper]

  7. Direct3d: Scalable image-to-3d generation via 3d latent diffusion transformer [Paper]

  8. Meshanything: Artist-created mesh generation with autoregressive transformers [Paper]

  9. Llama-mesh: Unifying 3d mesh generation with language models [Paper]

5.4 LLM for Heterogeneous Data Analysis

5.4.1 LLM for Modality Alignment

  1. Unicorn: a unified multi-tasking matching model [Paper]

  2. Symphony: Towards Natural Language Query Answering over Multi-modal Data Lakes. [Paper]

5.4.2 LLM for Heterogeneous Data Retrieval

  1. Lotus: Enabling semantic queries with llms over tables of unstructured and structured data [Paper]

  2. Towards Operationalizing Heterogeneous Data Discovery [Paper]

  3. CAESURA: Language Models as Multi-Modal Query Planners [Paper]

5.4.2 Heterogeneous Data Analysis Agents

  1. Explainable Multi-Modal Data Exploration in Natural Language via LLM Agent [Paper]

  2. An Interactive Multi-modal Query Answering System with Retrieval-Augmented Large Language Models [Paper]

  3. Must: An effective and scalable framework for multimodal search of target modality [Paper]