Lists (1)
Sort Name ascending (A-Z)
Stars
Code repo for the paper "LLM-QAT Data-Free Quantization Aware Training for Large Language Models"
Official PyTorch implementation of DistiLLM: Towards Streamlined Distillation for Large Language Models (ICML 2024)
Repo for the EMNLP'24 Paper "Dual-Space Knowledge Distillation for Large Language Models". A general white-box KD framework for both same-tokenizer and cross-tokenizer LLM distillation.
Awesome LLM pruning papers all-in-one repository with integrating all useful resources and insights.
[NeurIPS 2023] LLM-Pruner: On the Structural Pruning of Large Language Models. Support Llama-3/3.1, Llama-2, LLaMA, BLOOM, Vicuna, Baichuan, TinyLlama, etc.
GRAIN: Gradient-based Intra-attention Pruning on Pre-trained Language Models
This repository collects papers for "A Survey on Knowledge Distillation of Large Language Models". We break down KD into Knowledge Elicitation and Distillation Algorithms, and explore the Skill & V…
Survey on LLM Agents (Published on CoLing 2025)
EfficientQAT: Efficient Quantization-Aware Training for Large Language Models
Fully open reproduction of DeepSeek-R1
SciQAG is a novel framework for automatically generating high-quality science question-answer pairs from a large corpus of scientific literature using large language models.
Chat GPT autoblogger for Wordpress websites
Challenging BIG-Bench Tasks and Whether Chain-of-Thought Can Solve Them
[ACL 2023] PromptRank: Unsupervised Keyphrase Extraction Using Prompt
🔬 A curated list of awesome LLMs & deep learning strategies & tools in financial market.
Awesome LLM Papers and repos on very comprehensive topics.
Awesome list of Korean Large Language Models.
list of papers, code, and other resources
[ACL 2021] Code for our ACL 2021 paper "One2Set: Generating Diverse Keyphrases as a Set"
Large, curated set of benchmark datasets for evaluating automatic keyphrase extraction algorithms.
🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support
Code for "WR-One2Set: Towards Well-Calibrated Keyphrase Generation" (EMNLP 2022)
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Data for ArXiv 2023 paper "Is ChatGPT A Good Keyphrase Generator? A Preliminary Study".
Source code for "SimCKP: Simple Contrastive Learning of Keyphrase Representations", Findings of EMNLP 2023