YAYI 2 是中科闻歌研发的新一代开源大语言模型,采用了超过 2 万亿 Tokens 的高质量、多语言语料进行预训练。(Repo for YaYi 2 Chinese LLMs)
-
Updated
Apr 7, 2024 - Python
YAYI 2 是中科闻歌研发的新一代开源大语言模型,采用了超过 2 万亿 Tokens 的高质量、多语言语料进行预训练。(Repo for YaYi 2 Chinese LLMs)
Foundation Architecture for (M)LLMs
A curated list of pretrained sentence and word embedding models
An optimized deep prompt tuning strategy comparable to fine-tuning across scales and tasks
A plug-and-play library for parameter-efficient-tuning (Delta Tuning)
中文法律LLaMA (LLaMA for Chinese legel domain)
Code associated with the Don't Stop Pretraining ACL 2020 paper
Live Training for Open-source Big Models
Papers and Datasets on Instruction Tuning and Following. ✨✨✨
ACL'2023: DiffusionBERT: Improving Generative Masked Language Models with Diffusion Models
MWPToolkit is an open-source framework for math word problem(MWP) solvers.
[NeurIPS 2023] This is the code for the paper `Large Language Model as Attributed Training Data Generator: A Tale of Diversity and Bias`.
[NeurIPS 2021] COCO-LM: Correcting and Contrasting Text Sequences for Language Model Pretraining
On Transferability of Prompt Tuning for Natural Language Processing
BERT4ETH: A Pre-trained Transformer for Ethereum Fraud Detection (WWW23)
The official code for "TEMPO: Prompt-based Generative Pre-trained Transformer for Time Series Forecasting (ICLR 2024)". TEMPO is one of the very first open source Time Series Foundation Models for forecasting task v1.0 version.
[WWW 2022] Topic Discovery via Latent Space Clustering of Pretrained Language Model Representations
[KDD22] Official PyTorch implementation for "Towards Unified Conversational Recommender Systems via Knowledge-Enhanced Prompt Learning".
CODER: Knowledge infused cross-lingual medical term embedding for term normalization. [JBI, ACL-BioNLP 2022]
Official implementation of the ACL 2024: Scientific Inspiration Machines Optimized for Novelty
Add a description, image, and links to the pretrained-language-model topic page so that developers can more easily learn about it.
To associate your repository with the pretrained-language-model topic, visit your repo's landing page and select "manage topics."