-
BUPT
- Beijin Province, China
Highlights
- Pro
Lists (1)
Sort Name ascending (A-Z)
Stars
Bioregulatory Event Extraction using Large Language Models: A Case Study of Rice Literature
A curated list of practical guide resources of Medical LLMs (Medical LLMs Tree, Tables, and Papers)
This project aim to reproduce Sora (Open AI T2V model), we wish the open source community contribute to this project.
We unified the interfaces of instruction-tuning data (e.g., CoT data), multiple LLMs and parameter-efficient methods (e.g., lora, p-tuning) together for easy use. We welcome open-source enthusiasts…
An Open-sourced Knowledgable Large Language Model Framework.
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
MNBVC(Massive Never-ending BT Vast Chinese corpus)超大规模中文语料集。对标chatGPT训练的40T数据。MNBVC数据集不但包括主流文化,也包括各个小众文化甚至火星文的数据。MNBVC数据集包括新闻、作文、小说、书籍、杂志、论文、台词、帖子、wiki、古诗、歌词、商品介绍、笑话、糗事、聊天记录等一切形式的纯文本中文数据。
Example models using DeepSpeed
BELLE: Be Everyone's Large Language model Engine(开源中文对话大模型)
FlagAI (Fast LArge-scale General AI models) is a fast, easy-to-use and extensible toolkit for large-scale model.
⭐️ NLP Algorithms with transformers lib. Supporting Text-Classification, Text-Generation, Information-Extraction, Text-Matching, RLHF, SFT etc.
Making large AI models cheaper, faster and more accessible
microsoft / Megatron-DeepSpeed
Forked from NVIDIA/Megatron-LMOngoing research training transformer language models at scale, including: BERT & GPT-2
Chinese Pre-Trained Language Models (CPM-LM) Version-I
stable diffusion webui colab
Code repository of the paper "CKConv: Continuous Kernel Convolution For Sequential Data" published at ICLR 2022. https://arxiv.org/abs/2102.02611
Implementation of BERT that could load official pre-trained models for feature extraction and prediction
Code for "Locate and Label: A Two-stage Identifier for Nested Named Entity Recognition", accepted at ACL 2021.
BERT-for-BioNLP-OST2019-AGAC-Task2
Named Entity Recognition as Dependency Parsing
karypis / DRKG
Forked from gnn4dr/DRKGA knowledge graph and a set of tools for drug repurposing
A free and unlimited python API for google translate.
PyTorch Re-Implementation of "The Sparsely-Gated Mixture-of-Experts Layer" by Noam Shazeer et al. https://arxiv.org/abs/1701.06538