The Wayback Machine - https://web.archive.org/web/20210821112811/https://github.com/topics/gpt-2
Here are
344 public repositories
matching this topic...
Updated
Aug 8, 2021
Shell
An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.
Updated
Aug 14, 2021
Python
Chinese version of GPT2 training code, using BERT tokenizer.
Updated
Aug 1, 2021
Python
Toolkit for Machine Learning, Natural Language Processing, and Text Generation, in TensorFlow. This is part of the CASL project:
http://casl-project.ai/
Updated
Sep 17, 2020
Python
Kashgari is a production-level NLP Transfer learning framework built on top of tf.keras for text-labeling and text-classification, includes Word2Vec, BERT, and GPT2 Language Embedding.
Updated
Jul 9, 2021
Python
Open Source Pre-training Model Framework in PyTorch & Pre-trained Model Zoo
Updated
Aug 20, 2021
Python
GPT2 for Chinese chitchat/用于中文闲聊的GPT2模型(实现了DialoGPT的MMI思想)
Updated
Jun 16, 2021
Python
Large-scale pretraining for dialogue
Updated
Jul 1, 2021
Python
GPT2 for Multiple Languages, including pretrained models. GPT2 多语言支持, 15亿参数中文预训练模型
Updated
May 21, 2021
Python
🛸 Use pretrained transformers like BERT, XLNet and GPT-2 in spaCy
Updated
Aug 21, 2021
Python
Updated
Aug 19, 2021
TypeScript
A Large-scale Chinese Short-Text Conversation Dataset and Chinese pre-training dialog models
Updated
Feb 28, 2021
Python
Rust native ready-to-use NLP pipelines and transformer-based models (BERT, DistilBERT, GPT2,...)
Updated
Aug 20, 2021
Rust
Updated
Jan 12, 2021
Python
Simple Text-Generator with OpenAI gpt-2 Pytorch Implementation
Updated
Jul 8, 2019
Python
A curated list of NLP resources focused on BERT, attention mechanism, Transformer networks, and transfer learning.
Updated
Aug 15, 2021
Python
Medical Q&A with Deep Language Models
Updated
May 29, 2021
Jupyter Notebook
Awesome Pretrained Chinese NLP Models,高质量中文预训练模型集合
✍🏻 gpt2-client: Easy-to-use TensorFlow Wrapper for GPT-2 117M, 345M, 774M, and 1.5B Transformer Models 🤖 📝
Updated
Jun 24, 2021
Python
TextBox is an open-source library for building text generation system.
Updated
Aug 1, 2021
Python
Text-generation API via GPT-2 for Cloud Run
Updated
Jul 18, 2021
HTML
Your new Telegram buddy powered by transformers
Updated
Jan 25, 2021
Jupyter Notebook
EMNLP 2020: "Dialogue Response Ranking Training with Large-Scale Human Feedback Data"
Updated
Mar 2, 2021
Python
Updated
Aug 19, 2021
Java
Python script to download public Tweets from a given Twitter account into a format suitable for AI text generation.
Updated
May 21, 2020
Python
OpenAI GPT2 pre-training and sequence prediction implementation in Tensorflow 2.0
Updated
Jun 9, 2021
Python
a bot that generates realistic replies using a combination of pretrained GPT-2 and BERT models
Updated
Dec 2, 2020
Jupyter Notebook
Improve this page
Add a description, image, and links to the
gpt-2
topic page so that developers can more easily learn about it.
Curate this topic
Add this topic to your repo
To associate your repository with the
gpt-2
topic, visit your repo's landing page and select "manage topics."
Learn more
You can’t perform that action at this time.
You signed in with another tab or window. Reload to refresh your session.
You signed out in another tab or window. Reload to refresh your session.
I'm playing around with this wonderful code but I'm running into a curious issue when I try to train the model with my own data.
I replicated the
personachat_self_original.jsonfile structure and added my own data. I deleteddataset_cache_OpenAIGPTTokenizerfile but when I try to train, I get this error: