-
Updated
Jul 8, 2022 - Python
nlu
Here are 476 public repositories matching this topic...
-
Updated
Jul 8, 2022 - TypeScript
-
Updated
Sep 7, 2021 - Python
-
Updated
Jul 4, 2022 - JavaScript
-
Updated
Nov 17, 2021 - Python
-
Updated
Jun 29, 2022 - Python
-
Updated
Jul 5, 2022 - Python
-
Updated
May 26, 2022 - Python
-
Updated
May 30, 2019 - Clojure
-
Updated
Jun 22, 2022 - C#
-
Updated
Jun 13, 2021
-
Updated
Jul 6, 2022 - JavaScript
-
Updated
Oct 5, 2017 - OpenEdge ABL
-
Updated
Jul 8, 2022 - TypeScript
Description
Randomly the matrix connector will crash when connected to a public room.
Steps to Reproduce
Create a skill and point it to #geeklab:linuxdelta.com and wait for people t
-
Updated
Jun 21, 2022 - Python
-
Updated
Jun 23, 2022 - C#
-
Updated
Jul 9, 2022 - Python
-
Updated
Jun 1, 2022 - Java
-
Updated
Jul 5, 2022 - Python
-
Updated
Feb 20, 2019 - TypeScript
-
Updated
Jun 1, 2022
-
Updated
Dec 9, 2021 - C++
-
Updated
Apr 20, 2022 - Python
-
Updated
Jun 15, 2022 - Emacs Lisp
-
Updated
Dec 20, 2019
-
Updated
Jun 14, 2022 - Python
Improve this page
Add a description, image, and links to the nlu topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the nlu topic, visit your repo's landing page and select "manage topics."

Formed in 2009, the Archive Team (not to be confused with the archive.org Archive-It Team) is a rogue archivist collective dedicated to saving copies of rapidly dying or deleted websites for the sake of history and digital heritage. The group is 100% composed of volunteers and interested parties, and has expanded into a large amount of related projects for saving online and digital history.

Description
While using tokenizers.create with the model and vocab file for a custom corpus, the code throws an error and is not able to generate the BERT vocab file
Error Message
ValueError: Mismatch vocabulary! All special tokens specified must be control tokens in the sentencepiece vocabulary.
To Reproduce
from gluonnlp.data import tokenizers
tokenizers.create('spm', model_p