The Wayback Machine - https://web.archive.org/web/20200618142505/https://github.com/topics/natural-language-understanding
Skip to content
#

natural-language-understanding

Here are 388 public repositories matching this topic...

transformers
dhruvsakalley
dhruvsakalley commented Mar 2, 2020

When you look at the variables in the pretrained base uncased BERT the varibles look like list 1. When you do the training from scratch, 2 additional variables per layer are introduced, with suffixes adam_m and adam_v. It would be nice for someone to explain what these variables are? and what is their significance to the process of training?
If one were to manually initialize variables from a pri

ludwig
ZeroAurora
ZeroAurora commented Mar 6, 2020

Is your feature request related to a problem? Please describe.
Other related issues: #408 #251
I trained a Chinese model for spaCy, linked it to [spacy's package folder]/data/zh (using spacy link) and want to use that for ludwig. However, when I tried to set the config for ludwig, I received an error, which tell me that there is no way to load the Chinese model.

ValueError: Key ch
dipanjan77
dipanjan77 commented Aug 12, 2019

Description

Add a ReadMe file in the GitHub folder.
Explain usage of the Templates

Other Comments

Principles of NLP Documentation
Each landing page at the folder level should have a ReadMe which explains -
○ Summary of what this folder offers.
○ Why and how it benefits users
○ As applicable - Documentation of using it, brief description etc
Scenarios folder:

tokenizers
sai-prasanna
sai-prasanna commented Jan 23, 2020

Spacy has customizable word level tokenizers with rules for multiple languages. I think porting that to rust would add nicely to this package. Having a customizable uniform word level tokenization across platforms (client web, server) and languages would be beneficial. Currently, idk any clean way or whether it's even possible to write bindings for spacy cython.

Spacy Tokenizer Code

https:

stellarspot
stellarspot commented Apr 26, 2019

Run the code:

(use-modules
 (opencog)
 (opencog nlp)
 (opencog nlp relex2logic)
 (opencog ghost)
 (opencog ghost procedures)
 (opencog exec))

(define-public (where-somebody-work who-list)
 (display (string-append "argument: " (cog-name (car (cog-outgoing-set who-list))) "\n"))
 (Word "SomeCompany"))

;; Ghost Rules
(ghost-parse "
r: (where do _* work) '_0 work in ^where-
spark-nlp
ansorre
ansorre commented Jul 24, 2019

On home page of website: https://nlp.johnsnowlabs.com/ I read "Full Python, Scala, and Java support"

Unfortunately it's 3 days now I'm trying to use Spark NLP in Java without any success.

  • I cannot find Java API (JavaDoc) of the framework.
  • not event a single example in Java is available
  • I do not know Scala, I do not know how to convert things like:
    val testData = spark.createDataFrame(
lamthuy
lamthuy commented Apr 1, 2020

Hi,
When we try to tokenize the following sentence:

If we use spacy

a = spacy.load('en_core_web_lg')

doc = a("I like the link http://www.idph.iowa.gov/ohds/oral-health-center/coordinator")

list(doc)

We got

[I, like, the, link, http://www.idph.iowa.gov, /, ohds, /, oral, -, health, -, center, /, coordinator]

But if we use the Spacy transformer tokenizer:

agrimm
agrimm commented Mar 8, 2018

Quite a few issues or offers of patches seem to be related to people wanting to do "developer access token" requests, as opposed to "client access token" requests.

Although you mention the word "client" in the name of the repo and as one of the class names, it could easily be mistaken as merely the opposite of "server" rather than the opposite of "developer". Would it be possible to add a sente

Improve this page

Add a description, image, and links to the natural-language-understanding topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the natural-language-understanding topic, visit your repo's landing page and select "manage topics."

Learn more

You can’t perform that action at this time.