natural-language-processing
Natural language processing (NLP) is a field of computer science that studies how computers and humans interact. In the 1950s, Alan Turing published an article that proposed a measure of intelligence, now called the Turing test. More modern techniques, such as deep learning, have produced results in the fields of language modeling, parsing, and natural-language tasks.
Here are 10,037 public repositories matching this topic...
-
Updated
Jun 13, 2022 - Python
-
Updated
May 9, 2022 - Python
-
Updated
May 13, 2022 - Jupyter Notebook
-
Updated
Jun 12, 2022 - Python
-
Updated
Jun 13, 2022 - Python
-
Updated
Jun 7, 2022 - Python
-
Updated
May 26, 2022
-
Updated
Jun 12, 2017
Change tensor.data to tensor.detach() due to
pytorch/pytorch#6990 (comment)
tensor.detach() is more robust than tensor.data.
-
Updated
Jun 13, 2022 - Python
Although the results look nice and ideal in all TensorFlow plots and are consistent across all frameworks, there is a small difference (more of a consistency issue). The result training loss/accuracy plots look like they are sampling on a lesser number of points. It looks more straight and smooth and less wiggly as compared to PyTorch or MXNet.
It can be clearly seen in chapter 6([CNN Lenet](ht
As mentioned in huggingface/datasets#2552 it would be nice to improve the error message when a dataset fails to build because there are duplicate example keys.
The current one is
datasets.keyhash.DuplicatedKeysError: FAILURE TO GENERATE DATASET !
Found duplicate Key: 48
Keys should be unique and deterministic in natureand we could have something
-
Updated
Jun 11, 2022
In gensim/models/fasttext.py:
model = FastText(
vector_size=m.dim,
vector_size=m.dim,
window=m.ws,
window=m.ws,
epochs=m.epoch,
epochs=m.epoch,
negative=m.neg,
negative=m.neg,
# FIXME: these next 2 lines read in unsupported FB FT modes (loss=3 softmax or loss=4 onevsall,
# or model=3 supervi-
Updated
Apr 28, 2022
-
Updated
Jun 10, 2022 - Python
Is your feature request related to a problem? Please describe.
I typically used compressed datasets (e.g. gzipped) to save disk space. This works fine with AllenNLP during training because I can write my dataset reader to load the compressed data. However, the predict command opens the file and reads lines for the Predictor. This fails when it tries to load data from my compressed files.
-
Updated
Jul 25, 2021 - Jupyter Notebook
Checking the Python files in NLTK with "python -m doctest" reveals that many tests are failing. In many cases, the failures are just cosmetic discrepancies between the expected and the actual output, such as missing a blank line, or unescaped linebreaks. Other cases may be real bugs.
If these failures could be avoided, it would become possible to improve CI by running "python -m doctest" each t
-
Updated
Dec 22, 2020 - Python
-
Updated
Apr 10, 2022 - HTML
Add T9 decoder
Hey Hackers of this spoopy month!
Welcome to the Ciphey repo(s)!
This issue requires you to add a decoder.
This wiki section walks you through EVERYTHING you need to know, and we've added some more links at the bottom of this issue to detail more about the decoder.
https://github.com/Ciphey/Ciphey/wiki#adding-your-own-crackers--decoders
-
Updated
Jun 11, 2022 - Java
-
Updated
Jun 13, 2022 - Python
-
Updated
Jul 9, 2021 - Python
-
Updated
Jun 9, 2022 - Python
-
Updated
Jun 13, 2022 - Java
Created by Alan Turing
- Wikipedia
- Wikipedia


Feature request
Is the addition of the 'OPTforSequenceClassification' class scheduled?
Is someone handling it?
When adding these functions, I wonder if it is possible to PR one by one, or if I have to PR all classes supported by other models.
Motivation
Added function of OPT class, which is being actively discussed recently
Your contribution
I personally use the forSequenceCla