word-embedding Questions

0

I have a csv file, it has 2 columns: class and text_data. I first extract biGram and TriGrams, then tried to use SVM on my data for classification. But it shows "TypeError: sequence item 0: ex...
Lotty asked 23/7, 2021 at 4:35

2

What's the right way to return a limited number of layers using the longformer API? Unlike this case in basic BERT, it's not clear to me from the return type how to get only the last N layers. So, ...
Alleged asked 5/10, 2020 at 23:57

2

Solved

I am using the SentenceTransformers library (here: https://pypi.org/project/sentence-transformers/#pretrained-models) for creating embeddings of sentences using the pretrained model bert-base-nli-m...
Coh asked 20/7, 2020 at 18:55

4

Solved

How do I load a pre-trained word-embedding into a Keras Embedding layer? I downloaded the glove.6B.50d.txt (glove.6B.zip file from https://nlp.stanford.edu/projects/glove/) and I'm not sure how t...
Telekinesis asked 8/2, 2018 at 3:30

3

Assuming that I want to update a pre-trained word-embedding matrix during training, is there a way to update only a subset of the word embedding matrix? I have looked into the Tensorflow API page ...
Bunkmate asked 4/3, 2016 at 18:26

3

Solved

I have a large pretrained Word2Vec model in gensim from which I want to use the pretrained word vectors for an embedding layer in my Keras model. The problem is that the embedding size is enormou...
Glace asked 18/6, 2018 at 17:32

3

I am trying use an embeddings module from tensorflow hub as servable. I am new to tensorflow. Currently, I am using Universal Sentence Encoder embeddings as a lookup to convert sentences to embeddi...
Underglaze asked 10/6, 2018 at 21:10

3

Does anyone know how to load a tsv file with embeddings generated from StarSpace into Gensim? Gensim documentation seems to use Word2Vec a lot and I couldn't find a pertinent answer. Thanks, Amul...
Diadromous asked 3/3, 2018 at 20:5

4

I am using gensim library for loading pre-trained word vectors from GoogleNews dataset. this dataset contains 3000000 word vectors each of 300 dimensions. when I want to load GoogleNews dataset, I ...
Hyperactive asked 23/5, 2018 at 0:2

1

Solved

I would like to use state-of-the-art LM T5 to get sentence embedding vector. I found this repository https://github.com/UKPLab/sentence-transformers As I know, in BERT I should take the first token...
Maureen asked 28/10, 2020 at 18:35

2

Solved

What is the difference between word2vec and glove? Are both the ways to train a word embedding? if yes then how can we use both?
Spinet asked 10/5, 2019 at 6:10

1

I am trying to do document embedding using BERT. The code I use is a combination of two sources. I use BERT Document Classification Tutorial with Code, and BERT Word Embeddings Tutorial. Below is t...

2

Are there any latest pre-trained multilingual word embeddings (multiple languages are jointly mapped to a same vector space)? I have looked at the following but they don't fit my needs: FastText...

1

Solved

I have trained fasttext model with Gensim over the corpus of very short sentences (up to 10 words). I know that my test set includes words that are not in my train corpus, i.e some of the words in ...
Eschalot asked 5/7, 2020 at 16:39

1

Solved

I am using BERT Word Embeddings for sentence classification task with 3 labels. I am using Google Colab for coding. My problem is, since I will have to execute the embedding part every time I resta...

2

I have sentences that I vectorize using sentence_vector() method of BiobertEmbedding python module (https://pypi.org/project/biobert-embedding/). For some group of sentences I have no problem but f...

1

Solved

I'm coming from Keras to PyTorch. I would like to create a PyTorch Embedding layer (a matrix of size V x D, where V is over vocabulary word indices and D is the embedding vector dimension) with Glo...
Combustible asked 9/6, 2020 at 20:28

1

Solved

So I'm trying to build a word embedding model but I keep getting this error. During training, the accuracy does not change and the val_loss remains "nan" The raw shape of the data is x.shape, y.s...
Verdieverdigris asked 7/5, 2020 at 11:17

3

Solved

I'm loading a language model from torch hub (CamemBERT a French RoBERTa-based model) and using it do embed some french sentences: import torch camembert = torch.hub.load('pytorch/fairseq', 'camem...
Preuss asked 25/11, 2019 at 11:36

2

Solved

I've been reading some NLP with Deep Learning papers and found Fine-tuning seems to be a simple but yet confusing concept. There's been the same question asked here but still not quite clear....
Meany asked 31/10, 2016 at 15:41

4

Solved

I am working on a recurrent language model. To learn word embeddings that can be used to initialize my language model, I am using gensim's word2vec model. After training, the word2vec model holds ...
Thrush asked 23/10, 2017 at 12:44

2

Solved

I thought mask_zero=True will output 0's when the input value is 0, so the following layers could skip computation or something. How does mask_zero works? Example: data_in = np.array([ [1, 2,...
Daffy asked 25/11, 2017 at 11:3

2

I am trying to get textual representation(or the closest word) of given word embedding using BERT. Basically I am trying to get similar functionality as in gensim: >>> your_word_vector = ...
Gabrielagabriele asked 22/1, 2020 at 18:0

1

Solved

I have seen both terms used while reading papers about BERT and ELMo so I wonder if there is a difference between them.
Xavler asked 23/1, 2020 at 11:20

2

Solved

I've been following Towards Data Science's tutorial about word2vec and skip-gram models, but I stumbled upon a problem that I cannot solve, despite searching about it for hours and trying a lot of ...
Pickerelweed asked 27/9, 2018 at 17:30

© 2022 - 2024 — McMap. All rights reserved.