word-embedding Questions
0
I have a csv file, it has 2 columns: class and text_data. I first extract biGram and TriGrams, then tried to use SVM on my data for classification. But it shows "TypeError: sequence item 0: ex...
Lotty asked 23/7, 2021 at 4:35
2
What's the right way to return a limited number of layers using the longformer API?
Unlike this case in basic BERT, it's not clear to me from the return type how to get only the last N layers.
So, ...
Alleged asked 5/10, 2020 at 23:57
2
Solved
I am using the SentenceTransformers library (here: https://pypi.org/project/sentence-transformers/#pretrained-models) for creating embeddings of sentences using the pretrained model bert-base-nli-m...
Coh asked 20/7, 2020 at 18:55
4
Solved
How do I load a pre-trained word-embedding into a Keras Embedding layer?
I downloaded the glove.6B.50d.txt (glove.6B.zip file from https://nlp.stanford.edu/projects/glove/) and I'm not sure how t...
Telekinesis asked 8/2, 2018 at 3:30
3
Assuming that I want to update a pre-trained word-embedding matrix during training, is there a way to update only a subset of the word embedding matrix?
I have looked into the Tensorflow API page ...
Bunkmate asked 4/3, 2016 at 18:26
3
Solved
I have a large pretrained Word2Vec model in gensim from which I want to use the pretrained word vectors for an embedding layer in my Keras model.
The problem is that the embedding size is enormou...
Glace asked 18/6, 2018 at 17:32
3
I am trying use an embeddings module from tensorflow hub as servable. I am new to tensorflow. Currently, I am using Universal Sentence Encoder embeddings as a lookup to convert sentences to embeddi...
Underglaze asked 10/6, 2018 at 21:10
3
Does anyone know how to load a tsv file with embeddings generated from StarSpace into Gensim? Gensim documentation seems to use Word2Vec a lot and I couldn't find a pertinent answer.
Thanks,
Amul...
Diadromous asked 3/3, 2018 at 20:5
4
I am using gensim library for loading pre-trained word vectors from GoogleNews dataset. this dataset contains 3000000 word vectors each of 300 dimensions. when I want to load GoogleNews dataset, I ...
Hyperactive asked 23/5, 2018 at 0:2
1
Solved
I would like to use state-of-the-art LM T5 to get sentence embedding vector.
I found this repository https://github.com/UKPLab/sentence-transformers
As I know, in BERT I should take the first token...
Maureen asked 28/10, 2020 at 18:35
2
Solved
What is the difference between word2vec and glove?
Are both the ways to train a word embedding? if yes then how can we use both?
Spinet asked 10/5, 2019 at 6:10
1
I am trying to do document embedding using BERT. The code I use is a combination of two sources. I use BERT Document Classification Tutorial with Code, and BERT Word Embeddings Tutorial. Below is t...
Scotch asked 1/8, 2020 at 20:52
2
Are there any latest pre-trained multilingual word embeddings (multiple languages are jointly mapped to a same vector space)?
I have looked at the following but they don't fit my needs:
FastText...
Bundestag asked 15/6, 2020 at 9:13
1
Solved
I have trained fasttext model with Gensim over the corpus of very short sentences (up to 10 words). I know that my test set includes words that are not in my train corpus, i.e some of the words in ...
Eschalot asked 5/7, 2020 at 16:39
1
Solved
I am using BERT Word Embeddings for sentence classification task with 3 labels. I am using Google Colab for coding. My problem is, since I will have to execute the embedding part every time I resta...
Limbo asked 3/7, 2020 at 7:51
2
I have sentences that I vectorize using sentence_vector() method of BiobertEmbedding python module (https://pypi.org/project/biobert-embedding/). For some group of sentences I have no problem but f...
Terrilyn asked 26/6, 2020 at 15:36
1
Solved
I'm coming from Keras to PyTorch. I would like to create a PyTorch Embedding layer (a matrix of size V x D, where V is over vocabulary word indices and D is the embedding vector dimension) with Glo...
Combustible asked 9/6, 2020 at 20:28
1
Solved
So I'm trying to build a word embedding model but I keep getting this error.
During training, the accuracy does not change and the val_loss remains "nan"
The raw shape of the data is
x.shape, y.s...
Verdieverdigris asked 7/5, 2020 at 11:17
3
Solved
I'm loading a language model from torch hub (CamemBERT a French RoBERTa-based model) and using it do embed some french sentences:
import torch
camembert = torch.hub.load('pytorch/fairseq', 'camem...
Preuss asked 25/11, 2019 at 11:36
2
Solved
I've been reading some NLP with Deep Learning papers and found Fine-tuning seems to be a simple but yet confusing concept. There's been the same question asked here but still not quite clear....
Meany asked 31/10, 2016 at 15:41
4
Solved
I am working on a recurrent language model. To learn word embeddings that can be used to initialize my language model, I am using gensim's word2vec model.
After training, the word2vec model holds ...
Thrush asked 23/10, 2017 at 12:44
2
Solved
I thought mask_zero=True will output 0's when the input value is 0, so the following layers could skip computation or something.
How does mask_zero works?
Example:
data_in = np.array([
[1, 2,...
Daffy asked 25/11, 2017 at 11:3
2
I am trying to get textual representation(or the closest word) of given word embedding using BERT. Basically I am trying to get similar functionality as in gensim:
>>> your_word_vector = ...
Gabrielagabriele asked 22/1, 2020 at 18:0
1
Solved
I have seen both terms used while reading papers about BERT and ELMo so I wonder if there is a difference between them.
Xavler asked 23/1, 2020 at 11:20
2
Solved
I've been following Towards Data Science's tutorial about word2vec and skip-gram models, but I stumbled upon a problem that I cannot solve, despite searching about it for hours and trying a lot of ...
Pickerelweed asked 27/9, 2018 at 17:30
© 2022 - 2024 — McMap. All rights reserved.