From e70ec252663fca45cba7a3c8a6505a0f27ff9a5f Mon Sep 17 00:00:00 2001 From: Ahmed Nabil <72295771+AI-Ahmed@users.noreply.github.com> Date: Thu, 20 Oct 2022 16:00:41 +0200 Subject: [PATCH] Error in the `Word2Vec` model's parameters. Gensim Word2Vec doesn't have `size` as a parameter. Instead, it has `vector_size` for the embedding vector size. --- week05_nlp/seminar.ipynb | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/week05_nlp/seminar.ipynb b/week05_nlp/seminar.ipynb index c3de6e7bd..23b3e2e1a 100644 --- a/week05_nlp/seminar.ipynb +++ b/week05_nlp/seminar.ipynb @@ -118,9 +118,9 @@ "source": [ "from gensim.models import Word2Vec\n", "model = Word2Vec(data_tok, \n", - " size=32, # embedding vector size\n", - " min_count=5, # consider words that occured at least 5 times\n", - " window=5).wv # define context as a 5-word window around the target word" + " vector_size=32, # embedding vector size\n", + " min_count=5, # consider words that occured at least 5 times\n", + " window=5).wv # define context as a 5-word window around the target word" ] }, {