[('afternoon', 0.8329864144325256), ('weekend', 0.7690818309783936), ('evening', 0.7469204068183899),
('saturday', 0.7191835045814514), ('night', 0.7091601490974426), ('friday', 0.6764787435531616),
('sunday', 0.6380082368850708), ('newspaper', 0.6365975737571716), ('summer', 0.6268560290336609),
('season', 0.6137701272964478)]

[('mail', 0.7432783842086792), ('contact', 0.6995242834091187), ('address', 0.6547545194625854),
('replies', 0.6502780318260193), ('mailed', 0.6334187388420105), ('request', 0.6262195110321045),
('sas', 0.6220622658729553), ('send', 0.6207413077354431), ('listserv', 0.617364227771759),
('compuserve', 0.5954489707946777)]






class Word2Vec(BaseWordEmbeddingsModel):
    """Train, use and evaluate neural networks described in
    Once you're finished training a model (=no more updates, only querying)
    store and use only the :class:`~gensim.models.keyedvectors.KeyedVectors` instance in `self.
     wv` to reduce memory.
    The model can be stored/loaded via its :meth:`` 
    :meth:`~gensim.models.word2vec.Word2Vec.load` methods.
    The trained word vectors can also be stored/loaded from a format compatible with the
    original word2vec implementation via `self.wv.save_word2vec_format`
    and :meth:`gensim.models.keyedvectors.KeyedVectors.load_word2vec_format`.
    Some important attributes are the following:
    wv : :class:`~gensim.models.keyedvectors.Word2VecKeyedVectors`
    This object essentially contains the mapping between words and embeddings. After 
     training, it can be used
    directly to query those embeddings in various ways. See the module level docstring for 
    vocabulary : :class:'~gensim.models.word2vec.Word2VecVocab'
    This object represents the vocabulary (sometimes called Dictionary in gensim) of the 
    Besides keeping track of all unique words, this object provides extra functionality, such as
    constructing a huffman tree (frequent words are closer to the root), or discarding 
     extremely rare words.
    trainables : :class:`~gensim.models.word2vec.Word2VecTrainables`
    This object represents the inner shallow neural network used to train the embeddings. The 
     semantics of the
    network differ slightly in the two available training modes (CBOW or SG) but you can think 
     of it as a NN with
    a single projection and hidden layer which we train on the corpus. The weights are then 
     used as our embeddings
    (which means that the size of the hidden layer is equal to the number of features `self.size`).
    def __init__(self, sentences=None, size=100, alpha=0.025, window=5, min_count=5, 
        max_vocab_size=None, sample=1e-3, seed=1, workers=3, min_alpha=0.0001, 
        sg=0, hs=0, negative=5, ns_exponent=0.75, cbow_mean=1, hashfxn=hash, iter=5, 
        trim_rule=None, sorted_vocab=1, batch_words=MAX_WORDS_IN_BATCH, 
         compute_loss=False, callbacks=(), 

        sentences : iterable of iterables, optional
            The `sentences` iterable can be simply a list of lists of tokens, but for larger corpora,
            consider an iterable that streams the sentences directly from disk/network.
            See :class:`~gensim.models.word2vec.BrownCorpus`, :class:`~gensim.models.
            or :class:`~gensim.models.word2vec.LineSentence` in :mod:`~gensim.models.
             word2vec` module for such examples.
            See also the `tutorial on data streaming in Python
            If you don't supply `sentences`, the model is left uninitialized -- use if you plan to 
             initialize it
            in some other way.
        size : int, optional
            Dimensionality of the word vectors.
        window : int, optional
            Maximum distance between the current and predicted word within a sentence.
        min_count : int, optional
            Ignores all words with total frequency lower than this.
        workers : int, optional
            Use these many worker threads to train the model (=faster training with multicore 
        sg : {0, 1}, optional
            Training algorithm: 1 for skip-gram; otherwise CBOW.
        hs : {0, 1}, optional
            If 1, hierarchical softmax will be used for model training.
            If 0, and `negative` is non-zero, negative sampling will be used.
        negative : int, optional
            If > 0, negative sampling will be used, the int for negative specifies how many "noise 
            should be drawn (usually between 5-20).
            If set to 0, no negative sampling is used.
        ns_exponent : float, optional
            The exponent used to shape the negative sampling distribution. A value of 1.0 
             samples exactly in proportion
            to the frequencies, 0.0 samples all words equally, while a negative value samples low-
             frequency words more
            than high-frequency words. The popular default value of 0.75 was chosen by the 
             original Word2Vec paper.
            More recently, in, Caselles-Dupré, Lesaint, & Royo-
             Letelier suggest that
            other values may perform better for recommendation applications.
        cbow_mean : {0, 1}, optional
            If 0, use the sum of the context word vectors. If 1, use the mean, only applies when 
             cbow is used.
        alpha : float, optional
            The initial learning rate.
        min_alpha : float, optional
            Learning rate will linearly drop to `min_alpha` as training progresses.
        seed : int, optional
            Seed for the random number generator. Initial vectors for each word are seeded with 
             a hash of
            the concatenation of word + `str(seed)`. Note that for a fully deterministically-
             reproducible run,
            you must also limit the model to a single worker thread (`workers=1`), to eliminate 
             ordering jitter
            from OS thread scheduling. (In Python 3, reproducibility between interpreter launches 
             also requires
            use of the `PYTHONHASHSEED` environment variable to control hash randomization).
        max_vocab_size : int, optional
            Limits the RAM during vocabulary building; if there are more unique
            words than this, then prune the infrequent ones. Every 10 million word types need 
             about 1GB of RAM.
            Set to `None` for no limit.
        max_final_vocab : int, optional
            Limits the vocab to a target vocab size by automatically picking a matching min_count. 
             If the specified
            min_count is more than the calculated min_count, the specified min_count will be 
            Set to `None` if not required.
        sample : float, optional
            The threshold for configuring which higher-frequency words are randomly 
            useful range is (0, 1e-5).
        hashfxn : function, optional
            Hash function to use to randomly initialize weights, for increased training 
        iter : int, optional
            Number of iterations (epochs) over the corpus.
        trim_rule : function, optional
            Vocabulary trimming rule, specifies whether certain words should remain in the 
            be trimmed away, or handled using the default (discard if word count < min_count).
            Can be None (min_count will be used, look to :func:`~gensim.utils.keep_vocab_item`),
            or a callable that accepts parameters (word, count, min_count) and returns either
            :attr:`gensim.utils.RULE_DISCARD`, :attr:`gensim.utils.RULE_KEEP` or :attr:`gensim.utils.
            The rule, if given, is only used to prune vocabulary during build_vocab() and is not 
             stored as part of the

            The input parameters are of the following types:
                * `word` (str) - the word we are examining
                * `count` (int) - the word's frequency count in the corpus
                * `min_count` (int) - the minimum count threshold.

        sorted_vocab : {0, 1}, optional
            If 1, sort the vocabulary by descending frequency before assigning word indexes.
            See :meth:`~gensim.models.word2vec.Word2VecVocab.sort_vocab()`.
        batch_words : int, optional
            Target size (in words) for batches of examples passed to worker threads (and
            thus cython routines).(Larger batches will be passed if individual
            texts are longer than 10000 words, but the standard cython code truncates to that 
        compute_loss: bool, optional
            If True, computes and stores loss value which can be retrieved using
        callbacks : iterable of :class:`~gensim.models.callbacks.CallbackAny2Vec`, optional
            Sequence of callbacks to be executed at specific stages during training.

        Initialize and train a :class:`~gensim.models.word2vec.Word2Vec` model

        >>> from gensim.models import Word2Vec
        >>> sentences = [["cat", "say", "meow"], ["dog", "say", "woof"]]
        >>> model = Word2Vec(sentences, min_count=1)

        self.max_final_vocab = max_final_vocab
        self.callbacks = callbacks
        self.load = call_on_class_only
        self.wv = Word2VecKeyedVectors(size)
        self.vocabulary = Word2VecVocab(
            max_vocab_size=max_vocab_size, min_count=min_count, sample=sample, 
            null_word=null_word, max_final_vocab=max_final_vocab, ns_exponent=ns_exponent)
        self.trainables = Word2VecTrainables(seed=seed, vector_size=size, hashfxn=hashfxn)
        super(Word2Vec, self).__init__(sentences=sentences, workers=workers, 
         vector_size=size, epochs=iter, callbacks=callbacks, batch_words=batch_words, 
         trim_rule=trim_rule, sg=sg, alpha=alpha, window=window, seed=seed, hs=hs, 
         negative=negative, cbow_mean=cbow_mean, min_alpha=min_alpha, 
         compute_loss=compute_loss, fast_version=FAST_VERSION)
    def _do_train_job(self, sentences, alpha, inits):
        """Train the model on a single batch of sentences.

        sentences : iterable of list of str
            Corpus chunk to be used in this training batch.
        alpha : float
            The learning rate used in this batch.
        inits : (np.ndarray, np.ndarray)
            Each worker threads private work memory.

        (int, int)
             2-tuple (effective word count after ignoring unknown words and sentence length 
              trimming, total word count).

        work, neu1 = inits
        tally = 0
            tally += train_batch_sg(self, sentences, alpha, work, self.compute_loss)
            tally += train_batch_cbow(self, sentences, alpha, work, neu1, self.compute_loss)
        return tally, self._raw_word_count(sentences)
    def _clear_post_train(self):
        """Remove all L2-normalized word vectors from the model."""
        self.wv.vectors_norm = None
    def _set_train_params(self, **kwargs):
        if 'compute_loss' in kwargs:
            self.compute_loss = kwargs['compute_loss']
        self.running_training_loss = 0
    def train(self, sentences, total_examples=None, total_words=None, 
        epochs=None, start_alpha=None, end_alpha=None, word_count=0, 
        queue_factor=2, report_delay=1.0, compute_loss=False, callbacks=()):
        """Update the model's neural weights from a sequence of sentences.

        To support linear learning-rate decay from (initial) `alpha` to `min_alpha`, and accurate
        progress-percentage logging, either `total_examples` (count of sentences) or 
         `total_words` (count of
        raw words in sentences) **MUST** be provided. If `sentences` is the same corpus
        that was provided to :meth:`~gensim.models.word2vec.Word2Vec.build_vocab` earlier,
        you can simply use `total_examples=self.corpus_count`.

        To avoid common mistakes around the model's ability to do multiple training passes 
         itself, an
        explicit `epochs` argument **MUST** be provided. In the common and recommended 
        where :meth:`~gensim.models.word2vec.Word2Vec.train` is only called once, you can 
         set `epochs=self.iter`.

        sentences : iterable of list of str
            The `sentences` iterable can be simply a list of lists of tokens, but for larger corpora,
            consider an iterable that streams the sentences directly from disk/network.
            See :class:`~gensim.models.word2vec.BrownCorpus`, :class:`~gensim.models.
            or :class:`~gensim.models.word2vec.LineSentence` in :mod:`~gensim.models.
             word2vec` module for such examples.
            See also the `tutorial on data streaming in Python
        total_examples : int, optional
            Count of sentences. Used to decay the `alpha` learning rate.
        total_words : int, optional
            Count of raw words in sentences. Used to decay the `alpha` learning rate.
        epochs : int, optional
            Number of iterations (epochs) over the corpus.
        start_alpha : float, optional
            Initial learning rate. If supplied, replaces the starting `alpha` from the constructor,
            for this one call to`train()`.
            Use only if making multiple calls to `train()`, when you want to manage the alpha 
             learning-rate yourself
            (not recommended).
        end_alpha : float, optional
            Final learning rate. Drops linearly from `start_alpha`.
            If supplied, this replaces the final `min_alpha` from the constructor, for this one call to 
            Use only if making multiple calls to `train()`, when you want to manage the alpha 
             learning-rate yourself
            (not recommended).
        word_count : int, optional
            Count of words already trained. Set this to 0 for the usual
            case of training on all words in sentences.
        queue_factor : int, optional
            Multiplier for size of queue (number of workers * queue_factor).
        report_delay : float, optional
            Seconds to wait before reporting progress.
        compute_loss: bool, optional
            If True, computes and stores loss value which can be retrieved using
        callbacks : iterable of :class:`~gensim.models.callbacks.CallbackAny2Vec`, optional
            Sequence of callbacks to be executed at specific stages during training.

        >>> from gensim.models import Word2Vec
        >>> sentences = [["cat", "say", "meow"], ["dog", "say", "woof"]]
        >>> model = Word2Vec(min_count=1)
        >>> model.build_vocab(sentences)  # prepare the model vocabulary
        >>> model.train(sentences, total_examples=model.corpus_count, epochs=model.iter)  
         # train word vectors
        (1, 30)

        return super(Word2Vec, self).train(
            sentences, total_examples=total_examples, total_words=total_words, 
            epochs=epochs, start_alpha=start_alpha, end_alpha=end_alpha, 
            queue_factor=queue_factor, report_delay=report_delay, 
             compute_loss=compute_loss, callbacks=callbacks)
    def score(self, sentences, total_sentences=int(1e6), chunksize=100, queue_factor=2, 
        """Score the log probability for a sequence of sentences.
        This does not change the fitted model in any way (see :meth:`~gensim.models.word2vec.
         Word2Vec.train` for that).

        Gensim has currently only implemented score for the hierarchical softmax scheme,
        so you need to have run word2vec with `hs=1` and `negative=0` for this to work.

        Note that you should specify `total_sentences`; you'll run into problems if you ask to
        score more than this number of sentences but it is inefficient to set the value too high.

        See the `article by Matt Taddy: "Document Classification by Inversion of Distributed 
         Language Representations"
        <>`_ and the
        `gensim demo <https://github.
         com/piskvorky/gensim/blob/develop/docs/notebooks/deepir.ipynb>`_ for examples of
        how to use such scores in document classification.

        sentences : iterable of list of str
            The `sentences` iterable can be simply a list of lists of tokens, but for larger corpora,
            consider an iterable that streams the sentences directly from disk/network.
            See :class:`~gensim.models.word2vec.BrownCorpus`, :class:`~gensim.models.
            or :class:`~gensim.models.word2vec.LineSentence` in :mod:`~gensim.models.
             word2vec` module for such examples.
        total_sentences : int, optional
            Count of sentences.
        chunksize : int, optional
            Chunksize of jobs
        queue_factor : int, optional
            Multiplier for size of queue (number of workers * queue_factor).
        report_delay : float, optional
            Seconds to wait before reporting progress.

        if FAST_VERSION < 0:
            warnings.warn("C extension compilation failed, scoring will be slow. "
                "Install a C compiler and reinstall gensim for fastness.")"scoring sentences with %i workers on %i vocabulary and %i features, "
            "using sg=%s hs=%s sample=%s and negative=%s", 
            self.workers, len(self.wv.vocab), self.trainables.layer1_size,, self.hs, self.
             vocabulary.sample, self.negative)
        if not self.wv.vocab:
            raise RuntimeError("you must first build vocabulary before scoring new data")
        if not self.hs:
            raise RuntimeError(
                "We have currently only implemented score for the hierarchical softmax scheme, "
                "so you need to have run word2vec with hs=1 and negative=0 for this to work.")
        def worker_loop():
            """Compute log probability for each sentence, lifting lists of sentences from the jobs 
            work = zeros(1, dtype=REAL) # for sg hs, we actually only need one memory loc 
             (running sum)
            neu1 = matutils.zeros_aligned(self.trainables.layer1_size, dtype=REAL)
            while True:
                job = job_queue.get()
                if job is None: # signal to finish
                ns = 0
                for sentence_id, sentence in job:
                    if sentence_id >= total_sentences:
                        score = score_sentence_sg(self, sentence, work)
                        score = score_sentence_cbow(self, sentence, work, neu1)
                    sentence_scores[sentence_id] = score
                    ns += 1
                progress_queue.put(ns) # report progress
        start, next_report = default_timer(), 1.0 # buffer ahead only a limited number of jobs.. 
         this is the reason we can't simply use ThreadPool :(
        job_queue = Queue(maxsize=queue_factor * self.workers)
        progress_queue = Queue(maxsize=(queue_factor + 1) * self.workers)
        workers = [threading.Thread(target=worker_loop) for _ in xrange(self.workers)]
        for thread in workers:
            thread.daemon = True # make interrupting the process with ctrl+c easier
        sentence_count = 0
        sentence_scores = matutils.zeros_aligned(total_sentences, dtype=REAL)
        push_done = False
        done_jobs = 0
        jobs_source = enumerate(utils.grouper(enumerate(sentences), chunksize))
        # fill jobs queue with (id, sentence) job items
        while True:
                job_no, items = next(jobs_source)
                if (job_no - 1) * chunksize > total_sentences:
                    logger.warning("terminating after %i sentences (set higher total_sentences if you 
                     want more).", total_sentences)
                    job_no -= 1
                    raise StopIteration()
                logger.debug("putting job #%i in the queue", job_no)
            except StopIteration:
      "reached end of input; waiting to finish %i outstanding jobs", job_no - 
                 done_jobs + 1)
                for _ in xrange(self.workers):
                    job_queue.put(None) # give the workers heads up that they can finish -- no more 
                push_done = True
                while done_jobs < (job_no + 1) or not push_done:
                    ns = progress_queue.get(push_done) # only block after all jobs pushed
                    sentence_count += ns
                    done_jobs += 1
                    elapsed = default_timer() - start
                    if elapsed >= next_report:
              "PROGRESS: at %.2f%% sentences, %.0f sentences/s", 100.0 * 
                         sentence_count, sentence_count / elapsed)
                        next_report = elapsed + report_delay # don't flood log, wait report_delay 
                    break # loop ended by job count; really done
            except Empty:
                pass # already out of loop; continue to next push
        elapsed = default_timer() - start
        self.clear_sims()"scoring %i sentences took %.1fs, %.0f sentences/s", sentence_count, 
         elapsed, sentence_count / elapsed)
        return sentence_scores[:sentence_count]
    def clear_sims(self):
        """Remove all L2-normalized word vectors from the model, to free up memory.

        You can recompute them later again using the :meth:`~gensim.models.word2vec.
         Word2Vec.init_sims` method.

        self.wv.vectors_norm = None
    def intersect_word2vec_format(self, fname, lockf=0.0, binary=False, encoding='utf8', 
        """Merge in an input-hidden weight matrix loaded from the original C word2vec-tool 
        where it intersects with the current vocabulary.

        No words are added to the existing vocabulary, but intersecting words adopt the file's 
         weights, and
        non-intersecting words are left alone.

        fname : str
            The file path to load the vectors from.
        lockf : float, optional
            Lock-factor value to be set for any imported word-vectors; the
            default value of 0.0 prevents further updating of the vector during subsequent
            training. Use 1.0 to allow further training updates of merged vectors.
        binary : bool, optional
            If True, `fname` is in the binary word2vec C format.
        encoding : str, optional
            Encoding of `text` for `unicode` function (python2 only).
        unicode_errors : str, optional
            Error handling behaviour, used as parameter for `unicode` function (python2 only).

        overlap_count = 0"loading projection weights from %s", fname)
        with utils.smart_open(fname) as fin:
            header = utils.to_unicode(fin.readline(), encoding=encoding)
            vocab_size, vector_size = (int(x) for x in header.split()) # throws for invalid file format
            if not vector_size == self.wv.vector_size:
                raise ValueError("incompatible vector size %d in file %s" % (vector_size, fname)) # 
                 TOCONSIDER: maybe mismatched vectors still useful enough to merge (truncating/padding)?
            if binary:
                binary_len = dtype(REAL).itemsize * vector_size
                for _ in xrange(vocab_size): # mixed text and binary: read text first, then binary
                    word = []
                    while True:
                        ch =
                        if ch == b' ':
                        if ch != b'\n': # ignore newlines in front of words (some binary files have)
                    word = utils.to_unicode(b''.join(word), encoding=encoding, 
                    weights = fromstring(, dtype=REAL)
                    if word in self.wv.vocab:
                        overlap_count += 1
                        self.wv.vectors[self.wv.vocab[word].index] = weights
                        self.trainables.vectors_lockf[self.wv.vocab[word].index] = lockf # lock-factor: 0.0
                         =no changes
                for line_no, line in enumerate(fin):
                    parts = utils.to_unicode(line.rstrip(), encoding=encoding, errors=unicode_errors).
                     split(" ")
                    if len(parts) != vector_size + 1:
                        raise ValueError("invalid vector on line %s (is this really the text format?)" % 
                    word, weights = parts[0], [REAL(x) for x in parts[1:]]
                    if word in self.wv.vocab:
                        overlap_count += 1
                        self.wv.vectors[self.wv.vocab[word].index] = weights
                        self.trainables.vectors_lockf[self.wv.vocab[word].index] = lockf # lock-factor: 0.0
                         =no changes
       "merged %d vectors into %s matrix from %s", overlap_count, self.wv.vectors.
         shape, fname)
    @deprecated("Method will be removed in 4.0.0, use self.wv.__getitem__() instead")
    def __getitem__(self, words):
        """Deprecated. Use `self.wv.__getitem__` instead.
        Refer to the documentation for :meth:`~gensim.models.keyedvectors.

        return self.wv.__getitem__(words)
    @deprecated("Method will be removed in 4.0.0, use self.wv.__contains__() instead")
    def __contains__(self, word):
        """Deprecated. Use `self.wv.__contains__` instead.
        Refer to the documentation for :meth:`~gensim.models.keyedvectors.

        return self.wv.__contains__(word)
    def predict_output_word(self, context_words_list, topn=10):
        """Get the probability distribution of the center word given context words.

        context_words_list : list of str
            List of context words.
        topn : int, optional
            Return `topn` words and their probabilities.

        list of (str, float)
            `topn` length list of tuples of (word, probability).

        if not self.negative:
            raise RuntimeError(
                "We have currently only implemented predict_output_word for the negative 
                 sampling scheme, "
                "so you need to have run word2vec with negative > 0 for this to work.")
        if not hasattr(self.wv, 'vectors') or not hasattr(self.trainables, 'syn1neg'):
            raise RuntimeError("Parameters required for predicting the output words not found.")
        word_vocabs = [self.wv.vocab[w] for w in context_words_list if w in self.wv.vocab]
        if not word_vocabs:
            warnings.warn("All the input context words are out-of-vocabulary for the current 
            return None
        word2_indices = [word.index for word in word_vocabs]
        l1 = np_sum(self.wv.vectors[word2_indices], axis=0)
        if word2_indices and self.cbow_mean:
            l1 /= len(word2_indices)
        # propagate hidden -> output and take softmax to get probabilities
        prob_values = exp(dot(l1, self.trainables.syn1neg.T))
        prob_values /= sum(prob_values)
        top_indices = matutils.argsort(prob_values, topn=topn, reverse=True) # returning the 
         most probable output words with their probabilities
        return [(self.wv.index2word[index1], prob_values[index1]) for index1 in top_indices]
    def init_sims(self, replace=False):
        """Deprecated. Use `self.wv.init_sims` instead.
        See :meth:`~gensim.models.keyedvectors.Word2VecKeyedVectors.init_sims`.

        if replace and hasattr(self.trainables, 'syn1'):
            del self.trainables.syn1
        return self.wv.init_sims(replace)
    def reset_from(self, other_model):
        """Borrow shareable pre-built structures from `other_model` and reset hidden layer 

        Structures copied are:
            * Vocabulary
            * Index to word mapping
            * Cumulative frequency table (used for negative sampling)
            * Cached corpus length

        Useful when testing multiple models on the same corpus in parallel.

        other_model : :class:`~gensim.models.word2vec.Word2Vec`
            Another model to copy the internal structures from.

        self.wv.vocab = other_model.wv.vocab
        self.wv.index2word = other_model.wv.index2word
        self.vocabulary.cum_table = other_model.vocabulary.cum_table
        self.corpus_count = other_model.corpus_count
        self.trainables.reset_weights(self.hs, self.negative, self.wv)
    def log_accuracy(section):
        """Deprecated. Use `self.wv.log_accuracy` instead.
        See :meth:`~gensim.models.word2vec.Word2VecKeyedVectors.log_accuracy`.

        return Word2VecKeyedVectors.log_accuracy(section)
    @deprecated("Method will be removed in 4.0.0, use self.wv.evaluate_word_analogies() 
    def accuracy(self, questions, restrict_vocab=30000, most_similar=None, 
        """Deprecated. Use `self.wv.accuracy` instead.
        See :meth:`~gensim.models.word2vec.Word2VecKeyedVectors.accuracy`.

        most_similar = most_similar or Word2VecKeyedVectors.most_similar
        return self.wv.accuracy(questions, restrict_vocab, most_similar, case_insensitive)
    def __str__(self):
        """Human readable representation of the model's state.

            Human readable representation of the model's state, including the vocabulary size, 
             vector size
            and learning rate.

        return "%s(vocab=%s, size=%s, alpha=%s)" % (
            self.__class__.__name__, len(self.wv.index2word), self.wv.vector_size, self.alpha)
    def delete_temporary_training_data(self, replace_word_vectors_with_normalized=False):
        """Discard parameters that are used in training and scoring, to save memory.

        Use only if you're sure you're done training a model.

        replace_word_vectors_with_normalized : bool, optional
            If True, forget the original (not normalized) word vectors and only keep
            the L2-normalized word vectors, to save even more memory.

        if replace_word_vectors_with_normalized:
    def save(self, *args, **kwargs):
        """Save the model.
        This saved model can be loaded again using :func:`~gensim.models.word2vec.
         Word2Vec.load`, which supports
        online training and getting vectors for vocabulary words.

        fname : str
            Path to the file.

        # don't bother storing the cached normalized vectors, recalculable table
        kwargs['ignore'] = kwargs.get('ignore', ['vectors_norm', 'cum_table'])
        super(Word2Vec, self).save(*args, **kwargs)
    def get_latest_training_loss(self):
        """Get current value of the training loss.

            Current training loss.

        return self.running_training_loss
        "Method will be removed in 4.0.0, keep just_word_vectors = model.wv to retain just the 
         KeyedVectors instance")
    def _minimize_model(self, save_syn1=False, save_syn1neg=False, 
        if save_syn1 and save_syn1neg and save_vectors_lockf:
        if hasattr(self.trainables, 'syn1') and not save_syn1:
            del self.trainables.syn1
        if hasattr(self.trainables, 'syn1neg') and not save_syn1neg:
            del self.trainables.syn1neg
        if hasattr(self.trainables, 'vectors_lockf') and not save_vectors_lockf:
            del self.trainables.vectors_lockf
        self.model_trimmed_post_training = True
    def load_word2vec_format(
        cls, fname, fvocab=None, binary=False, encoding='utf8', unicode_errors='strict', 
        limit=None, datatype=REAL):
        """Deprecated. Use :meth:`gensim.models.KeyedVectors.load_word2vec_format` 
        raise DeprecationWarning("Deprecated. Use gensim.models.KeyedVectors.
         load_word2vec_format instead.")
    def save_word2vec_format(self, fname, fvocab=None, binary=False):
        """Deprecated. Use `model.wv.save_word2vec_format` instead.
        See :meth:`gensim.models.KeyedVectors.save_word2vec_format`.

        raise DeprecationWarning("Deprecated. Use model.wv.save_word2vec_format instead.")
    def load(cls, *args, **kwargs):
        """Load a previously saved :class:`~gensim.models.word2vec.Word2Vec` model.

        See Also
            Save model.

        fname : str
            Path to the saved file.

            Loaded model.

            model = super(Word2Vec, cls).load(*args, **kwargs)
        # for backward compatibility for `max_final_vocab` feature
            if not hasattr(model, 'max_final_vocab'):
                model.max_final_vocab = None
                model.vocabulary.max_final_vocab = None
            return model
        except AttributeError:
  'Model saved using code from earlier Gensim Version. Re-loading old 
             model in a compatible way.')
            from gensim.models.deprecated.word2vec import load_old_word2vec
            return load_old_word2vec(*args, **kwargs)