gensim|gensim word2vec实践

语料下载地址

# -*- coding: utf-8 -*-import jieba import jieba.analyse# suggest_freq调节单个词语的词频,使其能(或不能)被分出来 jieba.suggest_freq('沙瑞金', True) jieba.suggest_freq('田国富', True) jieba.suggest_freq('高育良', True) jieba.suggest_freq('侯亮平', True) jieba.suggest_freq('钟小艾', True) jieba.suggest_freq('陈岩石', True) jieba.suggest_freq('欧阳菁', True) jieba.suggest_freq('易学习', True) jieba.suggest_freq('王大路', True) jieba.suggest_freq('蔡成功', True) jieba.suggest_freq('孙连城', True) jieba.suggest_freq('季昌明', True) jieba.suggest_freq('丁义珍', True) jieba.suggest_freq('郑西坡', True) jieba.suggest_freq('赵东来', True) jieba.suggest_freq('高小琴', True) jieba.suggest_freq('赵瑞龙', True) jieba.suggest_freq('林华华', True) jieba.suggest_freq('陆亦可', True) jieba.suggest_freq('刘新建', True) jieba.suggest_freq('刘庆祝', True)with open('./in_the_name_of_people.txt', 'rb') as f: document = f.read() document_cut = jieba.cut(document) result = ' '.join(document_cut) result = result.encode('utf-8') with open('./in_the_name_of_people_segment.txt', 'wb+') as f2: f2.write(result)f.close() f2.close()

读分词后的文件到内存,这里使用了word2vec提供的LineSentence类来读文件,然后使用word2vec的模型
  • min_count:忽略总频率低于此值的所有单词
  • size:指定了训练时词向量维度,默认为100
  • window:句中当前词与预测词之间的最大距离
  • hs:If 1, hierarchical softmax .If 0 negative sampling.
# import modules & set up logging import logging import os from gensim.models import word2veclogging.basicConfig(format='%(asctime)s : %(levelname)s : %(message)s', level=logging.INFO)sentences = word2vec.LineSentence('./in_the_name_of_people_segment.txt')model = word2vec.Word2Vec(sentences, hs=1, min_count=1, window=3, size=100)

2019-05-14 17:13:22,538 : INFO : collecting all words and their counts 2019-05-14 17:13:22,540 : INFO : PROGRESS: at sentence #0, processed 0 words, keeping 0 word types 2019-05-14 17:13:22,593 : INFO : collected 17878 word types from a corpus of 161343 raw words and 2311 sentences 2019-05-14 17:13:22,594 : INFO : Loading a fresh vocabulary 2019-05-14 17:13:22,673 : INFO : effective_min_count=1 retains 17878 unique words (100% of original 17878, drops 0) 2019-05-14 17:13:22,674 : INFO : effective_min_count=1 leaves 161343 word corpus (100% of original 161343, drops 0) 2019-05-14 17:13:22,724 : INFO : deleting the raw counts dictionary of 17878 items 2019-05-14 17:13:22,724 : INFO : sample=0.001 downsamples 38 most-common words 2019-05-14 17:13:22,725 : INFO : downsampling leaves estimated 120578 word corpus (74.7% of prior 161343) 2019-05-14 17:13:22,738 : INFO : constructing a huffman tree from 17878 words 2019-05-14 17:13:23,069 : INFO : built huffman tree with maximum node depth 17 2019-05-14 17:13:23,097 : INFO : estimated required memory for 17878 words and 100 dimensions: 33968200 bytes 2019-05-14 17:13:23,098 : INFO : resetting layer weights 2019-05-14 17:13:23,271 : INFO : training model with 3 workers on 17878 vocabulary and 100 features, using sg=0 hs=1 sample=0.001 negative=5 window=3 2019-05-14 17:13:23,457 : INFO : worker thread finished; awaiting finish of 2 more threads 2019-05-14 17:13:23,458 : INFO : worker thread finished; awaiting finish of 1 more threads 2019-05-14 17:13:23,470 : INFO : worker thread finished; awaiting finish of 0 more threads 2019-05-14 17:13:23,471 : INFO : EPOCH - 1 : training on 161343 raw words (120329 effective words) took 0.2s, 613072 effective words/s 2019-05-14 17:13:23,655 : INFO : worker thread finished; awaiting finish of 2 more threads 2019-05-14 17:13:23,658 : INFO : worker thread finished; awaiting finish of 1 more threads 2019-05-14 17:13:23,676 : INFO : worker thread finished; awaiting finish of 0 more threads 2019-05-14 17:13:23,677 : INFO : EPOCH - 2 : training on 161343 raw words (120484 effective words) took 0.2s, 592001 effective words/s 2019-05-14 17:13:23,865 : INFO : worker thread finished; awaiting finish of 2 more threads 2019-05-14 17:13:23,866 : INFO : worker thread finished; awaiting finish of 1 more threads 2019-05-14 17:13:23,882 : INFO : worker thread finished; awaiting finish of 0 more threads 2019-05-14 17:13:23,883 : INFO : EPOCH - 3 : training on 161343 raw words (120571 effective words) took 0.2s, 589983 effective words/s 2019-05-14 17:13:24,065 : INFO : worker thread finished; awaiting finish of 2 more threads 2019-05-14 17:13:24,075 : INFO : worker thread finished; awaiting finish of 1 more threads 2019-05-14 17:13:24,084 : INFO : worker thread finished; awaiting finish of 0 more threads 2019-05-14 17:13:24,085 : INFO : EPOCH - 4 : training on 161343 raw words (120615 effective words) took 0.2s, 600460 effective words/s 2019-05-14 17:13:24,273 : INFO : worker thread finished; awaiting finish of 2 more threads 2019-05-14 17:13:24,274 : INFO : worker thread finished; awaiting finish of 1 more threads 2019-05-14 17:13:24,277 : INFO : worker thread finished; awaiting finish of 0 more threads 2019-05-14 17:13:24,279 : INFO : EPOCH - 5 : training on 161343 raw words (120605 effective words) took 0.2s, 631944 effective words/s 2019-05-14 17:13:24,279 : INFO : training on a 806715 raw words (602604 effective words) took 1.0s, 598553 effective words/s

与某个词最相近的3个字的词
req_count = 5 for key in model.wv.similar_by_word('李达康', topn=100): if len(key[0]) == 3: req_count -= 1 print(key[0], key[1]) if req_count == 0: break

2019-05-14 17:13:27,276 : INFO : precomputing L2-norms of word weight vectors赵东来 0.9634759426116943 陆亦可 0.9602197408676147 蔡成功 0.9589439034461975 王大路 0.9569779634475708 祁同伟 0.9561013579368591

req_count = 5 for key in model.wv.similar_by_word('赵东来', topn=100): if len(key[0]) == 3: req_count -= 1 print(key[0], key[1]) if req_count == 0: break

李达康 0.9634760618209839 陆亦可 0.9614400863647461 易学习 0.9584609866142273 祁同伟 0.9565587639808655 王大路 0.9549983739852905

req_count = 5 for key in model.wv.similar_by_word('高育良', topn=100): if len(key[0]) == 3: req_count -= 1 print(key[0], key[1]) if req_count == 0: break

沙瑞金 0.9721000790596008 侯亮平 0.9408242702484131 祁同伟 0.9268442392349243 李达康 0.9241408705711365 季昌明 0.913619339466095

req_count = 5 for key in model.wv.similar_by_word('沙瑞金', topn=100): if len(key[0]) == 3: req_count -= 1 print(key[0], key[1]) if req_count == 0: break

高育良 0.9721001386642456 李达康 0.9424692392349243 易学习 0.9424353241920471 无表情 0.9378770589828491 祁同伟 0.9351213574409485

计算两个词向量的相似度
print(model.wv.similarity('沙瑞金', '高育良')) print(model.wv.similarity('李达康', '王大路'))

0.9721002 0.95697814

计算某个词的相关列表
try: sim3 = model.most_similar(u'侯亮平',topn =20) print(u'和 侯亮平 与相关的词有:\n') for key in sim3: print(key[0],key[1]) except: print(' error')

和 侯亮平 与相关的词有:祁同伟 0.9691112041473389 陆亦可 0.9684256911277771 季昌明 0.9582957625389099 李达康 0.952505886554718 她 0.9482855200767517 他们 0.9475176334381104 易学习 0.9456426501274109 陈岩石 0.9433715343475342 马上 0.941593587398529 高育良 0.9408242702484131 郑西坡 0.9396289587020874 王大路 0.9381627440452576 沙瑞金 0.9350594282150269 赵东来 0.9322312474250793 陈海 0.9311630725860596 司机 0.9282065033912659 蔡成功 0.9281994104385376 他 0.92684006690979 组织 0.9237431287765503 大家 0.9234919548034668E:\Anaconda3\envs\sklearn\lib\site-packages\ipykernel_launcher.py:2: DeprecationWarning: Call to deprecated `most_similar` (Method will be removed in 4.0.0, use self.wv.most_similar() instead).

【gensim|gensim word2vec实践】找出不同类的词
print(model.wv.doesnt_match(u"沙瑞金 高育良 李达康 刘庆祝".split()))

刘庆祝

保留模型,方便重用
model.save(u'人民的名义.model')

2019-05-14 17:13:39,338 : INFO : saving Word2Vec object under 人民的名义.model, separately None 2019-05-14 17:13:39,338 : INFO : not storing attribute vectors_norm 2019-05-14 17:13:39,339 : INFO : not storing attribute cum_table 2019-05-14 17:13:39,906 : INFO : saved 人民的名义.model

加载模型
model_2 = word2vec.Word2Vec.load('人民的名义.model')

2019-05-14 17:13:42,714 : INFO : loading Word2Vec object from 人民的名义.model 2019-05-14 17:13:42,942 : INFO : loading wv recursively from 人民的名义.model.wv.* with mmap=None 2019-05-14 17:13:42,943 : INFO : setting ignored attribute vectors_norm to None 2019-05-14 17:13:42,943 : INFO : loading vocabulary recursively from 人民的名义.model.vocabulary.* with mmap=None 2019-05-14 17:13:42,944 : INFO : loading trainables recursively from 人民的名义.model.trainables.* with mmap=None 2019-05-14 17:13:42,944 : INFO : setting ignored attribute cum_table to None 2019-05-14 17:13:42,945 : INFO : loaded 人民的名义.model

try: sim3 = model_2.most_similar(u'侯亮平',topn =20) print(u'和 侯亮平 与相关的词有:\n') for key in sim3: print(key[0],key[1]) except: print(' error')

E:\Anaconda3\envs\sklearn\lib\site-packages\ipykernel_launcher.py:2: DeprecationWarning: Call to deprecated `most_similar` (Method will be removed in 4.0.0, use self.wv.most_similar() instead).2019-05-14 17:14:02,083 : INFO : precomputing L2-norms of word weight vectors和 侯亮平 与相关的词有:祁同伟 0.9691112041473389 陆亦可 0.9684256911277771 季昌明 0.9582957625389099 李达康 0.952505886554718 她 0.9482855200767517 他们 0.9475176334381104 易学习 0.9456426501274109 陈岩石 0.9433715343475342 马上 0.941593587398529 高育良 0.9408242702484131 郑西坡 0.9396289587020874 王大路 0.9381627440452576 沙瑞金 0.9350594282150269 赵东来 0.9322312474250793 陈海 0.9311630725860596 司机 0.9282065033912659 蔡成功 0.9281994104385376 他 0.92684006690979 组织 0.9237431287765503 大家 0.9234919548034668


转载于:https://www.cnblogs.com/chenxiangzhen/p/10863344.html

    推荐阅读