• python 用jieba分词统计关于红楼梦的高频词


    import jieba
    excludes = {"什么","一个","我们","那里","你们","如今","说道","知道","起来","姑娘","这里","出来","他们","众人","自己",
                "一面","只见","怎么","两个","没有","不是","不知","这个","听见","这样","进来","咱们","告诉","就是",
                "东西","袭人","回来","只是","大家","只得","老爷","丫头","这些","不敢","出去","所以","不过","的话","不好",
                "姐姐","探春","鸳鸯","一时","不能","过来","心里","如此","今日","银子","几个","答应","二人","还有","只管",
                "这么","说话","一回","那边","这话","外头","打发","自然","今儿","罢了","屋里","那些","听说","小丫头","不用","如何"}
    
    txt = open("红楼梦.txt","r",encoding='utf-8').read()
    
    words = jieba.lcut(txt)
    
    #利用jieba库将红楼梦的所有语句分成词汇
    
    counts = {}
    
    #创建的一个空的字典
    
    for word in words:
        if len(word) == 1:      #如果长度是一,可能是语气词之类的,应该删除掉
            continue
        else:
            counts[word] = counts.get(word,0) + 1
    for word in excludes:
        del(counts[word])#这一步:如果列出的干扰词汇在分完词后的所有词汇中那么删除
    
    items = list(counts.items())
    #把保存[姓名:个数]的字典转换成列表
    
    items.sort(key=lambda x:x[1],reverse = True)
    
     #对上述列表进行排序,'True'是降序排列
    
    for i in range(20):
        word,count = items[i]
        print("{0:<10}{1:>5}".format(word,count))

    运行结果:

  • 相关阅读:
    九.Protobuf3特殊类型
    八.Protobuf3更新消息类型(添加新的字段)
    七.Protobuf3 嵌套类型
    六.Protobuf3引入其他.proto文件
    五.Protobuf3 枚举
    四.Protobuf3 缺省值
    VC 在调用main函数之前的操作
    Windows下的代码注入
    C 堆内存管理
    VC++ 崩溃处理以及打印调用堆栈
  • 原文地址:https://www.cnblogs.com/Zmanqing/p/14021573.html
Copyright © 2020-2023  润新知