2018北京积分落户数据,用pyspark、pyecharts大数据可视化分析,按用户所在单位分析

xiaoxiao2021-10-27  101

2018北京积分落户数据,用pyspark、pyecharts大数据可视化分析,按用户所在单位分析。

按用户所在单位分组统计,取前50个。

#导入积分落户人员名单数据 df = spark.read.csv('jifenluohu.csv', header='true', inferSchema='true') df.cache() df.createOrReplaceTempView("jflh") #df.show() spCount = spark.sql("select unit as name,count(1) as ct from jflh group by unit order by ct desc limit 50").collect() name = [row.name for row in spCount] count = [row.ct for row in spCount] #图表展示 from pyecharts import Bar bar = Bar("2018北京积分落户用户数据分析", "按单位汇总统计用户数量") bar.add("用户数量", name, count) bar

转载请注明原文地址: https://www.6miu.com/read-4832016.html

最新回复(0)