🔥 这两年开始毕业设计和毕业答辩的要求和难度不断提升,传统的毕设题目缺少创新和亮点,往往达不到毕业答辩的要求,这两年不断有学弟学妹告诉学长自己做的项目系统达不到老师的要求。
为了大家能够顺利以及最少的精力通过毕设,学长分享优质毕业设计项目,今天要分享的是
🚩 基于大数据的抖音短视频数据分析与可视化
🥇学长这里给一个题目综合评分(每项满分5分)
本项目是大数据—基于抖音用户数据集的可视化分析。抖音作为当下非常热门的短视频软件,其背后的数据有极高的探索价值。本项目根据1737312条用户行为数据,利用python工具进行由浅入深的内容分析,目的是挖掘其中各类信息,更好地进行内容优化、产品运营。
数据信息查看
简单看一下前5行数据,确定需要进一步预处理的内容:数据去重、删除没有意义的第一列,部分列格式转换、异常值检测。
# 读取数据 df = pd.read_csv('data.csv') df.head()
df.info()
数据去重
无重复数据
print('去重前:',df.shape[0],'行数据') print('去重后:',df.drop_duplicates().shape[0],'行数据')
缺失值查看
print(np.sum(df.isnull()))
变量类型转换
real_time 和 date 转为时间变量,id、城市编码转为字符串,并把小数点去掉
df['date'] = df['date'].astype('datetime64[ns]') df['real_time'] = df['real_time'].astype('datetime64[ns]') df['uid'] = df['uid'].astype('str') df['user_city'] = df['user_city'].astype('str') df['user_city'] = df['user_city'].apply(lambda x:x[:-2]) df['item_id'] = df['item_id'].astype('str') df['author_id'] = df['author_id'].astype('str') df['item_city'] = df['item_city'].astype('str') df['item_city'] = df['item_city'].apply(lambda x:x[:-2]) df['music_id'] = df['music_id'].astype('str') df['music_id'] = df['music_id'].apply(lambda x:x[:-2])
df.info()
基本信息的可视化,面向用户、创作者以及内容这三个维度进行,构建成分画像,便于更好地针对用户、创作者进行策略投放、内容推广与营销。
user_city_count = user_info.groupby(['user_city']).count().sort_values(by=['uid'],ascending=False) x1 = list(user_city_count.index) y1 = user_city_count['uid'].tolist() len(y1)
不同地区用户数量分布图
#柱形图代码 chart = Bar() chart.add_xaxis(x1) chart.add_yaxis('地区使用人数', y1, color='#F6325A', itemstyle_opts={'barBorderRadius':[60, 60, 20, 20]}, label_opts=opts.LabelOpts(position='top')) chart.set_global_opts(datazoom_opts=opts.DataZoomOpts( range_start=0,range_end=5,orient='horizontal',type_='slider',is_zoom_lock=False, pos_left='1%' ), visualmap_opts=opts.VisualMapOpts(is_show = False,type_='opacity',range_opacity=[0.2, 1]), title_opts=opts.TitleOpts(title="不同地区用户数量分布图",pos_left='40%'), legend_opts=opts.LegendOpts(pos_right='10%',pos_top='2%')) chart.render_notebook()
覆盖到了387个城市,其中编号为99的城市用户比较多超过2000人,6、129、109、31这几个城市的使用人数也超过了1000。
h_num = round((df.groupby(['H']).count()['uid']/10000),1).to_list() h = list(df.groupby(['H']).count().index)
不同时间观看数量分布图
chart = Line() chart.add_xaxis(h) chart.add_yaxis('观看数/(万)',h_num, areastyle_opts=opts.AreaStyleOpts(color = '#1AF5EF',opacity=0.3), itemstyle_opts=opts.ItemStyleOpts(color='black'), label_opts=opts.LabelOpts(font_size=12)) chart.set_global_opts(legend_opts=opts.LegendOpts(pos_right='10%',pos_top='2%'), title_opts=opts.TitleOpts(title="不时间观看数量分布图",pos_left='40%'),) chart.render_notebook()
去掉时差后
根据不同时间的观看视频数量来看,11-18,20-21,尤其是13-16是用户使用的高峰期
点赞/完播率分布图
left = df.groupby(['H']).sum()[['finish','like']] right = df.groupby(['H']).count()['uid'] per = pd.concat([left,right],axis=1) per['finish_radio'] = round(per['finish']*100/per['uid'],2) per['like_radio'] = round(per['like']*100/per['uid'],2) x = list(df.groupby(['H']).count().index) y1 = per['finish_radio'].to_list() y2 = per['like_radio'].to_list()
#建立一个基础的图形 chart1 = Line() chart1.add_xaxis(x) chart1.add_yaxis('完播率/%',y1,is_smooth=True,label_opts=opts.LabelOpts(is_show=False),is_symbol_show = False, linestyle_opts=opts.LineStyleOpts(color='#F6325A',opacity=.7,curve=0,width=2,type_= 'solid' )) chart1.set_global_opts(yaxis_opts = opts.AxisOpts(min_=25,max_=45)) chart1.extend_axis(yaxis=opts.AxisOpts(min_=0.4,max_=3)) #叠加折线图 chart2 = Line() chart2.add_xaxis(x) chart2.add_yaxis('点赞率/%',y2,yaxis_index=1,is_smooth=True,label_opts=opts.LabelOpts(is_show=False),is_symbol_show = False, linestyle_opts=opts.LineStyleOpts(color='#1AF5EF',opacity=.7,curve=0,width=2,type_= 'solid' )) chart1.overlap(chart2) chart1.set_global_opts(legend_opts=opts.LegendOpts(pos_right='10%',pos_top='2%'), title_opts=opts.TitleOpts(title="点赞/完播率分布图",pos_left='40%'),) chart1.render_notebook()
关注到点赞率和完播率,这两个与用户粘性、创作者收益有一定关系的指标。可以看到15点是两个指标的小高峰,2、4、20、23完播较高,8、13、18、20点赞率较高。但结合观看数量与时间段的分布图,大致猜测15点深度用户较多。
df['weekday'] = df['date'].dt.weekday week = df.groupby(['weekday']).count()['uid'].to_list() df_pair = [['周一', week[0]], ['周二', week[1]], ['周三', week[2]], ['周四', week[3]], ['周五', week[4]], ['周六', week[5]], ['周日', week[6]]]
chart = Pie() chart.add('', df_pair,radius=['40%', '70%'],rosetype='radius',center=['45%', '50%'],label_opts=opts.LabelOpts(is_show=True,formatter = '{b}:{c}次')) chart.set_global_opts(visualmap_opts=[opts.VisualMapOpts(min_=200000,max_=300000,type_='color', range_color=['#1AF5EF', '#F6325A', '#000000'],is_show=True,pos_top='65%')], legend_opts=opts.LegendOpts(pos_right='10%',pos_top='2%',orient='vertical'), title_opts=opts.TitleOpts(title="一周内播放分布图",pos_left='35%'),) chart.render_notebook()
在统计的时间内周一到周三观看人数较多,但总体观看次数基本在20-30w之间。
df.groupby(['channel']).count()['uid']
观看途径主要以1为主,初步猜测为App。3途径也有部分用户使用,可能为浏览器。
author_info = df.drop_duplicates(['author_id','item_city'])[['author_id','item_city']] author_info.info() author_city_count = author_info.groupby(['item_city']).count().sort_values(by=['author_id'],ascending=False) x1 = list(author_city_count.index) y1 = author_city_count['author_id'].tolist()
df.drop_duplicates(['author_id']).shape[0]
不同城市创作者分布图
chart = Bar() chart.add_xaxis(x1) chart.add_yaxis('地区创作者人数', y1, color='#F6325A', itemstyle_opts={'barBorderRadius':[60, 60, 20, 20]}) chart.set_global_opts(datazoom_opts=opts.DataZoomOpts( range_start=0,range_end=5,orient='horizontal',type_='slider',is_zoom_lock=False, pos_left='1%' ), visualmap_opts=opts.VisualMapOpts(is_show = False,type_='opacity',range_opacity=[0.2, 1]), legend_opts=opts.LegendOpts(pos_right='10%',pos_top='2%'), title_opts=opts.TitleOpts(title="不同城市创作者分布图",pos_left='40%')) chart.render_notebook()
观看用户地区分布和创作者分布其实存在不对等的情况。4地区创作者最多,超5k人,33、42、10地区创作者也较多。
time = df.drop_duplicates(['item_id'])[['item_id','duration_time']] time = time.groupby(['duration_time']).count() x1 = list(time.index) y1 = time['item_id'].tolist()
不同时长作品分布图
chart = Bar() chart.add_xaxis(x1) chart.add_yaxis('视频时长对应视频数', y1, color='#1AF5EF', itemstyle_opts={'barBorderRadius':[60, 60, 20, 20]}, label_opts=opts.LabelOpts(font_size=12, color='black')) chart.set_global_opts(datazoom_opts=opts.DataZoomOpts( range_start=0,range_end=50,orient='horizontal',type_='slider'), visualmap_opts=opts.VisualMapOpts(max_=100000,min_=200,is_show = False,type_='opacity',range_opacity=[0.4, 1]), legend_opts=opts.LegendOpts(pos_right='10%',pos_top='2%'), title_opts=opts.TitleOpts(title="不同时长作品分布图",pos_left='40%')) chart.render_notebook()
视频时长主要集中在9-10秒,符合抖音“短”视频的特点。
like_per = 100*np.sum(df['like'])/len(df['like']) finish_per = 100*np.sum(df['finish'])/len(df['finish'])
gauge = Gauge() gauge.add("",[("视频互动率", like_per),['完播率',finish_per]],detail_label_opts=opts.LabelOpts(is_show=False,font_size=18), axisline_opts=opts.AxisLineOpts(linestyle_opts=opts.LineStyleOpts( color=[(0.3, "#1AF5EF"), (0.7, "#F6325A"), (1, "#000000")],width=20))) gauge.render_notebook()
内容整体完播率非常接近40%,点赞率在1%左右
df_cor = df[['finish','like','duration_time','H']] # 只选取部分 cor_table = df_cor.corr(method='spearman') cor_array = np.array(cor_table) cor_name = list(cor_table.columns) value = [[i, j, cor_array[i,j]] for i in [3,2,1,0] for j in [0,1,2,3]]
heat = HeatMap() heat.add_xaxis(cor_name) heat.add_yaxis("",cor_name,value,label_opts=opts.LabelOpts(is_show=True, position="inside")) heat.set_global_opts(visualmap_opts=opts.VisualMapOpts(is_show=False, max_=0.08, range_color=["#1AF5EF", "#F6325A", "#000000"])) heat.render_notebook()
因为变量非连续,采取spearman相关系数,制作相关性热力图。由于数据量比较大的缘故,几个数量性变量之间的相关性都比较小,其中看到finish和点赞之间的相关系数稍微大一些,可以一致反映用户对该视频的偏好。
pv/uv
temp = df['date'].to_list() puv = df.groupby(['date']).agg({'uid':'nunique','item_id':'count'}) uv = puv['uid'].to_list() pv = puv['item_id'].to_list() time = puv.index.to_list()
chart1 = Line() chart1.add_xaxis(time) chart1.add_yaxis('uv',uv,is_smooth=True,label_opts=opts.LabelOpts(is_show=False),is_symbol_show = False, linestyle_opts=opts.LineStyleOpts(color='#1AF5EF',opacity=.7,curve=0,width=2,type_= 'solid' )) chart1.add_yaxis('pv',pv,is_smooth=True,label_opts=opts.LabelOpts(is_show=False),is_symbol_show = False, linestyle_opts=opts.LineStyleOpts(color='#F6325A',opacity=.7,curve=0,width=2,type_= 'solid' )) chart1.render_notebook()
在2019.10.18进入用户使用高峰阶段,目标用户单人每天浏览多个视频。
7/10 留存率
lc = [] for i in range(len(time)-7): bef = set(list(df[df['date']==time[i]]['uid'])) aft = set(list(df[df['date']==time[i+7]]['uid'])) stay = bef&aft per = round(100*len(stay)/len(bef),2) lc.append(per) lc1 = [] for i in range(len(time)-1): bef = set(list(df[df['date']==time[i]]['uid'])) aft = set(list(df[df['date']==time[i+1]]['uid'])) stay = bef&aft per = round(100*len(stay)/len(bef),2) lc1.append(per)
x7 = time[0:-7] chart1 = Line() chart1.add_xaxis(x7) chart1.add_yaxis('七日留存率/%',lc,is_smooth=True,label_opts=opts.LabelOpts(is_show=False),is_symbol_show = False, linestyle_opts=opts.LineStyleOpts(color='#F6325A',opacity=.7,curve=0,width=2,type_= 'solid' )) chart1.set_global_opts(legend_opts=opts.LegendOpts(pos_right='10%',pos_top='2%'), title_opts=opts.TitleOpts(title="用户留存率分布图",pos_left='40%'),) chart1.render_notebook()
用户留存率保持在40%+,且没有跌破30%,说明获取到的数据中忠实用户较多。
通过已观看数、完播率、点赞率进行用户聚类,价值判断
df1 = df.groupby(['uid']).agg({'item_id':'count','like':'sum','finish':'sum'}) df1['like_per'] = df1['like']/df1['item_id'] df1['finish_per'] = df1['finish']/df1['item_id'] ndf1 = np.array(df1[['item_id','like_per','finish_per']])#.shape
kmeans_per_k = [KMeans(n_clusters=k).fit(ndf1) for k in range(1,8)] inertias = [model.inertia_ for model in kmeans_per_k] chart = Line(init_opts=opts.InitOpts(width='560px',height='300px')) chart.add_xaxis(range(1,8)) chart.add_yaxis("",inertias,label_opts=opts.LabelOpts(is_show=False), linestyle_opts=opts.LineStyleOpts(color='#F6325A',opacity=.7,curve=0,width=3,type_= 'solid' )) chart.render_notebook()
n_cluster = 4 cluster = KMeans(n_clusters=n_cluster,random_state=0).fit(ndf1) y_pre = cluster.labels_ # 查看聚好的类 from sklearn.metrics import silhouette_score from sklearn.metrics import silhouette_samples silhouette_score(ndf1,y_pre)
n_cluster = 3 cluster = KMeans(n_clusters=n_cluster,random_state=0).fit(ndf1) y_pre = cluster.labels_ # 查看聚好的类 from sklearn.metrics import silhouette_score from sklearn.metrics import silhouette_samples silhouette_score(ndf1,y_pre)
比较三类、四类的轮廓系数,确定聚为3类
c_ = [[],[],[]] c_[0] = [87.998,9.1615,39.92] c_[1] = [13.292,12.077,50.012] c_[2] = [275.011,8.125,28.751]
bar = Bar(init_opts=opts.InitOpts(theme='macarons',width='1000px',height='400px')) # 添加分类(x轴)的数据 bar.add_xaxis(['播放数','点赞率(千分之)','完播率(百分之)']) bar.add_yaxis('0', [round(i,2) for i in c_[0]], stack='stack0') bar.add_yaxis('1',[round(i,2) for i in c_[1]], stack='stack1') bar.add_yaxis('2',[round(i,2) for i in c_[2]], stack='stack2') bar.render_notebook()
可以大致对三类的内容做一个描述。
🧿 选题指导, 项目分享: https://gitee.com/yaa-dc/warehouse-1/blob/master/python/README.md