Python公众号开发部分代码开源
2016-06-08 12:57
477 查看
这是一年前写的代码了,看看就好了,很多不规范的代码,并且完全没有面向对象(这是因为新浪sae限制太多)等等。由于在这个问题下面回答了之后很多人问我要源代码,于是分享出来。你有哪些用计算机技能解决生活问题的经历? - 路人甲的回答
主要功能目录:
python基于新浪sae开发的微信公众平台,实现功能:
输入任意内容---会有小黄人机器人聊天
输入段子---回复笑话
输入开源+文章---发送消息到开源中国
输入快递+订单号---查询快递信息
输入天气---查询南京最近五天天气状况
输入微博热点---回复微博当前热门话题
输入电影+名称---回复百度云盘中搜索的链接
以及抓取美拍、秒拍、新浪热门视频在最后。
如何下载以下源码以及获取更多编程资源?只需要简单两步:
1、关注订阅号:smcode2016
2、回复关键词公众号源码即可获得
最后如果你想学习如果做这个订阅号,可以加我个人微信:18362983803 , 最近正在筹办爬虫零基础课程,不完全免费,谢谢!
提取视频:
如何下载以上源码以及获取更多编程资源?只需要简单两步:
1、关注订阅号:smcode2016
2、回复关键词公众号源码即可获得
最后看了这么多源码?难道你不想关注我的专栏,看更多的源码学习编程?:学习编程 - 知乎专栏
如果你想了解我,点击这里:路人甲
作者:路人甲
链接:https://zhuanlan.zhihu.com/p/21284127
来源:知乎
主要功能目录:
python基于新浪sae开发的微信公众平台,实现功能:
输入任意内容---会有小黄人机器人聊天
输入段子---回复笑话
输入开源+文章---发送消息到开源中国
输入快递+订单号---查询快递信息
输入天气---查询南京最近五天天气状况
输入微博热点---回复微博当前热门话题
输入电影+名称---回复百度云盘中搜索的链接
以及抓取美拍、秒拍、新浪热门视频在最后。
如何下载以下源码以及获取更多编程资源?只需要简单两步:
1、关注订阅号:smcode2016
2、回复关键词公众号源码即可获得
最后如果你想学习如果做这个订阅号,可以加我个人微信:18362983803 , 最近正在筹办爬虫零基础课程,不完全免费,谢谢!
# -*- coding: utf-8 -*- import hashlib import web import lxml import time import os import urllib2,json import urllib import re import random import hashlib import cookielib from urllib import urlencode from lxml import etree from smtplib import SMTP_SSL from email.header import Header from email.mime.text import MIMEText class WeixinInterface: def __init__(self): self.app_root = os.path.dirname(__file__) self.templates_root = os.path.join(self.app_root, 'templates') self.render = web.template.render(self.templates_root) def GET(self): #获取输入参数 data = web.input() signature=data.signature timestamp=data.timestamp nonce=data.nonce echostr=data.echostr #自己的token token="weixin9047" #这里改写你在微信公众平台里输入的token #字典序排序 list=[token,timestamp,nonce] list.sort() sha1=hashlib.sha1() map(sha1.update,list) hashcode=sha1.hexdigest() #sha1加密算法 #如果是来自微信的请求,则回复echostr if hashcode == signature: return echostr def POST(self): str_xml = web.data() #获得post来的数据 xml = etree.fromstring(str_xml)#进行XML解析 content=xml.find("Content").text#获得用户所输入的内容 msgType=xml.find("MsgType").text fromUser=xml.find("FromUserName").text toUser=xml.find("ToUserName").text if(content == u"天气"): url = "http://m.ip138.com/21/nanjing/tianqi/" headers = { 'Connection': 'Keep-Alive', 'Accept': 'text/html, application/xhtml+xml, */*', 'Accept-Language': 'en-US,en;q=0.8,zh-Hans-CN;q=0.5,zh-Hans;q=0.3', 'User-Agent': 'Mozilla/5.0 (Windows NT 6.3; WOW64; Trident/7.0; rv:11.0) like Gecko'} req = urllib2.Request(url, headers = headers) opener = urllib2.urlopen(req) html = opener.read() rex = r'(?<=img src="/image/s[0-9].gif" alt=").{1,6}(?=" />)' rexx = r'(?<=div class="temperature">).{5,15}(?=</div>)' n = re.findall(rex,html) m = re.findall(rexx,html) str_wether = "" for (i,j) in zip(m,n): str_wether = str_wether + j + " " +i + "\n" return self.render.reply_text(fromUser,toUser,int(time.time()),"最近五天天气:\n"+str_wether) elif(content[0:2] == u"电影"): keyword = urllib.quote(content[2:].encode("utf-8")) url = "http://www.wangpansou.cn/s.php?q="+keyword headers = { 'Connection': 'Keep-Alive', 'Accept': 'text/html, application/xhtml+xml, */*', 'Accept-Language': 'en-US,en;q=0.8,zh-Hans-CN;q=0.5,zh-Hans;q=0.3', 'User-Agent': 'Mozilla/5.0 (Windows NT 6.3; WOW64; Trident/7.0; rv:11.0) like Gecko'} req = urllib2.Request(url, headers = headers) opener = urllib2.urlopen(req) html = opener.read() rex = r'https?://pan.baidu.com.*\?uk=[0-9]{10}.*[\d+?]"' m = re.findall(rex,html) string = u"" for i in m: string = string + i + "\n" return self.render.reply_text(fromUser,toUser,int(time.time()),u"以下是电影链接:\n"+string) elif(u"段子" in content): url_8 = "http://www.qiushibaike.com/" url_24 = "http://www.qiushibaike.com/hot/" headers = { 'Connection': 'Keep-Alive', 'Accept': 'text/html, application/xhtml+xml, */*', 'Accept-Language': 'en-US,en;q=0.8,zh-Hans-CN;q=0.5,zh-Hans;q=0.3', 'User-Agent': 'Mozilla/5.0 (Windows NT 6.3; WOW64; Trident/7.0; rv:11.0) like Gecko'} req_8 = urllib2.Request(url_8, headers = headers) req_24 = urllib2.Request(url_24,headers = headers) opener_8 = urllib2.urlopen(req_8) opener_24 = urllib2.urlopen(req_24) html_8 = opener_8.read() html_24 = opener_24.read() rex = r'(?<=div class="content">).*?(?=<!--)' m_8 = re.findall(rex,html_8,re.S) m_24 = re.findall(rex, html_24, re.S) m_8.extend(m_24) random.shuffle(m_8) return self.render.reply_text(fromUser,toUser,int(time.time()),m_8[0].replace('<br/>','')) elif(content[0:2] == u"开源"): url = "https://www.oschina.net/action/user/hash_login" urll = "http://www.oschina.net/action/tweet/pub" username = "904727147@qq.com" passw = ""#密码肯定不会给你们的 password = hashlib.sha1(passw).hexdigest() cj = cookielib.CookieJar() opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cj)) opener.addheaders = [('User-agent','Mozilla/5.0 (X11; Linux x86_64; rv:38.0) Gecko/20100101 Firefox/38.0 Iceweasel/38.3.0')] urllib2.install_opener(opener) data = {'email':username,'pwd':password} data_post = urllib.urlencode(data) opener.open(url, data_post) user = "2391943" msg = content[2:].encode("utf-8") user_code = "lPFz26r3ZIa1e3KyIWlzPNpJlaEmZqyh6dAWAotd" post = {'user_code':user_code,'user':user,'msg':msg} msg_post = urllib.urlencode(post) html = urllib2.urlopen(urll,msg_post).read() return self.render.reply_text(fromUser,toUser,int(time.time()),u"发送到开源中国动弹成功!") elif(content[0:2] == u"快递"): keyword = content[2:] url = "http://www.kuaidi100.com/autonumber/autoComNum?text="+keyword cj = cookielib.CookieJar() opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cj)) opener.addheaders = [('User-agent','Mozilla/5.0 (X11; Linux x86_64; rv:38.0) Gecko/20100101 Firefox/38.0 Iceweasel/38.3.0')] urllib2.install_opener(opener) html = urllib2.urlopen(url).read() jo = json.loads(html) typ = jo["auto"][0]['comCode'] if(typ is None): return self.render.reply_text(fromUser,toUser,int(time.time()),u"请检查你的定单号!") urll = "http://www.kuaidi100.com/query?type="+typ+"&postid="+keyword html_end = urllib2.urlopen(urll).read() jo_end = json.loads(html_end) if(jo_end["status"] == "201"): return self.render.reply_text(fromUser,toUser,int(time.time()),u"订单号输入有误,请重新输入!") text = jo_end["data"] string = u"" for i in text: string = string + i["time"] + i["context"] + "\n" return self.render.reply_text(fromUser,toUser,int(time.time()),string) elif(content == u"微博热点"): url = "http://weibo.cn/pub/?tf=5_005" headers = { 'Connection': 'Keep-Alive', 'Accept': 'text/html, application/xhtml+xml, */*', 'Accept-Language': 'en-US,en;q=0.8,zh-Hans-CN;q=0.5,zh-Hans;q=0.3', 'User-Agent': 'Mozilla/5.0 (Windows NT 6.3; WOW64; Trident/7.0; rv:11.0) like Gecko'} req = urllib2.Request(url, headers = headers) opener = urllib2.urlopen(req) html = opener.read().decode("utf-8") rex = r'(?<=div class="c"><a href=").{60,79}(?=</a>)' ss = re.findall(rex,html) string = u"" for i in ss: string = string + i.replace('>','\n')+"\n" return self.render.reply_text(fromUser,toUser,int(time.time()),string.replace('"','')) elif(content == u"知乎信息"): username = '18362983803' password = ''#这是以前的密码别尝试了 _xsrf='558c1b60725377c5810ae2484b26781e' url = r'https://www.zhihu.com/login/phone_num' cj = cookielib.CookieJar() opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cj)) opener.addheaders = [('User-agent','Mozilla/5.0 (X11; Linux x86_64; rv:38.0) Gecko/20100101 Firefox/38.0 Iceweasel/38.3.0')] data = urllib.urlencode({"phone_num":username,"password":password,'_xsrf':_xsrf}) opener.open(url,data) html = opener.open('https://www.zhihu.com/noti7/new?r=1454793308655').read() jo = json.loads(html) data = jo[1] string = "增长了:"+str(data[0])+"个评论"+str(data[1])+"个粉丝"+str(data[2])+"个赞同" return self.render.reply_text(fromUser,toUser,int(time.time()),string) elif(content[0:2] == u"闹钟"): string = str(time.strftime("%H:%M", time.localtime())) if(string == content[2:]): mail_info = { "from": "904727147@qq.com", "to": "904727147@qq.com", "hostname": "smtp.qq.com", "username": "904727147@qq.com", "password": "himnbtwxa", "mail_subject": "懒猪起床!", "mail_text": "起床了,猪", "mail_encoding": "utf-8" } smtp = SMTP_SSL(mail_info["hostname"]) smtp.set_debuglevel(1) smtp.ehlo(mail_info["hostname"]) smtp.login(mail_info["username"], mail_info["password"]) msg = MIMEText(mail_info["mail_text"], "plain", mail_info["mail_encoding"]) msg["Subject"] = Header(mail_info["mail_subject"], mail_info["mail_encoding"]) msg["from"] = mail_info["from"] msg["to"] = mail_info["to"] i = 0 while(i<20): j = 0 while(j<2): smtp.sendmail(mail_info["from"], mail_info["to"], msg.as_string()) j = j + 1 i = i + 1 time.sleep(10) smtp.quit() return self.render.reply_text(fromUser,toUser,int(time.time()),string) return self.render.reply_text(fromUser,toUser,int(time.time()),string+u"879") elif(u"钟志远" in content): return self.render.reply_text(fromUser,toUser,int(time.time()),u"你想找全世界最帅的人干嘛?如果你是妹子,请加微信18362983803!汉子绕道!") elif(u"使用" in content): return self.render.reply_text(fromUser,toUser,int(time.time()),u"搜电影:电影+电影名,最近天气:天气,微博热门:微博热点,知乎信息:知乎信息,快递查询:快递+单号,看笑话:段子,发送动弹到开源中国:开源+内容") else: url = r'http://www.xiaohuangji.com/ajax.php' cj = cookielib.CookieJar() opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cj)) opener.addheaders = [('User-agent','Mozilla/5.0 (X11; Linux x86_64; rv:38.0) Gecko/20100101 Firefox/38.0 Iceweasel/38.3.0')] string = urllib.quote(content.encode("utf-8")) try: data = urllib.urlencode({"para":string}) html = opener.open(url,data).read() string = html+"\n----[回复[使用]]" return self.render.reply_text(fromUser,toUser,int(time.time()),string) except Exception,ex: return self.render.reply_text(fromUser,toUser,int(time.time()),u"我不想理你了~")
提取视频:
#encoding:utf-8 import urllib2 import cookielib import json import re def search(): url = "http://www.miaopai.com/miaopai/index_api?cateid=2002&per=20&page=1" url2 = "http://www.meipai.com/medias/hot" cj = cookielib.CookieJar() opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cj)) opener.addheaders = [('User-agent','Mozilla/5.0 (X11; Linux x86_64; rv:38.0) Gecko/20100101 Firefox/38.0 Iceweasel/38.3.0')] urllib2.install_opener(opener) html = urllib2.urlopen(url).read() html2 = urllib2.urlopen(url2).read() rex = r'http://mvvideo2.meitudata.com/.*?mp4' rexx = r'http://mvimg1.meitudata.com/.*?320' value = re.findall(rex, html2) value2 = re.findall(rexx, html2) jo = json.loads(html) f = open('/root/Desktop/sp.html','wb') text = jo["result"] f.write('<!DOCTYPE html><html lang="zh-CN"><head><meta charset="utf-8"><title>24小时最热视频</title><meta http-equiv="X-UA-Compatible" content="IE=edge"><meta name="viewport" content="width=device-width, initial-scale=1"> <!-- 上述3个meta标签*必须*放在最前面,任何其他内容都*必须*跟随其后! --> <title>全网24小时最热视频</title> <!-- Bootstrap --> <link href="dist/css/bootstrap.min.css" rel="stylesheet"> <link href="css/css.css" rel="stylesheet"> <!-- HTML5 shim and Respond.js for IE8 support of HTML5 elements and media queries --> <!-- WARNING: Respond.js doesnt work if you view the page via file:// --> <!--[if lt IE 9]> <script src="//cdn.bootcss.com/html5shiv/3.7.2/html5shiv.min.js"></script> <script src="//cdn.bootcss.com/respond.js/1.4.2/respond.min.js"></script><![endif]--><style>body {font-family: "Helvetica Neue",Helvetica,Arial,sans-serif;background:#F4F2ED none repeat scroll 0% 0%;}</style></head><body class = "home-tempate"><div class = "container">') f.write('<center><div class="gradient"><div class="header"><h2>路人甲的视频小站</h2><p>以下视频收集新浪、美拍、秒拍网24小时内最热视频,如有侵权必删</p><div class="clearfix"><a href="http://stchat.cn/zhihu.html" class="btn btn-success btn-lg">Try it now!</a></div></div></div><br><div class="container-fluid">') for i in text: f.write('<div class="row-fluid">') f.write('<video src="'+i["channel"]["stream"]["base"]+'" controls="controls" width="320" height="240"' + 'poster="' + i["channel"]["pic"]["base"] + '.jpg"></video></div>') for (i,j) in zip(value,value2): f.write('<div class="row-fluid">') f.write('<video src="'+i+'" controls="controls" width="320" height="240"' + 'poster="' + j+ '"></video></div>') f.write("</div><center></div></html>") f.flush() f.close() if __name__=='__main__': search()
如何下载以上源码以及获取更多编程资源?只需要简单两步:
1、关注订阅号:smcode2016
2、回复关键词公众号源码即可获得
最后看了这么多源码?难道你不想关注我的专栏,看更多的源码学习编程?:学习编程 - 知乎专栏
如果你想了解我,点击这里:路人甲
作者:路人甲
链接:https://zhuanlan.zhihu.com/p/21284127
来源:知乎
相关文章推荐
- android 代码实现控件之间的间距
- Python动态类型的学习---引用的理解
- Python3写爬虫(四)多线程实现数据爬取
- 垃圾邮件过滤器 python简单实现
- 介绍一款信息管理系统的开源框架---jeecg
- 下载并遍历 names.txt 文件,输出长度最长的回文人名。
- 源码被倒卖,大厂薅羊毛,开源真的只能被予取予求?
- install and upgrade scrapy
- Scrapy的架构介绍
- Centos6 编译安装Python
- 使用Python生成Excel格式的图片
- 让Python文件也可以当bat文件运行
- [Python]推算数独
- [Android]在代码里运行另一个程序的方法
- Python中zip()函数用法举例
- Python中map()函数浅析
- Python将excel导入到mysql中
- 专家解读:开源软件项目是否会被限制出口?