代码一:
#!/usr/bin/python # -*- coding: utf-8 -*- #encoding=utf-8 import threading import Queue import sys import urllib2 import re import MySQLdb # # 数据库变量设置 # DB_HOST = '127.0.0.1' DB_USER = "XXXX" DB_PASSWD = "XXXXXXXX" DB_NAME = "xxxx" # # 变量设置 # THREAD_LIMIT = 3 jobs = Queue.Queue(5) singlelock = threading.Lock() info = Queue.Queue() def workerbee(inputlist): for x in xrange(THREAD_LIMIT): print 'Thead {0} started.'.format(x) t = spider() t.start() for i in inputlist: try: jobs.put(i, block=True, timeout=5) except: singlelock.acquire() print "The queue is full !" singlelock.release() # Wait for the threads to finish singlelock.acquire() # Acquire the lock so we can print print "Waiting for threads to finish." singlelock.release() # Release the lock jobs.join() # This command waits for all threads to finish. # while not jobs.empty(): # print jobs.get() def getTitle(url,time=10): response = urllib2.urlopen(url,timeout=time) html = response.read() response.close() reg = r'<title>(.*?)</title>' title = re.compile(reg).findall(html) # title = title[0].decode('gb2312','replace').encode('utf-8') title = title[0] return title class spider(threading.Thread): def run(self): while 1: try: job = jobs.get(True,1) singlelock.acquire() title = getTitle(job[1]) info.put([job[0],title], block=True, timeout=5) # print 'This {0} is {1}'.format(job[1],title) singlelock.release() jobs.task_done() except: break; if __name__ == '__main__': con = None urls = [] try: con = MySQLdb.connect(DB_HOST,DB_USER,DB_PASSWD,DB_NAME) cur = con.cursor() cur.execute('SELECT id,url FROM `table_name` WHERE `status`=0 LIMIT 10') rows = cur.fetchall() for row in rows: # print row urls.append([row[0],row[1]]) workerbee(urls) while not info.empty(): print info.get() finally: if con: con.close()
代码二:
#!/usr/bin/python # -*- coding: utf-8 -*- #encoding=utf-8 #Filename:robot.py import threading,Queue,sys,urllib2,re # # 变量设置 # THREAD_LIMIT = 3 #设置线程数 jobs = Queue.Queue(5) #设置队列长度 singlelock = threading.Lock() #设置一个线程锁,避免重复调用 urls = ['http://games.sina.com.cn/w/n/2013-04-28/1634703505.shtml','http://games.sina.com.cn/w/n/2013-04-28/1246703487.shtml','http://games.sina.com.cn/w/n/2013-04-28/1028703471.shtml','http://games.sina.com.cn/w/n/2013-04-27/1015703426.shtml','http://games.sina.com.cn/w/n/2013-04-26/1554703373.shtml','http://games.sina.com.cn/w/n/2013-04-26/1512703346.shtml','http://games.sina.com.cn/w/n/2013-04-26/1453703334.shtml','http://games.sina.com.cn/w/n/2013-04-26/1451703333.shtml','http://games.sina.com.cn/w/n/2013-04-26/1445703329.shtml','http://games.sina.com.cn/w/n/2013-04-26/1434703322.shtml','http://games.sina.com.cn/w/n/2013-04-26/1433703321.shtml','http://games.sina.com.cn/w/n/2013-04-26/1433703320.shtml','http://games.sina.com.cn/w/n/2013-04-26/1429703318.shtml','http://games.sina.com.cn/w/n/2013-04-26/1429703317.shtml','http://games.sina.com.cn/w/n/2013-04-26/1409703297.shtml','http://games.sina.com.cn/w/n/2013-04-26/1406703296.shtml','http://games.sina.com.cn/w/n/2013-04-26/1402703292.shtml','http://games.sina.com.cn/w/n/2013-04-26/1353703286.shtml','http://games.sina.com.cn/w/n/2013-04-26/1348703284.shtml','http://games.sina.com.cn/w/n/2013-04-26/1327703275.shtml','http://games.sina.com.cn/w/n/2013-04-26/1239703265.shtml','http://games.sina.com.cn/w/n/2013-04-26/1238703264.shtml','http://games.sina.com.cn/w/n/2013-04-26/1231703262.shtml','http://games.sina.com.cn/w/n/2013-04-26/1229703261.shtml','http://games.sina.com.cn/w/n/2013-04-26/1228703260.shtml','http://games.sina.com.cn/w/n/2013-04-26/1223703259.shtml','http://games.sina.com.cn/w/n/2013-04-26/1218703258.shtml','http://games.sina.com.cn/w/n/2013-04-26/1202703254.shtml','http://games.sina.com.cn/w/n/2013-04-26/1159703251.shtml','http://games.sina.com.cn/w/n/2013-04-26/1139703233.shtml'] def workerbee(inputlist): for x in xrange(THREAD_LIMIT): print 'Thead {0} started.'.format(x) t = spider() t.start() for i in inputlist: try: jobs.put(i, block=True, timeout=5) except: singlelock.acquire() print "The queue is full !" singlelock.release() # Wait for the threads to finish singlelock.acquire() # Acquire the lock so we can print print "Waiting for threads to finish." singlelock.release() # Release the lock jobs.join() # This command waits for all threads to finish. # while not jobs.empty(): # print jobs.get() def getTitle(url,time=10): response = urllib2.urlopen(url,timeout=time) html = response.read() response.close() reg = r'<title>(.*?)</title>' title = re.compile(reg).findall(html) title = title[0].decode('gb2312','replace').encode('utf-8') return title class spider(threading.Thread): def run(self): while 1: try: job = jobs.get(True,1) singlelock.acquire() title = getTitle(job) print 'This {0} is {1}'.format(job,title) singlelock.release() jobs.task_done() except: break; if __name__ == '__main__': workerbee(urls)
本文向大家介绍Python实现多线程下载文件的代码实例,包括了Python实现多线程下载文件的代码实例的使用技巧和注意事项,需要的朋友参考一下 实现简单的多线程下载,需要关注如下几点: 1.文件的大小:可以从reponse header中提取,如“Content-Length:911”表示大小是911字节 2.任务拆分:指定各个线程下载的文件的哪一块,可以通过request header中添加“R
本文向大家介绍用python实现的线程池实例代码,包括了用python实现的线程池实例代码的使用技巧和注意事项,需要的朋友参考一下 python3标准库里自带线程池ThreadPoolExecutor和进程池ProcessPoolExecutor。 如果你用的是python2,那可以下载一个模块,叫threadpool,这是线程池。对于进程池可以使用python自带的multiprocessing
本文向大家介绍pandas apply多线程实现代码,包括了pandas apply多线程实现代码的使用技巧和注意事项,需要的朋友参考一下 一、多线程化选择 并行化一个代码有两大选择:multithread 和 multiprocess。 Multithread,多线程,同一个进程(process)可以开启多个线程执行计算。每个线程代表了一个 CPU 核心,这么多线程可以访问同
本文向大家介绍实例代码讲解Python 线程池,包括了实例代码讲解Python 线程池的使用技巧和注意事项,需要的朋友参考一下 大家都知道当任务过多,任务量过大时如果想提高效率的一个最简单的方法就是用多线程去处理,比如爬取上万个网页中的特定数据,以及将爬取数据和清洗数据的工作交给不同的线程去处理,也就是生产者消费者模式,都是典型的多线程使用场景。 那是不是意味着线程数量越多,程序的执行效率就越快呢
本文向大家介绍Python实现线程池代码分享,包括了Python实现线程池代码分享的使用技巧和注意事项,需要的朋友参考一下 原理:建立一个任务队列,然多个线程都从这个任务队列中取出任务然后执行,当然任务队列要加锁,详细请看代码
本文向大家介绍Python多进程编程multiprocessing代码实例,包括了Python多进程编程multiprocessing代码实例的使用技巧和注意事项,需要的朋友参考一下 在 多线程与多进程的比较 这一篇中记录了多进程编程的一种方式. 下面记录一下多进程编程的别一种方式,即使用multiprocessing编程 以上就是本文的全部内容,希望对大家的学习有所帮助,也希望大家多多支持呐喊教