| 
									
										
										
										
											2014-09-01 14:38:59 +02:00
										 |  |  | ## Bing (News) | 
					
						
							| 
									
										
										
										
											2014-12-07 16:37:56 +01:00
										 |  |  | # | 
					
						
							| 
									
										
										
										
											2014-09-01 14:38:59 +02:00
										 |  |  | # @website     https://www.bing.com/news | 
					
						
							| 
									
										
										
										
											2014-12-07 16:37:56 +01:00
										 |  |  | # @provide-api yes (http://datamarket.azure.com/dataset/bing/search), | 
					
						
							|  |  |  | #              max. 5000 query/month | 
					
						
							|  |  |  | # | 
					
						
							| 
									
										
										
										
											2014-09-01 14:38:59 +02:00
										 |  |  | # @using-api   no (because of query limit) | 
					
						
							|  |  |  | # @results     HTML (using search portal) | 
					
						
							|  |  |  | # @stable      no (HTML can change) | 
					
						
							|  |  |  | # @parse       url, title, content, publishedDate | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2014-03-04 13:10:04 +01:00
										 |  |  | from urllib import urlencode | 
					
						
							|  |  |  | from cgi import escape | 
					
						
							|  |  |  | from lxml import html | 
					
						
							| 
									
										
										
										
											2014-09-01 14:38:59 +02:00
										 |  |  | from datetime import datetime, timedelta | 
					
						
							|  |  |  | from dateutil import parser | 
					
						
							|  |  |  | import re | 
					
						
							| 
									
										
										
										
											2014-03-04 13:10:04 +01:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2014-09-01 14:38:59 +02:00
										 |  |  | # engine dependent config | 
					
						
							| 
									
										
										
										
											2014-03-04 13:10:04 +01:00
										 |  |  | categories = ['news'] | 
					
						
							|  |  |  | paging = True | 
					
						
							|  |  |  | language_support = True | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2014-09-01 14:38:59 +02:00
										 |  |  | # search-url | 
					
						
							|  |  |  | base_url = 'https://www.bing.com/' | 
					
						
							|  |  |  | search_string = 'news/search?{query}&first={offset}' | 
					
						
							| 
									
										
										
										
											2014-03-04 13:10:04 +01:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2014-09-02 17:13:44 +02:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2014-09-01 14:38:59 +02:00
										 |  |  | # do search-request | 
					
						
							| 
									
										
										
										
											2014-03-04 13:10:04 +01:00
										 |  |  | def request(query, params): | 
					
						
							|  |  |  |     offset = (params['pageno'] - 1) * 10 + 1 | 
					
						
							| 
									
										
										
										
											2014-09-01 14:38:59 +02:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2014-03-04 13:10:04 +01:00
										 |  |  |     if params['language'] == 'all': | 
					
						
							|  |  |  |         language = 'en-US' | 
					
						
							|  |  |  |     else: | 
					
						
							|  |  |  |         language = params['language'].replace('_', '-') | 
					
						
							| 
									
										
										
										
											2014-09-01 14:38:59 +02:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2014-03-04 13:10:04 +01:00
										 |  |  |     search_path = search_string.format( | 
					
						
							|  |  |  |         query=urlencode({'q': query, 'setmkt': language}), | 
					
						
							|  |  |  |         offset=offset) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     params['cookies']['SRCHHPGUSR'] = \ | 
					
						
							|  |  |  |         'NEWWND=0&NRSLT=-1&SRCHLANG=' + language.split('-')[0] | 
					
						
							| 
									
										
										
										
											2014-09-01 14:38:59 +02:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2014-03-04 13:10:04 +01:00
										 |  |  |     params['url'] = base_url + search_path | 
					
						
							|  |  |  |     return params | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2014-09-01 14:38:59 +02:00
										 |  |  | # get response from search-request | 
					
						
							| 
									
										
										
										
											2014-03-04 13:10:04 +01:00
										 |  |  | def response(resp): | 
					
						
							|  |  |  |     results = [] | 
					
						
							| 
									
										
										
										
											2014-09-01 14:38:59 +02:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2014-03-04 13:10:04 +01:00
										 |  |  |     dom = html.fromstring(resp.content) | 
					
						
							| 
									
										
										
										
											2014-09-01 14:38:59 +02:00
										 |  |  | 
 | 
					
						
							|  |  |  |     # parse results | 
					
						
							|  |  |  |     for result in dom.xpath('//div[@class="sn_r"]'): | 
					
						
							|  |  |  |         link = result.xpath('.//div[@class="newstitle"]/a')[0] | 
					
						
							| 
									
										
										
										
											2014-03-04 13:10:04 +01:00
										 |  |  |         url = link.attrib.get('href') | 
					
						
							|  |  |  |         title = ' '.join(link.xpath('.//text()')) | 
					
						
							| 
									
										
										
										
											2014-12-16 17:26:16 +01:00
										 |  |  |         contentXPath = result.xpath('.//div[@class="sn_txt"]/div' | 
					
						
							|  |  |  |                                     '//span[@class="sn_snip"]//text()') | 
					
						
							| 
									
										
										
										
											2014-12-07 16:37:56 +01:00
										 |  |  |         if contentXPath is not None: | 
					
						
							| 
									
										
										
										
											2014-09-07 18:10:05 +02:00
										 |  |  |             content = escape(' '.join(contentXPath)) | 
					
						
							| 
									
										
										
										
											2014-12-07 16:37:56 +01:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2014-09-01 14:38:59 +02:00
										 |  |  |         # parse publishedDate | 
					
						
							| 
									
										
										
										
											2014-12-16 17:26:16 +01:00
										 |  |  |         publishedDateXPath = result.xpath('.//div[@class="sn_txt"]/div' | 
					
						
							|  |  |  |                                           '//span[contains(@class,"sn_ST")]' | 
					
						
							|  |  |  |                                           '//span[contains(@class,"sn_tm")]' | 
					
						
							|  |  |  |                                           '//text()') | 
					
						
							| 
									
										
										
										
											2014-12-07 16:37:56 +01:00
										 |  |  |         if publishedDateXPath is not None: | 
					
						
							| 
									
										
										
										
											2014-09-07 18:10:05 +02:00
										 |  |  |             publishedDate = escape(' '.join(publishedDateXPath)) | 
					
						
							| 
									
										
										
										
											2014-03-04 13:10:04 +01:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2014-09-01 14:38:59 +02:00
										 |  |  |         if re.match("^[0-9]+ minute(s|) ago$", publishedDate): | 
					
						
							|  |  |  |             timeNumbers = re.findall(r'\d+', publishedDate) | 
					
						
							|  |  |  |             publishedDate = datetime.now()\ | 
					
						
							|  |  |  |                 - timedelta(minutes=int(timeNumbers[0])) | 
					
						
							|  |  |  |         elif re.match("^[0-9]+ hour(s|) ago$", publishedDate): | 
					
						
							|  |  |  |             timeNumbers = re.findall(r'\d+', publishedDate) | 
					
						
							|  |  |  |             publishedDate = datetime.now()\ | 
					
						
							|  |  |  |                 - timedelta(hours=int(timeNumbers[0])) | 
					
						
							| 
									
										
										
										
											2014-12-16 17:26:16 +01:00
										 |  |  |         elif re.match("^[0-9]+ hour(s|)," | 
					
						
							|  |  |  |                       " [0-9]+ minute(s|) ago$", publishedDate): | 
					
						
							| 
									
										
										
										
											2014-09-01 14:38:59 +02:00
										 |  |  |             timeNumbers = re.findall(r'\d+', publishedDate) | 
					
						
							|  |  |  |             publishedDate = datetime.now()\ | 
					
						
							|  |  |  |                 - timedelta(hours=int(timeNumbers[0]))\ | 
					
						
							|  |  |  |                 - timedelta(minutes=int(timeNumbers[1])) | 
					
						
							| 
									
										
										
										
											2014-09-07 18:10:05 +02:00
										 |  |  |         elif re.match("^[0-9]+ day(s|) ago$", publishedDate): | 
					
						
							|  |  |  |             timeNumbers = re.findall(r'\d+', publishedDate) | 
					
						
							|  |  |  |             publishedDate = datetime.now()\ | 
					
						
							|  |  |  |                 - timedelta(days=int(timeNumbers[0])) | 
					
						
							| 
									
										
										
										
											2014-09-01 14:38:59 +02:00
										 |  |  |         else: | 
					
						
							| 
									
										
										
										
											2014-09-07 18:10:05 +02:00
										 |  |  |             try: | 
					
						
							|  |  |  |                 # FIXME use params['language'] to parse either mm/dd or dd/mm | 
					
						
							|  |  |  |                 publishedDate = parser.parse(publishedDate, dayfirst=False) | 
					
						
							|  |  |  |             except TypeError: | 
					
						
							|  |  |  |                 # FIXME | 
					
						
							|  |  |  |                 publishedDate = datetime.now() | 
					
						
							| 
									
										
										
										
											2014-12-07 16:37:56 +01:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2014-09-01 14:38:59 +02:00
										 |  |  |         # append result | 
					
						
							| 
									
										
										
										
											2014-12-07 16:37:56 +01:00
										 |  |  |         results.append({'url': url, | 
					
						
							|  |  |  |                         'title': title, | 
					
						
							| 
									
										
										
										
											2014-09-01 14:38:59 +02:00
										 |  |  |                         'publishedDate': publishedDate, | 
					
						
							|  |  |  |                         'content': content}) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     # return results | 
					
						
							| 
									
										
										
										
											2014-03-04 13:10:04 +01:00
										 |  |  |     return results |