aboutsummaryrefslogtreecommitdiff
path: root/pyaggr3g470r
diff options
context:
space:
mode:
authorCédric Bonhomme <cedric@cedricbonhomme.org>2015-02-04 21:22:57 +0100
committerCédric Bonhomme <cedric@cedricbonhomme.org>2015-02-04 21:22:57 +0100
commit62ebba3399832a24a77c2b737e9d909e83e29020 (patch)
tree26f8d63e53e9fa0868deab466e3547af71719535 /pyaggr3g470r
parentTest if we effectively have retrieved some articles. (diff)
downloadnewspipe-62ebba3399832a24a77c2b737e9d909e83e29020.tar.gz
newspipe-62ebba3399832a24a77c2b737e9d909e83e29020.tar.bz2
newspipe-62ebba3399832a24a77c2b737e9d909e83e29020.zip
Get the feeds with aiohttp.
Diffstat (limited to 'pyaggr3g470r')
-rw-r--r--pyaggr3g470r/crawler.py17
1 files changed, 14 insertions, 3 deletions
diff --git a/pyaggr3g470r/crawler.py b/pyaggr3g470r/crawler.py
index e6daecd4..27f5a200 100644
--- a/pyaggr3g470r/crawler.py
+++ b/pyaggr3g470r/crawler.py
@@ -27,6 +27,7 @@ __copyright__ = "Copyright (c) Cedric Bonhomme"
__license__ = "AGPLv3"
import asyncio
+import aiohttp
import logging
import requests
import feedparser
@@ -47,13 +48,20 @@ logger = logging.getLogger(__name__)
#
@asyncio.coroutine
+def get(*args, **kwargs):
+ kwargs["connector"] = aiohttp.TCPConnector(verify_ssl=False)
+ response = yield from aiohttp.request('GET', *args, **kwargs)
+ return (yield from response.read_and_close(decode=False))
+
+@asyncio.coroutine
def fetch(user, feed):
"""
Fetch a feed.
"""
logger.info("Fetching the feed: " + feed.title)
print("Fetching the feed: " + feed.title)
- a_feed = feedparser.parse(feed.link)
+ data = yield from get(feed.link)
+ a_feed = feedparser.parse(data)
if a_feed['bozo'] == 1:
logger.error(a_feed['bozo_exception'])
if a_feed['entries'] == []:
@@ -177,9 +185,12 @@ def insert_database(user, feed):
def done(feed):
print('done {}'.format(feed.title))
+sem = asyncio.Semaphore(5)
+
@asyncio.coroutine
def process_data(user, feed):
- data = yield from asyncio.async(insert_database(user, feed))
+ with (yield from sem):
+ data = yield from asyncio.async(insert_database(user, feed))
print('inserted articles for {}'.format(feed.title))
def retrieve_feed(user, feed_id=None):
@@ -190,7 +201,7 @@ def retrieve_feed(user, feed_id=None):
# 1 - Get the list of feeds to fetch
user = User.query.filter(User.email == user.email).first()
- feeds = [feed for feed in user.feeds if feed.enabled]
+ feeds = [feed for feed in user.feeds if feed.enabled][:20]
if feed_id is not None:
feeds = [feed for feed in feeds if feed.id == feed_id]
bgstack15