From e841afdabe76d370da85da0cd34c323f0e2b0c5d Mon Sep 17 00:00:00 2001 From: Cédric Bonhomme Date: Tue, 22 Nov 2016 08:40:13 +0100 Subject: Renamed crawler name. --- src/manager.py | 13 ++----------- 1 file changed, 2 insertions(+), 11 deletions(-) (limited to 'src/manager.py') diff --git a/src/manager.py b/src/manager.py index 46f8fe10..60e4c4f1 100755 --- a/src/manager.py +++ b/src/manager.py @@ -40,22 +40,13 @@ def db_create(): UserController(ignore_context=True).create(**admin) -@manager.command -def fetch(limit=100, retreive_all=False): - "Crawl the feeds with the client crawler." - from crawler.http_crawler import CrawlerScheduler - scheduler = CrawlerScheduler(conf.API_LOGIN, conf.API_PASSWD) - scheduler.run(limit=limit, retreive_all=retreive_all) - scheduler.wait() - - @manager.command def fetch_asyncio(user_id=None, feed_id=None): "Crawl the feeds with asyncio." import asyncio with application.app_context(): - from crawler import classic_crawler + from crawler import default_crawler filters = {} filters['is_active'] = True filters['automatic_crawling'] = True @@ -73,7 +64,7 @@ def fetch_asyncio(user_id=None, feed_id=None): start = datetime.now() loop = asyncio.get_event_loop() for user in users: - classic_crawler.retrieve_feed(loop, user, feed_id) + default_crawler.retrieve_feed(loop, user, feed_id) loop.close() end = datetime.now() -- cgit