diff options
author | François Schmidts <francois.schmidts@gmail.com> | 2015-12-12 21:14:28 +0100 |
---|---|---|
committer | François Schmidts <francois.schmidts@gmail.com> | 2015-12-17 09:42:56 +0100 |
commit | b35e9773198ef2d8b37c4ca223f08147db47de0b (patch) | |
tree | ba4b1b171b3c1ab9414a96ad264c47b0f9d1246b /manager.py | |
parent | Updated link to Heroku deploy button on the About page. (diff) | |
download | newspipe-b35e9773198ef2d8b37c4ca223f08147db47de0b.tar.gz newspipe-b35e9773198ef2d8b37c4ca223f08147db47de0b.tar.bz2 newspipe-b35e9773198ef2d8b37c4ca223f08147db47de0b.zip |
moving the root of source code from / to /src/
Diffstat (limited to 'manager.py')
-rwxr-xr-x | manager.py | 74 |
1 files changed, 0 insertions, 74 deletions
diff --git a/manager.py b/manager.py deleted file mode 100755 index 5aedbe18..00000000 --- a/manager.py +++ /dev/null @@ -1,74 +0,0 @@ -#! /usr/bin/env python -# -*- coding: utf-8 -*- - -from bootstrap import application, db, populate_g, conf -from flask.ext.script import Manager -from flask.ext.migrate import Migrate, MigrateCommand - -import web.models - -Migrate(application, db) - -manager = Manager(application) -manager.add_command('db', MigrateCommand) - -@manager.command -def db_empty(): - "Will drop every datas stocked in db." - with application.app_context(): - populate_g() - web.models.db_empty(db) - -@manager.command -def db_create(): - "Will create the database from conf parameters." - with application.app_context(): - populate_g() - web.models.db_create(db) - -@manager.command -def fetch(limit=100, retreive_all=False): - "Crawl the feeds with the client crawler." - from web.lib.crawler import CrawlerScheduler - scheduler = CrawlerScheduler(conf.API_LOGIN, conf.API_PASSWD) - scheduler.run(limit=limit, retreive_all=retreive_all) - scheduler.wait() - -@manager.command -def fetch_asyncio(user_id, feed_id): - "Crawl the feeds with asyncio." - import asyncio - - with application.app_context(): - populate_g() - from flask import g - from web.models import User - from web import crawler - users, feed_id = [], None - try: - users = User.query.filter(User.id == int(user_id)).all() - except: - users = User.query.all() - finally: - if users == []: - users = User.query.all() - - try: - feed_id = int(feed_id) - except: - feed_id = None - - loop = asyncio.get_event_loop() - for user in users: - if user.activation_key == "": - print("Fetching articles for " + user.nickname) - g.user = user - feed_getter = crawler.retrieve_feed(loop, g.user, feed_id) - loop.close() - -from scripts.probes import ArticleProbe, FeedProbe -manager.add_command('probe_articles', ArticleProbe()) -manager.add_command('probe_feeds', FeedProbe()) - -if __name__ == '__main__': - manager.run() |