From 6ee91b3713a29df2dc6fc7d4f417ae4dbefa4972 Mon Sep 17 00:00:00 2001 From: Cédric Bonhomme Date: Thu, 23 Apr 2015 08:41:12 +0200 Subject: Automatically use the good Python executable for the classic crawler. --- conf/conf.cfg-sample | 17 ++++++++--------- 1 file changed, 8 insertions(+), 9 deletions(-) (limited to 'conf') diff --git a/conf/conf.cfg-sample b/conf/conf.cfg-sample index 76a91323..b141e93c 100644 --- a/conf/conf.cfg-sample +++ b/conf/conf.cfg-sample @@ -1,24 +1,23 @@ +[webserver] +host = 0.0.0.0 +port = 5000 +secret = a secret only you know [misc] -platform_url = http://127.0.0.1:5000/ +platform_url = http://0.0.0.0:5000/ admin_email = recaptcha_public_key = recaptcha_private_key = log_path = ./pyaggr3g470r/var/pyaggr3g470r.log -python = python3.3 nb_worker = 5 log_level = info [database] database_url = postgres://pgsqluser:pgsqlpwd@127.0.0.1:5432/aggregator -[feedparser] +[crawler] +crawling_method = classic +default_max_error = 6 http_proxy = user_agent = pyAggr3g470r (https://bitbucket.org/cedricbonhomme/pyaggr3g470r) resolve_article_url = false -default_max_error = 6 -crawling_method = classic -[webserver] -host = 0.0.0.0 -port = 5000 -secret = a secret only you know [notification] email = pyAggr3g470r@no-reply.com host = smtp.googlemail.com -- cgit