freqtrade_origin/freqtrade/configuration.py

420 lines
17 KiB
Python
Raw Normal View History

"""
This module contains the configuration class
"""
import json
2018-03-25 19:37:14 +00:00
import logging
2018-07-04 07:31:35 +00:00
import os
2019-03-29 19:12:44 +00:00
import sys
from argparse import Namespace
2019-03-29 19:12:44 +00:00
from logging.handlers import RotatingFileHandler
2019-03-29 19:16:41 +00:00
from typing import Any, Dict, List, Optional
2018-07-04 07:31:35 +00:00
2019-04-08 01:23:29 +00:00
from jsonschema import Draft4Validator, validators
from jsonschema.exceptions import ValidationError, best_match
from freqtrade import OperationalException, constants
from freqtrade.exchange import is_exchange_supported, supported_exchanges
2019-02-19 12:14:47 +00:00
from freqtrade.misc import deep_merge_dicts
2019-03-29 19:12:44 +00:00
from freqtrade.state import RunMode
2019-02-19 12:14:47 +00:00
2018-07-19 18:43:41 +00:00
logger = logging.getLogger(__name__)
2018-07-18 20:52:57 +00:00
2018-07-19 19:12:27 +00:00
def set_loggers(log_level: int = 0) -> None:
"""
Set the logger level for Third party libs
:return: None
"""
logging.getLogger('requests').setLevel(logging.INFO if log_level <= 1 else logging.DEBUG)
logging.getLogger("urllib3").setLevel(logging.INFO if log_level <= 1 else logging.DEBUG)
logging.getLogger('ccxt.base.exchange').setLevel(
logging.INFO if log_level <= 2 else logging.DEBUG)
logging.getLogger('telegram').setLevel(logging.INFO)
2019-04-08 01:23:29 +00:00
def _extend_with_default(validator_class):
validate_properties = validator_class.VALIDATORS["properties"]
def set_defaults(validator, properties, instance, schema):
for prop, subschema in properties.items():
if "default" in subschema:
instance.setdefault(prop, subschema["default"])
for error in validate_properties(
validator, properties, instance, schema,
):
yield error
return validators.extend(
validator_class, {"properties": set_defaults},
)
ValidatorWithDefaults = _extend_with_default(Draft4Validator)
class Configuration(object):
"""
Class to read and init the bot configuration
Reuse this class for the bot, backtesting, hyperopt and every script that required configuration
"""
def __init__(self, args: Namespace, runmode: RunMode = None) -> None:
self.args = args
2018-05-30 20:38:09 +00:00
self.config: Optional[Dict[str, Any]] = None
self.runmode = runmode
def load_config(self) -> Dict[str, Any]:
"""
Extract information for sys.argv and load the bot configuration
:return: Configuration dictionary
"""
2019-02-19 12:14:47 +00:00
config: Dict[str, Any] = {}
# Now expecting a list of config filenames here, not a string
for path in self.args.config:
logger.info('Using config: %s ...', path)
# Merge config options, overwriting old values
config = deep_merge_dicts(self._load_config_file(path), config)
if 'internals' not in config:
config['internals'] = {}
logger.info('Validating configuration ...')
self._validate_config_schema(config)
2019-03-16 09:38:25 +00:00
self._validate_config_consistency(config)
2018-03-27 16:20:15 +00:00
# Set strategy if not specified in config and or if it's non default
if self.args.strategy != constants.DEFAULT_STRATEGY or not config.get('strategy'):
config.update({'strategy': self.args.strategy})
2018-03-25 14:28:04 +00:00
if self.args.strategy_path:
config.update({'strategy_path': self.args.strategy_path})
# Load Common configuration
config = self._load_common_config(config)
# Load Backtesting
config = self._load_backtesting_config(config)
2018-11-14 11:53:20 +00:00
# Load Edge
config = self._load_edge_config(config)
# Load Hyperopt
config = self._load_hyperopt_config(config)
# Set runmode
if not self.runmode:
# Handle real mode, infer dry/live from config
self.runmode = RunMode.DRY_RUN if config.get('dry_run', True) else RunMode.LIVE
config.update({'runmode': self.runmode})
return config
def _load_config_file(self, path: str) -> Dict[str, Any]:
"""
Loads a config file from the given path
:param path: path as str
:return: configuration as dictionary
"""
try:
with open(path) as file:
conf = json.load(file)
except FileNotFoundError:
raise OperationalException(
2018-06-14 05:31:29 +00:00
f'Config file "{path}" not found!'
' Please create a config file or check whether it exists.')
2019-02-19 12:14:47 +00:00
return conf
def _load_common_config(self, config: Dict[str, Any]) -> Dict[str, Any]:
"""
Extract information for sys.argv and load common configuration
:return: configuration as dictionary
"""
# Log level
if 'loglevel' in self.args and self.args.loglevel:
2018-07-19 19:12:27 +00:00
config.update({'verbosity': self.args.loglevel})
else:
config.update({'verbosity': 0})
2019-03-29 19:12:44 +00:00
# Log to stdout, not stderr
2019-03-29 19:16:41 +00:00
log_handlers: List[logging.Handler] = [logging.StreamHandler(sys.stdout)]
2019-03-29 19:12:44 +00:00
if 'logfile' in self.args and self.args.logfile:
config.update({'logfile': self.args.logfile})
# Allow setting this as either configuration or argument
if 'logfile' in config:
log_handlers.append(RotatingFileHandler(config['logfile'],
maxBytes=1024 * 1024, # 1Mb
backupCount=10))
2018-07-19 19:12:27 +00:00
logging.basicConfig(
level=logging.INFO if config['verbosity'] < 1 else logging.DEBUG,
format='%(asctime)s - %(name)s - %(levelname)s - %(message)s',
2019-03-29 19:12:44 +00:00
handlers=log_handlers
2018-07-19 19:12:27 +00:00
)
set_loggers(config['verbosity'])
logger.info('Verbosity set to %s', config['verbosity'])
# Support for sd_notify
if self.args.sd_notify:
config['internals'].update({'sd_notify': True})
# Add dynamic_whitelist if found
if 'dynamic_whitelist' in self.args and self.args.dynamic_whitelist:
2018-12-03 19:00:18 +00:00
# Update to volumePairList (the previous default)
2018-12-04 19:23:03 +00:00
config['pairlist'] = {'method': 'VolumePairList',
'config': {'number_assets': self.args.dynamic_whitelist}
}
2018-12-03 19:00:18 +00:00
logger.warning(
'Parameter --dynamic-whitelist has been deprecated, '
'and will be completely replaced by the whitelist dict in the future. '
'For now: using dynamically generated whitelist based on VolumePairList. '
'(not applicable with Backtesting and Hyperopt)'
)
if self.args.db_url and self.args.db_url != constants.DEFAULT_DB_PROD_URL:
config.update({'db_url': self.args.db_url})
logger.info('Parameter --db-url detected ...')
if config.get('dry_run', False):
logger.info('Dry run is enabled')
if config.get('db_url') in [None, constants.DEFAULT_DB_PROD_URL]:
# Default to in-memory db for dry_run if not specified
config['db_url'] = constants.DEFAULT_DB_DRYRUN_URL
else:
if not config.get('db_url', None):
config['db_url'] = constants.DEFAULT_DB_PROD_URL
logger.info('Dry run is disabled')
if config.get('forcebuy_enable', False):
logger.warning('`forcebuy` RPC message enabled.')
# Setting max_open_trades to infinite if -1
if config.get('max_open_trades') == -1:
config['max_open_trades'] = float('inf')
2018-06-14 05:31:29 +00:00
logger.info(f'Using DB: "{config["db_url"]}"')
2018-03-30 20:14:35 +00:00
# Check if the exchange set by the user is supported
self.check_exchange(config)
2018-03-30 20:14:35 +00:00
return config
2019-01-01 13:07:40 +00:00
def _create_datadir(self, config: Dict[str, Any], datadir: Optional[str] = None) -> str:
if not datadir:
# set datadir
exchange_name = config.get('exchange', {}).get('name').lower()
datadir = os.path.join('user_data', 'data', exchange_name)
if not os.path.isdir(datadir):
os.makedirs(datadir)
logger.info(f'Created data directory: {datadir}')
return datadir
2019-04-24 19:13:57 +00:00
def _args_to_config(self, config, argname, logstring, logfun=None) -> None:
"""
logfun is applied to the configuration entry before passing
that entry to the log string using .format().
sample: logfun=len (prints the length of the found configuration instead of the content)
"""
2019-04-24 18:44:36 +00:00
if argname in self.args and getattr(self.args, argname):
2019-04-24 19:13:57 +00:00
config.update({argname: getattr(self.args, argname)})
if logfun:
2019-04-24 19:13:57 +00:00
logger.info(logstring.format(logfun(config[argname])))
else:
2019-04-24 19:13:57 +00:00
logger.info(logstring.format(config[argname]))
2019-04-24 18:44:36 +00:00
2019-04-24 19:12:08 +00:00
def _load_backtesting_config(self, config: Dict[str, Any]) -> Dict[str, Any]:
"""
Extract information for sys.argv and load Backtesting configuration
:return: configuration as dictionary
"""
# This will override the strategy configuration
if 'ticker_interval' in self.args and self.args.ticker_interval:
config.update({'ticker_interval': self.args.ticker_interval})
2018-03-25 19:37:14 +00:00
logger.info('Parameter -i/--ticker-interval detected ...')
logger.info('Using ticker_interval: %s ...', config.get('ticker_interval'))
2019-04-24 19:13:57 +00:00
self._args_to_config(config, argname='live',
logstring='Parameter -l/--live detected ...')
2019-04-24 19:13:57 +00:00
self._args_to_config(config, argname='position_stacking',
logstring='Parameter --enable-position-stacking detected ...')
2018-07-17 19:05:03 +00:00
if 'use_max_market_positions' in self.args and not self.args.use_max_market_positions:
config.update({'use_max_market_positions': False})
logger.info('Parameter --disable-max-market-positions detected ...')
logger.info('max_open_trades set to unlimited ...')
elif 'max_open_trades' in self.args and self.args.max_open_trades:
config.update({'max_open_trades': self.args.max_open_trades})
2019-04-14 08:17:06 +00:00
logger.info('Parameter --max_open_trades detected, '
'overriding max_open_trades to: %s ...', config.get('max_open_trades'))
2018-07-17 19:05:03 +00:00
else:
logger.info('Using max_open_trades: %s ...', config.get('max_open_trades'))
2019-04-24 19:13:57 +00:00
self._args_to_config(config, argname='stake_amount',
logstring='Parameter --stake_amount detected, '
'overriding stake_amount to: {} ...')
2019-04-24 19:13:57 +00:00
self._args_to_config(config, argname='timerange',
logstring='Parameter --timerange detected: {} ...')
if 'datadir' in self.args and self.args.datadir:
2019-01-01 13:07:40 +00:00
config.update({'datadir': self._create_datadir(config, self.args.datadir)})
else:
2019-01-01 13:07:40 +00:00
config.update({'datadir': self._create_datadir(config, None)})
logger.info('Using data folder: %s ...', config.get('datadir'))
2019-04-24 19:13:57 +00:00
self._args_to_config(config, argname='refresh_pairs',
logstring='Parameter -r/--refresh-pairs-cached detected ...')
2019-04-24 19:13:57 +00:00
self._args_to_config(config, argname='strategy_list',
logstring='Using strategy list of {} Strategies', logfun=len)
2018-07-27 21:00:50 +00:00
2019-04-24 19:13:57 +00:00
self._args_to_config(config, argname='ticker_interval',
logstring='Overriding ticker interval with Command line argument')
2018-07-27 21:00:50 +00:00
2019-04-24 19:13:57 +00:00
self._args_to_config(config, argname='export',
logstring='Parameter --export detected: {} ...')
2019-04-24 19:13:57 +00:00
self._args_to_config(config, argname='exportfilename',
logstring='Storing backtest results to {} ...')
2018-06-03 12:52:03 +00:00
return config
2018-11-14 11:53:20 +00:00
def _load_edge_config(self, config: Dict[str, Any]) -> Dict[str, Any]:
"""
Extract information for sys.argv and load Edge configuration
:return: configuration as dictionary
"""
2019-04-24 19:13:57 +00:00
self._args_to_config(config, argname='timerange',
2019-04-24 19:12:08 +00:00
logstring='Parameter --timerange detected: {} ...')
2018-11-14 11:53:20 +00:00
2018-11-14 15:31:23 +00:00
if 'stoploss_range' in self.args and self.args.stoploss_range:
txt_range = eval(self.args.stoploss_range)
config['edge'].update({'stoploss_range_min': txt_range[0]})
config['edge'].update({'stoploss_range_max': txt_range[1]})
config['edge'].update({'stoploss_range_step': txt_range[2]})
logger.info('Parameter --stoplosses detected: %s ...', self.args.stoploss_range)
2019-04-24 19:13:57 +00:00
self._args_to_config(config, argname='refresh_pairs',
2019-04-24 19:12:08 +00:00
logstring='Parameter -r/--refresh-pairs-cached detected ...')
2018-11-14 11:53:20 +00:00
return config
def _load_hyperopt_config(self, config: Dict[str, Any]) -> Dict[str, Any]:
"""
Extract information for sys.argv and load Hyperopt configuration
:return: configuration as dictionary
"""
2019-04-24 19:13:57 +00:00
self._args_to_config(config, argname='hyperopt',
2019-04-24 19:12:08 +00:00
logstring='Using Hyperopt file {}')
2019-04-24 19:13:57 +00:00
self._args_to_config(config, argname='epochs',
2019-04-24 19:12:08 +00:00
logstring='Parameter --epochs detected ... '
'Will run Hyperopt with for {} epochs ...'
)
2019-04-24 19:13:57 +00:00
self._args_to_config(config, argname='spaces',
2019-04-24 19:12:08 +00:00
logstring='Parameter -s/--spaces detected: {}')
2019-04-24 19:13:57 +00:00
self._args_to_config(config, argname='print_all',
2019-04-24 19:12:08 +00:00
logstring='Parameter --print-all detected ...')
2019-04-24 19:13:57 +00:00
self._args_to_config(config, argname='hyperopt_jobs',
2019-04-24 19:12:08 +00:00
logstring='Parameter -j/--job-workers detected: {}')
2019-04-24 19:13:57 +00:00
self._args_to_config(config, argname='refresh_pairs',
2019-04-24 19:12:08 +00:00
logstring='Parameter -r/--refresh-pairs-cached detected ...')
2019-04-24 19:12:08 +00:00
self._args_to_config(config, argname='hyperopt_random_state',
logstring='Parameter --random-state detected: {}')
return config
def _validate_config_schema(self, conf: Dict[str, Any]) -> Dict[str, Any]:
"""
Validate the configuration follow the Config Schema
:param conf: Config in JSON format
:return: Returns the config if valid, otherwise throw an exception
"""
try:
2019-04-08 01:23:29 +00:00
ValidatorWithDefaults(constants.CONF_SCHEMA).validate(conf)
return conf
except ValidationError as exception:
2018-05-30 20:38:09 +00:00
logger.critical(
'Invalid configuration. See config.json.example. Reason: %s',
exception
)
raise ValidationError(
best_match(Draft4Validator(constants.CONF_SCHEMA).iter_errors(conf)).message
)
2019-03-16 09:38:25 +00:00
def _validate_config_consistency(self, conf: Dict[str, Any]) -> None:
"""
Validate the configuration consistency
:param conf: Config in JSON format
2019-03-16 09:38:25 +00:00
:return: Returns None if everything is ok, otherwise throw an OperationalException
"""
# validating trailing stoploss
self._validate_trailing_stoploss(conf)
def _validate_trailing_stoploss(self, conf: Dict[str, Any]) -> None:
# Skip if trailing stoploss is not activated
if not conf.get('trailing_stop', False):
return
tsl_positive = float(conf.get('trailing_stop_positive', 0))
tsl_offset = float(conf.get('trailing_stop_positive_offset', 0))
tsl_only_offset = conf.get('trailing_only_offset_is_reached', False)
if tsl_only_offset:
if tsl_positive == 0.0:
raise OperationalException(
2019-03-16 09:38:25 +00:00
f'The config trailing_only_offset_is_reached needs '
'trailing_stop_positive_offset to be more than 0 in your config.')
if tsl_positive > 0 and 0 < tsl_offset <= tsl_positive:
raise OperationalException(
2019-03-16 09:38:25 +00:00
f'The config trailing_stop_positive_offset needs '
'to be greater than trailing_stop_positive_offset in your config.')
def get_config(self) -> Dict[str, Any]:
"""
Return the config. Use this method to get the bot config
:return: Dict: Bot config
"""
if self.config is None:
self.config = self.load_config()
return self.config
2018-03-30 20:14:35 +00:00
def check_exchange(self, config: Dict[str, Any]) -> bool:
2018-03-30 20:14:35 +00:00
"""
Check if the exchange name in the config file is supported by Freqtrade
:return: True or raised an exception if the exchange if not supported
"""
exchange = config.get('exchange', {}).get('name').lower()
if not is_exchange_supported(exchange):
2018-03-30 20:14:35 +00:00
exception_msg = f'Exchange "{exchange}" not supported.\n' \
f'The following exchanges are supported: ' \
f'{", ".join(supported_exchanges())}'
2018-03-30 20:14:35 +00:00
2018-03-25 19:37:14 +00:00
logger.critical(exception_msg)
2018-03-30 20:14:35 +00:00
raise OperationalException(
exception_msg
)
2018-03-25 19:37:14 +00:00
logger.debug('Exchange "%s" supported', exchange)
2018-03-30 20:14:35 +00:00
return True